Seminar "Selected Topics in Human Language Technology and Pattern Recognition"

In the Winter Semester 2019/20, the Lehrstuhl Informatik 6 will host a seminar entitled "Selected Topics in Human Language Technology and Pattern Recognition" for Bachelor and for Master level.

Registration for the seminar

Registration for the seminar is only possible online via the central registration page.

Prerequisites for Participation in the Seminar

General Goals of the Seminar

The goal of the seminar is to autonomously acquire knowledge and critical comprehension of an assigned topic, and present this topic both in writing and verbally.

This includes:

Seminar Format and Important Dates

The seminar will be started with a kick-off meeting, which will take place shortly after the central registration for the seminars in the Computer Science Department. The exact date of the kick-off meeting will be communicated directly the seminar participants selected in the central registration.

Please note the following deadlines during the seminar:

Note: failure to meet deadlines, absence without permission from compulsory sessions (presentations and preliminary meeting as announced by email to each participating student), or dropping out of the seminar after more than 3 weeks after the preliminary meeting/topic distribution results in the grade 5.0/not appeared.

The deadline for de-registration from the seminar is within three weeks after the distribution of the topics in during the kick-off meeting. After this deadline, seminar participation is confirmed and will be graded.



Topics, Initial References Defining the Topics, Participants, and Supervisors

In general, selected topics from the following general areas of Human Language Technology and Pattern Recognition will be offered: Below, you find exemplary topics. However, note that topics are subject to change/updates. The final topics will be presented in a kick-off meeting which will be announced to the seminar participants selected in the central registration for the seminar.
  1. Speaker Diarization

    1. Methods for Speaker Diarization (NN; Supervisor: Wilfried Michel)
      Initial References:

    2. Applications and Challenges for Speaker Diarization (NN; Supervisor: Wilfried Michel)
      Initial References:

  2. Speaker Separation


    1. Permutation-Invariant Training (NN; Supervisor: Tobias Menne)
      Initial References:

    2. Speaker-Dependent Speaker Separation (NN; Supervisor: Tobias Menne)
      Initial References:

  3. Natural Language Understanding

    1. Pre-training Language Representation Models for NLP (Choi; Supervisor: Kazuki Irie)
      Initial References:

    2. Natural Language Understanding (Zhan; Supervisor: Kazuki Irie)
      Presentation: Wed, Feb 26, 2020, 14h

      Initial References:

    3. Cross-lingual Word Embedding (Vanvinckenroye; Supervisor: Yunsu Kim)
      Presentation: Wed, Feb 26, 2020, 15h

      Initial References:

    4. Cross-lingual Sentence Embedding (NN; Supervisor: Yunsu Kim)
      Initial References:

  4. Sentiment Analysis

    1. Emotion Detection (Hugenroth; Supervisor: Eugen Beck)
      Initial References:

    2. Multimodal Sentiment Analysis (Mangel; Supervisor: Eugen Beck)
      Presentation: Wed, Feb 26, 2020, 16h

      Initial References:

  5. Language Identification

    1. State-of-the-Art Language Identification (Rompelberg; Supervisor: Markus Kitza)
      Presentation: Wed, Feb 26, 2020, 17h

      Initial References:

    2. Fusion based Native Language Identification (NN; Supervisor: Markus Kitza)
      Initial References:

  6. Text-to-Speech

    1. Auto-regressive Models (NN; Supervisor: Yingbo Gao)
      Initial References:

    2. Inverse Autoregressive Flows (NN; Supervisor: Yingbo Gao)
      Initial References:

    3. End-to-end Text-to-speech (NN; Supervisor: Peter Vieting)
      Initial References:


  7. Speech-to-Text Translation

    1. End-to-end Speech-to-text Translation (Saeed; Supervisor: Parnia Bahar)
      Initial References:


  8. Reinforcement Learning

    1. Minimum Expected Loss Training (Gerstenberger; Supervisor: Albert Zeyer)
      Initial References:

    2. Modern Policy Learning Methods for Games (El Qoraichi; Supervisor: Albert Zeyer)
      Initial References:

    3. Memory Augmented Networks for Reinforcement Learning (Petrick; Supervisor: Christoph Lüscher)
      Presentation: Thur, Feb 27, 2020, 14h

      Initial References:

  9. Text Summarization

    1. Extractive Text Summarization (Swoboda; Supervisor: Jan Rosendahl)
      Presentation: Thur, Feb 27, 2020, 15h

      Initial References:

    2. Abstractive Text Summarization (with Deep Learning) (Wynands; Supervisor: Christian Herold)
      Presentation: Thur, Feb 27, 2020, 16h

      Initial References:

  10. Named Entity Recognition

    1. Named Entity Recognition (Becks; Supervisor: Weiyue Wang)
      Presentation: Thur, Feb 27, 2020, 17h

      Initial References:

    2. Entity Linking (Das; Supervisor: Weiyue Wang)
      Initial References:

  11. Constituency Parsing

    1. Neural Network-based Parsing (NN; Supervisor: Parnia Bahar)
      Initial References:
      • A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin: "Attention Is All You Need," in Advances in Neural Information Processing Systems 30, Annual Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, Dec. 2017, arXiv:1706.03762.
      • C. Dyer, A. Kuncoro, M. Ballesteros, N. A. Smith: "Recurrent neural network grammars," in Proc. 15th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), San Diego, CA, Jun. 2016, arXiv:1602.03762.
      • M.-T. Luong, Q. V. Le, I. Sutskever, O. Vinyals, L. Kaiser: "Multi-task Sequence to Sequence Learning", in International Conference on Learning Representations (ICLR) , San Juan, Puerto Rico, May 2016, arXiv:1511.06114.
    2. Overview of Topics

        C2. Natural Language Understanding - Presentation: Wed, Feb 26, 2020, 14h
        C3. Cross-lingual Word Embedding - Presentation: Wed, Feb 26, 2020, 15h
        D2. Multimodal Sentiment Analysis - Presentation: Wed, Feb 26, 2020, 16h
        E1. State-of-the-Art Language Identification - Presentation: Wed, Feb 26, 2020, 17h
        H3. Memory Augmented Networks for Reinforcement Learning - Presentation: Thur, Feb 27, 2020, 14h
        I1. Extractive Text Summarization - Presentation: Thur, Feb 27, 2020, 15h
        I2. Abstractive Text Summarization (with Deep Learning) - Presentation: Thur, Feb 27, 2020, 16h
        J1. Named Entity Recognition - Presentation: Thur, Feb 27, 2020, 17h


    Article and Presentation Format

    The roughly 20-page article together with the slides (between 20 & 30) for the presentation should be prepared in LaTeX format. Presentations will consist of 30 to 40 minutes presentation time & 15 minutes discussion time. Document templates for both the article and the presentation slides are provided below along with links to LaTeX documentation available online. The article and the slides should be prepared in LaTeX format and submitted electronically in pdf format. Other formats will not be accepted.

    Detailed Guidelines:

    Some Tips:

    Time management is crucial for a successful seminar:
    Successful seminar articles/presentations typically:
    While reading papers, it might be useful to keep the following questions in mind:

    Contact

    Questions regarding the content of the assigned seminar topics should be directed to the respective topic's supervisors.

    General and administrative inquiries should be directed to:

    Parnia Bahar
    RWTH Aachen University
    Lehrstuhl Informatik 6
    Ahornstr. 55
    52074 Aachen

    Room 6125b
    Tel: 0241 80 21632

    E-Mail: bahar@cs.rwth-aachen.de