Seminar "Selected Topics in Human Language Technology and Pattern Recognition"

In the Summer Semester 2020, the Lehrstuhl Informatik 6 will host a seminar entitled "Selected Topics in Human Language Technology and Pattern Recognition" for Bachelor and for Master level.

Registration for the seminar

Registration for the seminar is only possible online via the central registration page.

Prerequisites for Participation in the Seminar

General Goals of the Seminar

The goal of the seminar is to autonomously acquire knowledge and critical comprehension of an assigned topic, and present this topic both in writing and verbally.

This includes:

Seminar Format and Important Dates

The seminar will be started with a kick-off meeting, which will take place shortly after the central registration for the seminars in the Computer Science Department. The exact date of the kick-off meeting will be communicated directly to the seminar participants selected in the central registration.

Please note the following deadlines during the seminar:

Note: failure to meet deadlines, absence without permission from compulsory sessions (presentations and preliminary meeting as announced by email to each participating student), or dropping out of the seminar after more than 3 weeks after the preliminary meeting/topic distribution results in the grade 5.0/not appeared.

The deadline for de-registration from the seminar is Monday, March 30, 2020, 23:59h, i.e. within three weeks after the distribution of the topics. After this deadline, seminar participation is confirmed and will be graded.



Topics, Initial References Defining the Topics, Participants, and Supervisors

In general, selected topics from the following general areas of Human Language Technology and Pattern Recognition will be offered: Below, you find exemplary topics. However, note that topics are subject to change/updates. The final topics will be presented in a kick-off meeting which will be announced to the seminar participants selected in the central registration for the seminar.
  1. Speaker Diarization

    1. Methods for Speaker Diarization (Schupp; Supervisor: Wilfried Michel)
      Presentation: Mon, Aug 10, 2020, 10h

      Initial References:

    2. Applications and Challenges for Speaker Diarization (NN; Supervisor: Wilfried Michel)
      Initial References:

  2. Speaker Separation


    1. Permutation-Invariant Training (NN; Supervisor: Tobias Menne)
      Initial References:

    2. Speaker-Dependent Speaker Separation (NN; Supervisor: Tobias Menne)
      Initial References:

  3. Speech Enhancement

    1. Speech Enhancement for Human Listeners (NN; Supervisor: Peter Vieting)
      Initial References:

    2. Speech Enhancement for ASR (Benaicha; Supervisor: Peter Vieting)
      Initial References:

  4. Natural Language Understanding

    1. Pre-training Language Representation Models for NLP (NN; Supervisor: David Thulke)
      Initial References:

    2. Joint Intent Classification and Slot Filling (NN; Supervisor: David Thulke)
      Initial References:

    3. Cross-lingual Word Embedding (Zhang; Supervisor: Yunsu Kim)
      Presentation: Mon, Aug 10, 2020, 11h

      Initial References:

    4. Cross-lingual Sentence Embedding (Scherer; Supervisor: Yunsu Kim)
      Presentation: Mon, Aug 10, 2020, 12h

      Initial References:

  5. Automatic Speech Recognition

    1. Multilingual ASR (Jain; Supervisor: Eugen Beck)
      Presentation: Tue, Aug 11, 2020, 10h

      Initial References:

    2. Domain Adaptation/Expansion (NN; Supervisor: Eugen Beck)
      Initial References:

  6. Language Identification

    1. State-of-the-Art Language Identification (Qamar; Supervisor: Markus Kitza)
      Presentation: Tue, Aug 11, 2020, 11h

      Initial References:

    2. Fusion based Native Language Identification (Sarwar; Supervisor: Markus Kitza)
      Presentation: Tue, Aug 11, 2020, 12h

      Initial References:

  7. Text-to-Speech

    1. Vocoders (NN; Supervisor: Yingbo Gao)
      Initial References:

    2. Neural Text-to-speech (Kumar; Supervisor: Nick Rossenbach)
      Presentation: Thur, Aug 13, 2020, 12h

      Initial References:


  8. Speech-to-Text Translation

    1. End-to-end Speech-to-text Translation (Chugh; Supervisor: Parnia Bahar)
      Presentation: Wed, Aug 12, 2020, 13h

      Initial References:


  9. Reinforcement Learning

    1. Minimum Expected Loss Training (Ali; Supervisor: Mohammad Zeineldeen)
      Presentation: Thur, Aug 13, 2020, 10h

      Initial References:

    2. Modern Policy Learning Methods for Games (NN; Supervisor: Mohammad Zeineldeen)
      Initial References:

    3. Memory Augmented Networks for Reinforcement Learning (NN; Supervisor: Christoph Lüscher)
      Initial References:

  10. Text Summarization

    1. Extractive Text Summarization (Ahmed; Supervisor: Jan Rosendahl)
      Presentation: Wed, Aug 12, 2020, 12h

      Initial References:

    2. Abstractive Text Summarization (with Deep Learning) (Balyan; Supervisor: Christian Herold)
      Presentation: Thur, Aug 13, 2020, 11h

      Initial References:

  11. Named Entity Recognition

    1. Named Entity Recognition (NN; Supervisor: Weiyue Wang)
      Initial References:


  12. Constituency Parsing

    1. Neural Network-based Parsing (Asghar; Supervisor: Parnia Bahar)
      Initial References:
      • A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, I. Polosukhin: "Attention Is All You Need," in Advances in Neural Information Processing Systems 30, Annual Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, Dec. 2017, arXiv:1706.03762.
      • C. Dyer, A. Kuncoro, M. Ballesteros, N. A. Smith: "Recurrent neural network grammars," in Proc. 15th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), San Diego, CA, Jun. 2016, arXiv:1602.03762.
      • M.-T. Luong, Q. V. Le, I. Sutskever, O. Vinyals, L. Kaiser: "Multi-task Sequence to Sequence Learning", in International Conference on Learning Representations (ICLR) , San Juan, Puerto Rico, May 2016, arXiv:1511.06114.
    2. Overview of Topics

        A1. Methods for Speaker Diarization
        A2. Applications and Challenges for Speaker Diarization
        B1. Permutation-Invariant Training
        B2. Speaker-Dependent Speaker Separation
        C1. Speech Enhancement for Human Listeners
        C2. Speech Enhancement for ASR
        D1. Pre-training Language Representation Models for NLP
        D2. Joint Intent Classification and Slot Filling
        D3. Cross-lingual Word Embedding
        D4. Cross-lingual Sentence Embedding
        E1. Multilingual ASR
        E2. Domain Adaptation/Expansion
        F1. State-of-the-Art Language Identification
        F2. Fusion based Native Language Identification
        G1. Vocoders
        G2. Neural Text-to-speech
        H1. End-to-end Speech-to-text Translation
        I1. Minimum Expected Loss Training
        I2. Modern Policy Learning Methods for Games
        I3. Memory Augmented Networks for Reinforcement Learning
        J1. Extractive Text Summarization
        J2. Abstractive Text Summarization (with Deep Learning)
        K1. Named Entity Recognition
        L1. Neural Network-based Parsing


    Article and Presentation Format

    The roughly 20-page article together with the slides (between 20 & 30) for the presentation should be prepared in LaTeX format. Presentations will consist of 30 to 40 minutes presentation time & 15 minutes discussion time. Document templates for both the article and the presentation slides are provided below along with links to LaTeX documentation available online. The article and the slides should be prepared in LaTeX format and submitted electronically in pdf format. Other formats will not be accepted.

    Detailed Guidelines:

    Some Tips:

    Time management is crucial for a successful seminar:
    Successful seminar articles/presentations typically:
    While reading papers, it might be useful to keep the following questions in mind:

    Contact

    Questions regarding the content of the assigned seminar topics should be directed to the respective topic's supervisors.

    General and administrative inquiries should be directed to:

    Parnia Bahar
    RWTH Aachen University
    Lehrstuhl Informatik 6
    Ahornstr. 55
    52074 Aachen

    Room 6125b
    Tel: 0241 80 21632

    E-Mail: bahar@cs.rwth-aachen.de