Seminar "Large Scale Language Models and Generative Pretrained Transformers"

In the Summer Semester 2023, the Lehrstuhl Informatik 6 will host a seminar entitled "Large Scale language models and Generative pretrained transformers" for the Bachelor level.

Registration for the seminar

Registration for the seminar is only possible online via the central registration page.

Prerequisites for Participation in the Seminar

General Goals of the Seminar

The goal of the seminar is to autonomously acquire knowledge and critical comprehension of an assigned topic, and present this topic both in writing and verbally.

This includes:

Seminar Format and Important Dates

The seminar will be started with a kick-off meeting, which will take place on 17.03.2023. Deatils are communicated directly to the seminar participants selected in the central registration.

Please note the following deadlines during the seminar:

Note: failure to comply with the ethical guidelines, failure to meet deadlines, absence without permission from compulsory sessions (presentations and preliminary meeting as announced by email to each participating student), or dropping out of the seminar after more than 3 weeks after the preliminary meeting/topic distribution results in the grade 5.0/not appeared.

The deadline for de-registration from the seminar is TBA, i.e. within three weeks after the distribution of the topics. After this deadline, seminar participation is confirmed and will be graded.



Topics, Initial References Defining the Topics, Participants, and Supervisors

In general, selected topics from the following general areas of Human Language Technology and Machine Learning will be offered: Below, you find exemplary topics. The final topics will be presented in a kick-off meeting which will be announced to the seminar participants selected in the central registration .
  1. Principles of Language Modeling (1): Count-Based and Continous Space (Student: Khan, Supervisor: Raissi)
    Initial References:

    This topic will look into background of LM before the recurrent based approaches.

  2. Principles of Language Modeling (2): Recurrent NN based Language Models (Students: Pletschko, Supervisor: Raissi)
    Initial References:

    In this topic the student will look into principles of statistical language modeling using recurrent neural networks. Compared to count-based n-gram LMs or feed-forward networks with fixed-length context, recurrent neural network can capture long-time dependencies from past and used that for the prediction of the next word. Moreover LSTM based LMs address the well-known vanishing gradient problem

  3. Extractive Question Answering (Student: Vogelbacher, Supervisor: Rossenbach)
    Initial References:

    Given a question and a paragraph of text, the task in extractive question answering is to highlight the answer to the question in the paragraph. The goal of this seminar is to give an overview of different approaches to solve this task starting with task specific architectures up to current approach utilising large language models.

  4. Prompt Engineering (Student: Diepers, Supervisor: Thulke)
    Initial References:

    When it comes to adapting pretrained transformer models to downstream tasks, the conventional approach involves fine-tuning all model parameters. However, a surprisingly effective alternative is called prompting. This technique involves adapting the model by inputting a task description, optionally along with a set of examples, while keeping all parameters frozen. The purpose of this topic is to provide an overview of the advantages and disadvantages of this method, as well as explore potential extensions.

  5. Large Language Models for Machine Translation (Students: Lu, Supervisor: Gao)
    Initial References:

    This topic discusses different approaches to how large language models can be utilised for Machine Translation (MT). This can cover pre-training individual components or zero-/few-shot prompting of very large language models.



Article and Presentation Format

The roughly 20-page article together with the slides (between 20 & 30) for the presentation should be prepared in LaTeX format. Presentations will consist of 30 to 40 minutes presentation time & 15 minutes discussion time. Document templates for both the article and the presentation slides are provided below along with links to LaTeX documentation available online. The article and the slides should be prepared in LaTeX format and submitted electronically in pdf format. Other formats will not be accepted.

Detailed Guidelines:

Some Tips:

Time management is crucial for a successful seminar:
Successful seminar articles/presentations typically:
While reading papers, it might be useful to keep the following questions in mind:

Contact

Questions regarding the content of the assigned seminar topics should be directed to the respective topic's supervisors.

General and administrative inquiries should be directed to:

Tina Raissi
RWTH Aachen University
Lehrstuhl Informatik 6
Theaterstrasse 35-39
52074 Aachen

Room 025
Tel: 0241 80 21630

E-Mail: raissi@cs.rwth-aachen.de