Advanced Natural Language Processing

M.Sc course, University of Debrecen, Department of Data Science and Visualization, 2024

This course delves into advanced concepts of Natural Language Processing (NLP) and Machine Learning (ML) with a strong focus on modern deep learning techniques. It covers foundational topics such as tokenization, text representation, and pipelines, as well as cutting-edge research in large language models (LLMs), transformers, and their applications. The course emphasizes both theoretical understanding and practical implementation, preparing students to tackle real-world NLP challenges, including security, privacy, and human-centered design. During the semester, students will also have the opportunity to test and train these architectures on real data using cloud-based services (Google Collab).

======

Requirements

  • Attendance sheet: Fewer absences than allowed. Active participation in classes.
  • Create a working application, solve a real problem, and present it as a video using the solutions and models learned in class.
    • It must be uploaded to Github and shared.
    • Maximum length of video is 5-10 minutes.
    • In the video, each creator must present their own contribution. (for 3-8 minutes)
    • The application must be shown in action at the end of the video. (for 1-2 minutes)
  • Organizing into teams (2-4 people) or working individually.
  • If the creator(s) uses a service based on a generative language model to complete the task, they must attach the prompt log to the completed project as additional material.
  • It is not certain that the team members receive a uniform grade, but they get grades proportionate to the task they have completed in the project.
  • Submission deadline: 2025.05.31
  • Submission form

Lecture

Labor

Submitted

  1. Jurafsky, Daniel, and James H. Martin. “Speech and language processing (draft).” Chapter A: Hidden Markov Models (Draft of September 11, 2018). Retrieved March 19 (2018): 2019.
  2. Eisenstein, Jacob. “Introduction to natural language processing.” MIT press, 2019.
  3. Goldberg, Yoav. “A primer on neural network models for natural language processing.” Journal of Artificial Intelligence Research 57 (2016): 345-420.
  4. Francois Chollet. “Deep Learning with Python”
  5. Hugging Face NLP Course
  6. MIT Introduction to Deep Learning
  7. Visual Guide to Transformer Neural Networks - (Episode 1)
  8. Visual Guide to Transformer Neural Networks - (Episode 2)
  9. Visual Guide to Transformer Neural Networks - (Episode 3)

Key Words

  1. Tokenization
    • Byte-Pair Encoding (BPE)
    • Byte-level BPE
    • WordLevel
    • WordPiece
    • Unigram
    • SentencePiece
  2. Embbeding
    • Skip-Gram
    • CBOW
    • GLOVE
    • Word2Vec
  3. Position Embedding
  4. (Multi-Head) Attention
  5. Neural Network (Feed Foward layer)
  6. Normalization
  7. Transoformer
  8. Pre-Trained
  9. Large Language Model
  10. NLP Tasks
    • Summarization
    • Translate
    • Generation
    • Q&A
    • Named Entity Recognition
    • Sentiment analysis
  11. Multimodal architectures
  12. Huggingface
  13. Keras
  14. Tensorflow
  15. Pytorch
  16. Python
  17. Pipline
  18. Notebook
  19. Google Colab

Usefull Publications

[1] Attention Is All You Need

[2] Improving Language Understanding by Generative Pre-Training

[3] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

[4] Efficient Estimation of Word Representations in Vector Space

[5] Global Vectors for Node Representations