Thesis

Interested in writing your MSc or BSc thesis with us? We offer several BSc and MSc thesis topics at MaiNLP.

Currently, the following research vectors characterize broad topics in which we offer MSc and BSc thesis projects. We provide a (non-exhaustive) list of research projects within each vector. We are also interested to supervise projects related to our research projects. You are welcome to send us your own project proposal. We recommend to check out the suggested/selected publications to get inspired.

Unless otherwise specified, all projects can be either at the MSc or BSc level. The exact scope of the project will be determined in a meeting before the start of the project.

Important note: We currently do not supervise industrial MSc/BSc thesis projects (Industrieabschlussarbeiten).

Regularly check back here for updates on thesis project suggestions.

News:

  • 2026, Jan 15: MSc/BSc project applications deadlines posted. New topics currently under development, stay tuned!

Legend:

  • :hourglass_flowing_sand: reserved
  • strikethrough topic no longer available

How to apply for a BSc and MSc thesis project

Important information for LMU students: You need to apply for a MSc/BSc thesis project the latest three weeks before the thesis project registration date.

Deadlines for the summer semester 2025-2026:

  • MSc students apply before Monday February 9, 2026
  • BSc students apply before Monday February 16, 2026

To apply, please send your application material with subject “[BSc (or MSc) thesis project at MaiNLP - inquiry [Name and which semester]” to: [email protected]

It should contain a single pdf with the following information:

  • CV, your study program, full grade transcript
  • Level: BSc or MSc thesis project
  • Which theme or project interests you (optional: we are open to project proposals related to the research vectors or on-going research projects). If you are interested in multiple, list your preferences for up to four (ranked list: first priority, second priority, third priority, fourth priority)
  • Languages you speak
  • Your favorite project so far, and why
  • Your knowledge and interest in data annotation, data analysis and machine learning/deep learning (including which toolkits you are familiar with)
  • Whether you have access to GPU resources (and which)
  • A term project report or your BSc thesis if you apply for a MSc thesis (optional)

Reach out if you have questions, using the email above.

MSc/BSc thesis research vectors:

V1: NLP for Dialects, Low-resource and Multilinguality

This research vector covers methods and resources for processing dialectal, low-resource, and multilingual language data. It focuses on improving robustness, fairness, and coverage of NLP systems across languages and varieties, including cross-lingual transfer and data-efficient learning.

Thesis projects

  • Computational Dialectology. Language usage often differs based on sociodemographic background, where linguistic differences based on the geographical origin of the speaker are typically studied in the field of dialectology. While qualitative studies into dialectal differences have yielded valuable insights into language variation, such studies often rely on labor-intensive data collection, annotation, and analysis. As such, computational approaches to dialect differences have emerged as a possible method towards the large-scale study of dialects. For students interested in this project, multiple directions are possible, including (but not limited to): (a) interpretability of what features dialect models rely on for differentiation, (b) creation of (parallel) resources for dialect continua, (c) development of new methods to quantify dialectal or sociolinguistic variation, (d) adapting existing models to better accommodate dialect variation. References: Bartelds & Wieling 2022, Bafna et al. 2025, Shim et al. 2026. Level: BSc or MSc.

  • Methods for mining low-resource parallel corpora. Parallel corpora are critical for developing and evaluating dedicated machine translation systems, as well as general-purpose large language models capable of translation. One strategy for obtaining such corpora is to mine unstructured text corpora (typically web crawls) for parallel sentences. However, standard methods typically score candidate sentences via cosine distance of their sentence embeddings, a method which requires strong sentence encoders. Such sentence encoders are typically weaker for very low-resource languages, including language varieties such as dialects. Strategies include: bootstrapping, building classifiers, coming up with simple heuristics such as word-edit distance, or relying on meta-data like HTML tags. Depending on the student’s interest and academic level, this project can focus more or less on specific directions such as: evaluating the impact of different methods, methods for scoring candidate sentences, or strategies for obtaining candidate sentences. References: Improving Parallel Sentence Mining for Low-Resource and Endangered Languages - ACL Anthology, Obtaining Parallel Sentences in Low-Resource Language Pairs with Minimal Supervision - PMC Level: BSc or MSc.

  • Synthetic language variation for robust NLP. Robust NLP entails models that can process human language variation, such as dialects and other language varieties. These varieties are typically characterized by high variation due to their orthography, lexicography, syntax, each of which present challenges to NLP. Furthermore, these varieties are typically low-resourced, such that we widely rely on transfer from standard language data to language varieties in building NLP models. One strategy for improving robustness to linguistic variation is to introduce synthetic variation. This can range from naive perturbation of characters in order to induce more varied tokenization of standard training data, to targeted de-standardization of training data. References: Improving Zero-Shot Cross-lingual Transfer Between Closely Related Languages by Injecting Character-Level Noise - ACL Anthology Neural, Text Normalization for Luxembourgish Using Real-Life Variation Data - ACL Anthology Level: BSc or MSc (scope adjusted by languages covered, and complexity of the approaches).

  • Universal Dependencies for Underrepresented Language Varieties. Universal Dependencies (UD) provides a cross-linguistically consistent framework for syntactic annotation, yet many language varieties and genres remain underrepresented. This thesis explores how existing UD annotations for a related standard language can be adapted to support annotation of a lower-resource variety or underrepresented genre (e.g., a regional variety, informal writing, or domain-specific text). The project will involve selecting a small corpus, analyzing where standard-language annotation guidelines break down, and producing a limited but linguistically principled UD annotation or conversion strategy. Additionally, one would analyze annotation challenges and experiment with downstream UD parsing. Level: BSc or MSc (scope adjusted by dataset size and depth of linguistic analysis and experiments).

  • Cross-lingual Semantic Disagreement in Human Annotations. Human judgments about meaning often diverge across languages. This thesis investigates such cross-lingual semantic disagreement by comparing annotations produced independently in two languages spoken by the student (e.g., German–English, Mandarin–English, Russian–German). One could choose to work on NLP tasks such as natural language inference, sentiment or stance classification, discourse relation labeling, etc. The project will involve selecting a small parallel or comparable dataset (or conducting the translation yourself), designing a controlled annotation setup, and analyzing where and why label distributions diverge across languages. The goal is not to resolve disagreements, but to characterize systematic patterns of variation and relate them to linguistic differences such as tense, aspect, definiteness, evidentiality, or pragmatic conventions. Level: BSc or MSc (scope adjusted by dataset size and depth of linguistic analysis and experiments).

  • Dialect Variation Dictionaries and Evaluation. A large amount of culture-specific knowledge is captured in dialect corpora. However, dialects pose a challenge due to their limited resources and high variability, with words frequently spelled differently, reflecting regional pronunciations. This thesis extends our previous research on the Bavarian dialect (Litschko et al., 2025) by exploring another dialect. The project consists of two main components: 1) Annotating pairs of German and dialect words to determine whether they are equivalent translations. A dataset will be provided for this purpose. 2a) Evaluating the ability of LLMs to "understand" the dialect through translation and word-pair classification tasks, or 2b) building and evaluating a rule-based lexical normalization model (Millour and Fort, 2019; Weissweiler and Fraser, 2018). To successfully complete this thesis, it is essential that the student has a strong understanding of one of the following dialects: Alemannic (Alemannisch), Palatinate (Pfälzisch), North Frisian (Nordfriesisch), Saterland Frisian (Saterfriesisch), Low German (Niederdeutsch), Colognian (Kölsch). Dialects of other languages can potentially also be considered, please reach out to us. Level: BSc or MSc (adjusted scope includes dialect-specific model adaptation).

V2: NLP Applications, Summarization and Information Retrieval

This theme covers applied NLP methods for impactful real-world domains (e.g., climate change, labor markets, education) and core applications such as summarization, information retrieval.

Thesis projects

  • NLP for Computational Job Market Analysis. Job postings and career paths are a rich resource to understand the dynamics of the labor market. For example, to track how skills demands change, how career paths are affected and how educational demands may shift. Such changes have large social implications, as they can inform strategic long-term decisions of governments to react to changing structural demands in the labor force. Recently, the emerging line of work on computational job market analysis (also known as NLP for human resources) has started to provide data resources and models for automatic job market analysis, such as the identification and extraction of skills in job postings, or the prediction of career paths. For students interested in real-world applications, this theme provides multiple thesis projects for two application domains: i) understanding skills in job postings (e.g. cross-lingual or cross-domain skill and knowledge extraction to data sources like job postings, patents or scientific articles) See references of MultiSkill project. See also Bhola et al., 2020, Gnehm et al. 2021, our own ESCOXLM-R model or ii) career path prediction, which is the task of predicting a person’s next occupation based on their resume. See the Karrierewege paper. Level: BSc or MSc.

  • Climate Change Insights through NLP. Climate change is a pressing global issue that is receiving more and more attention. It is influencing regulations and decision-making in various parts of society such as politics, agriculture, business, and it is discussed extensively on social media. For students interested in real-world societal applications, this project aims to contribute insights on the discussion surrounding climate change. Example projects: Analyzing social media data. The data will have to be collected (potentially from existing sources), cleaned, and analyzed using NLP techniques to examine various aspects or features of interest such as stance, sentiment, the extraction of key players, etc. References: Luo et al., 2020, Stede & Patz, 2021, Vaid et al., 2022. Level: BSc or MSc.

  • Multi agent debate for summarization or simplification. Automatic summarization (or simplification) is often performed in an end-to-end manner, using a single model (e.g., an LLM). Recent work on multi-agent systems suggests that “interaction” between LLMs can improve reasoning and reduce errors. This project explores whether multi-agent debate can improve summarization (or simplification) quality by having agents summarize the same document, critique each other’s summaries, and provide a final version. The project might investigate, e.g., the effect of prompting different agents with different priorities (factuality, conciseness, etc), whether debate improves performance compared to single-pass summaries and/or whether certain aggregation strategies (rounds of critique, voting, consensus-building) outperform others. References: Du et al, ICLR 2024; Koupaee et al, NAACL 2025; Wan et al., NAACL 2025 Level: MSc (preferred); adaptation to BSc is also possible. Other projects in summarization or simplification (e.g., resource building, multilinguality) are also possible depending on the student interests.

  • Improving the Cross-Lingual Alignment of IR Models. Machine Translation (MT) and Cross-Lingual Information Retrieval (CLIR) are two interconnected Natural Language Processing (NLP) tasks. In CLIR, MT is typically used to translate queries at retrieval time (translate test) or to translate training data (translate train). Recent studies have proposed a novel approach: aligning the representation spaces of MT models with those of large language models to improve their performance in multilingual NLP tasks (Acharya et al., 2025; Schmidt et al., 2024; Yoon et al., 2024). The goal of this project is to investigate the effectiveness of aligning internal representations within LLM-based IR models, and compare their performance to translation-based methods for CLIR across both high- and low-resource languages. Depending on the student’s interests, the focus of this project can be on cross-lingual retrieval or reranking. Level: MSc.

  • Evaluating the quality and difficulty of exam questions. When high-stakes educational assessments (like university entrance exams) are being developed, they usually need to be piloted with a large sample of test takers to make sure that the questions are of appropriate quality and difficulty. Recently, researchers have tried to use LLMs to predict these properties automatically, either directly or by simulating test takers at various levels of ability. However, out-of-the box LLMs are not very good at this. A student project could experiment with fine-tuning LLMs to match different ability levels or to directly evaluate properties such as item difficulty, discrimination, or guessability. References: Yaneva et al. (2024), Acquaye et al (2025), Liu et al. (2025), Gorgun & Bulut (2024), Laverghetta et al. (2022) Level: MSc.

  • Medical NLP. NLP studies on medical texts often rely on synthetic data due to privacy and accessibility constraints. If you are interested in working in this domain, please check this entry again in a couple of days for more information. Level: MSc.

V3: Natural Language Understanding, Semantic & Pragmatics, Computational Social Science

This theme covers aspects of uncertainty and ambiguity in human language and perception, including human label variation, subjective judgments, and cognitive factors affecting interpretation. It explores how NLP systems can model, reason about, and interact with uncertainty across text and multimodal settings, or how gaze data can inform modeling or evaluation, with an emphasis on human-centered evaluation and design.

Thesis projects

  • Characterizing Ambiguity and Errors in Open-Domain QA Datasets. Benchmark datasets are central to progress in open-domain question answering (QA), yet they often contain unresolved ambiguities or annotation errors (cf. Klie et al. 2023, Weber-Genzel et al. 2024) that affect both model training and evaluation. This project focuses on a systematic data analysis of open-QA benchmarks like AmbigQA to characterize common sources of ambiguity (e.g., entity, event, and temporal ambiguity or underspecification, see for example Tang et al., 2025) and may identify possible annotation errors in existing QA datasets. BSc-level: Using exploratory data analysis, annotation and predictive modeling with lightweight traditional interpretable classifiers, to investigate whether surface-level features can be used to identify and predict different ambiguity or error types. The outcome is a taxonomy of ambiguity and errors in open-QA datasets, empirical insights into how frequently these phenomena occur, and a prototype classifier to identify the ambiguity categories. MSc-level extensions include using LLMs for ambiguity characterization and generation and analysis of multiple interpretations (Saparina & Lapata, 2025), analyze LLM model output behavior (e.g. like in language-vision Testoni et al., 2025 or text-based LLMs Sedova et al, 2025), or alternatively, connect ambiguity characterizations identified in the first step to model-internal mechanisms like uncertainty or probing. The outcome is a taxonomy of ambiguity and errors in open-QA datasets, and an evaluation of how LLM output behavior (types of answer) or model-internal mechanisms relate to input ambiguity. Level: BSc or MSc.

  • Interpreting Visual Ambiguity: Humans and Vision-Language Models. The objective of this thesis is to investigate ambiguity in images and to compare how humans and Vision-Language Models perceive and resolve such ambiguity. Ambiguous images often contain multiple salient elements or support multiple plausible interpretations. While humans show attention shifts depending on context, expectations, and individual differences, what models do? Depending on the student’s interests, this thesis can focus more on modeling approaches or on human strategies for processing ambiguity (e.g., through behavioral or eye-tracking experiments). References: Hindennach et al., 2024, Testoni et al., 2025 Level: MSc.

  • Beyond Probability Metrics: Evaluating Free-Form Rationales for Disagreement: Current approaches to Human Label Variation (HLV) mostly treat the problem as a distribution-matching task. We evaluate models using metrics like KL-divergence or Jensen-Shannon Distance to see if the predicted probabilities align with the crowd. However, getting the numbers right doesn’t mean the model understands the disagreement. A model might predict a 60/40 split for the wrong reasons, or hallucinate a conflict where none exists. As we move towards "Glass-Box" NLP, we need models that can explicitly justify why a sample is ambiguous through Explanations (Chen et al. 2025a) or Chain-of-Thought (CoT, Chen et al. 2025b). While we have metrics for label distributions, we lack a robust framework for evaluating the quality of the reasoning behind the variation. How do we judge if a free-text explanation accurately captures a pragmatic ambiguity versus a semantic one? Does the generated CoT truly reflect the linguistic nuance that causes humans to disagree? This thesis aims to design an evaluation framework for Free-Form HLV Rationales. The student will move beyond standard overlaps (like BLEU/ROUGE) and develop metrics—potentially LLM-based or taxonomy-guided (e.g., using LITEX, Hong et al. 2025a)—to assess: 1. Faithfulness: Does the explanation actually align with the predicted label distribution? 2. Coverage: Does the CoT capture all valid perspectives (the "Yes" view and the "No" view) or does it collapse into a single viewpoint? 3. Linguistic Validity: Can we quantify the "quality" of the ambiguity detection? Level: MSc.

  • Ambiguity or Preference? Disentangling Variation Sources via Personalized Modeling: When humans disagree on an NLP task, it usually stems from two distinct sources (Hong, 2025b): Linguistic Ambiguity (the text itself is unclear/vague) or Annotator Preference (the human has specific biases, background knowledge, or subjectivity). Current HLV models tend to mix these into a single "noise" distribution. However, true Human-Centric NLU requires a model to know who is speaking. A sentence might be ambiguous to everyone (global uncertainty), or it might only be interpreted differently by people with specific cultural backgrounds (local preference). We currently struggle to separate "what is in the text" from "what is in the head." Can we simulate specific human Personas to model preference? If we condition a model on a specific persona, does the variation disappear (implying the variation was preference-based) or persist (implying the variation is intrinsic ambiguity)? This thesis explores Personalized HLV to disentangle ambiguity from preference. The student will investigate: 1. Persona Injection: Can we prompt LLMs with specific annotator profiles (e.g., "skeptical linguist" vs. "imaginative writer") to replicate specific human biases? (Sorensen et al. 2025,Li et al. 2025) 2. Decomposition: By measuring how much the output distribution shifts when the persona changes, can we quantify the "Subjectivity Score" of a dataset? 3. Disentanglement: The goal is to propose a method that separates instances where clarification is needed (ambiguity) from instances where personalization is needed (preference). Level: MSc.

  • Aggregating Individual Opinions through Discourse: Human vs. Chain-of-Thought Reasoning. Public discourse often consists of multiple, partially conflicting individual opinions that may be synthesized into a coherent collective interpretation. This thesis investigates how humans and language models differ in aggregating such opinions, with a particular focus on the role of discourse structure and reasoning strategies. These opinion pieces could come from different news agencies, Wikipedia articles, political debates, discussion forums, etc. The student will examine whether humans and LLMs preserve disagreement (e.g., minority viewpoints) and how they build argumentative structure. The study is well-suited for students interested in discourse analysis, evaluation of language models, and the representation of multiple voices. Level: MSc.

  • Synthetic data for metrics meta evaluation. Automatic metrics (including LLMs-as-judges) are typically meta-evaluated by measuring their correlation with human scores or preferences. However, this evaluation is often global (making it difficult to diagnose when and why a metric fails), and human judgments are expensive and complex to collect. An alternative is behavioural testing, e.g., checklist-like approaches (Ribeiro et al., ACL 2020), where targeted perturbations are designed to probe sensitivity to specific phenomena. This approach uses challenge sets to better understand metrics failure modes and test for specific biases. The goal of this project is to explore methods for automatically generating synthetic test sets for metric meta-evaluation, to assess their validity and limitations, and to use them to systematically benchmark evaluation metrics (including LLM-based judges). Depending on student interests, the project may focus on multilingual settings, specific evaluation dimensions (e.g., factuality or societal bias), robustness to perturbations, or other task- or domain-specific aspects. References include: Sai et al., EMNLP 2021, Ye et al., ICLR 2025. Level: BSc or MSc.

  • Pseudoword generation. Pseudowords are words that look and sound like they could exist in a particular language, but don’t actually have any meaning. Pseudowords are frequently used in (psycho)linguistic studies to investigate how humans learn and process language (e.g., lexical decision). Often, pseudowords need to fulfill certain criteria, e.g., they should appear like a specific part of speech. However, it is quite difficult to come up with good pseudowords manually. Previous approaches have often been based on phonotactic templates or Markov chains. Neural networks also have the potential to work well. A student project could involve implementing and evaluating a new pseudoword generator for a less-studied language. Depending on interest and available resources, the approach could be rule-based, or based on statistical or neural models. References: Keuleers & Brysbaert (2010), New et al. (2024) Level: BSc or MSc.

  • Interpreting Irony in Multimodal Memes. Internet memes are a widely used form of online communication (Shifman, 2013) and often express irony, where the literal meaning of the text or image differs from the intended message. This mismatch makes automatic meme understanding difficult, as multimodal large language models (MLLMs) may assign incorrect sentiment or intent labels when they rely on surface-level cues (Fersini et al., 2022, Nguyen et al., 2025). Previous research (Ilic et al., 2018) has studied irony and sarcasm in text and has addressed meme classification tasks (Kiela et al., 2020, Liu et al., 2022), but the specific ways in which irony arises from text, images, or their interaction are not yet fully understood. The goal of this thesis is to analyze how irony is constructed in multimodal memes and to evaluate how MLLMs detect and explain ironic meaning. Possible research directions include defining or refining a taxonomy of irony in memes, annotating or analyzing existing meme datasets, and assessing model performance in irony detection and explanation, including consistency between predicted labels and generated explanations as well as common error patterns. Level: BSc, adaptation to MSc possible.

  • Gaze data for NLP. The way in which our eyes move when reading a text can tell us a lot about the cognitive processes required for language understanding. For example, longer reading times indicate higher processing difficulty. In the past 10 years, a line of research has emerged that attempts to use gaze data obtained by eye tracking to improve NLP models for various tasks (e.g., Barrett et al., 2016, Hollenstein & Zhang, 2019, Deng et al., 2023, Alaçam et al., 2024). A student project could involve investigating under which circumstances and for which NLP tasks gaze data can be beneficial, or whether we can achieve the same effect with artificially synthesized gaze data. References: Hollenstein et al. (2019), Sood et al. (2020), Khurana et al. (2023), Bolliger et al. (2023). Level: MSc.

  • Tracking eye movements during annotation. Data annotation in NLP often involves subjective judgements or depends on how the annotators understand a given text. Eye tracking can be used to measure which parts of a text people pay most attention to. We can use this information to try to explain why different annotators choose different labels for the same text, which can also inform NLP models. A student project could involve collecting eye-tracking data for an annotation task where humans are known to disagree sometimes, and/or using existing eye-tracking data to analyze and explain annotation behavior. References: Alaçam et al. (2024), Joshi et al. (2014) Level: MSc, BSc possible (using existing gaze data).

  • (Eye-)Tracking the processing of visual representations of abstract concepts. Visual representations of concrete concepts (e.g., banana) tend to be more consistent across images and interpretations than those of abstract concepts (e.g., joy), which show a lot of variability in both visual representations and semantic interpretations. This thesis investigates how humans process visual representations of abstract concepts using eye tracking techniques. By analyzing gaze patterns and attention distributions, this project aims to highlight various processing strategies associated with abstractness. References: Tater et al. 2024, Tater et al. 2025, McRae et al. 2018 Level: MSc.

V4: Understanding Complex Model Behavior (LLMs/VLMs, Agents)

Thesis projects