Disputation: Yongchao Wu
Disputation
Datum: måndag 16 december 2024
Tid: 09.00 – 12.00
Plats: L30, DSV, Borgarfjordsgatan 12, Kista
Välkommen till en disputation på DSV! Yongchao Wu utforskar i sin doktorsavhandling hur stora språkmodeller och AI kan användas i undervisning.
16 december 2024 presenterar Yongchao Wu sin doktorsavhandling på Institutionen för data- och systemvetenskap (DSV) vid Stockholms universitet. Titeln är ”Exploring the Educational Utility of Pretrained Language Models”.
Doktorand: Yongchao Wu, DSV
Opponent: Filip Ginter, University of Turku, Finland
Huvudhandledare: Aron Henriksson, DSV
Handledare: Jalal Nouri och Martin Duneld, DSV
Ladda ner avhandlingen från Diva
Kontaktuppgifter till Yongchao Wu
Disputationen genomförs i DSVs lokaler i Kista, med start klockan 09.00.
Hitta till DSV
Sammanfattning på engelska
The emergence of pretrained language models has profoundly reshaped natural language processing, serving as foundation models for a wide range of tasks. Over the past decade, pretrained language models have evolved significantly, leading to the development of different types of models and approaches for utilising them. This progression spans from static to contextual models and from smaller models to more powerful, generative large language models. The increasing capabilities of these models have, in turn, led to growing interest in exploring new use cases and applications across various domains, including education, where digitalisation has created opportunities for AI applications that leverage pretrained language models, particularly due to the abundance of text data in educational contexts.
This thesis explores the educational utility of pretrained language models, specifically by investigating how different paradigms of these models can be applied to address tasks in education. These paradigms include various methodologies for leveraging the knowledge embedded in pretrained language models, such as embeddings, fine-tuning, prompt-based learning, and in-context learning.
For collaborative learning group formation, a clustering approach based on pretrained embeddings is proposed, enabling the creation of either homogeneous or heterogeneous groups depending on the specific learning situation. For automated essay scoring, a pretrained language model is fine-tuned using both the essay instructions and the essay text as input; the proposed method also highlights key topical sentences that contribute to the predicted essay score. For educational question generation, a method based on prompt-based learning is introduced and shown to be more data-efficient than existing methods. Finally, for educational question answering, certain limitations of the in-context learning (or prompting) paradigm, such as a tendency of large language models to hallucinate or miscalculate, are addressed.
Specifically, workflows and prompting strategies based on retrieval-augmented generation and tool-augmented generation are proposed, allowing large language models to ground answers in specific learning materials and to leverage external tools, such as calculators and knowledge bases, within chain-of-thought reasoning processes. These strategies are shown to produce more reliable and transparent answers to complex questions.
Through five empirical studies, methodological innovations within each paradigm of pretrained language models are proposed and evaluated for specific educational use cases. In addition to contributing methodologically to natural language processing, the results demonstrate the potential utility of pretrained language models in educational AI applications, thereby advancing the field of technology enhanced learning. The proposed methods not only improve predictive performance on specific tasks but also aim to enhance the transparency of pretrained language models, which is essential for building reliable and trustworthy educational AI applications.
Nyckelord
Natural Language Processing, Technology Enhanced Learning, Pretrained Language Models, Large Language Models, Generative AI, Collaborative Learning, Automated Essay Scoring, Educational Question Generation, Educational Question Answering
Senast uppdaterad: 21 november 2024
Sidansvarig: Institutionen för data- och systemvetenskap, DSV