Παρακαλώ χρησιμοποιήστε αυτό το αναγνωριστικό για να παραπέμψετε ή να δημιουργήσετε σύνδεσμο προς αυτό το τεκμήριο:
http://artemis.cslab.ece.ntua.gr:8080/jspui/handle/123456789/19679
Τίτλος: | Open-Domain Dialogue Systems for Low-Resource Languages: The case of Greek |
Συγγραφείς: | Κούκουνας, Ανδρέας Ποταμιάνος Αλέξανδρος |
Λέξεις κλειδιά: | Οpen-domain dialogue generation, low-resource languages, transformers, BERT, GPT-2, mT5, XGLM, cross-lingual transfer learning, multitask learning, prompt learning, human evaluation |
Ημερομηνία έκδοσης: | 1-Ιου-2025 |
Περίληψη: | Human-machine conversation has been a critical and challenging task in AI and NLP. Recent years have seen rapid progress in open-domain dialogue generation. However, because vast conversational data are only available in English, the development of generation-based chatbots for non-English languages has lagged behind. This Diploma Thesis studies dialogue generation in a low-resource language, specifically Greek, where training data and pre-trained language models are limited. We present theoretical background on machine learning (ML), deep learning (DL), and natural language processing (NLP), then study the transformer-based models used: BERT, GPT-2, T5 and XGLM. We analyze how the corresponding Greek (GREEK-BERT and GPT- 2 Greek) or multilingual (mT5) models were created. Following analysis of research on low-resource dialogue generation, we conduct experiments and discuss results. To address the lack of a Greek dialogue dataset, we used machine translation (MT) to create a Greek version of the DailyDialog dataset. We fine-tune Greek monolingual models (GREEK-BERT and GPT-2 Greek) on the translated dataset, then conduct 4 experiments with multilingual models mT5 and XGLM: 1. Native training: Fine-tuned multilingual models exclusively on the translated dataset for direct comparison with monolingual models. 2. Cross-Lingual transfer learning: Fine-tuned models using the original English DailyDialog dataset, then further fine-tuned on limited manually translated Greek examples. 3. Multitask Learning: Trained models simultaneously on both languages, utilizing the complete English dataset alongside a subset of the translated Greek dataset. 4. Prompt based Learning: Enhanced both approaches with specific prompt templates shared across languages to facilitate knowledge transfer from English to Greek dialogues. We evaluated all models using multiple metrics: Perplexity, BLEU, and BertScore for response quality, and Distinct-n for lexical diversity. Results demonstrate that native training achieved superior performance, with GPT2-Greek as the best-performing model (perplexity:12.47, BLEU B-1: 25.93, Distinct-1: 23.13%, BertScore F-1: 71.37%). Among multilingual approaches, prompt-based training significantly enhanced XGLM performance (F-1: 69.12%), while multitask learning consistently outperformed cross-lingual transfer learning. Human evaluations assessed qualitative aspects that automated metrics might not capture. These revealed that our XGLM model trained using prompt-based multitask learning achieved the best performance among our approaches, ranking second only to the much larger Meltemi model trained on substantially more Greek data. This demonstrates effective cross-linguistic knowledge transfer despite using considerably less Greek training data. This thesis opens new avenues for exploring open-domain dialogue generation for low-resource languages and proposes future extensions for further research. |
URI: | http://artemis.cslab.ece.ntua.gr:8080/jspui/handle/123456789/19679 |
Εμφανίζεται στις συλλογές: | Διπλωματικές Εργασίες - Theses |
Αρχεία σε αυτό το τεκμήριο:
Αρχείο | Περιγραφή | Μέγεθος | Μορφότυπος | |
---|---|---|---|---|
andreas-koukounas-thesis-final.pdf | 3.52 MB | Adobe PDF | Εμφάνιση/Άνοιγμα |
Όλα τα τεκμήρια του δικτυακού τόπου προστατεύονται από πνευματικά δικαιώματα.