Harnessing Generalist LLMs for Diverse Objective and Subjective NLP Tasks

dc.contributor.authorSahu, Gaurav
dc.date.accessioned2024-12-17T14:42:11Z
dc.date.available2024-12-17T14:42:11Z
dc.date.issued2024-12-17
dc.date.submitted2024-12-13
dc.description.abstractRecent advances in natural language processing (NLP), particularly in the subspace of large language modeling, have led to a major paradigm shift. Large language models (LLMs), like the GPT and LLaMA family of models, are trained on a massive Internet corpus covering data from a gamut of diverse domains. In addition, the billions of parameters in these models also invoke emergent capabilities in them, leading to strong improvements across diverse NLP tasks without much task-specific tuning; however, effectively harnessing the knowledge of these generalist models for real-world data still remains a major challenge as the LLMs can produce inconsistent, biased, and unsatisfactory outputs. In this thesis, we propose task-specific strategies for effectively leveraging LLMs for a number of challenging NLP tasks, such as (low-resource) text classification, text summarization, modeling artistic preferences of creative individuals, and automated data analysis. Our results suggest that LLMs can serve as excellent data generators and data labelers for well-defined single-step tasks like classification and summarization, crucially in data-scarce settings, where models trained on LLM-generated data achieved competitive performance to oracle models trained on a much larger labeled training data. On the other hand, for more subjective tasks like modeling artistic preferences among creative individuals, we demonstrate that while LLMs might not be able to discern between the likes and dislikes of artists, they can be effective in extracting key linguistic and poetic properties from text that can later be employed to infer artistic preferences among different individuals. Lastly, we also evaluate the effectiveness of LLMs in multi-step tasks that require the LLM to perform multiple tasks in tandem without compromising performance for individual tasks. Overall, our work draws critical insights into the strengths and shortcomings of LLMs for a wide range of subjective and objective NLP tasks and includes meaningful suggestions for the research community to harness LLMs for those tasks effectively.
dc.identifier.urihttps://hdl.handle.net/10012/21255
dc.language.isoen
dc.pendingfalse
dc.publisherUniversity of Waterlooen
dc.subjectlarge language models (LLMs)
dc.subjectnatural language processing (NLP)
dc.subjecttext classification
dc.subjectintent classification
dc.subjectfew-shot text classification
dc.subjecttext summarization
dc.subjectextractive text summarization
dc.subjectsemi-supervised text summarization
dc.subjectdata augmentation
dc.subjectzero-shot text classification
dc.subjectartistic preference modeling
dc.subjectLLM-based exploratory data analysis
dc.subjectabstractive text summarization
dc.subjectGPT
dc.subjectLLaMA-3
dc.subjectLLaMA-2
dc.subjectBERT
dc.subjectDistilBERT
dc.subjectDistilBART
dc.subjectPreSumm
dc.subjectPromptMix
dc.subjectMixSumm
dc.titleHarnessing Generalist LLMs for Diverse Objective and Subjective NLP Tasks
dc.typeDoctoral Thesis
uws-etd.degreeDoctor of Philosophy
uws-etd.degree.departmentDavid R. Cheriton School of Computer Science
uws-etd.degree.disciplineComputer Science
uws-etd.degree.grantorUniversity of Waterlooen
uws-etd.embargo.terms0
uws.contributor.advisorVechtomova, Olga
uws.contributor.affiliation1Faculty of Mathematics
uws.peerReviewStatusUnrevieweden
uws.published.cityWaterlooen
uws.published.countryCanadaen
uws.published.provinceOntarioen
uws.scholarLevelGraduateen
uws.typeOfResourceTexten

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Sahu_Gaurav.pdf
Size:
8.38 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
6.4 KB
Format:
Item-specific license agreed upon to submission
Description: