DSCode Comparator: An Interactive Interface for Comparing Models and Evaluating Code for Data Science Tasks

dc.contributor.authorYu, Xinxin
dc.date.accessioned2026-04-17T15:24:00Z
dc.date.available2026-04-17T15:24:00Z
dc.date.issued2026-04-17
dc.date.submitted2026-03-20
dc.description.abstractCode-generating models are increasingly used to support data science tasks. However, reviewing and validating their outputs remains largely manual and time-consuming, requiring users to understand how generated code works and to assess its quality and correctness. Rather than eliminating effort, these models often shift user work from writing code to verifying it. This challenge is further compounded by the fact that different models frequently produce diverse solutions with varying levels of effectiveness, making systematic comparison and evaluation difficult. To address these challenges, this thesis presents DSCode Comparator, an interactive system designed to support code understanding, evaluation, refinement, and comparison in data science workflows. The system enables users to examine code at multiple levels of granularity, ranging from individual lines of code to complete solutions across different prompts and tasks. DSCode Comparator incorporates an automated annotation pipeline that analyzes generated code and provides structured, line-level explanations to facilitate rapid comprehension. In addition, the system evaluates code quality along multiple functional and pragmatic dimensions, including efficiency, readability, usability, and resource usage. Beyond individual code inspection, DSCode Comparator supports comparative analysis across models by aggregating annotations and evaluation results into compact summaries that highlight key differences in behavior and performance. Through a combination of empirical evaluation and user studies with data science practitioners, this thesis demonstrates that the proposed approach improves users’ ability to understand, compare, and refine code generated by large language models, reducing verification effort while supporting more informed decision-making in model-assisted programming.
dc.identifier.urihttps://hdl.handle.net/10012/23013
dc.language.isoen
dc.pendingfalse
dc.publisherUniversity of Waterlooen
dc.subjectlarge language models
dc.subjectcode generation
dc.subjecthuman-ai interaction
dc.subjectcode evaluation
dc.subjectdata science workflows
dc.titleDSCode Comparator: An Interactive Interface for Comparing Models and Evaluating Code for Data Science Tasks
dc.typeMaster Thesis
uws-etd.degreeMaster of Mathematics
uws-etd.degree.departmentDavid R. Cheriton School of Computer Science
uws-etd.degree.disciplineComputer Science
uws-etd.degree.grantorUniversity of Waterlooen
uws-etd.embargo.terms0
uws.contributor.advisorCrisan, Anamaria
uws.contributor.affiliation1Faculty of Mathematics
uws.peerReviewStatusUnrevieweden
uws.published.cityWaterlooen
uws.published.countryCanadaen
uws.published.provinceOntarioen
uws.scholarLevelGraduateen
uws.typeOfResourceTexten

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Yu_Xinxin.pdf
Size:
18.18 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
6.4 KB
Format:
Item-specific license agreed upon to submission
Description:

Collections