Quantitative Evaluation of using Large Language Models and Retrieval-Augmented Generation in Computer Science Education
This program is tentative and subject to change.
Generative artificial intelligence (GenAI) is transforming Computer Science education, and every instructor is reflecting on how AI will impact their courses. Instructors must determine how students may use AI for course activities and what AI systems they will support and encourage students to use. This task is challenging with the proliferation of large language models (LLMs) and related AI systems. The contribution of this work is an experimental evaluation of the performance of multiple open-source and commercial LLMs utilizing retrieval-augmented generation in answering questions for computer science courses and a cost-benefit analysis for instructors when determining what systems to use. A key factor is the time an instructor has to maintain their supported AI systems and the most effective activities for improving their performance. The paper offers recommendations for deploying, using, and enhancing AI in educational settings.
This program is tentative and subject to change.
Sat 1 MarDisplayed time zone: Eastern Time (US & Canada) change
10:45 - 12:00 | |||
10:45 18mTalk | ASM Visualizer: a Learning Tool for Assembly Programming Papers Tia Newhall Swarthmore College, Kevin Webb Swarthmore College, Isabel Romea Swarthmore College, Emma Stavis Swarthmore College, Suzanne Matthews United States Military Academy | ||
11:03 18mTalk | ezFS: A Pedagogical Linux File System Papers DOI | ||
11:22 18mTalk | Grading for Equity in a Hyflex Compiler Design Course Papers | ||
11:41 18mTalk | Quantitative Evaluation of using Large Language Models and Retrieval-Augmented Generation in Computer Science Education Papers Kevin Shukang Wang The University of British Columbia, Ramon Lawrence The University of British Columbia |