Search Results for author: Arto Hellas

Found 11 papers, 2 papers with code

Open Source Language Models Can Provide Feedback: Evaluating LLMs' Ability to Help Students Using GPT-4-As-A-Judge

1 code implementation8 May 2024 Charles Koutcheme, Nicola Dainese, Sami Sarsa, Arto Hellas, Juho Leinonen, Paul Denny

Inspired by recent work that has utilised very powerful LLMs, such as GPT-4, to evaluate the outputs produced by less powerful models, we conduct an automated analysis of the quality of the feedback produced by several open source models using a dataset from an introductory programming course.

Benchmarking Educational Program Repair

1 code implementation8 May 2024 Charles Koutcheme, Nicola Dainese, Sami Sarsa, Juho Leinonen, Arto Hellas, Paul Denny

The emergence of large language models (LLMs) has sparked enormous interest due to their potential application across a range of educational tasks.

Benchmarking Program Repair

"Like a Nesting Doll": Analyzing Recursion Analogies Generated by CS Students using Large Language Models

no code implementations14 Mar 2024 Seth Bernstein, Paul Denny, Juho Leinonen, Lauren Kan, Arto Hellas, Matt Littlefield Sami Sarsa, Stephen MacNeil

Grasping complex computing concepts often poses a challenge for students who struggle to anchor these new ideas to familiar experiences and understandings.

Can We Trust AI-Generated Educational Content? Comparative Analysis of Human and AI-Generated Learning Resources

no code implementations18 Jun 2023 Paul Denny, Hassan Khosravi, Arto Hellas, Juho Leinonen, Sami Sarsa

In this study, we investigated the potential for LLMs to produce learning resources in an introductory programming context, by comparing the quality of the resources generated by an LLM with those created by students as part of a learnersourcing activity.

Exploring the Responses of Large Language Models to Beginner Programmers' Help Requests

no code implementations9 Jun 2023 Arto Hellas, Juho Leinonen, Sami Sarsa, Charles Koutcheme, Lilja Kujanpää, Juha Sorva

At the same time, the results highlight the unreliability of LLMs: LLMs make some of the same mistakes that students do, perhaps especially when formatting output as required by automated assessment systems.

Computing Education in the Era of Generative AI

no code implementations5 Jun 2023 Paul Denny, James Prather, Brett A. Becker, James Finnie-Ansley, Arto Hellas, Juho Leinonen, Andrew Luxton-Reilly, Brent N. Reeves, Eddie Antonio Santos, Sami Sarsa

The computing education community has a rich history of pedagogical innovation designed to support students in introductory courses, and to support teachers in facilitating student learning.

Code Generation

Comparing Code Explanations Created by Students and Large Language Models

no code implementations8 Apr 2023 Juho Leinonen, Paul Denny, Stephen MacNeil, Sami Sarsa, Seth Bernstein, Joanne Kim, Andrew Tran, Arto Hellas

In this paper, we explore the potential of LLMs in generating explanations that can serve as examples to scaffold students' ability to understand and explain code.

Using Large Language Models to Enhance Programming Error Messages

no code implementations20 Oct 2022 Juho Leinonen, Arto Hellas, Sami Sarsa, Brent Reeves, Paul Denny, James Prather, Brett A. Becker

Large language models can be used to create useful and novice-friendly enhancements to programming error messages that sometimes surpass the original programming error messages in interpretability and actionability.

Automatic Generation of Programming Exercises and Code Explanations using Large Language Models

no code implementations3 Jun 2022 Sami Sarsa, Paul Denny, Arto Hellas, Juho Leinonen

Our analysis suggests that there is significant value in massive generative machine learning models as a tool for instructors, although there remains a need for some oversight to ensure the quality of the generated content before it is delivered to students.

Language Modelling Large Language Model +1

Empirical Evaluation of Deep Learning Models for Knowledge Tracing: Of Hyperparameters and Metrics on Performance and Replicability

no code implementations30 Dec 2021 Sami Sarsa, Juho Leinonen, Arto Hellas

To evaluate how different aspects of DLKT models influence model performance, we test input and output layer variations found in the compared models that are independent of the main architectures.

Knowledge Tracing

Cannot find the paper you are looking for? You can Submit a new open access paper.