Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate Learning

Software testing education is important for building qualified testing professionals. To ensure that software testing graduates are ready for real-world challenges, it is necessary to integrate modern tools and technologies into the curriculum. With the emergence of Large Language Models (LLMs), the...

Full description

Saved in:
Bibliographic Details
Main Authors: Susmita Haldar, Mary Pierce, Luiz Fernando Capretz
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10904141/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849392042031972352
author Susmita Haldar
Mary Pierce
Luiz Fernando Capretz
author_facet Susmita Haldar
Mary Pierce
Luiz Fernando Capretz
author_sort Susmita Haldar
collection DOAJ
description Software testing education is important for building qualified testing professionals. To ensure that software testing graduates are ready for real-world challenges, it is necessary to integrate modern tools and technologies into the curriculum. With the emergence of Large Language Models (LLMs), their potential use in software engineering has become a focus, but their application in software testing education remains largely unexplored. This study, conducted in the Capstone Project course of a postgraduate software testing program, was carried out over two semesters with two distinct groups of students. A custom-built Travel Application limited to a web platform was used in the first semester. In the second semester, a new set of students worked with an open-source application, offering a larger-scale, multi-platform experience across web, desktop, and mobile platforms. Students initially created preparatory testing artifacts manually as a group deliverable. Following this, they were assigned an individual assignment to generate the same artifacts using LLM tools such as ChatGPT 3.5 in the first semester and Microsoft Copilot in the second. This process directly compared manually created artifacts and those generated using LLMs, leveraging AI for faster outputs. After completion, they responded to a set of assigned questions. The students’ responses were assessed using an integrated methodology, including quantitative and qualitative assessments, sentiment analysis to understand emotions, and a thematic approach to extract deeper insights. The findings revealed that while LLMs can assist and augment manual testing efforts, they cannot entirely replace the need for manual testing. By incorporating innovative technology into the curriculum, this study highlights how Generative AI can support active learning, connect theoretical concepts with practical applications, and align educational practices with industry needs.
format Article
id doaj-art-3afd4695ae4b44fcbf0ec861fe27ae6d
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-3afd4695ae4b44fcbf0ec861fe27ae6d2025-08-20T03:40:51ZengIEEEIEEE Access2169-35362025-01-0113460704609010.1109/ACCESS.2025.354588210904141Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate LearningSusmita Haldar0https://orcid.org/0009-0000-4566-1433Mary Pierce1Luiz Fernando Capretz2https://orcid.org/0000-0001-6966-2369School of Information Technology, Fanshawe College, London, ON, CanadaFaculty of Business, Information Technology and Part-Time Studies, Fanshawe College, London, ON, CanadaDepartment of Electrical and Computer Engineering, Western University, London, ON, CanadaSoftware testing education is important for building qualified testing professionals. To ensure that software testing graduates are ready for real-world challenges, it is necessary to integrate modern tools and technologies into the curriculum. With the emergence of Large Language Models (LLMs), their potential use in software engineering has become a focus, but their application in software testing education remains largely unexplored. This study, conducted in the Capstone Project course of a postgraduate software testing program, was carried out over two semesters with two distinct groups of students. A custom-built Travel Application limited to a web platform was used in the first semester. In the second semester, a new set of students worked with an open-source application, offering a larger-scale, multi-platform experience across web, desktop, and mobile platforms. Students initially created preparatory testing artifacts manually as a group deliverable. Following this, they were assigned an individual assignment to generate the same artifacts using LLM tools such as ChatGPT 3.5 in the first semester and Microsoft Copilot in the second. This process directly compared manually created artifacts and those generated using LLMs, leveraging AI for faster outputs. After completion, they responded to a set of assigned questions. The students’ responses were assessed using an integrated methodology, including quantitative and qualitative assessments, sentiment analysis to understand emotions, and a thematic approach to extract deeper insights. The findings revealed that while LLMs can assist and augment manual testing efforts, they cannot entirely replace the need for manual testing. By incorporating innovative technology into the curriculum, this study highlights how Generative AI can support active learning, connect theoretical concepts with practical applications, and align educational practices with industry needs.https://ieeexplore.ieee.org/document/10904141/Capstone projectChatGPTgenerative AIsoftware testing educationMicrosoft Copilotsentiment analysis
spellingShingle Susmita Haldar
Mary Pierce
Luiz Fernando Capretz
Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate Learning
IEEE Access
Capstone project
ChatGPT
generative AI
software testing education
Microsoft Copilot
sentiment analysis
title Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate Learning
title_full Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate Learning
title_fullStr Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate Learning
title_full_unstemmed Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate Learning
title_short Exploring the Integration of Generative AI Tools in Software Testing Education: A Case Study on ChatGPT and Copilot for Preparatory Testing Artifacts in Postgraduate Learning
title_sort exploring the integration of generative ai tools in software testing education a case study on chatgpt and copilot for preparatory testing artifacts in postgraduate learning
topic Capstone project
ChatGPT
generative AI
software testing education
Microsoft Copilot
sentiment analysis
url https://ieeexplore.ieee.org/document/10904141/
work_keys_str_mv AT susmitahaldar exploringtheintegrationofgenerativeaitoolsinsoftwaretestingeducationacasestudyonchatgptandcopilotforpreparatorytestingartifactsinpostgraduatelearning
AT marypierce exploringtheintegrationofgenerativeaitoolsinsoftwaretestingeducationacasestudyonchatgptandcopilotforpreparatorytestingartifactsinpostgraduatelearning
AT luizfernandocapretz exploringtheintegrationofgenerativeaitoolsinsoftwaretestingeducationacasestudyonchatgptandcopilotforpreparatorytestingartifactsinpostgraduatelearning