DIGITAL LIBRARY
TOWARDS A SUSTAINABLE OPERATING CONCEPT OF LLM GENERATED EXAMINATION SUPPORT IN HIGHER TECHNICAL EDUCATION
FHNW University of Applied Sciences and Arts Northwestern Switzerland (SWITZERLAND)
About this paper:
Appears in: ICERI2024 Proceedings
Publication year: 2024
Pages: 1064-1069
ISBN: 978-84-09-63010-3
ISSN: 2340-1095
doi: 10.21125/iceri.2024.0350
Conference name: 17th annual International Conference of Education, Research and Innovation
Dates: 11-13 November, 2024
Location: Seville, Spain
Abstract:
At the University of Applied Science Northwestern Switzerland (FHNW) it can be observed that the usage of preferred learning material shifted in the past years from books and articles to videos, summaries, flashcards and test exams. In a specialization module of the programme BSc Business Information Technology (BIT) at FHNW a Python-based prototype was created to semi-automatically generate MCQ (Multiple Choice Question) based exams and summaries using the - since 2023 – very popular large language models (LLM).

The current prototype software is based on the Python libraries langchain for the LLM framework, and streamlit for the user interfact (UI). As vector database ChromaDB is used and as LLM Vertex AI from Google. Currently the software is tested with selected documents of the course “Machine learning with Python” (which is part of a specialization within the BIT programme) in order to find out if the quality of generated exams and summaries match the expectations of the lecturers and students. The results will help to answer important questions regarding the practical usage of LLM in education in general.

Currently the AI community is flooded with an immense number of software prototypes for nearly every domain. However, a sustainable operational concept for AI in general and LLM in particular is rarely discussed. Our contribution seeks to address this issue for the domain of examination support in higher technical education. Specifically, the following topics are discussed: the cost/benefit ratio of using LLMs for MCQ-based exam generation, focussing on the operational aspects of prompt engineering, testing, quality management, confidentiality assurance and didactic strategy. These issues have to be answered before using the software in a broader context, namely other modules and programmes.
Keywords:
AI, LLM, Vertext AI, MCQ, Machine Learning.