1. Introduction
The attraction and appeal of large, or neural language models, commonly known as Generative Pre-Trained Transformers (GPTs), is found in their potential to infer and produce quality text responses to prompts from users. Pre-trained language models can handle datasets they are trained on, which enable a variety of Chatbot models to perform varieties of tasks. Such models are notable in commercial form by the release of ChatGPT, accessed through the OpenAi website, while alternative interfaces exist built on the same deep learning mechanisms used in order to process, sequence, organize and analyze data to assist in solving complex problems [
1]. Primarily online, these technologies can also run offline, with platforms now having transformer networks built into some models, including Acrobat PDFs, while Open Source alternatives such as GPT4ALL, through nomicAI, mean these tools can run on desktop central processing units without internet connections.
Uses vary for the application of the varying GPTs now proliferating, but educators have been quick to see benefits, with pedagogical affordances speculated for teachers, while expectations were conceived of how such models might save teachers time through the automation of labor-intensive tasks, including marking, lesson or course planning or burdensome administrative tasks. Issues surround the ethical implications of automating some of these tasks. Where bias appears to be coded into the training data [
2] of the language models, this can be passed across through responses to prompts, potentially discriminating against marginalised peoples through, for example, the widening use of data analytics [
3] to predict the success of learning outcomes. Equally, there are growing concerns around the disruption these technologies cause conventional means of assessment, with a corrosion of trust infiltrating students essay writing, since LLM interfaces can generate credible responses to tasks, and the plagiarism-detection software to catch them out is often flawed. This has led to general concerns about the impact of AI chatbots on academic integrity, particularly where it is related to classroom -based pedagogy [
3,
4].
The vocational sector, often called the skills sector in England, entails procedural knowledge underpinned by spatial abilities and psychomotor skills, including technical activities in the arts and media and with computers, musical ability, visual estimations, the relationships between objects, user and environment, often including the manipulation or handling of technical equipment, especially technologies. These types of skills, being redolent of the real world and work-based learning, have adequate fit with practical and experiential learning. Vocational Education (VE) tends to negotiate developments where it can in technological progresses and its training must be equally innovative and bold, where it can be. It is often a neglected sector, overlooked at the expense of schools and higher education and suffers from poor funding in England. The influence of GPTs on these domains and this sector remains to be seen, but this is a sector that, in reflecting real world learning and training, must be open to emergent technical tools that arise and seek to deploy them in teaching and learning contexts, as well as the opportunities they may represent in terms of efficiency in working conditions for its staff [
5]. Teachers, therefore, are at the vanguard of these changes and FE commentators have long made arguments around the necessity for students and staff alike to embrace digital innovation, as long as innovative technological knowledge is accompanied by pedagogical knowledge [
6]
Reports such as the 2014 FELTAG (Further Education Learning Technologies Action Group) inform the work of the vocational sector, with directives (prior to Covid 19) that proportions of teaching should be made online. While there are unanswered questions about the empirical evidence for learning gains around the uses of learning technologies, there can be no debate about the need for particular digital literacies for today’s students. This starts with the digital competencies of teachers, which is formally acknowledged in models such TPACK [
7], devised as the understanding of the overlay between ‘Technical, Pedagogical and Content’ knowledge to develop proficient educational experts. Frameworks and theoretical reference points and approaches are continually evolving [
8], with an emphasis on adaptiveness and reflection and competences of technological use being key for teacher education. Others [
9] use evaluation matrix to consider the perceived usefulness of technology in vocational education (VE), which is underpinned by the acceptance of technology (Technology Acceptance Model, or TAM) [
10], where ease of use is a main factor in take-up. Another potentially useful model for understanding how technology is utilized by teachers is SAMR, first proposed by Puentadura [
12], which was designed to identify the ways in which technologies were used at four principal levels and how technologies impacted on pedagogical classroom tasks, being a substitution, augmentation, modification of redefinition (SAMR) of previously existing practices, though issues in meta studies point to the lack of contextual information often used and a focus on product over process [
12]. The latter is an issues in VE, where product is certainly important, but focus on processes incorporates learning by doing, experimentation and risk-taking, and reflections on mistakes as paramount to students mastery of concepts.
GPTs certainly come with a plethora of problems: hype, inflated performance and operational capacity and a susceptibility to generating errors. They also carry ethical issues with built-in bias, glitches and even potential for exacerbating existing inequalities [
13]. But as potentially powerful tools in a state of interregnum development, they are often freely available and built into existing networks or virtual learning environments, meaning they are on hand and that platformatisation of education is likely here to stay [
14].
Against the previously noted need for competency and innovation are two socio-economic factors that are critical to draw into these debates: firstly, there is a growing teacher retention and recruitment crisis across most sectors in England, but seemingly internationally, with high workload, precarity [
15] and poor working conditions given as reasons for these twofold issues [
16]. Secondly, there is the possibility that the advent of General Artificial Intelligence creates automation in huge industries, posing stakeholders from policy Governance to local educational contexts to consider the purposes of vocational education. The UK Government has already started to investigate the ways in which AI may reduce workload or support with improved student feedback [
17], though this may come at the cost of teacher agency, which is integral to teacher professionalism [
18,
19]. Nevertheless, the arrival of GPTs entails a new form of imagining the potential for vocational educators to integrate these into their teaching and learning contexts, or alternatively to relieve some burden of the administrative duties that push teaching and learning priorities to the side at the expense of form filling. Proposals for technologies as solutions highlight issues within systems, since they engrain policy as frozen in silicon [
20], with government policymakers giving a heuristic understanding of the contemporary pressing issues. This is important, since research that addresses policy lays bare the issues inherent with the system that AIE is supposed to resolve and can shape critical pushback [
21] on potential solutions that may be ineffective or require further development. In this instance, we are interested in what the tools are being used for, in order that we can potentially forecast the needs of future teachers. The focus is on workload: as the previously referenced Government report states that AIE can help with:
creating educational resources
lesson and curriculum planning
tailored feedback and revision activities
administrative tasks
supporting personalised learning
and that, when used appropriately, generative AI has the potential to:
reduce workload across the education sector
free up teachers’ time, allowing them to focus on delivering excellent teaching
This paper aims to investigate the ways in which VE educators are utilizing AI up to now. It takes as a starting point the Pareto Principle, which is – admittedly - a slightly dated economic notion that explained how 80 per cent of a nation’s wealth was concentrated into the possession of 20 per cent of the population, a figure that’s likely fluctuated in the present era. The Principle has been applied in academic fields in varying ways, usually corresponding to the same equative result, i.e. that 80% of outcomes stem from 20% of the causes. Its relevance here is in terms of considering whether GPTs are potentially supporting teachers with heavy workloads. For example, in streamlining repetitive tasks. A significant portion of a teacher’s time is often consumed by repetitive tasks like creating quizzes, grading, or formatting materials. There may be potential in notions of offloading time-consuming repetitive tasks, freeing up time for the teacher to focus on customizing lessons to meet individual student needs, engaging directly with students in class discussions or activities, or designing more creative, interactive components of their lessons, etc. Since GPTs are stateless, that is they don’t retain data from preceding interactions or requests, they may miss much of the nuance of a classroom that a teacher gleans from day to day and over a career. This means that they capture static impressions of dynamic and multilayered shifting contexts, so their resources may be limited in value.
If GPTs are able to generate the bulk of lesson outlines, activities, and assessments, then teachers can refine these resources for their specific classroom context. This modification aspect of tweaking resources, materials and plans created by GPTs may be crucial to teachers development and agency in the classroom. Indeed, findings show that [
22] teachers with a depth of subject knowledge and mastery of effective prompt engineering rate GPTs as useful, whereas teachers with less secure subject knowledge and unfamiliarity with quality prompts do not appear to evaluate the benefits and use of AI highly. However, this does indicate that potentially in the future some offloading of creativity will occur and yet the quality of GPT generated resources remains questionable [
23]. More sophisticated uses of GPTs may focus on learning analytics, topic modelling, analysis of datasets or personalising individualised learning pathways to enhance student improvements. That was a study that focused purely on ChatGPT and on primary education teachers. What is less well known is how VE teachers are currently using GPTs of more varieties and for what purposes, so this guided our research design and instruments. It also leads to our research questions, which investigates the main ways these technologies and the application of them may be influencing teacher’s professional development and workload and delegation choices of what they offload to GPTs largely. We were interested in the potential choice architecture resident in such tools and this becomes a focus on the analysis discussion that follows. Although the sample is scattered across a variety of VE contexts, so does not reflect a concentrated and potentially cross-institutional approach, the variety in results potentially shows the need for a joined-up framework to guide GPT prompt engineering towards quality output.
2. Materials and Methods
This study presents findings from a survey of 60 vocational educators regarding their use of Artificial Intelligence (AI) tools in vocational educational settings. The authors wished to develop an understanding of how vocational educators were starting to use AI in their settings and contexts. This involved asking about types of tools, their application and purpose, their frequency of use and their perceptions of usefulness in helping with workload. Other contextual information about the educators’ length of teaching experience and subject specialism within teaching was drawn in to try to ascertain whether newly qualified or more experienced teachers had differing views of the uses of GPTs. Finally, an open question was added to the survey in order to elicit views and values about the emergence of AI into education (‘Is there anything else you would like to share about your experience with using AI Tools in education?’).
Quantitative analysis was used to identify patterns, and thematic insights were gathered for illustrative purposes. In the discussion that follows the presented results, we draw on the pareto principle as a means to identify the proportion of work tasks designated by teachers to GPTs. This is done in order to try to identify insights about the quantity of tasks most given to GPTs and whether some comprise a majority.
We acknowledge the limitations in this research design. The small sample size of 60 VE teachers was purely down to who responded to the survey. Survey responses were drawn from requests to participate from VE workplaces, through a learning technologies news bulletin and across social media channels by the authors and peers. This is not meant to be demonstrative or generalizable to VE educators nationally or further afield, but a contextual insight from who was available to us and responded to our call for participants. Further, we note that users of social media or readers of a learning technologies newsletter are more likely to be inclined to be innovative adapters and users of technologies in their teaching than is probably normal, but we count this as an advantage of the study rather than a limitation, since they may be more able to yield insights through the opportunity for open questions.
Participants generally ranged from across VE subject specialisms, including subject experts from Carpentry and Joinery, Maths, Counselling, Sports, English and ESOL (English as a Second Language) Business, Engineering, English, Computing, Bakery and catering, Travel and Tourism and Special Educational Needs support. This gives a varied and realistic representation of the typical VE subjects taught in English college systems.
The data was collected via a Microsoft Forms survey and was a mixture of information gathering questions, a Likert scale ranking and an open question giving qualitative insights into one combined dataset. The questions are shown below:
What is your role at the College and subject specialism?
How many years teaching experience do you have?
How often do you use Artificial Intelligence tools (such as TeacherMatic, Gemini Gamma, Chat GPT, etc) for work each month?
Which is the AI Tool that you use most frequently?
On a scale of 1–5 Stars (1 being 'Not at all useful' and 5 being 'Extremely useful'), how would you rate these tools in terms of supporting your workload?
Which functions of these AI Tools do you use most frequently? (Select up to three.)
How often do you need to adapt the output created by AI?
In general, do you think AI Tools are worthwhile for educators?
Is there anything else you would like to share about your experience with using AI Tools in education?
4. Discussion
Given the uneven levels of adoption of AI, it was interesting to see the range of tools being deployed as cited in the surveys. For the most part, this was predicated on Teachermatic and ChatGPT, which is unsurprising, since the former has licenses in many college institutions, while the latter is prolifically used as an early marketized tool. Other mentions were given to Co-Pilot, Perplexity, Gemini, and Notebook. A choice architecture reading of the data shows that these common ‘Big Tech’ tools are in the majority being used, so that where institutions and professional bodies, as well as Government, are heedlessly advocating for the embracing of ‘AI’, the main beneficiaries to this tend to be the market established companies whose products are visibly available, or built into existing platforms, like portable documents or browsers. This is problematic, since there are more varieties of tools available being overlooked. The choice of tool and its functionality determines the way it will be used and mainstream tools embedded into monopolistic eco-systems such as Microsoft and Google are likely to be prominently used, and thus shape educational practices, not necessarily in helpful ways but also harmful (i.e. the harvesting of student data [
23].
The data shows that GPT platforms hold potential for resource creation, but it could be argued that they appear to diminish subject specialism skills that would otherwise be honed in a labour intensive manner through the creation of original resources. Instead, something replaces the act of creativity: a refinement at best, as it is reported that these resources need careful scrutiny [
24] because they often contain mistakes that can go unchecked straight into teaching. Refinement or checking and modifying resources produced by GPTs, one could infer, is not always going to be practiced. We argue that certainly early career teachers may benefit from creating resources to develop knowledge of how to communicate and teach the content knowledge of their subject specialism. For example, beginning teachers typically need to design and develop schemes of work for an extended syllabus, that may contain twelve weeks of lessons and activities mapped to a curriculum. This is exhaustive and could be automated and processed by an interface such as Teachermatic, yet while this or other GPTs might create a standard twelve week lesson plan, they could also problematically overlook much of the classroom environment, knowledge of students needs and preferences, the diversity of students in the group, targets and other variations that are unexpected, hard to predict or control. Teachers noted this in their open answers:
“Often it is sold as reducing your workload but I'm not sure. Often the quality or robustness of the product it gives you requires more work to make it effective. I worry the impact it will have on student teachers and the lessons they will lose in their early career as they use AI.” Returning to the SAMR model mentioned in the earlier literature review, it is important to note the modification element; here it means something different. In order to stay as constituent central parts of the teaching and learning experience, as endorsed by [
24], teachers will need to be vigilant about ‘quality’ and be prepared to modify GPT-generated resources or lesson plans that have been created to ensure some alignment with their students and their curriculum content [
25,
26].
It may alternatively be argued that early career teachers would benefit from the agency and undertaking the ‘grunt work’ involved in planning a scheme of work to give them closer oversight of the scaffolding of the subject and its curriculum, the nuances of lessons, and a personalised understanding of their students’ learning needs that an interface may not be able to provide, given it may generate generalised responses. However, many teachers recognized that the first results were never appropriate and there is evidently skill in evaluating the resulting resources: “It’s a great tool to take the initial edge of bulk admin tasks. I find it easier being the editor and creator than the doer (thats the Ai’s job).”
Choice Architecture and GPTs
Choice architecture appears present in the ways that the platforms are used is apparent from this data, with the most common uses being to generate resources, with powerpoints and quizzes cited as part of that. Teachers also reported that they used GPTs to develop lesson plans. Most uses were for teaching/learning resource creation, though creating these is unlikely to be the source of teacher workload that leads to fatigue and burnout. AI guides responses to certain normative pedagogical choices, for instance direct instruction and grouped work, which might stifle teachers and students’ creativity and innovation. This determinism means that innovation according to individual needs may be stifled by the implementation, or prescriptivism, of the tools. For example, courses which require practical elements, including experiential learning, or where teachers want to advocate for risk-taking and experimentation, allowing for mistakes and problem solving through so-called ‘inquiry-based approaches. These are potentially inhibited by the choice architecture of the tools, as noted by an English teacher in their response to the open question: “
It [Teachermatic] has a good variety of resources that can be drawn from it; in terms of what it does, it tends to induce teachers to teach in certain ways by narrowing down the planning it suggests, and the resources almost always need some tweaking.” The systems appear to foreground certain functions in the user interface (e.g., template-based generators, “favourites”), lowering the cognitive and time cost for those tasks. By making content generation fast, visible, and low-risk, teachers may then be “nudged” or compelled towards these features without necessarily exploring a full pedagogical range or alternative options. Over time, this knowledge could dissipate. This was evident in responses to the open question, where some participants recognized that innovation is necessary, is guided by experience of prompting (“
Understanding that the output is only ever as good as the input is crucial to using AI effectively.”), and does not always result in the common pedagogical design resulting from resource or planning creation (that is, powerpoints or presentations and grouped work). “
Please use as innovation for learning and teaching to enhance experience and engagement. I recently visited one of our English classes where the teacher felt that students do not understand what it was like before technology. We organised something through Nichesss where students had a discussion with Jane Austen to ask what it was like before technology - by using technology. A fantastic bridge, and great lesson.” This is borne out by other studies, [
27] where 77% of teachers relied on platforms such as Teachermatic to produce resources, which they cited as quizzes and powerpoints. It’s plausible to suggest that without rich teacher education, knowledge exchange and continued professional development, mediation between teacher and students and students with their peers could diminish at the cost of the creation of staid resources. Alternatively, GPTs may provide ideas and examples for experiential learning and it’s possible that in the future other technologies emerge from this interregnum in development that supplement those. There already exist virtual and immersive reality, in the form of smartglasses which have been shown to enhance mobile assessment of vocational education, but this is reliant on highly imaginative teachers [
28].
Another concern about the ways that teachers are utilizing the platforms is in what we may see as an early stage novelty effect bias towards visible and tangible outputs. This is natural: teachers want products that would seemingly alleviate processes and efficiency in doing so. It’s shown in a lower uptake in lesson planning and administrative tasks that may result in less visually gratifying results. What is highly regarded tend to be resource outputs that can be put to use (quizzes, rubrics, worksheets). Teachers are more likely to adopt features that yield immediate, concrete artefacts that feel like a “win.” The architecture of these AI tools emphasizes ready-to-use, tangible outputs that subtly channels user engagement away from back-end or abstract tasks, where workload may be higher and more nuanced. As to the question of lowering workload, this is hard to discern, though on the surface level it would appear that the infrequent uses reported show staff are not taking the opportunities seemingly available to offload work to GPTs. Again, this is sometimes put down to resources and refinement of them, as one Maths teacher notes:
“I'm not sure if it's usefulness is lacking due to my subject. The quizzes it produces are simply adequate and more well though tout resources, particularly those inline with the principals of mastery, are ready made elsewhere on specialist sites, of which there are an abundance.” This modification, entailing extra work, is echoed by others, such as this Counselling Teacher: “
I find TeacherMatic extremely useful for giving me ideas and provides me with an advance in creating resources and assessment I need when preparing my lessons, more so when I have limited time. Another effective function is the Rubric, however just lately I have had to make significant amendments to make it sufficient.” While another Maths teacher was critical of this particular platform: “
Doesn't seem to work effectively with Maths - was looking to see whether it could generate examples with or without solutions but only seemed able to draw in definitions. Could generate problems to an extent, but incomplete without graphical elements (e.g. 'use the graph to define' without a graph).” These conflicting views seem to point to the need for a better understanding of the affordances that can be gained from better prompt engineering, harnessing the tools to get them to work for users specific needs. This, however, also depends on the development capacity of the tools to capture more nuance and, to another extent, improvements in teacher education, such that:
“the end result is only ever as good as the prompts you put into whichever AI generator you are using.” Elsewhere CPD workshops around these issues have been developed that yield positive results in being critically informed about AI, curriculum mapping to identify spaces where it can be integrated, while also exchanging and sharing knowledge in participatory sessions regarding the ways in which GPTs can enhance teaching and assessment [
29]. Thes affordances will potentially lead to time-saving and reduced workload if more advanced practitioners can support this knowledge with their experience.
Yet others in the survey reported categorically on its time-saving potential, as shown by a Functional Skills and English teacher: “Teachermatic has made 2 hour jobs 20 minute jobs; now I can use my time for more student facing work.” However, this is in the context of one tool being licensed across a college; where others find having to explore the full range of GPTs available to be time-consuming in itself: “There is so much and its hard to know what is best to use from a security, cost and output point of view. Significant time has to be spent exploring different tools “. This again might result in a choice architecture where users go to the centralized technologies available from Big Tech, since they are most prominent and available (and this statement from a Digital Innovation coach who uses Co-Pilot most commonly. Finally, another statement from an Arts lecturer is distinctly more negative about the impact of these tools, and they point to the issues prevalent in working systems, rather than technologies: “They are useful in the current teaching environment. But they shouldn't have to be if the workload was balanced correctly. If we continue to have to use it will it de-skill teachers. Will we run the risk of lesson be created by AI and the teacher not knowing or understanding how or if it meets the needs of learners.”