Assessment in humanities in time of LLMs requires, "if not a change of heart, two changes of mindset.
The first is an acceptance that LLMs – as the distinguished Berkeley psychologist Alison Gopnik puts it – are “cultural technologies”, like writing, print, libraries and internet search. In other words, they are tools for human augmentation, not replacement.
Second, and more importantly perhaps, is a need to reinforce in students’ minds the importance of writing as a process."
Technology evangelists and the big tech providers have long dreamed of disrupting education. That despite all the changes brought about by technology most education remains organised by institutions, many funded by the public sector is a source of frustration to the. One of the most prominent of teh change advocates is Salman Khan. Founder of the Salmon Academy, which boomed during the COVID 19 lock downs and advocate of video lessons and flipped learning, Salmon Khan was in London last week to publicize his new book, Brave New Words: how AI will revolutionise education (and why that’s a good thing), and while there he was interviewed for an article in the Times Education Supplement.
As Dan Meyer says in his newsletter Mathworlds, his remarks indicate "the edtech industry is starting to realize that the possibility of revolutionary impact with generative AI is small and the possibility of any impact will require them to operate as partners with institutions that many of them have disregarded."
Meyer points out that despite all the rhetoric "in reality Khan Academy has not transformed teaching like Khan hoped it might. In the US, as in the UK, students still typically sit at desks while a teacher delivers a lecture-style presentation, and then they complete tasks based on what they have learned." And Khan seems to agree. "“If you walk into a random classroom, for the most part it seems pretty similar to what we used to see,” he says. “If you asked me 10 years ago, I would have hoped… I mean, I’ve given TED Talks saying you shouldn’t need to give lectures any more, and everyone should be able to go at their own pace.”
And this raises the question of why the Khan Academy Hasn't ushered in a new era of education?
Well, the platform is designed to help students who are “trying to get through the [public school] system”, Khan says. “Either we support them in moments where they have a gap or we are used more systematically by their teacher, by their school, to improve the learning that goes on.”
And the public school system is far bigger than Khan Academy, he says.
The academy “needs to be pretty well integrated with the formal systems for it to have the maximum impact. That’s the journey that will keep us busy for decades to come”, its founder says.
It seems increasingly unlikely that AI alone is going to revolutionise education this t9ime round.
As a recent publication from the Universitat Oberta de Catalunya points out, Artificial Intelligence remains an opportunity (or an excuse) to transform assessment, curriculum, teaching, personalization and teaching competencies. This is especially so in relation to assessment with widespread concern in the academic world about the near impossibility of detecting whether or not a student has used generative AI in an assignment.
The Universitat Oberta de Catalunya article explores the potential of continuous assessment aimed at self-regulation of learning. It suggests changing the assessment approach, moving from criteria focused on the assessment of the result to criteria focused on the process of development of the activity by the students.
Furthermore it advocates designing continuous assessment activities as part of the same learning sequence, with relationships of dependency and complementarity, instead of discrete tests and focusing the activities on the development of competencies and the assessment of progress and reflection on the learning process of each student.
Leon Furze is a prolific contributor to LinkedIn and describes his work as "Guiding educators through the practical and ethical implications of GenAI. Consultant & Author | PhD Candidate."
Witting from the perspective of education in Australia he says:
When it comes to GenAI, much of the conversation in education has been focused on academic achievement, perceived threats to academic integrity, and the risk that this technology poses to written assessments. I think that vocational education actually offers some fantastic alternative forms of assessment which are less vulnerable to generative artificial intelligence. If you’re not familiar with vocational education, assessments are often incredibly rigorous, sometimes to the point where the paperwork on evaluation and assessment is significantly longer than the assessment itself.
Vocational training, by nature, is practical and geared around skills which are needed for the particular job role or discipline being studied. Mainstream education, by contrast, is focused predominately on subjects and content.
Furze provides examples of different types of assessment in vocational educati9n and training:
Observation checklists
Role plays
Scenarios
Workplace activities
Reports from employers
He has prublished a free 60 page ebook - Rethinking Assessment for GenAI which he says covers everything from ways to update assessments, to the reasons I advise against AI detection tools.
Alexa Steinbrück / Better Images of AI / Explainable AI / CC-BY 4.0
The European Digital Education Hub has put out a call for members for a working group (they call them "squads") on Explainable AI. They say:
As AI systems become increasingly influential in shaping teaching, educational outcomes and assessment, the demand for transparency and accountability in these systems has grown. Explainable AI aims to bridge the gap between complex AI algorithms on the one side and educators, learners, and administrators on the other by providing clear insights into how AI systems arrive at their conclusions.
This transparency is crucial as it fosters trust among users, who can see and understand the rationale behind AI-driven decisions, recommendations and actions. It also empowers educators to make informed decisions about integrating AI tools into their teaching strategies. Additionally, it ensures that AI systems uphold ethical standards, mitigating potential biases and promoting fairness in educational assessments.
I like the idea and have put in my application (membership is unpaid). But I got to thinking that we should pick up this in the AI Pioneers project. My idea is to try to write one 'explanation' about AI a week and also to publish it as a TIC TOC video. And of course I would love to involve AI Pioneer members in the whole process. As an aside we will be unveiling more interactive tools and activities for members of the network in the new few weeks.
But here for starters is to decide what topics related to AI and education need explaining. Here is a few I jotted down on a Google doc one evening (after, I have to admit, a couple of cooling glasses of white wine). What do you think and what have I missed out? Add your ideas on the Google doc here.
Maybe the panic over the impact of AI on assessment in education has died down a little, but it has been useful in that it has focused attention on the puropse of assessment and the pedagogic approaches to assessment. Simon Brookes, Executive Dean, Faculty of Creative & Cultural Industries, at the University of Portsmouth in the UK has started a new blog series on Rethinking Assessment in the Age of AI. His latest post features insights from the University of Melbourne's Centre for the Study of Higher Education. Their recent guide, "Rethinking Assessment in Response to AI" (pdf) offers a thoughtful approach to redesigning assessments that maintain academic integrity without sacrificing pedagogical value, he says.
The guide includes seven critical strategies for improving assessment design and integrity:
1. Shift from product to process: Focus on evaluating students' thinking processes and problem-solving approaches rather than just the final output. This could involve asking students to maintain learning journals, document their research process, or explain their reasoning in solving problems.
2. Incorporate evaluative judgement tasks: Ask students to review or evaluate work against set criteria, encouraging higher-order thinking skills. This might include peer review exercises, critiquing published works, or assessing case studies against industry standards.
3. Design nested or staged assessments: Create assignments that build on each other throughout the semester, allowing for feedback and adaptation. For example, a research project could be broken down into proposal, literature review, draft, and final submission stages, each informing the next.
4. Diversify assessment formats: Use various modalities, such as videos, blogs, podcasts, and animations, which are less susceptible to AI generation. This not only makes cheating more difficult but also allows students to develop a broader range of communication skills.
5. Create authentic, context-specific assignments: Design tasks that mirror real-world scenarios or are highly specific to the subject matter. This could involve analysing local case studies, solving problems specific to your discipline, or applying theories to current events.
6. Include more in-class and group assignments: Incorporate collaborative learning and reduce opportunities for individual cheating. This might involve group presentations, debates, or problem-solving sessions during class.
7. Use oral interviews: Test understanding through verbal responses to unpredictable prompts, making it difficult to use AI. This could range from viva voce examinations to informal discussions about a student's work process.
This website uses cookies to improve your experience. We'll assume you're ok with this, but you can opt-out if you wish. Cookie settingsACCEPT
Privacy & Cookies Policy
Privacy Overview
This website uses cookies to improve your experience while you navigate through the website. Out of these cookies, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may have an effect on your browsing experience.
Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.
Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. It is mandatory to procure user consent prior to running these cookies on your website.