X Close

Digital Education team blog

Home

Ideas and reflections from UCL's Digital Education team

Menu

Archive for the 'Digital Assessment' Category

An institutional environment that increases academic integrity

By Marieke Guy, on 25 November 2022

Over the last two years UCL has made major strides in its use, support and understanding of digital assessment at scale. Prior to the pandemic there had been interest in providing a centrally-managed environment for exams and other assessment type. Covid accelerated these plans resulting in the procurement of the AssessmentUCL platform.  The move to digital assessments brings many benefits for users (as evidenced by Jisc and others). Students can be assessed in an environment they are increasingly comfortable with; there are administration benefits, for example through system integrations and streamlined processes; and, pedagogically, digital assessments allow us to experiment with diverse, authentic, ways to appraise, and contribute to, learning.

However, at their core, digital assessment environments consist of a set of tools and technologies which can be used well or badly. In a recent Blog post,  ‘Are digital assessments bad?’ I discussed many of the reservations people have about their use including issues related to digital equity, technology limitations and academic misconduct. My conclusion was that “these issues are not insignificant, but all of them, perhaps with the exception of a preference for offline, are surmountable. The solutions are complex and require resource and commitment from institutions. They also require a change management approach that recognises both the importance of these issues, and how they impact on individuals. It would be a massive fail to brush them under the carpet.” We in the Digital Assessment Team stand by this conclusion.  The move to online assessment during Covid has surfaced a lot of pre-existing issues with ‘traditional’ assessment as identified by Sally Brown and Kay Sambell, and others.

In this post I’d like to look at one of the identified problem areas: Academic integrity. I want to consider the relationship between digital assessments and academic integrity, and then put forward some ideas that could better support academic integrity. Is the solution removing digital assessments and insisting on in-person handwritten exams? Or if it is something more complex and nuanced, like creating the right institutional environment to support academic integrity.

A very brief history of cheating

Cheating in assessments has been around since the origin of assessments. We can all remember fellow students who wrote on their arms, or stored items in their calculator memory. The reddit Museum of Cheating shares cheating stories and images from old Chinese dynasties to the present day. However, the move to digital assessments and online exams has not only shone a light on existing misconduct, but also presented new possibilities for collusion and cheating. Figures vary across the sector but one dramatic poll by Alpha Academic Appeal found that 1 in 6 students have cheated over the last year, most of whom were not caught by their institution. And “almost 8 out 10 believed that it was easier to cheat in online exams than in exam halls”. Clearly this is not acceptable. If we cannot ensure that our degree standards are being upheld this devalues both the Higher Education sector and the qualifications awarded.

We need to begin by accepting that issues exist, as is often the case with any technological progress. For example, think about online banking. It brings many benefits but also challenges as we adapt to online scams and fraud. Artificial Intelligence (AI) adds additional complexity to this picture. Deep learning tools such GPT-3 now have the power to write original text that can fool plagiarism checking software, but AI can also support streamline marking processes and personalise learning.

A multi-faceted, multi-stakeholder approach

So what is the answer? Is it moving all students back to in-person assessments and reinstating the sort of exams that we were running previously? Will this ensure that we create exceptional UCL graduates ready to deal with the challenges of the world we live in? Will it prevent all forms of cheating?

UCL has heavily invested in a digital assessment platform and digital assessment approaches.  How can we take the learning from the pandemic and the progress we have made in digital provision and build on that for the benefit of students and the university. See Peter Bryant’s recent post ‘…and the way that it ends is that the way it began’. While there are some cases where in-person assessment may be the best option temporarily, long-term we actually hope to take a more progressive and future-proof approach. This requires an institutional conversation around our assessment approaches and how our awards remain relevant and sustainable in an ever-uncertain environment.   If we are to stand by our commitments to an inclusive, accessible, developmental and rewarding learning experience for students, will this be achieved by returning to ‘traditional’ practice?  How can we support staff by managing workload and processes so that they can focus on learning and assessment that make a difference to students?

Changing an institutional environment or culture isn’t a straightforward process. It encompasses many ideas and methods, and requires a multi-stakeholder commitment.

Within the Digital Assessment Team we have begun to look to Phill Dawson’s academic integrity tiers list for reference. Phill Dawson is Associate Director of the Centre for Research in Assessment and Digital Learning at Deakin University in Australia. In his tier list exercise Phil is asking the question about assessment priorities. If the priority is academic integrity /preventing cheating what strategies will enable you to achieve this?  He often talks about an anti-cheating methodology that involves multiple approaches (the Swiss cheese approach), with programmatic assessment, central teams and student amnesties being the most effective mechanisms.

For reference Tier lists are a popular online classification system usually used to rank video games characters. Their adoption has increasingly extended to other areas including education. Within the tier system S stands for ‘superb’, or the best, A-F are the next levels in decreasing value.   You can see the full explanation of Phill Dawson’s tier list along with research to justify his placement of activities in his HERDSA keynote talk.

Phill Dawson’s academic integrity tiers list

A possible UCL approach?

At UCL we are already undertaking many of these approaches and more in order to build an institutional environment that increases academic integrity. You could say that we have our own Swiss cheese model. Let me give you a flavour of the type of activities that are already happening.

Assessment design – I have heard it said that that you can’t design out cheating, but you can inadvertently design it in. Focusing on assessment design is the first step in building better assessments that discourage students from academic misconduct. Within UCL we are bringing together Arena, the Digital Assessment Team and others to look assessment design early in the process, often during the programme design level, or at the point a programme or module design workshop is run. The aim is diverse, engaging, well-designed assessments (formative and summative) that are connected to learning outcomes and are often authentic and reflective. The UCL Academic Integrity toolkit supports academics in this design process. We are collecting case studies on our team blog.

Assessment support – We need to build student and staff capability with regard to academic integrity. We plan to rationalise and promote academic integrity resources and skills-building training for students. These resources need to give the right message to our students and be constructive, not punitive. They need to be co-developed in collaboration with faculty and students and be part of an ongoing dialogue. They also need to be accessible to our diverse student body including our international cohort who may join our institution with a different understanding of academic practice. There also needs to be well-communicated clarity around assessment expectations and requirements for students. We are supporting Arena with further training specifically for staff, for example on MCQ and rubric design, and with supporting their students in understanding this complex area. There is also scope for training on misconduct identification and reporting processes.  We are also working with Arena to review Academic Integrity courses for students.

Technical approaches – There is a stated need for assessments that focus on knowledge retention in certain discipline areas, which is why we are piloting the Wiseflow lockdown browser. The pilot has three different threads: (1) a Bring Your Own Device pilot with Pharmacy; (2) a pilot using hired laptops in the ExCel Centre, likely with modules from School of European Languages, Culture and Society (SELCS); and (3) a departmental pilot using iPads in the Medical School. In parallel, the Maths Department are using Crowdmark to combine in-person assessments with digital exams. And as a team we are building up our knowledge of detection platforms. such as Turnitin and their stylometry tool draft coach, and of the use of AI in relation to assessment.

Better understanding of academic misconduct – In order to tackle academic misconduct we need to understand more about why it takes place, not in a general way but at UCL on particular modules and particular assessments, and with particular cohorts or students. Some of this information can be identified through better analysis of data. For example, one recent initiative is the redevelopment of an assessment visualisation (‘Chart’) tool, which depicts the assessment load and patterns across modules. This can help us identify choking points for students. We are also involved in a mini-project working with the student changemakers team in which we will be running a series of workshops with students looking at the area of academic integrity.

For discussion

We are unlikely to stamp out all forms of cheating, regardless of where our assessments take place. Thinking we can do this is both naïve and short-sighted.

However, we can work to creating an institutional environment where misconduct is less likely. In this environment we develop students who are well-supported and fully understand good academic practice and its benefits, for themselves as students, and the community more widely. They are able to prepare well for their assessments and are not put under unnecessary pressure that doesn’t relate to learning outcomes of their course. They are provided with well-explained, well-designed, challenging and relevant assessments that engage their interest and build capability and skills. They are supported in understanding the misconduct process and can be honest and open about their academic behaviour throughout their time at UCL. If the assessment is heavily knowledge-based, then it takes place in an environment in which access to outside resources is limited, but it is non-invasive and students are still respected and trusted throughout. An approach like this is compassionate and supports student wellbeing. It is cost effective, by reducing academic misconduct panels and the need for large-scale in-person assessments. It is progressive and scalable with useful online audit trails.  It should work for academic and administrative staff.  The Digital Assessment Team is committed to working with academic teams to realise this vision.

What do others think? How does such a vision support our institutional strategy and could it form part of a UCL teaching and assessment framework?

What’s next for the Digital Assessment Team?

By Marieke Guy, on 10 October 2022

In his conversation with Mary Richardson at the UCL Education conference earlier this year Professor David Boud laid out the unavoidable truth  that “students are profoundly affected by how they are assessed, it provides messages to them about what really counts”. He explained that assessment “has to contribute to students learning” as student learning behaviour is both framed and driven by assessment. In essence good assessment is imperative in ensuring good education. At UCL we are lucky to have a dedicated Digital Assessment team based within the Digital Education team in ISD. This gives us a real opportunity to improve the assessment journey for both staff and students.

A little history

Back in 2020 the AssessmentUCL Project (or product) was established  to “procure and deliver an end-to-end digital assessment delivery platform to support online assessments”. The project successfully purchased and implemented Wiseflow – the first part of the platform known as AssessmentUCL. During the first exam period of 2020-21 more than 1,071 assessments were delivered in 48,742 sittings. This increased to 1,717 assessments at 57,590 sittings in 2021-22. Throughout the process the product team have listened to feedback on experiences from staff and students, and shaped regulations and policy accordingly.

To support the transition to online assessment at UCL, the Digital Assessment team was formed in September 2021 and led by Anisa Patel. The new team initially comprised of a team of five Digital Assessment Advisors (DAA) responsible for different faculties and a Learning Technologist. Over the last year the team has worked incredibly hard. They have developed hands-on training (running 45+ training sessions over the last year) and comprehensive guidance covering the AssessmentUCL platform, they have championed the capabilities of the platform and tested all aspects of functionality. They have also spoken at length with faculty colleagues to understand their assessment needs and requirements. In some sense they have been acting in an ‘ambassador’ role – making sure students, academics and professional staff are heard. This has resulted in positive enhancements for both the platform and wider assessment practice.

Plans for this year

The team continues to evolve and right now we have three DAAs: Eliot Hoving, Isobel Bowditch and Lene-Marie Kjems; one Learning Technologist: Nadia Hussain and one Senior Learning Technologist: John Spittles. You can read more about the team on the Digital Education Team site, and which faculty they are responsible for supporting on the Education support contacts page. I have recently taken on the position of Head of Digital Assessment at UCL, moving on from my previous DAA role.

Digital Assessment Team

Digital Education Team. Top row: Eliot Hoving, Marieke Guy, John Spittles. Bottom row: Lene-Marie Kjems, Isobel Bowditch. Nadia Hussain joined after this photo was taken.

In the 2022-23 academic year our approach two pronged.  Firstly, we plan to ensure scalable and effective support for AssessmentUCL. We will do this through improving our support model and training resources, continuing to push forward improvements to the AssessmentUCL platform and ensuring the effectiveness of current policies and practices. Some of the areas we will be looking at including the use of formative assessment and on-campus digital assessment using a lockdown browser. And secondly, we will look at other areas in which we can support good assessment design and delivery more widely. We want to explore and promote best practice and be leaders in the assessment area. For example, we have begun work on an academic integrity and misconduct project to help us better understand the issue from both staff and student perspectives, and to identify strategies to promote academic integrity and support in this area. Whenever possible we will be working in collaboration with Arena, our academic colleagues, the Central Assessment Teams and the AssessmentUCL product team. It is likely to be a busy year!

We have started a team blog to share our activities and support wider assessment discussion, and also plan to post more regularly on the Digi-Ed blog.

Assessment hackathon event – Collaboratively exploring the digital assessment challenge  

By Anisa Patel, on 28 March 2022

“In a digital era, how can we design assessments differently and better?” 

This is the question that a group of more than 35 key partners in UCL’s teaching and assessment community gathered to consider last week. Hosted jointly by ARENA and Digital Education, the group comprised academics, students, the Digital Assessment team, Faculty Learning Technologists, professional services staff and representatives from UNIwise (suppliers of our digital assessment platform, AssessmentUCL).  

Attendees were split into teams of mixed disciplines to share their experience of assessment at UCL and bring forward ideas and recommendations on how digital assessments could look in the future. 

The breadth of representation made for a rich and varied discussion and enabled each partner to express their principle areas of focus: 

  • From our student representatives we heard a genuine desire both for continued improvement enabled through assessment feedback not just marks and for assessments to test modern-day marketplace skills (e.g. distilling information, writing reports).   
  • Our academic representatives expressed their overarching concern to do the right thing by our students: seeking to understand and share methods and tools for designing assessments to help students in future life.
  • Our Digital Assessment team focussed on ways to build and share capabilities within faculties enabling us to connect academics, students and technologists and ways to understand and share how technology can support innovative assessment design both now and in the future.  
  • Like our students, our Faculty Learning Technologists focussed on the importance of feedback (or “feedforward”) which is easily accessible, timely and meaningful: enabling students to act upon it.  
  • Our Professional Services representatives focused on how to connect people to ensure that fantastic work around assessment design is shared widely and to ensure that academics and faculty teams are all aware of the tools and supporting resources available to them. 
  • Our UNIwise representatives were interested in all points raised: keen to consider how future enhancements to the AssessmentUCL platform might facilitate the continued evolvement of digital assessment. 

Group discussions were wide ranging and often raised more questions than answers but surfaced a clear desire to continue the conversations about the issues raised and to focus on how we share knowledge to maximise the depth of expertise in assessment design across departments. 

Next steps 

As a result of the discussion, Arena and the Digital Assessment team will focus their attention on the following key themes over the coming months:  

  1. Enabling assessment design knowledge sharing: Helping academics and others involved in assessment design to understand what is possible and how to achieve it. Ensuring clear information channels and networks are established to enable academics and others involved in assessment design to share experience and learn from the experience of others as well as raising awareness of existing resources and support.  
  2. Continuing to improve the markers journey: Exploring how to enable flexibility in marking where one may allocate all work to all markers and each takes one off the top of the pile and can go to the next unmarked (or not yet started to mark script), to ensure that there is no loss of marks when marking allocations change pre/post marking. 
  3. Continuing the conversation: Building on the foundations and links established during this event, we plan to set up a learning lab through which staff and students can continue the discussion around how we can design digital assessments differently and better using existing mechanisms like the Chart tool, as well as rethinking how assessments are delivered currently to make them more applicable to real-world situtations and careers.
  4. Working collaboratively with suppliers and academic colleagues to shape enhancements / design solutions to particular issues: We hope to connect key members in Departments with our suppliers Uniwise to workshop and work through desirability and current system functionalities.

If you would like to join events like this in the future, please let us know by contacting assessments@ucl.ac.uk.