Even as artificial intelligence becomes an integral tool for students and academics, less than half of the top 50 universities worldwide have developed publicly available guidelines for its use in academic settings, a study has found.
Only 23 of the 50 institutions, which were chosen based on their performance on Times Higher Education’s 2023 World University Rankings, have listed on their webpages clear directions for instructors on the use of generative AI (GAI) tools in assessments – a “concerning” finding, according to Benjamin Moorhouse, an assistant professor of education at Hong Kong Baptist University who carried out the study, and his co-authors.
“Without clear guidelines, instructors may take a defensive approach to GAI and adopt more in-class assessments or feel frustrated as they struggle to adapt their assessment practices without institutional guidance,” the researchers warn.
Believed to be the first such review of its kind, the study in Computers and Education Open examines the extent to which institutions have created guidance that helps “raise awareness about academic integrity and reduce academic misconduct” amid the broad uptake of game-changing new technology in higher education.
It comes as universities around the globe grapple with how to adapt to the arrival of ChatGPT, which has already changed the nature of coursework and assessments for students and those teaching them. In the 12 months since the introduction of the technology, which prompted bans at some institutions, such tools have become near-ubiquitous in academia.
Campus resource collection: AI transformers like ChatGPT are here, so what next?
Speaking with THE, Dr Moorhouse cautioned that lack of a clear stance on the tools can put universities “in a difficult position”, leaving faculty flying blind even as many learners use AI in their coursework.
“I was talking to colleagues at another institution, which has no firm guidelines, and they’re struggling to communicate with students on this,” he said.
So why do so many institutions still lack public guidelines on AI? Dr Moorhouse believed some may be taking a “wait and see” approach.
“It can be hard to be first – also people weren’t sure of the exact effect GAI was going to have until we went through one semester cycle of it,” he said.
Dr Moorhouse noted that there are some universities that have created guidelines, but do not make them publicly available, HKBU included. He said it was not always clear why institutions refrained from making internal policies public, but believed this could be to do with the “sensitivity” around assessment design.
“They might be concerned [about] things in docs they don’t want students to know about, things around AI detection tools,” he said.
While he recognised the need to proceed with caution, he hoped institutions would be more transparent about their policies.
“It’s going to be really important that university instructors become AI-literate,” he said. “We do need to gain those skill sets and our institutions need to support us.”