AI-powered chatbots, designed ethically, can support high-quality university teaching

AI-powered chatbots, designed ethically, can support high-quality university teaching
Chatbots could take over the majority of low-level guidance tasks fielded by staff in teaching and learning centres to free them up for where in-person support is most needed. Credit: Shutterstock

While COVID-19 forced an emergency transformation to online learning at universities, learning how to teach efficiently and effectively online using different platforms and tools is a positive addition to education and is here to stay.

To sustain this beneficial evolution and ensure quality education, universities should focus on supporting to embrace and lead the change.

The ethical and strategic use of artificial intelligence at centers of and learning, which support faculty in troubleshooting and innovating their online teaching practices, can help with this task. Centers of teaching and learning are responsible for educational technology support, teaching and learning support, as well as .

Expansive move to online education

Research conducted at 19 centers of teaching and learning and their equivalents from Canada, the United States, Lebanon, the United Kingdom and France published in August 2020 showed that staff in these centers deployed all available resources to support the rapid switch to online education.

Staff had been working 10- to 14-hour workdays during the first phase of the pandemic to meet the increase in faculty and staff needs. These centers also reported difficulty recruiting and training qualified candidates.

Used strategically, chatbots could take over repetitive low-level guidance tasks that teaching and learning centers field and help avoid overload. A chatbot, also called a conversational or virtual agent, is a software or computer system designed to communicate with humans using natural language processing.

This communication can be via text messages or voice commands.

Why a chatbot?

Chatbots offer a viable, win-win solution to teaching and learning centers and to faculty. They are available 24/7, can respond to thousands of simultaneous requests and provide instant and robust service support when needed.

Using chatbots could free teams for complex inquiries that require human interventions, such as transforming teaching approaches and collaborating to innovate solutions to respond to problems like improving equity and access in online teaching.

A collaboration between the centers' experts and technology could provide better services and support for faculty to improve the learning experiences they create for students. Chatbots can guide faculty towards appropriate and effective resources and professional development activities, such as how-to articles, tutorials and upcoming workshops. These would be tailored to suit faculties' individual needs, their varied digital skills levels and backgrounds in designing hybrid learning experiences.

Chatbot systems are already used in educational institutions for teaching and learning, to deliver administrative tasks, to advise students and assist them in research.

How would it work?

Two options are possible when it comes to chatbots' AI conversational ability:

  1. Artificial Intelligence Markup Language methodology: Programmers give the AI a library of questions/answers and keyword associations through a database. From there, the chatbot is able to give appropriate answers in a strictly defined frame.
  2. The Natural Language Processing approach: This allows for more flexibility. Once programmers build an initial dataset, the AI-powered tool will then learn from ongoing exchanges to find the best combination of answers to recurring questions asked by faculty members. The AI will then be able to identify keywords in a sentence and understand the context of a question.
AI-powered chatbots, designed ethically, can support high-quality university teaching
Chatbot implementation scenario for teaching and learning centres to support faculty. Credit: Nadia Naffi

That programmers would need to add data from the conversation to an ongoing dataset building throughout time is expected. When asked a question, the chatbot will respond based on its current knowledge database. If the conversation introduces a concept that it isn't programmed to understand, the chatbot can state it doesn't understand the question—or pass the communication to a human operator.

Either way, the chatbot will also learn from this interaction as well as future interactions. Thus, the will gradually grow in scope and gain relevance.

For chatbots' reliability and trustworthiness to increase, it should be effective in helping these centers supporting their faculty. Implementing and fine tuning chatbots so they are ready for use is important, even if it requires an investment of time and resources.

Ethical framework for AI in education

The Institute for Ethical AI in Education, based at the University of Buckingham in the United Kingdom, and funded by McGraw Hill, Microsoft Corporation, Nord Anglia Education and Pearson PLC, released The Ethical Framework for AI in Education in 2020. The framework argues AI systems should increase the capacity of organizations and the autonomy of learners while respecting human relationships and ensuring human control.

Chatbots in university settings should be ethical by design, meaning that they should be designed to be sensitive to values like security, safety and accountability and transparency. If used in centers of teaching and learning, users should be protected against all forms of harm or abuse. They also need to feel treated fairly and to always be provided the option to reach a human. Faculty members must know they are exchanging with an AI.

Chatbots can and should be accessible. Tolerating user errors and input variation, being designed for diverse abilities and allowing multilingual texting communication are examples of facilitating accessibility.

Do no harm: privacy

Chatbots should be designed to "do no harm," as per the UNESCO's recent recommendations on the ethics of artificial intelligence. When talking about non-maleficence, privacy should be addressed.

AI-powered tools come with data recording issues. Strong barriers in data collection and storage are needed. Following the European approach to data protection, centers should minimize data collection. Only required information should be stored, such as specific parts of conversations, but not the interlocutors' identity.

The transparency-based approach allows for users to agree on which personal data can be shared or not with the centers. This would help keep trust and usability of the tool high. In case of malfunction, faculty members would provide feedback on the problem, and centers would fix it.

Centers should consider anonymization of users, strong encryption of all data stored, and in-house storage when possible or by a trusted contracted third party following similar data privacy rules.

Addressing bias, environmental impact

The possible bias in the initial database needs to be addressed. Whether it relates to gender, ethnicity, language or other variables, the initial dataset needs to be cleaned and carefully analyzed prior to being used to train the AI, whether AI markup language or NLP methods are deployed. If the latter is applied, ongoing monitoring should be considered.

A green data storage solution needs to be addressed to reduce the CO2 cost of activity on the environment. For example, universities might investigate if water cooling systems could be used in server rooms instead of air conditioning.

The university of the future as anticipated by many scholars and policy makers has already started. Technology, if used ethically and strategically, can support faculty in their mission to prepare their students for the needs of our society and the future of work.

Provided by The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation: AI-powered chatbots, designed ethically, can support high-quality university teaching (2022, January 3) retrieved 8 February 2023 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

'Sorry, I don't understand that': The trouble with chatbots and how to use them better


Feedback to editors