This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

proofread

Social scientists recommend addressing ChatGPT's ethical challenges before using it for research

chatgpt
Credit: Unsplash/CC0 Public Domain

A new paper by researchers at Penn's School of Social Policy & Practice (SP2) and Penn's Annenberg School for Communication offers recommendations to ensure the ethical use of artificial intelligence resources such as ChatGPT by social work scientists.

Published in the Journal of the Society for Social Work and Research, the article was co-written by Dr. Desmond Upton Patton, Dr. Aviv Landau, and Dr. Siva Mathiyazhagan. Patton, a pioneer in the interdisciplinary fusion of social work, communications, and , holds joint appointments at Annenberg and SP2 as the Brian and Randi Schwartz University Professor.

Outlining challenges that ChatGPT and other large language models (LLMs) pose across bias, legality, ethics, , confidentiality, informed consent, and , the piece provides recommendations in five areas for ethical use of the technology:

  • Transparency: Academic writing must disclose how content is generated and by whom.
  • Fact-checking: Academic writing must verify information and cite sources.
  • Authorship: Social work scientists must retain authorship while using AI tools to support their work.
  • Anti-plagiarism: Idea owners and content authors should be located and cited.
  • Inclusion and social justice: Anti-racist frameworks and approaches should be developed to counteract potential biases of LMMs against authors who are Black, Indigenous, or people of color, and authors from the Global South.

Of particular concern to the authors are the limitations of artificial intelligence in the context of human rights and . "Similar to a bureaucratic system, ChatGPT enforces thought without compassion, reason, speculation, or imagination," the authors write.

Pointing to the implications of a model trained on existing content, they state, "This could lead to bias, especially if the text used to train it does not represent diverse perspectives or scholarship by under-represented groups. . . . Further, the model generates text by predicting the next word based on the previous words. Thus, it could amplify and perpetuate existing bias based on race, gender, sexuality, ability, caste, and other identities."

Noting ChatGPT's potential for use in research assistance, theme generation, data editing, and presentation development, the authors describe the chatbot as "best suited to serve as an assistive tech tool for scientists."

More information: Desmond Upton Patton et al, ChatGPT for Social Work Science: Ethical Challenges and Opportunities, Journal of the Society for Social Work and Research (2023). DOI: 10.1086/726042

Citation: Social scientists recommend addressing ChatGPT's ethical challenges before using it for research (2023, August 17) retrieved 27 April 2024 from https://phys.org/news/2023-08-social-scientists-chatgpt-ethical.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Tackling the ethical dilemma of responsibility in large language models

3 shares

Feedback to editors