Preprint
Communication

Children of AI: A Protocol for Managing the Born-Digital Ephemera Spawned by ChatGPT

Altmetrics

Downloads

659

Views

582

Comments

0

A peer-reviewed article of this preprint also exists.

This version is not peer-reviewed

Submitted:

28 July 2023

Posted:

31 July 2023

You are already at the latest version

Alerts
Abstract
The recent public release of the generative AI language model ChatGPT has captured the public imagination and has resulted a rapid uptake and widespread experimentation by the general public and academia alike. The number of academic publications focusing on the capabilities as well as practical and ethical implications of generative AI has been growing exponentially. One of the concerns with this unprecedented growth in scholarship related to generative AI, in particular ChatGPT, is that in most cases the raw data, that is the text of the original ‘conversations,’ have not been made available to the audience of the papers and thus cannot be drawn on to assess the veracity of the arguments made and the conclusions drawn therefrom. This paper provides a protocol for the documentation and archiving of these raw data.
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

1. Introduction

In recent months there has been wide-spread public attention regarding the use of artificial intelligence (AI) in various fields. The public releases of the image generator DALL-E and of the generative AI language model ChatGPT (Chat Generative Pre-trained Transformer) in early 2022 caught the public imagination. Since then, free-ranging debate emerged regarding the present and potential future abilities of generative AI, the dangers its may pose and the ethics of its usage. ChatGPT is type of deep learning model that uses transformer architecture to generate coherent and contextually relevant, human-like responses based on the input it receives [1].
Since its initial release in 2018, ChatGPT has undergone several revisions, mainly focusing on increased capabilities in providing longer segments of coherent text and contextual answering of questions including the addition of human preferences and feedback. The release of ChatGPT 2.0 in September 2019 relied on a training data set with 1.5 billion parameters, while the ChatGPT 3 (June 2020) was trained (by human trainers) on 175 billion parameters. ChatGPT 3.5 was released to the general public to encourage experimentation [2], with a temporal cut for the addition of its training data in September 2021.
ChatGPT has been shown to be capable of producing poetry [3], short stories and plays [4,5,6], English essays [7], as well as writing lines of code [8]. A growing number of studies has been looking at the effects of generative AI on education and academia, examining the level of knowledge and capabilities of ChatGPT as reflected in its responses to several field of academic endeavor, such as agriculture [9], archaeology [10], chemistry [11], computer programming [12], cultural heritage management [13], diabetes education [14], digital forensics [15], medicine [16,17,18,19,20,21,22,23], medical education [24], nursing education [25], and remote sensing [10]. ChatGPT is the typical double-edged sword presented by many new technologies [26]: both useful a potential tool to enhance student learning (e.g., [27,28,29]) given its ability to aid students in assignment writing and the associated potential misconduct (e.g., [30,31,32]).

2. The Problem

A primary mandate of academic publishing is ethical academic conduct. To ensure the integrity, transparency, and reproducibility of published research, the Committee on Publication Ethics (COPE) has issued a set of guidelines on the deposition and management of data and datasets that were used to explain and substantiate findings that are reported in academic publications [33]. In brief, this entails the deposition of research data in a common readable format in a curated, public, institutional or governmental data depository; the publication of the data and their collection methodology as a stand-alone data publication; or the appendage of the data as supplementary material to the article hosted on the publisher’s servers. This mandate is followed by and large, with some disciplines being more compliant than others (e.g. medical research).
At this point in time, academic research into ChatGPT, its abilities and limitations is expanding at rapid rate across most disciplines. The question arises how the research data associated with the publications are being managed.
The nature of ChatGPT as a generative AI tool means that each response to a given task will be different. While responses may be structurally and conceptually similar [34], they are not identical. Thus, it is not possible to recreate an identical or near identical response. Furthermore, all conversations with ChatGPT are deleted after 30 days to maintain server space [35]. Irrespective of this, all conversations with ChatGPT are virtual artefacts (sensu [36]) which, even when will eventually disappear due server upgrades or data warehouse restructuring. Consequently, the original conversations are equivalent to and should be treated like an experiment’s raw data and should be archived. At present, academic papers that have been written in relation to ChatGPT have taken five different approaches:
  • articles that include the entire conversation in the paper [9,10,11,21,24,34]
  • articles which quote extracts of the conversation in the text and provide the full text of the conversation as a supplementary document [13] or an appendix [37]
  • articles which quote extracts of the conversation in the text but does not provide access to the full text [15,22,23,25,29,31,38,39]
  • articles which discuss but not quote specific conversation(s) and do not provide access to the full text [30]; and
  • articles which discuss ChatGPT but do not refer to specific conversations but discuss the topic at a more abstract level [26,37].
Setting aside the first group, where the conversation makes up the core of the paper, and the second group where the full text is supplied as an appendix or a supplementary file, the other three groups do not allow readers to understand the full context of the conversation and cannot independently assess the validity of the author’s interpretation of the interaction.

3. Towards a Solution

3.1. Functional Considerations

While the conversations with ChatGPT have great similarity with formal interviews in anthropological, and ethnographical and sociological settings [40], they differ in on key respect. Human-to-human interviews are conducted in a linear fashion, with one question following on from, and building on, an answer. While, in theory, a human respondent could be asked to re-answer the question, this normally occurs with transitioning phrases [41] and a concomitant second-guessing by the respondent trying to taking cues from the interviewer why the previous answer was insufficient (else it would not have been asked in exactly the same way). ChatGPT, on the other hand, allows the human participant to request ChatGPT to answer the question again, generating a new response to the question. Thus any archiving of a conversation with ChatGPT needs to archive all “regenerations” of a question if they were conducted, while the paper needs to identify which instance of regeneration was being used or cited.
ChatGPT as a service is not static but is both being continually upgraded in terms of functionality and server performance and due to its ability to ‘learn.’ The latter is reinforced through user feedback which is solicited when a user tasks ChatGPT to regenerate a response and ChatGPT delivers the second version (Figure 1) and by OpenAI staff monitoring selected conversations (“Conversations may be reviewed by our Al trainers to improve our systems”).
Thus it can be posited that a ChatGPT is time-dependent, and that therefore the date and time of conversation should also be recorded, akin to the practice of stating the access date of web pages in standard referencing.

3.2. Ethical Considerations

Standard ethnographic research practice mandates that interviews are based on a mutual understanding of trust that the conversation is confidential and any interpretation of the conversation is carried out with expressed and informed consent, even though power dynamics and their changing nature need to be considered [42]. Irrespective, notebooks and interview transcripts were often deemed personal data, ‘owned’ by the respective researcher.
In recent years research ethics to mitigate against falsified research findings have led to the mandate to archive and make accessible the original or raw research data. In the space of qualitative research, this posed the ethical conundrum to allow access while at the same time maintaining the confidentiality of information provided in the interviews [43,44,45,46]. While this can be overcome by anonymizing or de-identifying the respondents, contextual information in the interviews may well allow for a re-identification of the informants [47,48,49]
In the current understanding, works created by generative AI do not accrue copyright for the AI system [50] as they fail to meet the human authorship requirement, but can generate copyright for the human actor in the interaction if the latter has substantive guiding involvement [51]. In the same vein, an ‘interview’ with ChatGPT differs from sociological or ethnographic interviews because generative AI is, at least at this point in time, not a sentient entity and thus cannot provide informed consent. It follows that, at lkeast at this point of legal understanding, ‘conversations’ with ChatGPT can be archived.

3.3. Protocol

We propose the following five-step process for ChatGPT data collection and archiving:
Step 1) record the metadata, comprised of version and version date, which can be found at the bottom of the interface (Figure 2), as well as the date and time the conversation occurred, using GMT as the standard.
Step 2) carry of the conversation as required.
Step 3) add the end time of the conversation to the metadata entry.
Step 4) Copy the text of the conversation into a text editor or word processor and save the file(s), making sure that all iterations of task generation (if any) are captured and identified as such (e.g. Regeneration 1/3, 2/3 etc.).
Step 5) Generate a complete data document that contains the metadata and text of each conversation.
Step 6) Submit the data document to a data repository of append it to the publication as a supplementary file.

4. Conclusions

Over the past decade, the development of generative artificial intelligence systems has accelerated dramatically, resulting in the recent public release of the generative AI language model ChatGPT. ChatGPT has captured the public imagination with widespread experimentation by academia and the general public alike. Numerous academic disciplines experimented with the capabilities of ChatGPT in relation to their research directions, examining its ability to provide accurate responses. The number of publications on the capabilities of ChatGPT and the practical and ethical implications of the use and abuse generative AI has been growing exponentially.
This unprecedented growth in scholarship related to generative AI, in particular ChatGPT, occurs in a large unregulated space, wherein in most cases the raw data, that is the text of the original ‘conversations,’ are not being made available to the audience of the papers. In consequence, they cannot be drawn on to assess the veracity of the arguments made in the publications and the conclusions drawn therefrom. This paper has provided a protocol for the documentation and archiving of these raw data.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. OpenAI. Introducing ChatGPT. Available online: https://openai.com/blog/chatgpt/ (accessed on June 28, 2023).
  2. Ray, P.P. ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope. Internet of Things and Cyber-Physical Systems 2023.
  3. Moons, P.; Van Bulck, L. ChatGPT: can artificial intelligence language models be of value for cardiovascular nurses and allied health professionals. European journal of cardiovascular nursing 2023. [Google Scholar] [CrossRef]
  4. Garrido-Merchán, E.C.; Arroyo-Barrigüete, J.L.; Gozalo-Brihuela, R. Simulating HP Lovecraft horror literature with the ChatGPT large language model. arXiv preprint, arXiv:2305.03429 2023.
  5. McGee, R.W. The Assassination of Hitler and Its Aftermath: A ChatGPT Short Story. Available at SSRN 4426338 2023. [Google Scholar]
  6. Landa-Blanco, M.; Flores, M.A.; Mercado, M. Human vs. AI Authorship: Does it Matter in Evaluating Creative Writing? A Pilot Study Using ChatGPT. 2023. [Google Scholar]
  7. Fitria, T.N. Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay. In Proceedings of the ELT Forum: Journal of English Language Teaching; 2023; pp. 44–58. [Google Scholar]
  8. Liu, J.; Xia, C.S.; Wang, Y.; Zhang, L. Is your code generated by chatgpt really correct? rigorous evaluation of large language models for code generation. arXiv preprint, arXiv:2305.01210 2023.
  9. Biswas, S. Importance of chat GPT in Agriculture: According to chat GPT. Available at SSRN 4405391 2023. [Google Scholar] [CrossRef]
  10. Agapiou, A.; Lysandrou, V. Interacting with the Artificial Intelligence (AI) Language Model ChatGPT: A Synopsis of Earth Observation and Remote Sensing in Archaeology. Heritage 2023, 6, 4072–4085. [Google Scholar] [CrossRef]
  11. Castro Nascimento, C.M.; Pimentel, A.S. Do Large Language Models Understand Chemistry? A Conversation with ChatGPT. Journal of Chemical Information and Modeling 2023, 63, 1649–1655. [Google Scholar] [CrossRef]
  12. Surameery, N.M.S.; Shakor, M.Y. Use chat gpt to solve programming bugs. International Journal of Information Technology & Computer Engineering (IJITC) ISSN: 2455-5290 2023, 3, 17–22. [Google Scholar]
  13. Spennemann, D.H.R. ChatGPT and the generation of digitally born “knowledge”: how does a generative AI language model interpret cultural heritage values? preprint.org 2023, 1–40. [Google Scholar] [CrossRef]
  14. Sng, G.G.R.; Tung, J.Y.M.; Lim, D.Y.Z.; Bee, Y.M. Potential and pitfalls of ChatGPT and natural-language artificial intelligence models for diabetes education. Diabetes Care 2023, 46, e103–e105. [Google Scholar] [CrossRef]
  15. Scanlon, M.; Breitinger, F.; Hargreaves, C.; Hilgert, J.-N.; Sheppard, J. ChatGPT for Digital Forensic Investigation: The Good, The Bad, and The Unknown. Preprints.org 2023. [Google Scholar] [CrossRef]
  16. King, M.R. The future of AI in medicine: a perspective from a Chatbot. Ann. Biomed. Eng. 2023, 51, 291–295. [Google Scholar] [CrossRef] [PubMed]
  17. Sarraju, A.; Bruemmer, D.; Van Iterson, E.; Cho, L.; Rodriguez, F.; Laffin, L. Appropriateness of Cardiovascular Disease Prevention Recommendations Obtained From a Popular Online Chat-Based Artificial Intelligence Model. JAMA 2023, 329, 842–844. [Google Scholar] [CrossRef] [PubMed]
  18. Bays, H.E.; Fitch, A.; Cuda, S.; Gonsahn-Bollie, S.; Rickey, E.; Hablutzel, J.; Coy, R.; Censani, M. Artificial intelligence and obesity management: An Obesity Medicine Association (OMA) Clinical Practice Statement (CPS) 2023. Obesity Pillars 2023, 6, 100065. [Google Scholar] [CrossRef]
  19. Grünebaum, A.; Chervenak, J.; Pollet, S.L.; Katz, A.; Chervenak, F.A. The exciting potential for ChatGPT in obstetrics and gynecology. Am. J. Obstet. Gynecol. 2023, 228, 696–705. [Google Scholar] [CrossRef]
  20. Rao, A.S.; Pang, M.; Kim, J.; Kamineni, M.; Lie, W.; Prasad, A.K.; Landman, A.; Dryer, K.; Succi, M.D. Assessing the utility of ChatGPT throughout the entire clinical workflow. medRxiv 2023, 2023.2002. 2021.23285886.
  21. Sabry Abdel-Messih, M.; Kamel Boulos, M.N. ChatGPT in Clinical Toxicology. JMIR Med Educ 2023, 9, e46876. [Google Scholar] [CrossRef] [PubMed]
  22. Zhu, Y.; Han, D.; Chen, S.; Zeng, F.; Wang, C. How Can ChatGPT Benefit Pharmacy: A Case Report on Review Writing. Preprints.org 2023. [Google Scholar] [CrossRef]
  23. Haver, H.L.; Ambinder, E.B.; Bahl, M.; Oluyemi, E.T.; Jeudy, J.; Yi, P.H. Appropriateness of Breast Cancer Prevention and Screening Recommendations Provided by ChatGPT. Radiology 2023, 307, e230424. [Google Scholar] [CrossRef]
  24. Eysenbach, G. The Role of ChatGPT, Generative Language Models, and Artificial Intelligence in Medical Education: A Conversation With ChatGPT and a Call for Papers. JMIR Med Educ 2023, 9, e46885. [Google Scholar] [CrossRef]
  25. Qi, X.; Zhu, Z.; Wu, B. The promise and peril of ChatGPT in geriatric nursing education: What We know and do not know. Aging and Health Research 2023, 3, 100136. [Google Scholar] [CrossRef]
  26. Malik, T.; Dwivedi, Y.; Kshetri, N.; Hughes, L.; Slade, E.L.; Jeyaraj, A.; Kar, A.K.; Baabdullah, A.M.; Koohang, A.; Raghavan, V. “So what if ChatGPT wrote it?” Multidisciplinary perspectives on opportunities, challenges and implications of generative conversational AI for research, practice and policy. International Journal of Information Management 2023, 71, 102642. [Google Scholar]
  27. Khan, R.A.; Jawaid, M.; Khan, A.R.; Sajjad, M. ChatGPT-Reshaping medical education and clinical management. Pakistan Journal of Medical Sciences 2023, 39, 605. [Google Scholar] [CrossRef] [PubMed]
  28. Lim, W.M.; Gunasekara, A.; Pallant, J.L.; Pallant, J.I.; Pechenkina, E. Generative AI and the future of education: Ragnarök or reformation? A paradoxical perspective from management educators. The International Journal of Management Education 2023, 21, 100790. [Google Scholar] [CrossRef]
  29. Rudolph, J.; Tan, S.; Tan, S. ChatGPT: Bullshit spewer or the end of traditional assessments in higher education? Journal of Applied Learning and Teaching 2023, 6. [Google Scholar]
  30. Ali, K.; Barhom, N.; Marino, F.T.; Duggal, M. The Thrills and Chills of ChatGPT: Implications for Assessments in Undergraduate Dental Education. Preprints.org 2023, 2023020513. [Google Scholar] [CrossRef]
  31. Currie, G.; Singh, C.; Nelson, T.; Nabasenja, C.; Al-Hayek, Y.; Spuur, K. ChatGPT in medical imaging higher education. Radiography 2023, 29, 792–799. [Google Scholar] [CrossRef]
  32. Stokel-Walker, C. AI bot ChatGPT writes smart essays-should academics worry? Nature 2022. [Google Scholar] [CrossRef]
  33. COPE. Data and reproducibility. Available online: https://publicationethics.org/data (accessed on.
  34. Spennemann, D.H.R. Exhibiting the Heritage of Covid-19—a Conversation with ChatGPT. preprint.org 2023, 1–20. [Google Scholar] [CrossRef]
  35. J, J. Data Controls FAQ. Available online: https://help.openai.com/en/articles/7730893-data-controls-faq (accessed on July 21, 2023).
  36. Spennemann, D.H.R. The Digital Heritage of the battle to contain COVID-19 in Australia and its implications for Heritage Studies. Heritage 2023, 6, 3864–3884. [Google Scholar] [CrossRef]
  37. Shen, Y.; Heacock, L.; Elias, J.; Hentel, K.D.; Reig, B.; Shih, G.; Moy, L. ChatGPT and Other Large Language Models Are Double-edged Swords. Radiology 2023, 307, e230163. [Google Scholar] [CrossRef]
  38. Gilson, A.; Safranek, C.W.; Huang, T.; Socrates, V.; Chi, L.; Taylor, R.A.; Chartash, D. How Does ChatGPT Perform on the United States Medical Licensing Examination? The Implications of Large Language Models for Medical Education and Knowledge Assessment. JMIR Med Educ 2023, 9, e45312. [Google Scholar] [CrossRef]
  39. Farhat, F. ChatGPT as a Complementary Mental Health Resource: A Boon or a Bane. Preprints.org 2023. [Google Scholar] [CrossRef] [PubMed]
  40. Denzin, N.K.; Lincoln, Y.S. The Sage handbook of qualitative research; sage: 2011.
  41. Sarantakos, S. Social research, 4th ed.; Macmillan International Higher Education: Basingstoke, 2012. [Google Scholar]
  42. Russell, L.; Barley, R. Ethnography, ethics and ownership of data. Ethnography 2020, 21, 5–25. [Google Scholar] [CrossRef]
  43. Feldman, S.; Shaw, L. The epistemological and ethical challenges of archiving and sharing qualitative data. American Behavioral Scientist 2019, 63, 699–721. [Google Scholar] [CrossRef]
  44. Kuula, A. Methodological and ethical dilemmas of archiving qualitative data. Iassist Quarterly 2011, 34, 12–12. [Google Scholar] [CrossRef]
  45. Reeves, J.; Treharne, G.J.; Ratima, M.; Theodore, R.; Edwards, W.; Poulton, R. A one-size-fits-all approach to data-sharing will not suffice in lifecourse research: a grounded theory study of data-sharing from the perspective of participants in a 50-year-old lifecourse study about health and development. BMC Medical Research Methodology 2023, 23, 118. [Google Scholar] [CrossRef] [PubMed]
  46. Richardson, J.C.; Godfrey, B.S. Towards ethical practice in the use of archived transcripted interviews. Int. J. Social Research Methodology 2003, 6, 347–355. [Google Scholar] [CrossRef]
  47. Cecaj, A.; Mamei, M.; Bicocchi, N. Re-identification of anonymized CDR datasets using social network data. In Proceedings of the 2014 IEEE International Conference on Pervasive Computing and Communication Workshops (PERCOM WORKSHOPS); 2014; pp. 237–242. [Google Scholar]
  48. Bandara, P.K.; Bandara, H.D.; Fernando, S. Evaluation of re-identification risks in data anonymization techniques based on population uniqueness. In Proceedings of the 2020 5th International Conference on Information Technology Research (ICITR); 2020; pp. 1–5. [Google Scholar]
  49. Larbi, I.B.C.; Burchardt, A.; Roller, R. Clinical Text Anonymization, its Influence on Downstream NLP Tasks and the Risk of Re-Identification. In Proceedings of the Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics: Student Research Workshop, 2023; pp. 105–111.
  50. Guadamuz, A. Artificial intelligence and copyright. Wipo Magazine 2017, 14–19. [Google Scholar]
  51. Copyright Office. Copyright Registration Guidance: Works Containing Material Generated by Artificial Intelligence. Available online: https://www.federalregister.gov/documents/2023/03/16/2023-05321/copyright-registration-guidance-works-containing-material-generated-by-artificial-intelligence (accessed on July 23, 2023).
Figure 1. Request for user feedback by ChatGPT following the provision of a regenerated response.
Figure 1. Request for user feedback by ChatGPT following the provision of a regenerated response.
Preprints 80816 g001
Figure 2. Version date (as shown in the footer of the interaction window)
Figure 2. Version date (as shown in the footer of the interaction window)
Preprints 80816 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated