Finds documents with both search terms in any word order, permitting "n" words as a maximum distance between them. Best choose between 15 and 30 (e.g. NEAR(recruit, professionals, 20)).
Finds documents with the search term in word versions or composites. The asterisk * marks whether you wish them BEFORE, BEHIND, or BEFORE and BEHIND the search term (e.g. lightweight*, *lightweight, *lightweight*).
This article delves into the use of AI, particularly large language models like ChatGPT, in genetics classrooms. It discusses the potential of AI as an educational tool, emphasizing its role in creative writing assignments. The study explores how AI can assist in student learning without replacing traditional methods, and how it can be used to address ethical concerns in genetics research. The research also examines student perceptions and concerns about AI, demonstrating the need for guidelines and ethical considerations in its implementation. The article provides insights into the impact of AI on student learning and the importance of fostering critical thinking skills in the age of AI.
AI Generated
This summary of the content was generated with the help of AI.
Abstract
Since it became available free to the public in November of 2022, ChatGPT and other large-language model AIs have impacted the higher education classroom. Some fear that this is the end of essay-based assignments, as these are easily generated by ChatGPT. Principles of Genetics, an entry level genetics class, has previously incorporated a creative writing assignment designed to prompt students to engage with the ethics surrounding genetics in a non-traditional manner. The overarching aims of this study are the following: (1) educate students on the ethical use of AI, (2) gauge how they feel AI impacts their education using pre and post surveys, and (3) develop a creative writing assignment that can co-exist with ChatGPT while allowing students to explore its strengths and weaknesses. To this end, we collected pre- and post-surveys from 191 students in the genetics class. We used a mixed method analysis (quantitative and qualitative) to analyze students’ ratings and explanations. We found that after the study, students were less likely to use AI to write an entire paper or essay, but students also reported more positivity toward AI as a tool. Additionally, we present suggestions from students as to how they feel AI should be used in the classroom. This study and the activity used in the classroom provide an exploration of AI use in the classroom for educators to expand upon.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Introduction
Artificial intelligence (AI) is a broad term, used to encompass a multitude of different complex systems. The name implies a higher level of thinking than these complex systems are currently capable of, as many of their functionalities are supported by humans on the back end. For the purposes of educational research, AI that use a large language model are the most relevant. A simplified explanation for how a large language model (LLM) works is that the model learns to assign statistical relationships between words using training data (Radford et al., 2019). By processing paragraphs, or larger pieces of text, as a single entity, it is able to assign context to words and then predict their usage in text. This is how AI such as Chat Generative Pre-Trained Transformer (ChatGPT) are able to produce text that sounds as if it has been written by a live person.
ChatGPT (OpenAI, 2023) is one such LLM based on the GPT-3 algorithm; however, there are many more like it as companies race to produce their own versions. Further iterations of the model, such as GPT-4 and WebGPT (OpenAI, 2023), which enables the chatbot interface to browse the web for answers, are already in development. These large language models are trained on data found on the internet, such as books, websites, and articles, which raises questions about copyright that has not yet been resolved in the USA or Europe at the time this study was published. A related development in AI is that of generative AI, which are AI models that produce novel outputs such as images, music, voices, or text (Gozalo-Brizuela & Garrido-Merchan, 2023). While this is sometimes considered separate from LLMs, the two are intertwined and will likely continue to be so as AI continues to rapidly develop.
Advertisement
When ChatGPT debuted, educators immediately decried the end of writing assignments as we know them (McMurtie, 2022; Rudolph et al., 2023). Many traditional types of assessments, such as essays, are easily generated with an LLM. Other educators have argued that perhaps, it is time to change the way we assess student learning (Stokel-Walker, 2022). This is not a shift that can happen immediately. Instructors are unsurprisingly hesitant to incorporate AI into the classroom (Kim & Kim, 2022), which would require overhauling assignments that they have perfected over many semesters. Implementing new assignments takes time and effort, and is subject to trial and error. This reflects an overall reluctance of some instructors to incorporate any sort of educational technology into the classroom (Tallvid, 2016; Wood et al., 2005). As of yet, the movement to develop assignments suitable for an age of AI is still in its infancy. The impact of ChatGPT is not, however, exclusively detrimental (Baidoo-Anu & Owusu Ansah, 2023; Crawford et al., 2023; Lo, 2023). ChatGPT also offers opportunities to assist in student learning in a way that does not preclude preexisting methods, such as by providing feedback, generating practice problems, and more (Fuchs, 2023; Kasneci et al., 2023).
Advancements in technology have impacted education previously, such as with the invention of calculators (Milou, 1999; Sheets, 2007; Simonsen & Dick, 1997; Wheatley, 1980). Trying to beat the technology with AI detectors or other similar programs creates an arms race between the AI and the detector that is untenable, and these AI detection programs have a high rate of type I error (Dalalah & Dalalah, 2023; Uzun, 2023). There has been progress on this front, as researchers have developed programs for detecting AI-generated scientific writing, but these programs are not widely applicable (Desaire et al., 2023). An additional challenge is that these AI detectors are biased toward non-native English writers (Liang et al., 2023). Many universities have backed off using software to detect AI (Ghaffary, 2023). Notably, one instructor erroneously declared that an entire class used ChatGPT for their final assignment after they pasted each essay into ChatGPT and asked the chat-bot if it had produced the essay (Uwa Ede-Osifo, 2023). Due to a shortcoming of the AI, ChatGPT confirmed that it had written each essay as it had not actually written all the essays. Exploiting loopholes, such as ChatGPT’s problems with generating correct references, is a short-term solution. LLMs will continue to improve, though they will likely never be perfect, and any attempt to use what is now a weakness is simply a band-aid solution. Rather than attempt to prevent its usage, finding ways to adapt our classrooms to AI with careful attention to its strengths and weaknesses is a better strategy (Cardona et al., 2023).
Educators are adaptable and have begun to develop new approaches for teaching with AI. One popular approach is to use AI as an exercise in detecting mistakes. Already literature exists on how one might use this approach and other similar exercises in the classroom (Bitzenbauer, 2023; Mogali, 2023; Naumova, 2023; Tlili et al., 2023; Yang, 2023). The basic principle is to prompt critical thinking in students by having them evaluate outputs of AI for errors. This encourages students not to take these outputs at face value, and instead determine for themselves whether all aspects of the output are accurate. Particularly important is ensuring that students know to look for bias in these outputs, as an AI is only as good as its training data, and there are many implicit biases (Srinivasan & Chander, 2021; Varsha, 2023). Assignments can be designed to point students toward some of these biases directly.
The best solution is to design assignments for which student outcomes do not change even if they use AI to complete the assignment (Ross, 2023). This is a challenging type of assignment to develop, however. In the short-term, it is better to utilize AI in the classroom in a way that allows students to incorporate it into their skillset as a tool, and give them the knowledge of how best to use it and the wisdom to know when it is not appropriate. When developing these assignments, instructors should keep in mind the greatest challenges AI presents to education, which are outlined by Fuchs (2023): accuracy that is dependent on the input data, overreliance on technology that results in a lack of critical thinking, and AI’s shortcomings in the areas of grammar and linguistics. In this study, we implemented a creative writing assignment in an introductory genetics course designed to highlight some of these challenges for students. This provides an early exploration of a still developing challenge in the classroom and a possible starting point for educators who are unsure of how to adapt.
Advertisement
Creative Writing as a Teaching Tool in STEM
Creativity, regardless of the way it is defined, is essential to science, and fostering this in students will set them up to better succeed in the future (Gaspar & Mabic, 2015; Georgiou et al., 2022). Creative writing offers students an opportunity to process knowledge in a non-traditional manner, such as connecting topics in a way that they might not have done with traditional teaching methods (Bansal, 2023; Ostrom et al., 2020; Watkins & Tehrani, 2020). Creative writing has also been found to be a motivating factor for students. Students tend to become more invested in work that they feel a connection to (Nicholes, 2022; Ostrom et al., 2020). While it has been found that creativity and creative writing in an academic setting can induce anxiety in students, especially those in the sciences who erroneously perceive there to be only right and wrong answers, there is value to encouraging students to overcome these limitations (Daker et al., 2020).
Creative writing is also an ideal arena for interrogating the relationship between science and society. Students are infrequently encouraged to bring their personal circumstances into traditional undergraduate coursework; however, it is these circumstances that inform their science. Additionally, creative writing provides students with an opportunity to investigate different perspectives of a problem and develop empathy for situations that are unfamiliar to them (Nicholes, 2020; Watkins & Tehrani, 2020). This aligns well with introducing STEM students to ethical issues, and has been utilized in areas such as engineering to achieve a level of connection between students and the ethical issue that was not well accomplished with case studies (Atwood & Read-Daily, 2015; Jedlicka, 2020). This opportunity for deeper ethical thinking in STEM classes has been expanded to include fictional frameworks and directed thinking (Hansen, 2021).
Studies have shown that science students perceive creative writing as being subjective and belonging more to the arts and humanities (Diluzio & Congdon, 2015; Nicholes, 2020). Students often perceive science writing as belonging to a category of its own, one unrelated to that of the writing typically associated with non-science areas (Nicholes, 2020). Despite student perceptions, creative writing and science writing are not completely independent and developing either kind of writing skill is beneficial. Furthermore, students seem to underestimate the role of writing in their future careers as scientists (Wheeler et al., 2023), and this is particularly true of biology students at the universities where this study was conducted.
In one study (Marbach-Ad et al., 2016, 2019), biological science seniors (N = 1324) were asked before graduation to rate the importance of acquiring different skills (such as creativity, scientific writing, group work) during their undergraduate studies on a 5-level Likert scale (1 = not important, 2 = slightly important; 3 = fairly important; 4 = important; 5 = very important). Only 67% of the respondents rated scientific writing as important or very important to acquire during their undergraduate studies. During interviews, one student explained that they were planning to pursue a medical career, and therefore, they would not need this skill. In comparison to students responses, 85% of the biological sciences instructors (N = 52) rated this skill as important or very important for students to acquire (Marbach-Ad et al., 2019). One would expect that all faculty members in the biological science would rate scientific writing as highly important. In a focus group, faculty members provided the following explanation: Incorporating writing assignments into lower-level STEM courses is challenging due to the large number of students in these courses, which limits the instructor’s ability to grade and provide feedback (Marbach-Ad et al., 2019).
Indeed, when students were asked to rate how often they were exposed to scientific writing in their undergraduate studies (1 = none of my courses; 2 = a few of my courses; 3 = some of my courses, 4 = most of my courses, or 5 = almost in all of my courses), only 67% responded that they were engaged in scientific writing in some, most, or almost in all of their courses. Despite this, it is clear that there is value in incorporating writing assignments into the curriculum (Gillen et al., 2020). Science communication is a critical component of a career in science. One must not only be able to conduct experiments, but also to convey results to a broader audience. The ability to communicate via writing is also valued by employers of all sectors (e.g., policy, industry, government) should students choose to pursue a non-academic career in science, but STEM students have been falling short of employer expectations (American Association of Advancement of Science [AAAS], 2011; Hart Research Associates, 2015). While scientific writing is important for long term student outcomes as shown in the studies above, it is just as critical that STEM students experience creativity and writing not only through general education or English courses but also in the context of their STEM area of interest. In this way, creative writing is an opportunity to span disciplines and produce more well-rounded students (Baram-Tsabari & Osborne, 2015; Diluzio & Congdon, 2015; Nicholes, 2022).
The creative writing assignment (Supplemental Material Section I) utilized for this study that was developed by the course instructor has several goals. The first is to encourage students to think critically about issues of ethics related to genetics. As a mandatory course for biological science students, this class contains many students who aspire to a medicine-related career, and who will perhaps in the future be confronted with some ethical issues directly. The second goal is for students to develop their science communication skills. Finally, the students are required to incorporate some aspect of genetics into their creative writing, which fulfills the purpose of fostering learning through creative writing.
Many theoretical papers have been published to try to address the impact of AI on the classroom (Cooper, 2023; Firat, 2023; Naumova, 2023; Yu, 2023); however, few empirical studies have been formerly conducted in the classroom. The aim of this study was to educate students on the ethical use of AI, gauge how they feel AI impacts their education using pre- and post-surveys, and develop a creative writing assignment that can co-exist with AI while allowing students to explore both its strengths and weaknesses, as well as interrogate its biases. This study provides preliminary insights into student perceptions of AI in the classroom, and whether they feel that it benefits or hurts their education. Additionally, it explores the functionality of AI as an educational tool and exposes weaknesses as well as opportunities to better student writing. We have framed these goals into the following research questions:
1)
To what extent did students report on awareness/knowledge/use of AI prior to the course?
2)
To what extent did students believe that AI is a useful tool prior to and after the class assignment? What are their concerns?
3)
To what extent did student perceptions of the use of AI change following the assignment?
4)
How effective was the assignment at promoting student learning about AI and ethical issues in genetics?
Methods
Context of the Study
This study was conducted at a large research-forward land grant university on the East Coast. For this study we recruited students in the BSCI222 Principles of Genetics class as subjects. Principles of Genetics is an entry level genetics class required for progression in the biology major and is taken primarily by sophomores, juniors, and seniors from biology and biology-adjacent majors. Only a small number of freshmen were enrolled, since in order to take the class, students must complete two biology and two chemistry courses as pre-requisites. This class has previously had a creative writing assignment designed to prompt students to engage with the ethics surrounding genetics outside of traditional lectures. This 4-credit class generally has 150–200 students enrolled. The class meets twice a week for 75 min lectures as well as in multiple sections once a week for a 120 min recitation/discussion section (in groups of 20–25 students) led by a teaching assistant (TA). For the semester in which this study took place, initial enrollment in the class was 194 students. Data from the registrar on the class at the beginning of the semester showed that it was predominantly female (78%, 151 students), and the most common ethnicities were white or Caucasian ethnicity (43%, 83 students), Asian (27%, 53 students), and Black or African American (16%, 31 students) (Table 1).
Table 1
Class-wide demographics
Gender
# of students
Female
151
Male
43
Race/ethnicity/citizenship
# of students
Asian
53
Black or African American
31
Hispanic/Latino
14
Two or more races
5
US nonresident
3
Unknown
5
White
83
The students were asked to participate in a pre-survey at the beginning of the semester after the drop/add period closed, to avoid recruiting students who would not remain in the class. The students were then asked to generate a first draft of their creative writing assignment using ChatGPT (or an AI like it, which will be inferred for the rest of this study). Once they generated their first draft, they were required to complete a self-reflection worksheet (Supplemental Material Section IV) designed to encourage them to think critically about the output from the chat-bot. This reflection worksheet was designed to have two purposes. The first was to guide student interactions with the AI by directing them to provide it multiple prompts, and to adjust these prompts to get different results. The second was to have students interrogate the AI output in terms of factual accuracy, quality of writing, ability to generate legitimate sources, and evidence of biases in the AI. Students then exchanged their writing during recitation and provided feedback to their peers using a guided worksheet (Supplemental Material Section V). This worksheet was similar to the self-reflection and focused on further assessing the quality of the output. These two worksheets served the purpose of guiding student use of AI as well as directing them to focus on areas where AI might fall short. During this time period, students were encouraged but not required to complete an informative online module on AI developed by the university’s teaching center and library. This module informed students on AI ethics issues, how to cite AI, and potential uses of AI tools.
After the writing was critiqued both by themselves and a peer, the students were then asked to re-write their writing assignment without using ChatGPT based on the feedback they received. This was then considered the final draft and graded by the course TAs. After the writing assignment was turned in, the students took the post-survey. Both the pre- and post-surveys were administered online using Qualtrics during their recitation section to encourage participation and to ensure that the data remained confidential. The study concluded with an in-class discussion with the students on the assignment and AI. The timeline and progression of the study are summarized in Fig. 1.
The data collected for this study was in the form of two surveys that included 12 or 13 questions each (Supplemental Material Sections I and II) and were administered before and after the creative writing assignment. Face validity of the novel survey was established using a science education faculty member (2nd author), a graduate research assistant (first author), the senior graduate teaching assistant, an education development specialist, and the instructor of the course (senior author). Survey questions were negotiated among these parties until one hundred percent agreement was reached. The pre-survey started with a five-sentence explanation of AI to ensure that students had at least some minimal background on the topic before proceeding with the survey. The surveys utilized both Likert-style and open-ended questions. Some questions overlapped between the two surveys to allow change in perception and knowledge to be assessed. For most questions students were prompted to provide further detail as to why they selected the Likert-style response. Other questions did not overlap and were designed to understand how the students interact with the course assignment as well as prompt further reflection. We compared change in perception in the class to see if education and hands on experience had an effect on students. In addition to the student surveys, we also had the opportunity to formally interview the graduate-level TAs for this class to assess student perceptions.
The pre-survey was completed by 191 students (98% of the students in class at the time of delivery) and the post-survey was completed by 173 students (93% of the students in class at the time of delivery, n = 186 students). This represented the majority of the class, and the distribution of students across majors for pre and post survey participants can be found in Table 2. As most students took both surveys, this is a representative sampling of the class demographics. The class is almost evenly split between biology majors that are broken into specific concentration areas, and public health science majors. The majority of the participants in both surveys were juniors (Table 3). Additionally, there are a small number of freshmen also enrolled in the class. The post-survey had slightly less participation that the pre-survey. The decrease in participation likely can be accounted for by students dropping the class, and student fatigue as this survey took place later in the semester.
Table 2
Distribution of responses by student-reported major
Major
Pre
Post
Biochemistry
5
4
Bioengineering
8
4
Cell Biology and Genetics
7
5
Chemistry
1
1
Ecology and Evolution
1
1
Environmental Science and Policy
2
3
Environmental Science and Technology
6
4
General Biology
30
29
Horticulture
1
2
Kinesiology
4
4
Letters and Sciences
3
3
Microbiology
3
2
Nutritional Science
1
0
Neuroscience
23
20
Physiology and Neurobiology
13
11
Plant Biology
0
1
Psychology
5
3
Public Health Science
77
76
Wildlife Ecology and Management
1
0
Total
191
173
Table 3
Distribution of responses by student year
Year
Pre (n = 191)
Post (n = 173)
Freshman
3 (2%)
2 (1%)
Sophomore
49 (26%)
48 (28%)
Junior
98 (51%)
87 (50%)
Senior
41 (21%)
36 (21%)
The TAs were interviewed at the end of the course to gain further insights into how well the assignments worked. As some TAs have taught the course with this instructor in previous years, they were able to speak to quality of writing assignments previously in comparison with the study’s version of the assignment. TAs graded the draft and final versions, and thus they have the best understanding of how using the AI on the first draft impacted final versions. Additionally, they had the most communication with students around the assignment and can illuminate any other challenges that arose.
Data Analysis
The first three questions were used to collect demographic information on students, such as student year and major, as well as an identifier that could be used to connect an individual’s pre- and post-survey. Mean and standard deviation were calculated for the Likert-style questions. To calculate a mean, choices were converted to a numerical scale ranging of 1–3, 1–4, or 1–5, where the lowest number is the strongest disagreement, and the highest number is strongest agreement. Yes/No questions were measured by calculating percentages of the total number of responses.
To determine change in student perception, several questions were included in both the pre- and post-survey. Responses for students who participated in both surveys were matched using the identifier, where any duplication of the identifier was resolved either with student major and/or student year data, or else it was removed from this part of the study. Once the dataset was cleaned and responses matched, Wilcoxon-Signed Rank tests assuming paired data were calculated for each question to determine if the distributions were significantly different. A Kruskal–Wallis test was used to compare the Likert-style results by student year and by student major where majors were grouped into Public Health Science or Non-Public Health Science as this is nearly an even split of the class. Year and major were considered the predictors of the response in this test.
The open-ended responses were utilized to illustrate the themes in the qualitative results. One author developed the codes for each question based on a subset of the responses. A second author then reviewed the codes and responses. The codes were additionally reviewed by a graduate assistant who was not involved in the study. Each set of codes was unique to the question, though for questions used in both surveys the same set of codes was used. These codes were developed to best represent the students’ responses. With this coding framework, it was possible for a response to be coded as falling under multiple codes. An example of this can be seen in Table 4. The total number of instances of each code were then summed for each question and percentage out of total responses to the question was assessed. From these responses the authors together also identified quotes that represent the data. Quotes have been lightly edited for conciseness and clarity; verbatim quotations are available upon request.
Table 4
Examples of coded responses from the post-survey question “Do you have concerns about ChatGPT (or AI like it) as a student?”
Student response
Code
Can provide inaccurate or misleading information, and devalues hard work done without AI
Don’t think it produces good writing/incorrect facts
AI models can reduce critical thinking skills, but can also save people more time to balance other things
Loss of student ability/not learning
Plagiarism, cheating, inaccurate information
Plagiarism/cheating (themselves)
Don’t want peers to cheat/unfair grading
Don’t think it produces good writing/incorrect facts
It could be used by others to cheat and therefore give a disadvantage to hard-working students
Don’t want peers to cheat/unfair grading
It is essential that AI is used in an ethical manner to prevent the disproportionate benefit or harm to any group of people
Concerns about ethics/biases
It isn’t very helpful
Don’t think it’s useful
Bold indicates key words used for coding
Interviewing the TAs served the purpose of illuminating trends or shortcomings that may not have been apparent from student-reported data. The TAs interviewed were all experienced graduate TAs who have taught more than one semester. For the purposes of this study, the undergraduate course TAs were not interviewed as they were first-time TAs. All three graduate TAs were asked the same set of questions, with the exception of the TA who has taught the course previously. This TA was asked an additional question based on their previous experiences. The answers to these questions were transcribed and compared between all three TAs. As with the open-ended responses, quotes from these interviews were lightly edited for clarity.
Results and Discussion
To What Extent Did Students Report on Awareness/Knowledge/Use of AI Prior to the Course?
The first two questions in the pre-survey asked students to report about their previous experience with AI in a classroom setting. In response to the first question, 69 students (36%) reported that they had learned about AI in one of their classes. To distinguish if there was an effect by student year, this was broken out further by percentage of each group that responded “yes” to learning about AI in one of their classes (one freshman student [33% of freshmen], 21 sophomores [43%], 37 juniors [38%], 10 seniors [24%]). This does not reflect the normal expectation, where seniors should have had the most opportunities for exposure, likely because AI has only been relevant in the class setting for approximately a year. Thus, this has not left much time for differences between student years to develop. In response to the second question, only 23 students (12%) reported that they had been prompted to use it by an instructor in the classroom.
When asked specifically about their use of ChatGPT (or AI like it) and Grammarly prior to this class, most students (118 students, 61.8%) reported that they had never used ChatGPT for an academic assignment (Fig. 2). A small subset (21 students, 11%) reported that they had only used ChatGPT once, and another subset (51 students, 26.7%) reported using it sometimes. Only one student reported that they used ChatGPT for almost every written assignment. The distribution of Grammarly use was slightly different. Students either reported never using it (43 students, 23%) or used it fairly frequently (reported as sometimes [77 students, 40%] or for almost every assignment [58 students, 30%]) to assist with their writing. This suggests a precedence for students using tools to improve their writing. Grammarly use has been generally accepted as meeting academic integrity policies, and for the authors’ institution is included in the university software catalog. However, Grammarly now has its own AI functionality and additionally has a feature called GrammarlyGO (Grammarly, 2023) that is classified as a generative AI.
Fig. 2
Student responses when asked if they have used ChatGPT (or an AI like it) for a class without being asked, and if they have ever used the writing tool Grammarly
Students were also asked in the pre-survey to report the ethical issues related to AI that they were aware of. The primary issue reported was academic integrity (71 students, 47.3%), which is unsurprising considering that this is also one of their main concerns. They also reported being concerned about copyright issues (36 students, 24%), biases or errors in AI training datasets (29 students, 19.3%), privacy and data use issues (18 students, 12%), the impact of AI on jobs (13 students, 8.7%), an impact on learning (9 students, 6%), and impacts on humans at a social level (19 students, 12.7%). A student expressed more in-depth knowledge of a social impact of AI, stating, “It’s powered by underpaid workers in 3rd world countries.” This appears to be a lesser-known ethical concern of AI, alongside the environmental impact. A thoughtful exploration of this issue can be found in the “Labor” chapter of Kate Crawford’s book Atlas of AI (Crawford, 2021).
At the beginning of this study, students reported they had little knowledge of AI (mean = 1.5 out of 3, SD = 0.55, Table 5). This was not statistically different between major or year, likely because AI is still very new. As AI becomes more prevalent, it is likely that students will have a greater baseline knowledge of AI, however the fact that AI technology keeps changing means that this may not be consistent. Many instructors fear that students are using AI frequently, which is not supported by these reported data. As this class consists almost solely of STEM students, it is recommended to conduct similar studies on other majors in order to address the prevalence of AI use in the student body as a whole. However, perhaps this will provide insight to instructors on the likelihood of student AI use in their own classrooms.
Table 5
Likert-style results from the pre-survey. To calculate a mean, choices were converted to a numerical scale ranging of 1–3, 1–4, or 1–5, where the lowest number is the strongest disagreement, and the highest number is highest agreement (these number represent all students who participate either in the pre or the post, not the matched data)
Question
# of choices
Mean \(\pm\)SD pre-survey (N = 191)
Mean \(\pm\)SD post-survey (N = 175)
Have you used ChatGPT (or AI like it) in other classes without being instructed to use it?
4
1.66 \(\pm\) 0.89
N/A
Have you ever used Grammarly or other similar writing-assistance software?
4
2.79 \(\pm\) 1.11
N/A
Do you think ChatGPT (or AI like it) helps [or could help?] you as a student?
5
3.8 \(\pm\) 0.96
4.19 \(\pm\) 0.85
Do you think ChatGPT (or AI like it) should be used in the classroom?
5
3.38 \(\pm\) 0.93
3.51 \(\pm\) 1
Would you use ChatGPT (or AI like it) to write the entirety of an essay/paper/assignment?
5
1.4 \(\pm\) 0.74
1.38 \(\pm\) 0.76
Would you use ChatGPT (or AI like it) to help you write essay/paper/assignment? (e.g., for checking grammar, developing an outline, etc.)
5
3.87 \(\pm\) 1.1
4 \(\pm\) 1.08
How aware are you of the ethical concerns surrounding AI?
3
2.19 \(\pm\) 0.69
2.55 \(\pm\) 0.58
How much do you know about large-language model AIs such as ChatGPT?
3
1.5 \(\pm\) 0.55
1.79 \(\pm\) 0.56
Do you feel that this assignment helped you better understand AI and its ethical uses?
5
N/A
4.14 \(\pm\) 0.8
Do you feel like this assignment challenged you to think about the ethics and social implications of genetics research?
5
N/A
3.91 \(\pm\) 0.97
Did your opinion on using AI change after completing this assignment?
5
N/A
3.3 \(\pm\) 0.99
To What Extent Did Students Believe that ChatGPT Is a Useful Tool Prior to and After the Class Assignment? What Are Their Concerns?
In the pre-survey, students felt fairly positively about the use of AI in the classroom (mean = 3.78 out of 5, SD = 0.96). This persisted in the post-survey, where student perception indicated a significant increase in favorability about AI as an assistant to learning (mean = 4.19 out of 5, SD = 0.85, p-value = 0.03). In regard to using AI as a learning tool, students stated, “It provides valuable explanations and help understanding material.” Another student provided multiple ways that AI could be helpful to them, “ChatGPT could help a student develop an outline for a writing assignment, stay organized, find basic background information, or could answer simple calculations.”
When asked to describe how they felt about use of AI in the classroom in the post-survey, a subset of students felt that AI should be used in the classroom, but only if used “right” (45 students, 31% of short-answer responses), and 38% (52 students) felt as if AI was helpful to them as a learning or writing tool. A prevailing theme is that students were positive about the use of AI but required that guidelines or restrictions be placed on its use. The vocabulary they used to describe this utilized the following words: right, proper, appropriate, and moderation. This terminology was frequently used, however not defined by any of the students. Students have a sense of what they want but were not able to explicitly define how it might work in practice. Some students (11, 7%) felt that use of AI in the classroom should be class or assignment dependent, indicating that there are instances where it is appropriate to use, and others where it is not. One student summarized several of these issues.
I think there is no harm to using chat GPT in the classroom as long as everyone knows how to use it appropriately and fact check. I also think it depends on the class. In a writing/discussion/opinion-based class it could definitely help but in a math or physics class that’s mostly factual it may be less helpful, but I am sure there are ways it could be used that I’m not considering.
Other students were still not convinced, “I don’t think it would be useful to use in class.”
The responses to this survey show that students are aware of the barriers to education that some of their peers face.
ChatGPT can provide students with instant access to information and feedback on various topics and subjects. Students can ask questions, request explanations, or seek guidance from ChatGPT at any time and place, without relying on teachers, tutors, or peers. This can be especially helpful for students who have limited resources, face barriers to education, or need extra support.
The vast array of responses to this question show that students are thinking deeply about how AI affects them and their learning.
Overall, students felt as if this assignment helped them become more informed about AI, and the ethical issues surrounding it (mean = 4.14 out of 5, SD = 0.8, Table 5). Many students had not used AI previous to this assignment, and thus had a more dramatic change in knowledge than students who were more informed at the start of this study. This was reflected in their short responses.
I never used ChatGPT before this assignment and never really had an idea how this sort of AI worked, this was a great opportunity to learn about this tool, and though I don’t completely trust it, I know I will be using it in the future.
Data collected from the pre-survey showed that slightly more than half (53%) of the students (102 out of 191) reported that they were concerned about AI’s impact on them as students. Students were again asked in the post-survey if they had concerns about the impact of AI on them as students, and 72% (124) responded that they were concerned, which is an increase from the pre-survey by 19%. When they were prompted to explain what those concerns were, 98 students responded to the open-ended question in the pre-survey and 121 responded to the post survey. The following four major themes emerged and were consistent for both the pre- and post-survey.
A.
Concerns that AI Tools Are not Accurate/Reliable
In the pre-survey, 27 (27%) students raised concerns about the accuracy of AI, demonstrating that they had some knowledge of the weakness of LLMs. Responses included the following statement: “Asking it [the AI tool] specific questions may not have accurate answers,” “It makes stuff up,” and “The information on there could be false and misinformed.” In the post survey, 63 students (52%) expressed concerns about the accuracy of what AI produces. There was an increased concern specifically in regard to the citations produced by the AI, which is something they learned from the self-reflection worksheet, “Some concerns that came up while completing the assignment were the fake citations.” This increased concern from the pre-survey suggests that completing the assignment increased students’ awareness of possible inaccuracies in the AI outputs, as well as the challenges ChatGPT faces in generating sources.
B.
Concern that the Use of AI May Negatively Affect Students Learning Skills
Twenty-six students (27%) expressed concern in the pre-survey that AI would be detrimental to their learning. Several students referred to writing and reading skills. As one wrote:
“… I also think reading comprehension and writing skills are important to practice, even if annoying or hard in the moment, because they give you the tools to be well-informed and less persuaded by misinformation. Using ChatGPT could make us worse at those necessary skills.”
Students also raised concerns that using ChatGPT “minimizes creativity” and makes you “rely on [the AI tool] completely.” Others were worried that students will become too dependent on AI and would not “struggle on their own to come up with new ideas” or make them “lazy in problem solving.” In the post-survey, 23 students (19%) expressed this same concern. “…ChatGPT could make gaps in the knowledge that I might need in my career.”
C.
Concern that AI Will Be Used by Others to Cheat
A related concern is that students felt strongly that their peers would cheat, and that this was unfair.
“I am concerned that my fellow students are using/would use it to be academically dishonest, which is concerning because I am always academically honest and want to be graded fairly against other people’s work, not an AI.”
In the pre-survey, 20 students (20%) expressed concerns about their peers cheating, and 24 (20%) had the same concerns in the post-survey. As a whole, students are against their peers receiving better grades for work they did not truly do.
D.
Fear of Being Accused in Academic Dishonesty or Plagiarism
Twenty students (20%) were concerned in the pre-survey that using AI would result in being accused of plagiarism, cheating, or some other aspect of academic dishonesty that would involve censure from the university. One student reported, “I am worried about plagiarism and being flagged within my classes,” while another indicated that they had already been accused of using an AI in their academic work.
“I have been reported for using an AI on exam before, when I did not use it. The only reason I was reported was because other students in the class were using AI. This caused the entire class to have to go through the Office of Student Conduct, which is a very stressful situation to be in when you did not use the AI or go against academic integrity.”
In the post-survey, 15 students (12%) expressed persisting concerns about academic honesty, and potential accusations.
The concerns students voiced in response to this question align with those expressed by instructors. Students are aware of the potential impact on their learning, which may prove helpful to instructors and facilitate cooperation.
To What Extent Did Student Perceptions of the Use of AI Change Following the Assignment?
We targeted the change in student perceptions in two different ways. First, we prompted students to self-report how they felt their opinions changed over the course of the study, and second, we connected pre- and post-survey responses of individual students. For self-reporting, students were asked if they felt that this assignment resulted in a change in their opinion on AI. The mean response was 3.3 out of 5 (Table 5); however, in the short answer, 54% (78 students) stated that they felt their opinion had not changed. “No, I still think that I’ll stay pretty clear of any AI or ChatGPT. The cons outweigh the benefits to me.” Interestingly, 30% (43 students) reported that this assignment gave them a more positive opinion on AI. For example, “Previously, I thought AI was just used to cheat but now I can see how it can be helpful!” An additional 10% (12 students) reported either feeling more negative about AI or being more critical of its outputs.
A total of 143 students responded to both the pre- and post-surveys. Six questions were shared between the two surveys to assess change in student perception. Five of these questions showed statistically significant change in the distribution of student response between the two surveys when assessed with the Wilcoxon Signed-Rank Test (Table 6). In the post-survey, students appeared more inclined to think that AI could help them academically than they did in the pre-survey. Additionally, they became somewhat more favorable to AI being used in a classroom setting. This was accompanied by short-answer feedback as to how this might be done, with a focus on a desire to be taught how to use AI in a way that is appropriate for a given class. For example, one student wrote, “Yes, instructors should teach students about the implications and proper use of ChatGPT and other AI learning platforms.” They also expressed a desire for clearer instructions and boundaries. After the assignment, students reported less inclination to use AI to write an entire assignment; however, their willingness to use AI as a writing tool/assistant did not change significantly. There was no significant difference between predictors such as major or year for any of the questions.
Table 6
Comparison of student change in perception between the pre and post survey calculated from matched data. Significance calculated from the Wilcoxon Signed Rank Test
Question (N = 143)
# of choices
Mean pre
Mean post
p-value
Do you think ChatGPT (or AI like it) helps [or could help?] you as a student?
5
3.78
4.2
2.85E-07***
Do you think ChatGPT (or AI like it) should be used in the classroom?
5
3.33
3.51
0.03*
Would you use ChatGPT (or AI like it) to write the entirety of an essay/paper/assignment?
5
3.78
1.36
2.20E-16***
Would you use ChatGPT (or AI like it) to help you write essay/paper/assignment? (e.g., for checking grammar, developing an outline, etc.)
5
3.86
4.02
0.06
How aware are you of the ethical concerns surrounding AI?
3
2.17
2.51
1.63E-08***
How much do you know about large-language model AIs such as ChatGPT?
3
1.5
1.8
1.20E-06***
*p < 0.05, **p < 0 .01, ***p < 0.001
Finally, students reported overall more knowledge of AI and its ethical issues; however, this mean was still relatively low (1.79 out of 3, Table 5). Thus, while the purpose of this assignment was to give students an introductory level of knowledge on AI, students still felt that there was more they could learn. This is compounded by the fact that AI technologies are changing at a rapid rate, and thus underwent advancements even over the course of this study. In order to stay abreast of the advancements, both students and instructors must invest time into keeping up to date. Unfortunately, this means that some assignment types may not maintain relevance the same way they have historically and could require year-to-year or semester-to-semester updates.
How Effective Was the Assignment at Promoting Student Learning About AI and Ethical Issues in Genetics?
As reported in the post-survey results (Table 5), more than half of students found this assignment to be helpful on educating them on AI (89 students, 62%) as well as prompting them to think more deeply about ethical issues in genetics (88 students, 63%). They reported that integrating AI into the assignment forced them to be critical of the ChatGPT outputs and do their own extensive research to validate sources. One student commented, “ChatGPT provided some reasons, but during my edits I had to think of more. Thus, I thought of it on my own as well as understood the reasons ChatGPT provided.” A small subset of students (8) reported that because they used ChatGPT for this assignment, they did not do as much research on their own as they might have if they had done the assignment in a traditional manner. As mentioned by one of the students, “I didn’t do much of my own research on the issue because ChatGPT did the bulk of it for me so I didn’t get as much out of it as I probably could have.”
The assignment also helped students further investigate ethical and social implications of genetics. Fifty-seven percent of the students who answered to the open-ended question (80 of 140) reported that the assignment itself helped improve their understanding of these issues. Some students reported that the assignment prompted them to think more in depth “It allowed me the space to consider an ethical genetics debate more in depth,” and prompted them to be more engaged because the assignment was something interesting to them. “[The assignment] allowed us to get creative on any topic related to genetics and I believe that gave us free reign to investigate issues that are interesting to ourselves and kept us engaged.” This aligns with what has reported previously for assignments like this, where creative freedom encourages a higher investment from students (Nicholes, 2020; Ostrom et al., 2020). It would be interesting to further disentangle how much this creative aspect motivates investment in the topic itself in comparison to their investment in their own writing. It seems possible that creative writing may be achieving both of these things.
Students found that having to address an ethical dilemma from a story-like or creative approach changed their perspective. One student reported, “[The assignment] allowed me to think about both sides of a story.” Another student described something similar, “I had to think of both viewpoints and think about how each person thinks/views a discussion.”
Creative writing provides students the opportunity to interrogate a problem from multiple perspectives, which in turn can change their own perspective on an issue (Nicholes, 2020; Watkins & Tehrani, 2020). The ability to evaluate a problem from multiple perspectives is an important skill that students need, regardless of major or desired profession.
Students expressed a desire for clearer directions or a rubric for this assignment should it be conducted again. For this assignment, it was not defined how much of the final draft should be original writing versus writing generated by the AI. Students found this confusing and were unsure how much they were supposed to edit or re-write the AI-generated writing. It is challenging to set a threshold for this that can be easily verified, and for this assignment the researchers did not want to be too stringent and thus restrict student creativity. One student suggested having AI-generated text and human generated text in different fonts or colors which might be one way to accomplish this. In another iteration of this assignment, it would be beneficial to be more explicit about how different the re-written assignment is expected to be from the AI-generated version. Overall, this reflects how critical it is for instructors to define clear boundaries or guidelines for the use of AI in the classroom.
In addition to the student responses, the course TAs provided valuable insight into the effectiveness of the assignment. When asked if significant improvement occurred between the draft and the final paper, one TA primarily reported, “I saw mostly either students that really re-worked it or did very little work with it.” Another TA further elaborated on this.
I think there were a couple students that made an effort, I think there were students who I hypothesize are stronger writers and felt bothered by the style that was output and, in their feedback and comments and stuff, they indicated that. They were the most likely to go in and really pare down what it was saying because first they had a more advanced understanding of writing style and what makes good writing, and that more words aren’t always better.
This suggests an expected spectrum of writing ability across students, which impacted their willingness or ability to revise the AI-generated content.
The suggestions made by students (Table 7) for alternate versions or expansions on this assignment were useful and reflected that they understand how to challenge AI tools. These suggestions primarily broke out into three categories: Further comparison of AI draft vs. human draft, further challenge or explore the AI’s responses, or change the type of paper written. These three categories represent areas where instructors can develop AI-based assignments. Adjustments to this assignment could be class-dependent and reflect the overall goals of that class. For example, a STEM-based course might want to emphasize problem-solving skills, whereas an English course might want to focus on how to hone writing or reading comprehension skills. Framing exercises in AI around course goals will result in assignments that are beneficial and relevant to students. This study shows that students can be a valuable resource for creating these assignments, as they have well-thought-out suggestions for how to utilize AI in the classroom in a meaningful way. Thus, the onus of coming up with these assignments is not solely on the instructors. Especially while AI is still new, expanding discussions to include students can improve AI-based assignments.
Table 7
Student suggestions for improving the assignment grouped by category
Categories
Students’ quotes
Further challenge or explore the AI’s responses (5 of 139)
I think we could explore more creative ways that ChatGPT works or falls short
I would probably ask more questions to challenge the AI
I find a lot of trouble with AI explaining math-based questions. That would be interesting, to find its mistakes
I would make it so we asked more questions to explore what responses we get
Show how AI can be used as a potential study tool and the potential drawbacks?
Change the type of paper written (2 of 139)
Have it write an academic paper
I would make the assignment more catered toward academic papers than creative assignment
Further comparison of AI draft vs. human draft (9 of 139)
I would have students do another self-evaluation after submitting the final draft so they can compare what they produced with what AI produced. I think this would help students really see what AI is useful for (giving suggestions) and what it should not be used for (completing entire assignments)
I would make it so that the student writes the essay on their own first, and then has AI make changes and revisions
Compare the initial AI made essay and the final edited essay
Maybe you could have students compare AI written and student written pieces to determine which one is better
Writing it in my own words before would have been an interesting way to see with AI can do and do differently than humans
I would use the AI to provide an outline and then from their use humanity to write the essay
Compare our own essay to a certain prompt with ChatGPT and then combine
If I could, I’d add another step where we add in a different font the changes, we would make to the essay just straight from the chat
I would require that the students add reliable evidence and include references to support what the AI computes. Anything that cannot be found on reliable sources can probably be deleted or changed
The TAs over all reported that they had little difficulty implementing the assignment. The biggest challenges they faced were students failing to understand that the assignment was meant to be creative, or students who did not have a good grasp on what an ethical problem was. In their recitations or via email, TAs were generally able to address these issues. TAs reported that students found the most challenging part of the assignment to be the peer review. Two of three TAs reflected that students were not as engaged in this task. “Some students were kind of sluggish with the peer review. I think some of them just weren’t as interested in the peer review questions, so they were like “yes” or “no” [rather than elaborating more],” while the third TA stated, “I have never seen, in the entire semester, such concentration and hard work from every single student in the room.” A suggestion for improving the peer review was as follows,
I think there could have been fewer peer review questions, but then if they also had to provide a statement responding to the peer review as well. Like ‘here’s what I felt the strengths and weaknesses were from my initial draft so here’s what I changed in the revised copy.’
Alongside this suggestion for improvement, another TA recommended to reduce the number of evaluation tasks (the self-evaluation or the peer review) as students disliked busy work. This represents a primary challenge of this sort of assignment. The evaluation portions of the assignment were designed to guide students’ thinking; however, this may not be something they find to be interesting or engaging. In the future, these worksheets could be streamlined to be less arduous for students while still accomplishing the same goals. Overall, the TAs were positive about the assignment.
I think [the AI portion of the assignment] was fantastic, I think they understood how to use it, they whined and complained about [the informative module] before they did it, but when they did it, they were proud of their learning. They feel in the know about this thing that their peers are talking about.
While both students and TAs encountered some challenges, it seems the assignment accomplished the goals outlined at the beginning of the study.
Study Limitations
This study is limited by being conducted only once in this course. Ideally, more than 1 year of data would be collected; however, with the urgency of adapting to AI in the classroom, we felt this data was most useful to other educators sooner rather than later. The other limitation of this study is that it is based on student reported data collected through voluntary surveys, with some additional data contributed from interviews of the course TAs. This could be further expanded on by interviewing students as well, though this is more challenging to recruit for. However, in the pre-survey, 99% of students enrolled in the class responded, and this did not drop off dramatically in the post-survey considering approximately 10 students dropped the class after the initial drop/add period.
Conclusions
AI is rapidly becoming an inescapable challenge of education at all levels (Cardona et al., 2023; Lo, 2023; Stokel-Walker, 2022). One major concern of educators is how this will affect assignments, particularly writing-based assignments that are most vulnerable to generation with AI (McMurtie, 2022; Rudolph et al., 2023). This has many implications for student learning, particularly in STEM majors where students already express anxiety about writing creatively (Daker et al., 2020; Marbach-Ad et al., 2016, 2019). Creative writing, and scientific writing in general, however, is a critical skill STEM students need to develop (American Association of Advancement of Science [AAAS], 2011; Hart Research Associates, 2015). In administering this creative writing assignment, we were able to accomplish several goals. The first was promoting student writing and reinforcing its relevance in their future careers. Second, students gained knowledge on AI and how to utilize it in their classwork. Finally, we summarize students concerns and their insights on how and when AI should be used in the classroom.
This study provides insights that will be useful to instructors as AI becomes more prevalent. Students have strong opinions on AI, especially as it pertains to academic integrity. They also expressed desire to have clear boundaries in academic settings that define how and when AI should be used. These two aspects are helpful for instructors to consider when they are deciding how to address AI use in their own classroom. Students also had thoughtful suggestions on how to utilize AI in the classroom that could be used to develop new assignments. Inclusion of students in further discussions of AI in the classroom could yield more thoughtful discussion of assignments suitable for this, as students have a unique perspective.
What we found is that, unsurprisingly, there are advantages and disadvantages to using AI in the classroom. It can be very helpful for grammar checking, generating outlines, providing prompts upon which to build an essay. We also found that educating students about AI and how to use it appropriately initiates a shift in student opinion on its use. This education component is important, and students desire to better understand a tool that will likely become more prevalent in their lifetime. Getting ahead of the curve and educating them about AI might mitigate some of the problems regarding academic dishonesty. Students were much less inclined to use AI to write a whole assignment or essay at the end of the study. An unexpected result of this assignment was that some students generated prompts using ChatGPT, and then gave that prompt to ChatGPT to get their first draft. This was not an anticipated approach to the assignment, but a useful way of utilizing ChatGPT, nonetheless. In future iterations of this assignment, it would be useful to have students write a reflection, though this was somewhat served by the second survey in this study.
Nevertheless, there are challenges in incorporating it into a classroom setting beyond the concerns of academic integrity. The most obvious are challenges that arise from questionable AI accuracy and biases of AI training datasets (Srinivasan & Chander, 2021; Varsha, 2023). This is a challenge that can be addressed by educating students, but still presents problems. Another is that the boundaries that are so important to define for the use of AI are not simple to isolate. One question we faced in this study was the following: “What is the correct percentage of the assignment that has to be a student’s original writing?”. Further questions we have come upon are the following: “How are instructors to decide if an AI-inclusive assignment is right for their class?” and “How does one create assignments that are not AI-inclusive, but are not easily completed with AI?”.
Writing and reading comprehension skills are important for STEM students, yet it becomes increasingly difficult to impart these skills on students over the course of their undergraduate education. Students have found ways to evade writing assignments via skilled plagiarism or paying others to write for them, but AI makes it much easier for them to avoid writing entirely. This is to their own detriment, as they will need those skills when they enter the workforce, regardless of their career trajectory (Marbach-Ad et al., 2016, 2019; Wheeler et al., 2023). Faculty in STEM fields face the challenge of grading and providing feedback for writing assignments in large courses (Marbach-Ad et al., 2019); however, utilization of peer review and an AI to check for grammar, etc., might provide an avenue for including student writing more frequently in a curriculum.
There are still many challenges to be faced by educators as we come to grips with the new realities of education at all levels. It is our hope that this study provides some potential avenues for incorporating AI into the classroom, as well as insight into the student perspective to inform instructor decision-making.
Acknowledgements
The authors would like to thank Mona Thompson and the Teaching and Learning Transformation Center at the University of Maryland for their assistance and insight on this manuscript, as well as their efforts developing the informative AI module. We would also like to thank the BSCI222 teaching assistants for their help in facilitating this study.
Declarations
Ethics Approval
The study was determined exempt by the Institutional Review Board of University of Maryland, MD (protocol code # 2071504–1 and 30 June 2023) for studies involving humans.
Informed Consent
Informed consent was obtained from all individual study participants.
Conflict of interest
The authors declare no competing interests.
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Atwood, S., & Read-Daily, B. (2015). Using a creative fiction assignment to teach ethics in a first-year introduction to engineering course. 2015 ASEE Annual Conference and Exposition Proceedings, 26.1649.1–26.1649.11. https://doi.org/10.18260/p.24985
Baidoo-Anu, D., & Owusu Ansah, L. (2023). Education in the era of generative artificial intelligence (AI): Understanding the Potential benefits of ChatGPT in promoting teaching and learning (SSRN Scholarly Paper 4337484). https://doi.org/10.2139/ssrn.4337484
Bansal, G. (2023). Competency-based assessments in STEM classrooms: Using creative story writing as a formative assessment. In Fostering Science Teaching and Learning for the Fourth Industrial Revolution and Beyond (pp. 62–85). IGI Global. https://doi.org/10.4018/978-1-6684-6932-3.ch003
Baram-Tsabari, A., & Osborne, J. (2015). Bridging science education and science communication research. Journal of Research in Science Teaching,52(2), 135–144. https://doi.org/10.1002/tea.21202CrossRef
Bitzenbauer, P. (2023). ChatGPT in physics education: A pilot study on easy-to-implement activities. Contemporary Educational Technology,15(3), ep430. https://doi.org/10.30935/cedtech/13176CrossRef
Cardona, M. A., Rodríguez, R. J., & Ishmael, K. (2023). Artificial intelligence and the future of teaching and learning. Office of Educational Technology, Insights and Recommendations. Retrieved September 2023 from https://digital.library.unt.edu/ark:/67531/metadc2114121/m1/2/
Cooper, G. (2023). Examining science education in ChatGPT: An exploratory study of generative artificial intelligence. Journal of Science Education and Technology,32(3), 444–452. https://doi.org/10.1007/s10956-023-10039-yCrossRef
Crawford, J., Cowling, M., & Allen, K.-A. (2023). Leadership is needed for ethical ChatGPT: Character, assessment, and learning using artificial intelligence (AI). Journal of University Teaching & Learning Practice, 20(3). https://doi.org/10.53761/1.20.3.02
Daker, R. J., Cortes, R. A., Lyons, I. M., & Green, A. E. (2020). Creativity anxiety: Evidence for anxiety that is specific to creative thinking, from STEM to the arts. Journal of Experimental Psychology. General,149(1), 42–57. https://doi.org/10.1037/xge0000630CrossRef
Dalalah, D., & Dalalah, O. M. A. (2023). The false positives and false negatives of generative AI detection tools in education and academic research: The case of ChatGPT. The International Journal of Management Education,21(2), 100822. https://doi.org/10.1016/j.ijme.2023.100822CrossRef
Desaire, H., Chua, A. E., Kim, M.-G., & Hua, D. (2023). Accurately detecting AI text when ChatGPT is told to write like a chemist. Cell Reports Physical Science,4(11), 101672. https://doi.org/10.1016/j.xcrp.2023.101672CrossRef
Diluzio, R., & Congdon, C. B. (2015). Infusing the creative-thinking process into undergraduate STEM education: An overview. IEEE Integrated STEM Education Conference,2015, 52–57. https://doi.org/10.1109/ISECon.2015.7119945CrossRef
Ede-Osifo, U. (2023). College instructor put on blast for accusing students of using ChatGPT on final assignments. NBC News.
Firat, M. (2023). How chat GPT can transform autodidactic experiences and open education? [Preprint]. Open Science Framework. https://doi.org/10.31219/osf.io/9ge8m
Fuchs, K. (2023). Exploring the opportunities and challenges of NLP models in higher education: Is Chat GPT a blessing or a curse? Frontiers in Education, 8. https://www.frontiersin.org/articles/https://doi.org/10.3389/feduc.2023.1166682
Georgiou, H., Turney, A., Matruglio, E., Jones, P., Gardiner, P., & Edwards-Groves, C. (2022). Creativity in higher education: A qualitative analysis of experts’ views in three disciplines. Education Sciences,12(3), 154. https://doi.org/10.3390/educsci12030154CrossRef
Gillen, C. M., Kerkhoff, A. J., Lynn, D. H., McFarlane, H. G., Niemiec, A. J., Petersen, S. C., & Reach, A. D. (2020). Does creative writing improve scientific writing and learning? The FASEB Journal,34(S1), 1–1. https://doi.org/10.1096/fasebj.2020.34.s1.05351CrossRef
Gozalo-Brizuela, R., & Garrido-Merchan, E. C. (2023). ChatGPT is not all you need. A state of the art review of large generative AI models (arXiv:2301.04655). arXiv. http://arxiv.org/abs/2301.04655
Hart Research Associates. (2015). Selected findings from online surveys of employers and college students conducted on behalf of the Association of American Colleges & Universities.
Jedlicka, S. (2020). Using creative writing as a tool for learning professional development in materials science and engineering. 2020 ASEE Virtual Annual Conference Content Access Proceedings, 35459. https://doi.org/10.18260/1-2--35459
Kasneci, E., Sessler, K., Küchemann, S., Bannert, M., Dementieva, D., Fischer, F., Gasser, U., Groh, G., Günnemann, S., Hüllermeier, E., Krusche, S., Kutyniok, G., Michaeli, T., Nerdel, C., Pfeffer, J., Poquet, O., Sailer, M., Schmidt, A., Seidel, T., … & Kasneci, G. (2023). ChatGPT for good? On opportunities and challenges of large language models for education. Learning and Individual Differences, 103, 102274. https://doi.org/10.1016/j.lindif.2023.102274
Kim, N. J., & Kim, M. K. (2022). Teacher’s perceptions of using an artificial intelligence-based educational tool for scientific writing. Frontiers in Education, 7. https://www.frontiersin.org/articles/https://doi.org/10.3389/feduc.2022.755914
Lo, C. K. (2023). What is the impact of ChatGPT on education? A rapid review of the literature. Education Sciences, 13(4), Article 4. https://doi.org/10.3390/educsci13040410
Marbach-Ad, G., Rietschel, C., & Thompson, K. V. (2016). Validation and application of the survey of teaching beliefs and practices for undergraduates (STEP-U): Identifying factors associated with valuing important workplace skills among biology students. CBE—Life Sciences Education,15(4), ar59. https://doi.org/10.1187/cbe.16-05-0164CrossRef
Marbach-Ad, G., Hunt, C., & Thompson, K. V. (2019). Exploring the values undergraduate students attribute to cross-disciplinary skills needed for the workplace: An analysis of five STEM disciplines. Journal of Science Education and Technology,28(5), 452–469. https://doi.org/10.1007/s10956-019-09778-8CrossRef
Mogali, S. R. (2023). Initial impressions of ChatGPT for anatomy education. Anatomical Sciences Education. https://anatomypubs.onlinelibrary.wiley.com/doi/full/https://doi.org/10.1002/ase.2261
Naumova, E. N. (2023). A mistake-find exercise: A teacher’s tool to engage with information innovations, ChatGPT, and their analogs. Journal of Public Health Policy,44(2), 173–178. https://doi.org/10.1057/s41271-023-00400-1CrossRef
Nicholes, J. (2022). Creative writing across the curriculum: Defining and illustrating a method for STEM/humanities integration. Routledge.CrossRef
Nicholes, J. (2020). Lab reports and horror stories: Exploring chemistry majors’ evaluations of scientific and creative writing. Journal for Learning through the Arts, 16(1). https://eric.ed.gov/?id=EJ1289956
Ostrom, R., Gotesman, M., & But, J. C. (2020). Poetry in biology: Enhancing science education with creative writing. In J. C. But (Ed.), Teaching college-level disciplinary literacy: Strategies and practices in STEM and professional studies (pp. 147–166). Springer International Publishing. https://doi.org/10.1007/978-3-030-39804-0_6
Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). Language models are unsupervised multitask learners. Open AI Blog.
Rudolph, J., Tan, S., & Tan, S. (2023). ChatGPT: Bullshit spewer or the end of traditional assessments in higher education? Journal of Applied Learning & Teaching, 6(1). https://doi.org/10.37074/jalt.2023.6.1.9
Simonsen, L. M., & Dick, T. P. (1997). Teachers’ perceptions of the impact of graphing calculators in the mathematics classroom. Journal of Computers in Mathematics and Science Teaching,16(2), 239–368.
Tallvid, M. (2016). Understanding teachers’ reluctance to the pedagogical use of ICT in the 1: 1 classroom. Education and Information Technologies,21(3), 503–519. https://doi.org/10.1007/s10639-014-9335-7CrossRef
Tlili, A., Shehata, B., Adarkwah, M. A., Bozkurt, A., Hickey, D. T., Huang, R., & Agyemang, B. (2023). What if the devil is my guardian angel: ChatGPT as a case study of using chatbots in education. Smart Learning Environments,10(1), 15. https://doi.org/10.1186/s40561-023-00237-xCrossRef
Varsha, P. S. (2023). How can we manage biases in artificial intelligence systems – A systematic literature review. International Journal of Information Management Data Insights,3(1), 100165. https://doi.org/10.1016/j.jjimei.2023.100165CrossRef
Watkins, A., & Tehrani, Z. (2020). Brave new worlds: Transcending the humanities/STEM divide through creative writing.Honors in Practice, 16, 29–51.
Wheeler, L. B., Thompson, K. V., Marbach-Ad, G., Sheehan, P., Bortiatynski, J. L., & Ghent, C. (2023). Factors predicting the extent to which STEM students value cross-disciplinary skills: A study across four institutions. CBE—Life Sciences Education,22(2), ar20. https://doi.org/10.1187/cbe.22-06-0101CrossRef
Wood, E., Mueller, J., Willoughby, T., Specht, J., & Deyoung, T. (2005). Teachers’ perceptions: Barriers and supports to using technology in the classroom. Education, Communication & Information,5(2), 183–206. https://doi.org/10.1080/14636310500186214CrossRef
Yu, H. (2023). Reflection on whether Chat GPT should be banned by academia from the perspective of education and teaching. Frontiers in Psychology, 14. https://www.frontiersin.org/articles/https://doi.org/10.3389/fpsyg.2023.1181712