Graduate students' increasing reliance on AI tools like ChatGPT may compromise their academic integrity, particularly by encouraging the use of shortcuts and reducing the frequency of fact-checking.
With the widespread adoption of AI tools like ChatGPT, students in higher education are finding more efficient ways to complete tasks such as research and writing. However, this efficiency may come at the cost of academic rigor. Research shows that students who frequently use these tools often prioritize speed over accuracy, neglecting crucial fact-checking steps (Aljanabi & ChatGPT, 2024). This behavior can diminish critical thinking skills, as students may rely heavily on AI-generated information without thorough verification (Krullaars et al., 2024). Additionally, AI systems sometimes produce inaccurate or misleading information, further exacerbating the risks of over-reliance (Schei et al., 2024). Addressing these issues is crucial to ensuring that AI enhances the academic experience without undermining integrity.
Responses:
1. To assist with research and gathering information: 9 (24.3%)
2. To help with writing: 4 (10.8%)
3. To clarify difficult concepts or explanations: 8 (21.6%)
4. To save time on tasks: 12 (32.4%)
6. I do not use AI tools for academic work: 3 (8.1%)
7. Other: 1 (2.7%)
To improve the survey design, we could start with an initial questionnaire that doesn’t exclude participants if the constructs being measured don’t apply to them. Instead, it should be more inclusive and ensure that everyone can continue regardless of their AI usage. Another improvement would be to create questions that are applicable to all respondents, such as asking about their general thoughts and opinions on AI usage for schoolwork, rather than focusing on specific actions related to AI.