HSC Guidance for the Use of AI – Teaching and Instruction
The impact of artificial intelligence (AI) on teaching, learning, and academic operations is substantial. To ensure the effective and responsible use of AI tools, it is essential to establish comprehensive guidelines. This guide offers a framework from the faculty and instruction perspective, providing practical advice on utilizing AI responsibly within the academic context. It aims to support educators in integrating AI tools effectively while maintaining academic integrity and enhancing the learning experience.
AI in teaching and instructional contexts can enhance educational delivery by automating administrative tasks, improving learning methods and experiences, and supporting instructional strategies. It can aid in curriculum development, analyzing educational trends and student performance data, and helping instructors refine their teaching methods and materials. Overall, AI often boosts efficiency, enriches the learning experience, and enables more effective, data-informed instruction. To ensure ethical AI use in education, HSC faculty should focus on transparency by clearly explaining AI tools’ roles and practices to students while ensuring program quality is maintained. Bias and fairness must be continuously monitored to prevent perpetuating inequalities, and informed consent should be obtained for data collection. Accountability involves maintaining human oversight and addressing concerns, while regular reviews and adherence to ethical standards will ensure AI tools are effective and equitable. Accessibility for all students, including those with disabilities, must also be ensured. As a university—HSC and its faculty are responsible for the quality, integrity and review of its programs.
AI has the opportunity to transform workflows at HSC. Recognizing the power that AI tools bring, this guidance aims to protect the integrity of HSC’s programs, services, and reputation. Utilizing AI outside these guidelines may expose you and the institution to additional liability. Examples of potential liability related to AI include:
- Reputational Damage: An individual’s improper use of AI can harm the reputations of institutions (HSC) as well as the individual. This means that not only could HSC’s credibility be damaged, but individual instructors’ careers could be damaged as well.
- Intellectual Property Infringement: Using AI-generated content that lacks proper sourcing and citation may lead to violations of intellectual property rights and expose private data, including students’ information, to unauthorized sources.
Instructors have the ability to impact the reputation of the institution and thus have responsibility to uphold the values and academic standards of HSC. This guidance exists to help mitigate risk and promote a culture of integrity related to the ethical use of artificial intelligence.
Key Strengths and Weaknesses of AI to Consider
Strengths | Weaknesses |
• Rephrasing: AI tools can rephrase existing content for the desired readability and tone. | • Lack of human touch: AI is not human and may not understand the nuances of human emotion and interaction in the way humans do. |
• Grammar: AI tools are generally trained on proper syntax, outputting | • Bias: AI is only as good as the data it is trained on, and that data may be susceptible to bias that influences the AI output. |
• Speed: AI is fast, generally providing output seconds after prompting. | • Privacy: Many AI tools record the prompts that users enter. |
• Accessibility: Some AI solutions may be valuable for students with disabilities. If and when AI is appropriate for providing accommodation to a student, that information will be communicated by the Office of Disability Access to the appropriate faculty. | • Timeliness: AI models generally rely on discrete sets of data. Current events and research findings may not be incorporated into the AI tool. |
• Workforce Development: As AI technology advances, it is increasingly integrated into the day-to-day operations of careers which HSC specifically trains our students to enter. It is appropriate to prepare our students to use the tools— including AI tools—that will be required of them in their careers. | • Paywall Avoidance: Many AI systems are trained with openly available resources. Journal articles or other information that requires paid access may not be represented in an AI tool. |
• Learning Analytics: AI tools can analyze large data sets of student engagement with course materials to identify trends and patterns in student performance and highlight opportunities for improvement in curriculum design and teaching. | • Accuracy: Because AI tools are trained on data generated by humans, and humans are flawed, responses from AI tools may also be flawed. In each instance, the accuracy of responses from AI tools should be reviewed. |
• Opportunity for Enhanced Personalization of Instruction/Learning: AI can help identify individual learning preferences and utilize adaptive learning platforms to tailor curriculum (e.g. pacing, real-time feedback) to support enhanced student learning. | |
- Consider your role. Guidance exists to support different roles at HSC (e.g. for students, instructors, staff, researchers). When operating in roles beyond an instructor (e.g. as a researcher), other guidance may be appropriate and helpful to consult.
In Student Coursework
- Determining and communicating allowed tools and resources: Instructors are responsible for determining which AI tools and resources that students are permitted to use for course assignments and assessments. It is crucial for instructors to clearly communicate these guidelines to students to ensure transparency and consistency.
To maintain clarity and fairness:
- Authority: Instructors have the authority to specify the tools and resources that can be used for assignment or assessment.
- Communication: These guidelines should be explicitly outlined in the course syllabus or included in the assignment instructions.
- Consistency: All students must be informed of the allowed tools and resources at the time the assignment or assessment is issued to prevent confusion and maintain academic integrity.
- Integrity: At HSC, unauthorized use of AI by students to complete assignments is considered cheating based on the Student Code of Conduct and Civility (HSC Policy 7.105). Note that this specifically means that their use must be unauthorized, which requires communication and clear distinction from faculty as to which uses are and are not
- Using generative AI tools in assignments: Instructors should clearly state their intentions regarding the use of generative AI tools by students. To ensure that students learn effectively and adhere to their expectations. It is essential for students to understand the boundaries set by the instructor for AI tool usage.
Recommendations:
- Disclosure: Instructors should explicitly outline the appropriateness or inappropriateness of generative AI tool use in their course syllabus.
- Assignment Instructions: Clearly indicate any restrictions or allowances for AI tools in individual assignment instructions.
- Consistency: Ensure that students are aware of these guidelines and expectations from the outset to support their learning process and maintain academic integrity.
- Managing AI tools in coursework: It is crucial for instructors to understand several forms of which students might use AI tools in relation to the coursework assigned. To integrate AI into the learning process or acknowledge its potential use by students:
- Encourage AI Use: If you intend for students to use AI tools for an assignment, assessment, or learning activity, evaluate how AI use aligns with the designed learning outcomes. Ensure that the use of AI supports students in achieving these outcomes rather than hindering their understanding.
- Restricting AI Use: If you prefer students not to use AI tools, consider how the task is assessed. Design assignments in a way that minimizes the effectiveness of AI tools and promotes the intended learning objectives. This may involve creating tasks that require critical thinking, personal reflection, or unique problem-solving skills that AI tools cannot easily replicate. Additionally, live in-person assessments, and assessments that use established tools for creating a proctored environment are valuable in restricting AI use.
- Using AI in developing educational tools. AI tools offer valuable support in enhancing educational content by assisting with the ideation and creation of lesson plans, questions, scenarios, images, and other course components. Utilizing AI can save time and reduce the creative burden associated with developing these elements. However, it is important to maintain accountability and ensure the quality and accuracy of the course content.
Recommendations:
- Responsibility for Content: Regardless of the AI tools used, you are responsible for the quality, accuracy, and appropriateness of the course content. Ensure that all materials meet the educational standards and learning objectives of your course.
- Disclosure of AI Use: It is recommended to disclose the use of AI tools in the development of educational materials to maintain transparency. This can be achieved by:
- Informing students: Instructors should mention in the syllabus or course documentation that AI tools have been used to assist in creating course components. This helps students understand the resources and technologies involved in their learning.
- Providing context: Explain how AI tools have contributed to the content, if relevant, provide the students insight into the development process and its impact on their learning experience.
- Using AI in assessing student work. When assessing student work, adherence to FERPA regulations is essential to protect student privacy. Uploading a student’s entire assignment to third-party websites, including AI tools or AI detectors, may violate their rights to privacy. Only third-party entities with which the university has specifically entered into a contract, and which have agreed to maintain the privacy of our student data are permitted for this function.
Recommendations:
- FERPA Compliance: Ensure that any use of AI tools in assessing student work complies with FERPA guidelines, which govern the handling and confidentiality of student records. Avoid uploading entire assignments to third-party platforms that could compromise student privacy.
- Disclosure of AI Use: If AI tools are used in the assessment process, it is best practice to disclose this to students. This includes:
- Informing Students: Clearly communicate in the syllabus or assessment instructions if AI tools will be used to evaluate their work. This transparency helps students understand how their submissions will be assessed and the role of AI in the evaluation process.
- Providing Details: Outline how AI tools will be used, what data will be analyzed, and the measures in place to protect student privacy.
- If an instructor does not plan to disclose the use of AI in assessing student work, they should refrain from using AI tools for this purpose.
- Accreditation: It is a condition of university accreditation that HSC provides qualified experts to teach and assess our students. Therefore, the use of AI tools alone to evaluate student work is not permissible. In each instance, if AI is used to assist in an instructor’s or TA’s grading or assessment workflow, the AI generated feedback must be reviewed by a qualified expert for accuracy and appropriateness.
HSC Guidance for the Use of AI for Students
AI in teaching and instructional contexts enhances educational delivery by automating administrative tasks, personalizing learning experiences, improving learning methods, and supporting instructional strategies. AI can assist faculty in improving the quality of their courses through the analysis of student achievement and progression. Overall, AI may boost efficiency, enrich the learning experience, and enable more effective, data-informed instruction.
Additionally, AI can support students by facilitating brainstorming, developing initial concepts for projects, and aiding in the drafting of academic materials. To ensure the ethical use of AI in education, HSC faculty should prioritize transparency by clearly explaining the roles and practices of AI tools to students. Continuous monitoring for bias and fairness is essential to prevent perpetuating inequalities, and informed consent must be obtained for data collection. Accountability requires maintaining human oversight and addressing any concerns, while regular reviews and adherence to ethical standards will help ensure that AI tools are both effective and equitable. Finally, it is crucial to ensure accessibility for all students, including those with disabilities.
AI has the opportunity to transform workflows at HSC and beyond. In recognition of the power that AI tools have, this guidance aims to protect students and HSC from AI misuse and the potential consequences. Examples of potential liability related to AI include:
- Reputational Damage: An individual’s improper use of AI can harm the reputations of institutions (HSC) as well as the individual. This means that not only could HSC’s credibility be damaged, but individual instructors’ careers could be damaged as well.
- Intellectual Property Infringement – Using content provided by AI that should be properly sourced and cited AND exposing private data, including the data of students to unauthorized sources.
This guidance exists to help mitigate risk and promote a culture of integrity related to the ethical use of artificial intelligence.
Key Strengths and Weaknesses of AI to Consider - Students
Strengths | Weaknesses |
• Rephrasing: AI tools can rephrase existing content for optimal readability and tone. | • Lack of human touch: AI is not human and may not understand the nuances of human emotion and interaction in the way humans do |
• Grammar: AI tools are generally trained on proper syntax, outputting grammatically correct content. | • Bias: AI is only as good as the data it is trained on, and that data may be susceptible to bias that influences the AI output. |
• Speed: AI is fast, generally providing output seconds after prompting. | • Privacy: Many AI tools record the prompts that users enter. |
• Accessibility: Some AI solutions may be valuable for students with disabilities. If and when AI is appropriate for providing accommodation to a student, that information will be communicated by the Office of Disability Access to the appropriate faculty. | • Timeliness: AI models generally rely on discrete sets of data. Current events and research findings may not be incorporated into the AI tool. |
• Workforce Development: As AI technology advances, it is increasingly integrated into the day-to-day operations of careers which HSC specifically trains our students to enter. It is appropriate to prepare our students to use the tools— including AI tools—that will be required of them in their careers. | • Paywall Avoidance: Many AI systems are trained with openly available resources. Journal articles or other information that requires paid access may not be represented in an AI tool. |
• Learning Analytics: AI tools can analyze large data sets of student engagement with course materials to identify trends and patterns in student performance and highlight opportunities for improvement in curriculum design and teaching. | • Accuracy: Because AI tools are trained on data generated by humans, and humans are flawed, responses from AI tools may also be flawed. In each instance, the accuracy of responses from AI tools should be reviewed. |
• Opportunity for Enhanced Personalization of Instruction/Learning: AI can help identify individual learning preferences and utilize adaptive learning platforms to tailor curriculum (e.g. pacing, real-time |
- Consider your role. Guidance exists to support different roles at HSC (students, instructors, staff, researchers). When operating in other roles (e.g. a TA, a researcher), other guidance may be appropriate and helpful to consult.
- You are responsible for your own work. Regardless of which resources a student chooses to use, students are responsible for the quality and correctness of the coursework they submit.
Students are responsible for evaluating the validity and reliability of the resources they use in their academic work, especially when utilizing AI tools. This involves verifying the credibility of sources, checking for accuracy, and ensuring that the information is current and relevant. AI tools may sometimes provide inaccurate or plagiarized citations, and in such cases, any penalties will primarily be the student’s responsibility. Additionally, students must uphold the quality and integrity of their final work, which includes conducting thorough research, providing proper citations, and adhering to academic standards.
- Instructors decide which tools are allowed. Instructors have the authority to decide what resources a student can use on an assignment or assessment. The HSC Student Code of Conduct and Civility (HSC Policy 7.105) reads that “[u]se of sources beyond those authorized by the instructor in writing papers, preparing reports, solving reports or completing assignments” constitutes cheating. Ideally, these expectations are communicated to students either in assignment instructions and/or in the syllabus. See HSC Policy 6.101 “Academic Freedom and Academic Responsibility” for more details.
Keep in mind, many Generative AI tools often exist inside comprehensive software packages (e.g. Grammarly, Microsoft Office, etc.). Note that if an instructor has indicated AI is not to be used, you may need to adjust your use of these software packages such that you do not utilize the generative components.
If you have doubts or questions about what resources are appropriate for the completion of an assignment or assessment, please review your syllabus or assignment instructions or contact your instructor.
- Use of AI should be open and cited. It is important for students to list the sources used in their work, as uncited resources can be easily confused for the student’s own work. If language or ideas come from AI, that information should be cited. Be mindful that your instructor may also require other documentation or stipulations for disclosing AI use.
- Understand the content. If you use AI to generate content that you do not understand, you may not be achieving the learning objectives of the class.
- Prepare for your career. With the growing capabilities of AI, many professions—including health professions—are seeing AI tools developed specifically for work within those professions. It is appropriate during your studies to prepare to use the kinds of AI tools that will be expected of you as you join the workforce. It is appropriate to rely on your faculty for guidance regarding the use of these tools, and when their use is not appropriate as you engage in learning designed to prepare you for a career.
Social media