What is Generative AI
What is Generative AI?
Generative artificial intelligence (GenAI) refers to AI models that can generate text, images, or other media, using predictive modeling. GenAI models are initially trained on large datasets. They use predictive algorithms to generate new content that mimics the structure and style of the training data. Text models learn linguistic patterns, image models learn visual features, and audio/video models learn both special and temporal dynamics to produce outputs. To learn more about the process, see visual explanation here: This is how it works
Try it for Yourself
Engaging in exploration of Generative AI tools allows educators to better understand the user experience, potential challenges, and creative possibilities. This firsthand experience enables faculty to provide more informed guidance to students. We invite you to explore more than one of these popular text-based tools to fully understand their potential ChatGPT, Google Bard, Microsoft Copilot, and Perplexity.
View specific tool details on our UHD Faculty Generative AI Libguide
ChatGPT | Quick Tour
Bard | Quick Tour
Copilot | Quick Tour
Perplexity | Quick Tour
Ethical Considerations
- Job loss, economic shifts, income inequality
- Over-reliance on AI, diminishing human capabilities
- Digital divides between those who have AI knowledge and access and those who do not
- New risks for civil and political rights as AI facilitates data collection and mass surveillance
- Global AI power imbalances as nations with advanced AI exert political, economic and military dominance over less developed countries.
- Loss of human control over AI systems
While many students are initially interested in using chatbots, the novelty can wear out pretty quickly, especially when they're surface-level interactions (Deng & Yu, 2023; Wu & Yu, 2024). These surface-level interactions where the responses seem vague or repetitive frustrate students and can cause them to disengage.
At this point, students may not use chatbots to learn more, but they could still rely on them as a shortcut to do their work for them rather than working through problems or thinking critically about the material (Aad & Hardey, 2025; Abbas et al., 2024).
Chatbots can also take the place of meaningful conversations with instructors and classmates, which are often what keep students interested and involved in their learning over time (Aad & Hardey, 2025). Without that human connection, students may begin to disengage and, again, only rely on the chatbot for answers (Abbas et al., 2024).
Other Risks and Limitations
- Hallucinations: Generative AI systems can sometimes produce false content that looks credible. It makes sense in the pattern of data but not in reality.
- Inaccurate sources: If the training data contains inaccuracies, the bot may replicate them. This can lead to more misinformation.
- Outdated information: Many Generative AI models are not trained on data of current events. It might generate information that is obsolete, leading to misinformation.
- Adversarial Use: Malicious actors can exploit generative AI for misinformation, impersonation, scams, or cyberattacks.
- Erosion of Trust: The proliferation of synthetic media (e.g., deepfakes) can undermine public trust in institutions, journalism, and democratic processes.
- Lack of Explainability: AI systems often operate as "black boxes," making it difficult to understand or justify how a particular output was generated.
Read more about 6 Risks of Generative AI & How to Mitigate Them in 2025
Opportunities
Generative AI can offer a variety of support materials and interaction methods tailored for students with disabilities, neurodiversity, multilingual backgrounds, and other challenges they may encounter. Tools with natural language processing and speech recognition can help support students with visual and auditory needs. Involving individuals with disabilities must be at the forefront of these tools' development to ensure accessibility and inclusivity