tl;dr - It's a computer you can talk to, and it (deceptively) appears to understand.
Generative artificial intelligence (GAI) is a form of computer application that receives data from users in highly accessible modes like text, images, and sound files, and uses predictive data-crunching to generate responses that plausibly imitate human modes of expression contained in their data. Whereas prior computer applications either required well-formed computer code or a structured user interface for users to interact with them, the built-in natural language processing (NLP) and natural language generation (NLG) capabilities of GAI tools enable users to talk and issue instructions to them as though interacting with a person, even though the affordances and limitations of GAIs and humans often differ in fundamental ways.
❖
tl;dr - It calculates mathematical relationships between words and uses that to predict what words are most likely to go with which. AI companies then pay human evaluators (RLHF) to rate its responses in order to steer the AI away from answers that might cause the AI company trouble or endanger users; but this is like playing whack-a-mole and often doesn't work.
Generative artificial intelligence models are trained using large datasets containing trillions of words and subwords, which are converted into smaller units called “tokens.” Each token is represented by a numerical vector known as an “embedding,” which captures its relationship to other tokens in the dataset. During training, the AI analyzes these embeddings to learn patterns and associations between tokens, predicting which tokens are likely to follow others. It does this by processing the vectorized tokens through a neural network, that is, a network of nodes.
Inside the AI, there are many sets of calculations called nodes. Each node takes input numbers from the vectors and multiplies them by weights assigned to that node. When the AI model is first created, these weights are random starting values assigned using a random number generator. Then, in each layer of its training, the AI is made to guess which token comes next in a sequence. It compares these guesses to the correct answers and adjusts the weights in the nodes of the next layer to make fewer mistakes. By repeating this across many nodes and layers, the model builds up increasingly sophisticated representations of the network of embedded tokens, and in turn, of the human language captured by the data it was trained on. This approach of allowing the AI to learn by testing and refining its next-word predictions across multiple stacked layers is called unsupervised or deep learning.
But learning models struggled to connect information from early in a sequence of tokens to information in later parts of the sequence. A solution was found by Google researchers in 2017 through the creation of the transformer, whose "attention mechanism" enables the model to consider all tokens in a sequence simultaneously and assess their relevance for generating predictions.
The complex network of relationships formed by these embeddings is known as the model’s “latent space.” After initial training, developers often refine the model’s behavior using a method called “reinforcement learning from human feedback” (RLHF), where human evaluators provide guidance to improve the model’s responses, particularly to address issues related to ethical considerations and user experience. (As has been true with social media content filtering on websites like Facebook and Instagram, much of this review labor is carried out for low pay by workers recruited from countries in the Global South, as well as through countless uncredited contributions by individuals with humanities expertise directly and indirectly employed by AI companies.)
❖
tl;dr - No. But (like Wikipedia) it can be a useful first step on your journey to better, more reliable sources.
When the RLHF safe-guards imposed on a generative AI fail—or when the user (intentionally or unintentionally) causes a successful jailbreak—then GAIs will frequently go along with any crazy premise or misinformation the user presents to it, regardless of plausibility or truth. This is because generative AI models have no awareness of what is true or what is not true, only what words are likely to go together.
Aside: You don't have to be trying to break the GAI either; they constantly get things wrong, and often by roleplaying or automatically mirroring your language: oh, I'm so beat by these exams! you might say. Yeah, I often feel the same way, the AI responds. No, it doesn't feel the same way—it is mirroring language that might be used by the type of person who might feel "beat by these exams." Its response is an artifact of the probabilistic co-occurrence of words, not real empathy.)
One result of this are what have come to be known as “hallucinations”: confidently stated information that is invented, inaccurate, or misleading. While researchers and corporate leaders in the AI industry often represent hallucinations as a problem that can be solved, others are of the view that hallucinations are a fundamental property of generative AIs—an inevitable result of how they go about generating text.
While AI companies try to use techniques like RLHF (reinforcement learning from human feedback) to patch the most obvious problems, these fixes are often temporary and incomplete—like a game of whack-a-mole, where new issues appear as quickly as old ones are suppressed. We can see this in how GAI companies are constantly tweaking their models in response to customer complaints, now responding to customers (like the characters in the story of Goldilocks and the Three Bears) who find certain models too dry and boring, now rushing to patch the problem when certain models are found “too sycophantic.” These examples illustrate that, on some level, the “alignment problem” is a result of the inevitable effects of averaging out language, rather than the fake issue of a non-thinking machine needing a moral adjustment.
It is less often observed that GAI outputs are rooted in societal genres--the typical forms of speech and writing that have been repeated countless times. What this generic bias in GAI outputs means is that you may often get results that are true (or at least reflected in language) a lot of the time, but not true in the specific circumstances you are applying them to (and not responsive to the specific assignment or class context you are attempting to use them for.)
But there are further problems with trusting GAIs that go beyond just how they are trained: There is the risk that commercial or political actors could manipulate these systems to serve specific agendas, intentionally shaping what information is promoted or suppressed. What happens if the owners of these LLMs intentionally train the models to censor a particular political viewpoint, or systematically align their outputs with the personal politics of corporate leaders? Even assuming we can trust the intentions of those who run AI companies (quite a big assumption), it has been frequently observed that the datasets used to train GAI models are saturated with many of the same social, cultural, and political biases which are to be found in their training data, which are inevitably reflected in the outputs.
What this means for your research papers: you should treat generative AI the way you might treat Wikipedia: a useful place to gather initial ideas or terms, but not an authoritative or citable secondary source. If your assignment is about how AI responds to questions, then its output can be your primary source. But for topics where factual accuracy matters, always use library databases and scholarly sources—and ask your librarian for help if you’re unsure how to find them.
❖
This is a thorny question that requires us to unpack multiple related, but distinct issues.
tl;dr - Each professor will determine how much AI use is acceptable (if any) and how you need to disclose it in their VWU class. (So ask them!) But even when AI use is allowed, information from GAIs always originates from somebody else; so it's up to us to find the real source (a person, not an AI) and give them credit.
❖
Let's start with the practical before treating this as a philosophical question; the Virginia Wesleyan University Honor Code provides institutional definitions of terms like cheating and plagiarism that we are bound to follow while we are members of the VWU community:
Cheating is the deliberate submission of work for a grade or credit that is not one's own or that violates professors' implied or stated instructions concerning the type and amount of aid permitted. The student who gives prohibited aid shall be considered as responsible as the student who receives it.
Plagiarism is the oral and/or written presentation of words, facts, or ideas belonging to another source without proper acknowledgment. This includes the reuse of the student’s own academic work from another class or assignment without the permission of the instructor.
Notably, these guidelines refer to professors' implied or stated instructions, and what this means is that what counts as acceptable or unacceptable use of generative AI is determined by the professor for each course. (So, is this way of using GAI acceptable or not acceptable in this class? Step one: ask the professor.)
❖
But beyond determining what the official policy is regarding using GAI for assignments in any given class, we can also have a broader discussion about whether different ways of using generative AI are ethical or not. Ethics is a branch of philosophy concerned with debating what is right or wrong. Outside the bounds of purely academic discussion, it is also a way of thinking about our personal values: who are my actions helping? Who is being hurt? Is what I am doing having an overall positive or negative effect on the world?
Imagine you created an amazing thing called X. Maybe X is a hilarious character you designed for a comic strip, a beautiful piece of music you spent hours writing and recording, or a brilliant idea for a science project that you developed through weeks of research, late-night brainstorming, and dozens of drafts. You poured your time, talent, and heart into making X, practicing, revising, and sometimes even sacrificing sleep or social plans to get it right.
Now, picture this: you start seeing X show up in unexpected places—a classmate posts your character on their Instagram feed and claims it as their own; someone you once chatted with is performing your song at an open mic and acting like they wrote it; a student in another class presents your science idea as their project, never mentioning you. Maybe a corporation, realizing that X is a hit, sells the film and TV rights and puts X on billboards. Suddenly, X is everywhere, but no one knows—or seems to care—that it was yours. Even worse, you now have no way to earn money or recognition for your work in creating X.
Fortunately, our legal system has invented a way to prevent this from happening: as soon as you create something that is uniquely and recognizably your own work, it becomes your copyrighted intellectual property. That means you have a right to be credited, and possibly compensated, for your creation. (And to take people to court if they try to profit off of your invention.)
But now, here's another plot twist: X starts popping up as a generic answer to questions online, delivered by generative AI. The AI makes X seem like a piece of “ambient information”—just another random fact floating around, with no hint that you ever existed or played the main role in creating it.
This is what’s at stake in the debate about AI and authorship: if creators aren’t credited or respected, their hard work can disappear into the background, and the meaning of authorship—and your right to be recognized—can vanish, too.
This problem is known as “documentation debt,” a term introduced by Bender et al. in their influential article, On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜 (2021). Documentation debt describes what happens when large language models (LLMs) like ChatGPT remix and reproduce information from countless sources, but do not keep track of where specific ideas, phrases, or techniques originally came from.
The “many-hands problem” is the challenge of figuring out how to fairly distribute credit when something—like a piece of writing generated using a GAI—is the result of contributions from many different people, not just the person attempting to claim it as just their work. In their 2024 article "Engaging the many-hands problem of generative-AI outputs: a framework for attributing credit," Donal Khosrowi, Finola Finn, and Elinor Clark tackle this problem by proposing the CCC (collective-centered creation) framework, which lays out key criteria for us to decide when and how different contributors should be credited.
The table below adapts Khosrowi, Finn, and Clark's framework, helping us think about what makes a contribution significant enough that we need to take further steps to ensure that an idea's original creators get credit.
Criterion | Description | Deserves Credit | Doesn’t Need to be Credited |
---|---|---|---|
Relevance/ (Non-)Redundancy and Control |
Did a source provide unique, irreplaceable content that shapes the AI’s answer? | An AI summarizes a distinctive argument from a journal article or book chapter that can be traced to a particular author; this idea should be cited, not just presented as general knowledge. | An AI lists the boiling point of water or recites the alphabet; these facts are universally known and don’t need individual credit. |
Originality | Was the information or idea novel or distinctive when first introduced in the training data? | An AI presents a newly-coined scientific theory or a unique artistic technique originally described in a specific publication; attribution is necessary. | The AI restates common proverbs or sayings that have become public domain knowledge over centuries. |
Time/Effort | Did a creator invest significant labor into developing the information in the training data? | An AI outputs a recipe or poem that took a chef or poet years to perfect and publish; this labor merits acknowledgment. | The AI repeats a basic definition (e.g., “photosynthesis is how plants make food”) from widely available textbooks. |
Leadership/Independence | Was the information developed independently or under the direction of a particular thinker, researcher, or artist? | The AI provides a step-by-step process for a therapeutic intervention designed and tested by a specific psychologist; proper credit is needed. | The AI offers a generic study tip, like “review your notes regularly,” found in countless sources. |
Directness | Was the AI’s output a direct reproduction or close paraphrase of the training data? | The AI generates a paragraph nearly identical to a section of a copyrighted novel or research paper; this should be flagged and attributed. | The AI generates a summary that blends information from dozens of sources into a generic, fact-based overview. |
❖
It depends. If your professor has explicitly prohibited the use of generative AI like ChatGPT, or if you use it without proper attribution, then yes, it may be considered plagiarism under the Virginia Wesleyan University Honor Code. Even when AI use is permitted, you are still responsible for ensuring that any distinctive ideas, arguments, or phrasing (especially those drawn from other people's work) are properly credited.
Using generative AI doesn't remove your responsibility to engage thoughtfully with the material or to give credit where it's due. When you fail to acknowledge the contributions of others, whether they are human authors or the original sources embedded in an AI's output, you risk erasing the effort, labor, and creativity behind those ideas.
Bottom line: Always check your professor’s policy first, disclose your use of generative AI when required, and be intentional about citing where ideas come from. Doing so not only avoids plagiarism; it also honors the intellectual work of others and allows you to take your rightful place in the ongoing scholarly conversation.
❖
tl;dr - Any information you put online—including prompts to an AI—can potentially be sold to a third party, turn up in a web crawl, or be exposed to hackers. And if you've been reading this far, you know that GAIs 1) don't know what truth is and 2) don't have real empathy, just words that generically go together.
Any information you put online—including prompts to an AI—can potentially be sold to a third party, turn up in a web crawl, or be exposed to hackers. Beyond privacy concerns, it’s important to remember that GAIs like ChatGPT do not actually know what truth is, nor do they possess real empathy. Instead, they generate responses by stringing together words and ideas that statistically go together, without any genuine understanding or emotional awareness. Therefore, it’s not wise to treat an AI like a therapist, confidant, or trusted professional, even if it sometimes sounds convincing.
Need a real, kind, and empathetic human being to talk to?
Reach out to the good folks in ★Counseling Services!★ (link)
LegalEagle is a popular Youtube channel among law school students and enthusiasts because it reviews movies and TV shows for how accurately they portray real legal practice. (There are a lot of other channels like this: historian rates historical movies, doctor rates TV hospital shows, etc.) Legal Eagle, who is a practicing lawyer, will often explain in his videos what a movie gets right and wrong about trials and lawyerly practice and rate pop culture for its "legal realism."
Aside: A Youtube video by someone claiming to be an expert should probably not be considered as authoritative a source as a peer-reviewed scholarly article or book on the same topic. Lateral reading is a fact-checking strategy that involves consulting multiple sources of information to determine the credibility of our original source, and is one effective strategy we can use to help prevent the spread of misinformation.
If we extend this idea to artificial intelligence, it’s important to realize that language models like ChatGPT are trained on huge collections of internet data—including not only textbooks and articles, but also fiction, movie and TV scripts, blogs, and online social media conversations.
This means that when ChatGPT “pretends” to be a lawyer, a doctor, a therapist, or a teacher, it’s often blending real-world information with pop culture and fictionalized accounts. Just like movies often exaggerate for dramatic effect, AIs can end up giving advice or explanations that sound convincing but don't actually reflect how things work in real life. For example, a real lawyer would know the actual process of filing a motion or the details of courtroom etiquette, while AI simply reproduces generic behavior from popular movies—leading to errors, omissions, or suggestions that are unrealistic or simply wrong.
This is why, if you asked real professionals—lawyers, therapists, teachers, doctors, or software engineers—to judge the “realism” of AI-generated advice or explanations, the AI would likely receive a low score, much like some Hollywood movies do. Professionals are trained to recognize nuance, understand context, follow professional codes of ethics, and apply knowledge to specific situations. AI, on the other hand, is simply mimicking patterns of language, not practicing with true understanding, skill, or responsibility.
This highlights why developing your own expertise is crucial, especially in college. The ability to discern truth from fiction, apply knowledge accurately, and communicate ideas clearly are skills you build through authentic engagement with your field, not by copying and pasting, or relying on an algorithm that “plays an expert” but is not one.
In summary: Don’t treat AI like a best friend or a trusted professional—use it as a tool, but remember that your privacy, your learning, and your ability to think independently matter much more. Developing genuine expertise takes time and effort, but it’s the only way to become the kind of real-world professional who can spot the difference between movie magic, AI-generated responses, and authentic, actionable knowledge.
❖
❖
AI Use Disclosure Guide |
|
When to NOT USE GAI |
|
When to write an AI USE DISCLOSURE STATEMENT |
|
When to CITE GAI AS A SOURCE |
|