11 Times You Should Absolutely Not Use ChatGPT
ChatGPT has quickly become the internet's most popular AI chatbot, and for good reason. In recent years, AI has transformed how we interact with the world, simplifying everything from planning a trip to saving money on groceries.
However, while AI is a powerful tool, it's crucial to understand its limitations. ChatGPT is great for brainstorming recipes or learning a new language, but you shouldn't give it free rein over your life. The chatbot can be surprisingly unreliable, sometimes inventing information and presenting it as fact—a phenomenon known as hallucination. It can be incredibly confident even when it's completely wrong, and this is a risk you can't afford in high-stakes situations.
Whether it's dealing with taxes, medical issues, or legal matters, knowing when to avoid ChatGPT is key. Here are 11 scenarios where you should choose another option.
(Disclosure: Ziff Davis, the parent company of CNET, in April filed a lawsuit against ChatGPT maker OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.)
1. Diagnosing Physical Health Issues
It can be tempting to enter your symptoms into ChatGPT out of curiosity, but the results can be terrifying and misleading. A simple query about a lump on the chest could yield responses ranging from dehydration to cancer. This was the author's experience, who was told by ChatGPT they might have cancer, only for a licensed doctor to diagnose a non-cancerous lipoma. AI can help you draft questions for your doctor or translate medical jargon, but it cannot examine you, order labs, or carry malpractice insurance. For a diagnosis, always trust a qualified medical professional.
2. Managing Your Mental Health
While ChatGPT can offer grounding techniques, it is not a substitute for a real therapist. It lacks lived experience, cannot read body language or tone, and has zero capacity for genuine empathy. A licensed therapist operates under professional codes that protect you, whereas ChatGPT does not. Its advice can misfire or overlook critical red flags. The hard, messy work of therapy should be left to a trained human. If you or someone you know is in crisis, please contact the 988 lifeline in the US or your local hotline.
3. Making Immediate Safety Decisions
If your carbon monoxide alarm goes off, do not waste time asking ChatGPT if you're in danger. Go outside first and ask questions later. An AI cannot smell gas, detect smoke, or call an emergency crew. In a crisis, every second you spend typing is a second you're not evacuating or dialing 911. Treat your chatbot as a tool for post-incident explanations, never as a first responder.
4. Getting Personalized Financial or Tax Advice
ChatGPT can explain financial concepts, but it doesn't know your personal financial situation, tax bracket, or retirement goals. Its training data may not include the latest tax laws or rate changes, making its advice potentially outdated and costly. Sharing sensitive information like your income or Social Security number is also a major privacy risk. When money and IRS penalties are on the line, call a professional, not an AI.
5. Handling Confidential or Regulated Data
Never paste sensitive information into ChatGPT. This includes client contracts, medical charts, or anything covered by privacy laws like HIPAA or GDPR. Once that data is in the prompt window, you lose control over where it's stored, who can see it, and whether it will be used to train future AI models. If you wouldn't post it in a public forum, don't paste it into ChatGPT.
6. Doing Anything Illegal
This should be self-explanatory. Do not use AI to assist with illegal activities.
7. Cheating on Schoolwork
Using AI to write your essays is a high-risk gamble. Plagiarism detectors like Turnitin are constantly improving their ability to spot AI-generated text, and professors can often recognize the distinct, unemotional 'ChatGPT voice'. The consequences, including suspension or expulsion, are not worth it. Use ChatGPT as a study partner to brainstorm or explain concepts, not as a ghostwriter to do the work for you.
8. Monitoring Breaking News
While newer versions of ChatGPT can search the web for current information, it is not designed for real-time news monitoring. It cannot provide a continuous stream of updates on its own; each refresh requires a new prompt. For fast-moving events, live data feeds, official news sites, and streaming coverage are still your most reliable sources.
9. Gambling
Don't rely on ChatGPT to predict the winner of the next big game. The author notes that while they once got lucky on a parlay, it was only after double-checking every piece of AI-generated information. ChatGPT is prone to hallucinating player stats, misreporting injuries, and getting win-loss records wrong. It can't see the future, so don't bet your money on its predictions.
10. Drafting Wills or Legal Contracts
ChatGPT is useful for understanding basic legal concepts, like a revocable living trust. However, asking it to draft an actual legal document is a terrible idea. Estate and family law vary significantly by state and even county. A missing signature or an incorrect clause can render the entire document invalid. Use the AI to prepare questions for your lawyer, then pay that lawyer to create a document that will actually hold up in court.
11. Making Art (An Ethical Consideration)
This is a matter of opinion, but the author argues that AI should not be used to create art that is then passed off as one's own. While AI is a fantastic tool for brainstorming ideas and supplementing the creative process, using it as a substitute for human creativity and originality is ethically questionable.