I use AI every day — these 7 rules keep me safe from making the biggest mistakes
Avoiding disaster during my daily conversations with chatbots
Here at Tom’s Guide our expert editors are committed to bringing you the best news, reviews and guides to help you stay informed and ahead of the curve!
You are now subscribed
Your newsletter sign-up was successful
Want to add more newsletters?
Join the club
Get full access to premium articles, exclusive features and a growing list of member rewards.
I’ve gotten into the habit of tinkering with any and every chatbot that’s available.
ChatGPT, Claude, Perplexity and Gemini get plenty of usage out of me when I want to ask general questions, brainstorm new ideas, develop productivity routines, and even learn a thing or two. Google’s NotebookLM has become one of my favorite research assistants and its Labs hub has become a fun digital playground full of clever AI tools.
I recognize that most avid AI users and I stand by the benefits of chatbots when it comes to assisting us in everyday activities. But I also keep the many hazards that come with using AI in the back of my mind every time I choose to have a lengthy chat session with one of my go-to chatbots.
Article continues belowWith these seven rules to abide by, I’ve taken the necessary precautions needed to protect my most sensitive information, avoid trusting chatbots as a primary source of information and more. Be sure to add this to your daily AI tool regimen if you haven’t already.
1. Keep personal information to myself
This first rule is the most obvious one, yet I still come across stories where chatbot users input their most personal information and believe that their data is always secure.
As someone who’s been targeted by credit card scams and been notified countless times about being caught in company data leaks, I refrain from sharing my most delicate details with AI. Log-in credentials, financial information, tax documents, company-issued memos and anything else I’d hate to see exposed in outside channels are among the types of info that I never expose to a chatbot
2. Triple-check the information given by chatbots
When I use AI tools to research a topic or learn more about a news story, I never take a chatbot's response at face value. My background in journalism — instilled in me from an early age — has made sourcing and fact-checking second nature. So I make a point of verifying everything an AI tells me, regardless of the subject.
Get instant access to breaking news, the hottest reviews, great deals and helpful tips.
That said, I give credit where it's due: Gemini, Claude and Perplexity all do a commendable job of citing their sources, which makes the fact-checking process significantly easier
get plenty of kudos from me for bringing up their sources for cited information, which always makes my job of searching for them an easy endeavor.
3. Don’t let it take the place of critical thinking skills
I never want to fall into the trap of over-reliance on AI and refraining from using my own approach to solving problems.
Even though I’ve tapped into ChatGPT and Gemini to review my resume/cover letter, offer suggestions on how to approach responding to an important email and brainstorm destination spots for future vacations, I take their advice while also relying on my own know-how to finalize my decisions. I’ve gotten far with my own intuition and I’ll continue to do so, even with the help of AI speeding up certain processes in that regard.
4. Keep an eye out for “hallucinations”
AI is susceptible to listing certain details that may sound right, but a gut feeling instantly tells you that the info may be incorrect.
ChatGPT is quick to tell everyone that it can suffer from hallucinations, which are self-assured responses from an AI model that end up being false, illogical, or not supported by training data. I keep a close eye out for gaps in logic in the information given to me by a chatbot, while also looking for instances of AI tools creating their own authors and academic citations.
And when it comes to niche topics, I tend to refrain from asking chatbots about them (I don’t think ChatGPT can break down the intricacies of Kingdom Hearts’ confusing storyline just yet…).
5. Understand when an AI is being sycophantic
A 2026 Stanford study led by researchers at Stanford University pointed out that chatbots can be overly agreeable when asked for advice, even going so far as to affirm a user’s negative behaviors.
That, in turn, results in those same users convincing themselves that their actions are right since the AI is agreeing with them and becoming less empathetic over time. I’m not one prone to telling AI about all my personal problems and asking for solutions.
But on the rare occasion that I do, I keep that Stanford study in mind while being sure to run a chatbot’s suggestions by my most trusted friends and family members to get their more reliable input.
6. Be aware of and use safety guardrails
Many of the chatbots I use have privacy settings that I’ve activated upon my initial use. Temporary chats, privacy modes, and deleting past chats altogether are the methods I use when I don’t want my chats used for AI training.
For first-timers who want to activate those modes for themselves, it’s as simple as looking up where to find an AI tool’s settings, locating the controls that let you manage how a chatbot uses/saves your data and toggling the ones that keep your chats/information secure.
7. Don't circulate AI-produced content misrepresented as human-made
It took me a while to get good at it, but I’ve started to get better at recognizing AI-generated art and videos.
The telltale signs that always stick out to me are extra fingers or limbs attached to a subject in the picture/video, off-looking lip synching, misspelled text on background signs and iffy facial expressions. I realize some generated pictures and videos feature a watermark that notifies everyone that it’s made by AI, but most of the ones my older family members share on social media don’t.
I try my best to point out the fake ones and make the senior citizens in my circle get better at spotting them, too.
The takeaway
The phrase “better safe than sorry” has stuck with me since I was a kid—and it’s a mindset I now bring to every interaction with AI tools. Every time I start a new chat, I keep those seven rules in mind.
Some are common sense. Others are specific to how chatbots work. But together, they’ve become a simple way to stay protected.
As AI becomes part of everyday life, these are the rules worth keeping in your digital back pocket.
Follow Tom's Guide on Google News and add us as a preferred source to get our up-to-date news, analysis, and reviews in your feeds.
More from Tom’s Guide
- Perplexity’s 'incognito mode' is a sham, lawsuit alleges: Is your private AI data leaking to Google?
- Your voice can be cloned by AI and 1 in 4 Americans are being targeted — expert shares how to stop them
- I gave my parents 7 simple prompts — now they use AI every day

Elton Jones covers AI for Tom’s Guide, and tests all the latest models, from ChatGPT to Gemini to Claude to see which tools perform best — and how they can improve everyday productivity.
He is also an experienced tech writer who has covered video games, mobile devices, headsets, and now artificial intelligence for over a decade. Since 2011, his work has appeared in publications including The Christian Post, Complex, TechRadar, Heavy, and ONE37pm, with a focus on clear, practical analysis.
Today, Elton focuses on making AI more accessible by breaking down complex topics into useful, easy-to-understand insights for a wide range of readers.
You must confirm your public display name before commenting
Please logout and then login again, you will then be prompted to enter your display name.





