AI Medical Scams: 9 Expert Tips for Avoiding Them

How to Not Fall for AI Medical Scams

How to Not Fall for AI Medical Scams
iStock

Artificial intelligence (AI) is reshaping healthcare, and scammers are taking full advantage. AI medical scams now use fake doctor videos, cloned voices, and bogus telehealth sites to look and sound real, luring people into sharing money or medical information.

“The risks are real financially, medically, and emotionally. Victims can lose money, have their health or insurance data stolen, or even take harmful or useless treatments,” says Peter Horadan, CEO of Vouched, a company that builds identity-verification tools for healthcare companies, based in Henderson, Nevada. “On a bigger level, these scams erode trust in healthcare providers and digital platforms that people rely on.”

That growing threat makes it more important than ever to know what these scams look like and how to protect yourself. Here’s how to recognize and avoid the most common AI scams in healthcare.

Pause and Evaluate Before You React

Scammers often create a false sense of urgency to make you feel you must respond immediately.

 If you receive a message, video, or offer that feels suspicious, don’t reply, says Abhishek Karnik, the Hillsboro, Oregon–based director of threat research at McAfee, a cybersecurity software company. “Stop contact right away and avoid responding to any further messages,” he advises.

Taking a moment to verify the source — and asking a trusted healthcare provider — can prevent costly mistakes, says Karnik.

Know the Red Flags

AI scams often follow familiar patterns that include some form of urgency, secrecy, and request for sensitive data.

 “Deepfakes can look convincing — go straight to the source if something feels off,” says Horadan. In other words, verify the claim on the provider’s official site or through their verified contact information.
According to Horadan and Karnik, common signs of an AI medical scam include:

  • Pressure to act immediately or keep it secret
  • Requests for personal, medical, or payment information
  • Offers of “miracle” results or prescription-free drugs
  • Unusual payment methods (gift cards, cryptocurrency, wire transfer)
  • Slightly distorted or unnatural-looking videos
  • Videos or voices with strange word choices or unnatural tone or cadence
  • Spelling errors
  • Slight variations in logos

Don’t Share Sensitive Information

Never share your Social Security number, insurance details, or full medical history through links, messages, or pop-ups, even if they look legitimate, per the Federal Deposit Insurance Corporation (FDIC).

 AI healthcare scams often mimic secure patient portals or provider messages to collect this data, says Horadan.

If a site or chatbot asks for sensitive information before verifying your identity, stop and contact your provider directly, adds Karnik. Legitimate clinics and pharmacies will never request personal data through unsecured channels or social media messages, he says.

Use Secure Payment Methods

Scammers often request payment through wire transfers, prepaid cards, or cryptocurrency because these are hard to trace or reverse, according to the Federal Trade Commission (FTC). If a healthcare service or online pharmacy insists on these options, it’s a red flag for AI-deceptive medical practices, says Horadan.

Before paying, confirm that the site is encrypted to prevent unauthorized access. Look for “https” in the URL in your web browser, which indicates a higher level of security, per the chief information officer of the United States.

Verify Medical Sources Before You Trust Them

Before believing any medical claim — especially one that comes through social media, text, or email — take a moment to confirm the source, says Karnik. Many AI medical scams look professional and trustworthy, borrowing logos, names, and even the likeness of real doctors to make fake information seem legitimate, he adds.

So before engaging, check the source. Karnik recommends visiting the provider’s official website to confirm their contact information, address, and licensing information. You can also confirm licenses through your state medical board or by using the Federation of State Medical Boards’ DocInfo tool.

Be Skeptical of Medical Endorsements

AI can make fake endorsements seem legit when they are not. “Scammers now use AI to create content that looks completely legitimate, from fake ads and pharmacy websites to deepfake videos of doctors,” says Karnik.

If you see a doctor or celebrity promoting a product online, confirm the endorsement on their verified site or social accounts. Karnik also advises being cautious of claims for “miracle” results or quick fixes — both hallmarks of AI deceptive medical practices.

Avoid Buying Medication From Unverified Online Stores

Fake online pharmacies are a common form of AI healthcare fraud. Scammers use AI-generated logos, testimonials, and “FDA-approved” badges to mimic real sites and sell counterfeit or nonexistent drugs, according to Karnik.

“We’ve seen this especially around weight loss drugs like Ozempic and Wegovy,” he says. “Scammers continue to exploit interest in these drugs by creating fake online stores, generating professional-looking images, and even cloning doctor voices to sell counterfeit or nonexistent products.”

Only purchase medication from licensed pharmacies listed on your state’s board of pharmacy or BeSafeRx, an online guide of state-licensed pharmacies from the U.S. Food and Drug Administration (FDA), Karnik adds.

He also recommends avoiding any site offering prescription drugs without requiring a prescription or promising to “ship worldwide” at a steep discount.

Watch Out for Fake Telehealth Platforms

Scammers now use AI to build convincing telehealth or insurance websites that look real but exist only to collect data, says Horadan.

Before booking an appointment or entering your details, verify that the web address matches your provider’s official domain, Horadan adds. He also recommends avoiding links shared through social media or email and searching for reviews or alerts about the service before engaging.

Be Cautious With Chatbots

Not every health chatbot has your best interests in mind. Some are built to steal your personal or payment data, says Horadan.

“Be cautious of medical advice or product recommendations from unverified sources, even if the content looks professional or comes from an AI chatbot,” Karnik adds. “There have been reports of people getting sick after following unsafe health advice generated by [AI].”

Always double-check health information with reputable sources, like the Centers for Disease Control and Prevention (CDC), National Institutes of Health (NIH), or your healthcare provider, before taking action.

What to Do if You Identify or Fall for an AI Medical Scam

If you spot a suspicious medical ad, chatbot, or telehealth site, stop interacting immediately and avoid sharing any more information, says Karnik. Take screenshots or notes about what you saw — they can help investigators later.

Report the scam and where you encountered it to the FTC at ReportFraud.ftc.gov, says Horadan. If the scam uses a real doctor’s name or likeness, alert that provider’s office so they can warn others, he adds.

If you shared financial details, Karnik says to contact your bank or credit card company immediately to block charges. If you entered personal or insurance information, visit IdentityTheft.gov to create a recovery plan and place a fraud alert with credit bureaus, he adds.

“Acting quickly helps limit both health and financial harm,” says Karnik. “The sooner you report and secure your accounts, the less damage these scams can do.”

The Takeaway

  • AI medical scams are becoming more sophisticated, using deepfakes, fake websites, and chatbots to mimic real healthcare providers and collect personal data. Stay alert to these evolving tactics.
  • Always verify medical information and online pharmacies through trusted sources, like your state medical board or the FDA’s BeSafeRx, before engaging or making a purchase.
  • Never share sensitive information or send payment through insecure channels. If you suspect a scam, report it to ReportFraud.ftc.gov and secure your accounts immediately.
  • It’s important to approach AI health tools with caution. Before acting on medical advice from such a tool, confirm the source and check in with your doctor to protect your privacy, finances, and health.

Resources We Trust

EDITORIAL SOURCES
Everyday Health follows strict sourcing guidelines to ensure the accuracy of its content, outlined in our editorial policy. We use only trustworthy sources, including peer-reviewed studies, board-certified medical experts, patients with lived experience, and information from top institutions.
Resources
  1. August Cybersecurity Awareness Tip: How Scammers Manipulate Current Events for Personal Gain. Yale University. December 11, 2025.
  2. Avoiding Scams and Scammers. Federal Deposit Insurance Corporation. October 2021.
  3. Did Someone Insist You Pay Them With Cryptocurrency? Federal Trade Commission. August 28, 2023.
  4. What Are the Signs of a Scam? Federal Trade Commission. March 13, 2025.
  5. The HTTPS-Only Standard. Chief Information Officer of the U.S.
  6. Information You Can Trust. Doc Info.
  7. Boards of Pharmacy. National Association of Boards of Pharmacy.
  8. Locate a State-Licensed Online Pharmacy. U.S. Food & Drug Administration. November 13, 2024.
Justin Laube

Justin Laube, MD

Medical Reviewer

Justin Laube, MD, is a board-certified integrative and internal medicine physician, a teacher, and a consultant with extensive expertise in integrative health, medical education, and trauma healing.

He graduated with a bachelor's in biology from the University of Wisconsin and a medical degree from the University of Minnesota Medical School. During medical school, he completed a graduate certificate in integrative therapies and healing practices through the Earl E. Bakken Center for Spirituality & Healing. He completed his three-year residency training in internal medicine at the University of California in Los Angeles on the primary care track and a two-year fellowship in integrative East-West primary care at the UCLA Health Center for East-West Medicine.

He is currently taking a multiyear personal and professional sabbatical to explore the relationship between childhood trauma, disease, and the processes of healing. He is developing a clinical practice for patients with complex trauma, as well as for others going through significant life transitions. He is working on a book distilling the insights from his sabbatical, teaching, and leading retreats on trauma, integrative health, mindfulness, and well-being for health professionals, students, and the community.

Previously, Dr. Laube was an assistant clinical professor at the UCLA Health Center for East-West Medicine and the David Geffen School of Medicine at UCLA, where he provided primary care and integrative East-West medical consultations. As part of the faculty, he completed a medical education fellowship and received a certificate in innovation in curriculum design and evaluation. He was the fellowship director at the Center for East-West Medicine and led courses for physician fellows, residents, and medical students.

Kathleen Ferraro

Author

Kathleen Ferraro is a writer and content strategist with a master’s degree in journalism from Northwestern University’s Medill School and nearly a decade of experience in health, wellness, and science storytelling across editorial, creative, and marketing roles. She specializes in mental health, sleep, and women's health content, blending evidence-based health reporting with person-first storytelling. She has served as a health editor at Livestrong.com and currently works as a freelance content strategist for health brands including Peloton, Everyday Health, and Verywell Health.