Planning for your Life, your Livelihood, your Legacy
Book an Initial Call
It seems like you could not go anywhere in 2025 without running into AI. Artificial intelligence has brought some convenience, but it has also given scammers powerful new tools to deceive people of all ages. These schemes can mimic a relative’s voice, impersonate legitimate companies or professionals, and create fake emergencies that feel urgent and real. Protecting yourself and the people you care about requires awareness, a plan, and open conversations about warning signs—especially in a state like Oregon, where regulators regularly warn about increasingly sophisticated AI‑driven scams. With the right strategies, Oregonians can stay confident and secure while using technology.

Voice cloning and “urgent” calls
AI tools can clone voices from short audio clips, making “relative in trouble” calls more convincing than ever. A scammer may say a loved one has been arrested, is stuck in another country, or needs money for an emergency, then pressure the person to act quickly and secretly. Its estimated that 22% of bad actors are now using AI to generate deepfake videos posing as loved ones or celebrities. Scammers are also able to take control of existing financial or social media accounts to steal funds or solicit money.
Polished messages and emails
AI helps scammers write grammatically correct, personalized phishing messages that look like they are from banks, delivery services, government agencies, or even Oregon lawyers and investment professionals. These messages often ask the recipient to click a link, download an attachment, move money, or share sensitive information.
Exploiting trust and urgency
Many scams rely on emotional pressure—fear, excitement, or a “limited‑time opportunity”—to override careful thinking. Oregon officials warn that criminals frequently impersonate government, law enforcement, financial experts, or attorneys, and emphasize that legitimate organizations rarely demand immediate payment or secrecy.
Long‑term “friend” or “companion” scams
Scammers now use AI‑driven chatbots, social‑media personas, and “virtual companions” to build relationships over weeks or months before asking for money or investment. These conversations may feel friendly, supportive, and highly personalized, which can make it harder to recognize red flags.
Microtransactions and emotional dependence
Some AI companion services or apps encourage frequent small payments that add up over time and can create emotional dependence, especially for people who feel isolated. This pattern can make users more vulnerable to additional offers, upsells, or external scams.
Warning signs to watch for
Oregon’s Department of Justice and Division of Financial Regulation specifically warn about “pig‑butchering” and other relationship‑based investment scams that use social media and messaging apps to build trust and then push fraudulent crypto or trading schemes.
Create a verification system
Strengthen digital habits
Use Oregon‑specific resources
Powers of attorney and oversight
A well‑drafted Oregon durable power of attorney can authorize a trusted person to monitor accounts, spot unusual activity, and act if someone is manipulated into sending money, while still respecting the individual’s autonomy when things are going well. Families should pair this with clear communication about when and how that authority will be used.
Trusted contacts and monitoring
Many banks and investment firms allow customers to name a “trusted contact” who can be reached if the institution sees suspicious activity. Encouraging all adults—especially those managing significant savings or retirement accounts—to add a trusted contact gives financial institutions another way to help detect and stop fraud.
Reporting and remedies in Oregon
If someone in Oregon is scammed or you suspect financial exploitation, you can:
By treating AI‑driven scams as a shared risk—and building habits, legal tools, and support systems around that reality—people of all ages in Oregon can stay connected, independent, and significantly safer online.
References:
Oregon Dept. Of Consumer & Business Services News via Flashalert.net
"What Are AI Scams? A Guide for Older Adults"
Other good resources about protecting loved ones from AI:
Privacy Law FAQs for Businesses – DOJ FAQ that touches on privacy and data use, a good background citation when you discuss how personal data can be misused by AI tools.
Oregon attorney general issues advice to businesses on use of AI – news release summarizing DOJ guidance on AI and existing Oregon laws; supports general statements about Oregon taking AI risks seriously.
OR AG Issues Guidance Regarding OR State Laws and AI – analysis of the AG’s guidance that emphasizes legal obligations around AI and consumer protection
1865 NW 169th Place, Suite 202
Beaverton, Oregon 97006
201 NE Park Plaza Drive, Suite 200
Vancouver, Washington 98684
