Criminals are exploiting AI to create more convincing scams

Delve into the world of cybercrime where AI chatbots are being weaponized by criminals to orchestrate more convincing scams, and learn the three main methods identified by the police to stay one step ahead of these sophisticated threats.

One of the many cool things about the new wave of Artificial Intelligence tools is their ability to sound convincingly human.

AI chatbots can be prompted to generate text you’d never know was written by a robot. And they can keep producing it – quickly and with minimal human intervention.

So it’s no surprise that cybercriminals have been using AI chatbots to make their lives easier.

Police have identified the three main ways crooks have been found to use the chatbot for malicious reasons.

  1. Better phishing emails
    Until now, terrible spelling and grammar have made it easy to spot many phishing emails. These are intended to trick you into clicking a link to download malware or steal information. AI-written text is way harder to spot simply because it isn’t riddled with mistakes.

    Worse, criminals can make every phishing email they send unique, making it harder for spam filters to spot potentially dangerous content.

  2. Spreading misinformation
    “Write me ten social media posts that accuse the CEO of the Acme Corporation of having an affair. Mention the following news outlets”. Spreading misinformation and disinformation may not seem like an immediate threat to you. Still, it could lead to your employees falling for scams, clicking malware links, or even damaging the reputation of your business or team members.

  3. Creating malicious code
    AI can already write pretty good computer code and is getting better constantly. Criminals could use it to create malware.

    It’s not the software’s fault – it’s just doing what it’s told – but it remains a potential threat until there’s a reliable way for the AI creators to safeguard against this.

The creators of AI tools are not the ones responsible for criminals taking advantage of their powerful software. ChatGPT creator OpenAI, for example, is working to prevent its tools from being used maliciously.

What this does show is the need to stay one step ahead of the cyber crooks in everything we do. That’s why we work so hard with our clients to keep them protected from criminal threats, and informed about what’s coming next.

If you’re concerned about your people falling for increasingly sophisticated scams, be sure to keep them updated about how the scams work and what to look out for.

If you need help with that, get in touch.

Published with permission from Your Tech Updates.