Holiday phishing scams have an unlikely ally in 2023 - generative artificial intelligence tools like ChatGPT.
As consumers and companies spend a record-breaking $270 billion in online shopping, generative AI is arming phishing attackers with a new weapon in their arsenal.
In response, cybersecurity professionals are also leveraging AI to build innovative machine-learning tools that combat growing instances of phishing fraud.
Cybercriminals have long considered online holiday fraud to be a bonanza. In 2022, more than $73 million was gobbled up by online con artists, according to the FBI.
Researchers conclude that AI contributed to a 1,265% increase in phishing email scams last year.
ChatGPT and its malicious offshoots like FraudGPT and WormGPT are making phishing as well as spear phishing campaigns easier and faster to deploy against vulnerable targets. Researchers from IBM conclude that ChatGPT can write convincing emails almost as well as social engineering experts. All this while taking a fraction of the time it takes human beings to do so.
AI tools can do a better job with:
Lack of obvious spelling and grammar mistakes that give away scam texts and messages
Mimicking legitimate websites and making them almost impossible to distinguish from the real thing.
Experts agree that generative AI is boosting both success rates and volumes of online scams this year by making use of tactics such as Amazon gift card scams, charity scams and fake delivery tracking links.
For its part, ChatGPT has certain safeguards built into it to forbid malicious applications. However, threat actors are using carefully worded prompts to sidestep those safeguards. Tools like WormGPT use their own custom language model and data sources which makes them ideal for cybercriminals.
As much as these cyber scams have gotten more conniving and convincing, cybersecurity tool makers are also rising to the threats with the help of AI.
In November, Google released an open-source RETVec (Resilient & Efficient Text Vectorizer) tool that trains spam filter AI models against text manipulation tactics used by phishers.
These text manipulation methods can be classified into:
Homoglyphs: Letters or numbers that look the same but mean different things, like the big letter O and the number zero (0).
Invisible characters: Empty spaces that sneaky people might use to spread out tricky words in messages to fool spam filters.
Keyword stuffing: Hiding extra words in emails to trick spam detectors into thinking the message is normal and not spam.
Google claims that its AI tool has:
Improved spam detection by 38%
Reduced false positives by 19.49%
Minimized false negatives by 17.71%
Cybersecurity software solutions provider Norton has released its free Norton Genie app. Part AI chatbot and part text analyzer. The app scans uploaded messages for signs of phishing, even offering answers to users’ questions about any content they find suspicious.
Norton says that Genie is always learning and has been trained on millions of scam messages that users have uploaded so far. It is as easy as copy-pasting text or uploading a screenshot, or sending a link to Norton’s free app.
Reports of new holiday phishing scams and AI-powered anti-phishing tools showcase how the fight against cybercrime continues to be an ongoing concern.