Connect with us


AI-generated spam may soon flood your inbox with scams

Improved AI spam filters could accurately detect and prevent unwanted spam while also allowing legitimate marketing emails to come through.

The image shows an email inbox with various folders containing emails from different senders. Full text: inbox (2) · sender · marie sent · sales drafts. Thomas king. John b spam (316) · don gutenberg trash (13) · frank zurtionze · ruthie overs. Monty. Jewell
Image: Getty Images

Each day, messages from Nigerian princes, peddlers of wonder drugs, and promoters of can’t-miss investments choke email inboxes.

Improvements to spam filters only seem to inspire new techniques to break through the protections.

Now, the arms race between spam blockers and spam senders is about to escalate with the emergence of a new weapon: generative artificial intelligence.

With recent advances in AI made famous by ChatGPT, spammers could have new tools to evade filters, grab people’s attention and convince them to click, buy or give up personal information.

As director of the Advancing Human and Machine Reasoning lab at the University of South Florida, I research the intersection of artificial intelligence, natural language processing, and human reasoning.

I have studied how AI can learn people’s individual preferences, beliefs, and personality quirks.

This can be used to understand better how to interact with people, help them learn, or provide helpful suggestions.

But this also means you should brace for smarter spam that knows your weak spots – and can use them against you.

Spam, spam, spam

New gmail spam
Image: Unsplash

So, what is spam? Spam is defined as unsolicited commercial emails sent by an unknown entity. The term is sometimes extended to text messages, direct messages on social media, and fake reviews on products.

Spammers want to nudge you toward action: buying something, clicking on phishing links, installing malware, or changing views.

Spam is profitable. One email blast can make $1,000 in only a few hours, costing spammers only a few dollars – excluding initial setup. An online pharmaceutical spam campaign might generate around $7,000 per day.

Legitimate advertisers also want to nudge you to action – buying their products, taking their surveys, and signing up for newsletters.

Still, whereas a marketer email may link to an established company website and contain an unsubscribe option in accordance with federal regulations, a spam email may not.

Spammers also lack access to mailing lists that users signed up for. Instead, spammers utilize counter-intuitive strategies such as the “Nigerian prince” scam.

A Nigerian prince claims to need your help to unlock an absurd amount of money, promising to reward you nicely.

Savvy digital natives immediately dismiss such pleas, but the absurdity of the request may actually select for naïveté or advanced age, filtering for those most likely to fall for the scams.

Advances in AI, however, mean spammers might not have to rely on such hit-or-miss approaches.

AI could allow them to target individuals and make their messages more persuasive based on easily accessible information, such as social media posts.

Future of spam

Chatgpt on laptop
Image: Pexels

Chances are you’ve heard about the advances in generative large language models like ChatGPT.

The task these generative LLMs perform is deceptively simple: given a text sequence, predict which token – think of this as a part of a word – comes next.

Then, predict which token comes after that. And so on, over and over.

Somehow, training on that task alone, when done with enough text on a large enough LLM, seems to be enough to imbue these models with the ability to perform surprisingly well on a lot of other tasks.

Multiple ways to use the technology have already emerged, showcasing the technology’s ability to adapt to, and learn about, individuals quickly.

For example, LLMs can write full emails in your writing style, given only a few examples of how you write. And there’s the classic example – now over a decade old – of Target figuring out a customer was pregnant before her father knew.

Spammers and marketers alike would benefit from being able to predict more about individuals with less data.

Given your LinkedIn page, a few posts, and a profile image or two, LLM-armed spammers might make reasonably accurate guesses about your political leanings, marital status, or life priorities.

Our research showed that LLMs could be used to predict which word an individual will say next with a degree of accuracy far surpassing other AI approaches in a word-generation task called the semantic fluency task.

We also showed that LLMs can take certain types of questions from tests of reasoning abilities and predict how people will respond to that question.

This suggests that LLMs already have some knowledge of what typical human reasoning ability looks like.

If spammers make it past initial filters and get you to read an email, click a link, or even engage in conversation, their ability to apply customized persuasion increases dramatically.

Here again, LLMs can change the game. Early results suggest that LLMs can be used to argue persuasively on topics ranging from politics to public health policy.

Good for the gander

Man looking at phone in front of his computer
Image: Unsplash

AI, however, doesn’t favor one side or the other. Spam filters should also benefit from AI spam advances, allowing them to erect new barriers to unwanted emails.

Spammers often try to trick filters with special characters, misspelled words, or hidden text, relying on the human propensity to forgive small text anomalies – for example, “c1îck h.ere n0w.”

But as AI better understands spam messages, filters could better identify and block unwanted spam. And maybe even let through wanted spam, such as marketing email you’ve explicitly signed up for.

Imagine a filter that predicts whether you’d want to read an email before you even read it.

Despite growing concerns about AI – as evidenced by Tesla, SpaceX, and Twitter CEO Elon Musk, Apple founder Steve Wozniak and other tech leaders calling for a pause in AI development – a lot of good could come from advances in the technology.

AI can help us understand how weaknesses in human reasoning might be exploited by bad actors and develop ways to counter malevolent activities.

All new technologies can result in both wonder and danger. The difference lies in who creates and controls the tools and how they are used.

This article was updated to indicate that it was a teenager’s father.

Have any thoughts on this? Drop us a line below in the comments, or carry the discussion over to our Twitter or Facebook.

Editors’ Recommendations:

Editor’s Note: This article was written by John Licato, Assistant Professor of Computer Science and Director of AMHR Lab, University of South Florida and republished from The Conversation under a Creative Commons license. Read the original article.

Follow us on Flipboard, Google News, or Apple News

The Conversation is a nonprofit, independent news organization dedicated to unlocking the knowledge of experts for the public good.

Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Deals of the Day

More in AI