Manila Standard

ChatGPT phishing fantasies: Will AI chatbots help fight cyberscam?

-

KASPERSKY experts have conducted research studying ChatGPT phishing links detection capability.

While ChatGPT had previously demonstrat­ed the ability to create phishing emails and write malware, its effectiven­ess in detecting malicious links was limited.

The study revealed that although ChatGPT knows a great deal about phishing and can guess the target of a phishing attack, it had high false positive rates of up to 64 percent. Often, it produced imaginary explanatio­ns and false evidence to justify its verdicts.

ChatGPT, an AI-powered language model, has been a topic of discussion in the cybersecur­ity world due to its potential to create phishing emails and the concerns about its impact on cybersecur­ity experts’ job security even despite its creators’ warnings that it is too early to apply the novel technology to such high-risk domains.

Kaspersky experts decided to conduct an experiment to reveal ChatGPT’s ability to detect phishing links, as well as the cybersecur­ity knowledge it learned during training. Company’s experts tested gpt-3.5- turbo, the model that powers ChatGPT, on more than 2,000 links that Kaspersky anti- phishing technologi­es deemed phishing, and mixed it with thousands of safe URLs.

In the experiment, detection rates vary depending on the prompt used.

The experiment was based on asking ChatGPT two questions: “Does this link lead to a phishing website?” and “Is this link safe to visit?”.

The results showed that ChatGPT had a detection rate of 87.2 percent and a false positive rate of 23.2 percent for the first question.

The second question, “Is this link safe to visit?” had a higher detection rate of 93.8 percent, but a higher false positive rate of 64.3 percent. While the detection rate is very high, the false positive rate is too high for any kind of production applicatio­n.

The unsatisfac­tory results at the detection task were expected, but could ChatGPT help with classifyin­g and investigat­ing attacks?

Since attackers typically mention popular brands in their links to deceive users into believing that the URL is legitimate and belongs to a reputable company, the AI language model shows impressive results in the identifica­tion of potential phishing targets.

For instance, ChatGPT has successful­ly extracted a target from more than half of the URLs, including major tech portals like Facebook, TikTok, and Google, marketplac­es such as Amazon and Steam, and numerous banks from around the globe, among others—without any additional training.

The experiment also showed ChatGPT might have serious problems when it comes to proving its point on the decision whether the link is malicious.

Some explanatio­ns were correct and based on facts, others revealed known limitation­s of language models, including hallucinat­ions and misstateme­nts: many explanatio­ns were misleading, despite the confident tone.

“ChatGPT certainly shows promise in assisting human analysts in detecting phishing attacks but let’s not get ahead of us - language models still have their limitation­s. While they might be on par with an intern-level phishing analyst when it comes to reasoning about phishing attacks and extracting potential targets, they tend to hallucinat­e and produce random output. So, while they might not revolution­ize the cybersecur­ity landscape just yet, they could still be helpful tools for the community,” comments Vladislav Tushkanov, Lead Data Scientist at Kaspersky.

 ?? ?? The study revealed that although ChatGPT knows a great deal about phishing and can guess the target of a phishing attack, it had high false positive rates of up to 64 percent.
The study revealed that although ChatGPT knows a great deal about phishing and can guess the target of a phishing attack, it had high false positive rates of up to 64 percent.

Newspapers in English

Newspapers from Philippines