When LLMs Go Online: The Emerging Threat of Web-Enabled LLMs

Subscribers:
351,000
Published on ● Video Link: https://www.youtube.com/watch?v=UAaOtOEH2Kg



Duration: 0:00
268 views
13


Recent advancements in Large Language Models (LLMs) have established them as agentic systems capable of planning and interacting with various tools. These LLM agents are often paired with web-based tools, enabling access to diverse sources and real-time information. Although these advancements offer significant benefits across various applications, they also increase the risk of malicious use, particularly in cyberattacks involving personal information. In this work, we investigate the risks associated with misuse of LLM agents in cyberattacks involving personal data. Specifically, we aim to understand: 1) how potent LLM agents can be when directed to conduct cyberattacks, 2) how cyberattacks are enhanced by web-based tools, and 3) how affordable and easy it becomes to launch cyberattacks using LLM agents. We examine three attack scenarios: the collection of Personally Identifiable Information (PII), the generation of impersonation posts, and the creation of spear-phishing emails. Our experiments reveal the effectiveness of LLM agents in these attacks: LLM agents achieved a precision of up to 95.9% in collecting PII, generated impersonation posts where 93.9% of them were deemed authentic, and boosted click rate of phishing links in spear phishing emails by 46.67%. Additionally, our findings underscore the limitations of existing safeguards in contemporary commercial LLMs, emphasizing the urgent need for robust security measures to prevent the misuse of LLM agents.

Speaker: Hanna Kim (KAIST)




Other Videos By Microsoft Research


2025-09-24Understanding How Users Prepare for and React to Smartphone Theft
2025-09-24When LLMs Go Online: The Emerging Threat of Web-Enabled LLMs
2025-09-24A Formal Analysis of Apple’s iMessage PQ3 Protocol
2025-09-24Email Spoofing with SMTP Smuggling: How the Shared Email Infrastructures Magnify this Vulnerability
2025-09-24A Framework for Abusability Analysis: The Case of Passkeys in Interpersonal Threat Models
2025-09-24‘Hey mum, I dropped my phone down the toilet’: Investigating Hi Mum and Dad SMS Scams in the UK
2025-09-24Dehumanizing machines: Making sense of AI systems that seem human
2025-09-24Scalable emulation of protein equilibrium ensembles with BioEmu
2025-09-24Disrupting the AI infrastructure with MicroLEDs
2025-09-24Dion: The distributed orthonormal update revolution is here
2025-09-24Pushing boundaries of complex reasoning in small language models
2025-09-22zk-promises: Anonymous Moderation, Reputation, & Blocking from Anonymous Credentials with Callbacks
2025-09-22More is Less: Extra Features in Contactless Payments Break Security
2025-09-18Sub-Population Identification of Multi-morbidity in Sub-Saharan African Populations
2025-09-03Echoes in GenAI generations
2025-08-27Six Years of Rowhammer: Breakthroughs and Future Directions
2025-08-25Sub-Population Identification of Multi-morbidity in Sub-Saharan African Populations
2025-08-19MindJourney: Test-Time Scaling with World Models for Spatial Reasoning
2025-08-11Medical Bayesian Kiosk (2010)
2025-08-07Reimagining healthcare delivery and public health with AI
2025-08-05VeriTrail: Detect hallucination and trace provenance in AI workflows