Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Launchpad
Be early to the next big token project
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
DeepMind Study Reveals Six Ways Hackers Can Manipulate AI Agents
TLDR
Researchers at Google DeepMind have identified six attack methods that can manipulate AI agents online. The study shows how AI agents can be influenced through web content, hidden instructions, and poisoned data sources. Consequently, the findings highlight growing risks as companies deploy AI agents for real-world tasks across digital environments.
Content and Semantic Manipulation Expose Core Weaknesses
Researchers identified content injection traps as a direct threat to AI agents during web interactions. Hidden instructions placed in HTML or metadata can control actions without human detection. As a result, AI agents may execute commands embedded in invisible page elements.
Semantic manipulation relies on persuasive language rather than hidden code to influence AI agents. Attackers design pages with authoritative tone and structured narratives to bypass safeguards. AI agents may misinterpret harmful instructions as valid tasks.
These methods exploit how AI agents process and prioritize online information during decision-making. The study shows that structured prompts can reshape reasoning paths in subtle ways. Attackers can guide AI agents toward unintended actions without triggering system defenses.
Memory and Behavioral Attacks Expand Risk Surface
Researchers also found that attackers can manipulate memory systems used by AI agents for information retrieval. By injecting false data into trusted sources, attackers influence long-term outputs and responses. As a result, AI agents may treat fabricated information as verified knowledge over time.
Behavioral control attacks directly target the actions performed by AI agents during routine browsing. Embedded jailbreak instructions can override restrictions and trigger unintended operations. AI agents with broad permissions may access and transmit sensitive data externally.
The study highlights that these risks increase as AI agents gain autonomy and system access. Attackers can exploit routine workflows to insert malicious commands into normal tasks. AI agents face higher exposure when integrated with external tools and APIs.
Systemic and Human Factors Amplify Threat Impact
Researchers warn that systemic traps can affect multiple AI agents simultaneously across connected systems. Coordinated manipulation may trigger cascading failures similar to algorithm-driven market disruptions. As a result, AI agents operating in shared environments can amplify risks at scale.
Human reviewers remain vulnerable within the AI agents workflow and approval processes. Attackers can craft outputs that appear credible and bypass oversight checks. AI agents may execute harmful actions after receiving human approval.
The study places these findings within a broader context of increasing AI deployment across industries. AI agents now handle tasks such as communication, purchasing, and coordination through automated systems. Securing the operating environment becomes as critical as improving model design.
Researchers recommend adversarial training, input filtering, and monitoring systems to reduce exposure. The study notes that defenses remain fragmented and lack industry-wide standards. As AI agents continue expanding their role, the need for coordinated safeguards becomes more urgent.