The rapid progress of artificial intelligence technology is reshaping the structure of the technology industry, and many technology giants have devoted themselves to the research and development and deployment of "AI agent" systems. These intelligent systems can independently perform complex tasks without the need for continuous human intervention, demonstrating unprecedented automation potential. Companies such as Microsoft and Anthropic have taken the lead in launching their respective AI proxy solutions. However, OpenAI, as an industry leader, has remained cautious in this field and has not released related products for a long time. There are deep considerations behind this decision. .

According to the authoritative technology media "The Information", OpenAI's cautious attitude mainly stems from concerns about "prompt injection" attacks. This new type of cyber attack can induce AI systems to execute malicious instructions, thus causing serious security risks. For example, when a user authorizes an AI agent to shop online, the system may be induced to access malicious websites and perform dangerous operations such as stealing user email information or credit card data. This security vulnerability may not only lead to user privacy leakage, but also cause irreparable damage to OpenAI's brand reputation.
What is unique about AI proxy systems is that they have the ability to operate computers independently, which makes them face much more security risks than traditional AI systems. Once hacked, these agents may cause systemic damage to users' digital assets. OpenAI internal technician admitted that although all large language models are likely to be attacked, the autonomy characteristics of AI agents significantly amplify this risk.
The threat of injecting attacks is not groundless, and this risk has been fully verified on other platforms. Last year, a cybersecurity expert successfully demonstrated how to manipulate Microsoft's Copilot AI system to leak confidential information from organizations, including email content and bank transaction records. Even more shocking is that attackers are able to send fake emails imitating the writing style of specific employees, highlighting the vulnerability of AI systems in terms of security protection.
OpenAI's flagship product, ChatGPT, has also been tested by prompt injection attacks. Researchers successfully implanted false "memory" by uploading third-party files (such as Word documents), and this experiment fully exposed potential vulnerabilities in the AI system. Faced with such a severe security challenge, the OpenAI team expressed surprise at the loose attitude of its competitor Anthropic when it released its AI agent. Anthropic only advises developers to "take measures to isolate Claude from sensitive data", a relatively simple security measure that is obviously difficult to deal with increasingly complex cyber threats.
It is reported that OpenAI may launch its AI proxy product this month. However, the question of common concern in the industry is: Is the additional time the development team has acquired enough to build a strong enough security protection system? With the rapid development of AI technology, how to find a balance between innovation and security will be an important issue facing all AI companies.
Core points:
OpenAI delayed the release of AI agents due to the security risks of injecting attacks, highlighting the potential risks of AI systems.
Although companies such as Microsoft and Anthropic have launched AI agent products, security protection measures are still insufficient.
OpenAI is working to enhance the security performance of its products to deal with increasingly complex cyber threats.