安全
使用集合让一切井井有条
根据您的偏好保存内容并对其进行分类。
AI 安全包括一组设计和操作技术,旨在避免和控制可能有意或无意造成伤害的行为。例如,AI 系统是否会按预期运行,即使在发生安全漏洞或遭到有针对性攻击时也是如此?AI 系统是否足够稳健,即使在受到干扰时也能安全运行?您如何提前规划以防范或避免风险?AI 系统在压力下是否可靠且稳定?
其中一种安全技术就是对抗性测试,即尝试“破坏”您自己的应用,以了解在用户恶意或无意输入有害内容时,应用将如何应对。Responsible Generative AI 工具包详细介绍了安全评估,包括对抗性测试。如需详细了解 Google 在此领域的工作和所学教训,请参阅关键字博文 Google 的 AI 红队:让 AI 更安全的道德黑客,或参阅 SAIF:Google 的安全 AI 指南。
如未另行说明,那么本页面中的内容已根据知识共享署名 4.0 许可获得了许可,并且代码示例已根据 Apache 2.0 许可获得了许可。有关详情,请参阅 Google 开发者网站政策。Java 是 Oracle 和/或其关联公司的注册商标。
最后更新时间 (UTC):2025-07-27。
[null,null,["最后更新时间 (UTC):2025-07-27。"],[[["\u003cp\u003eAI safety encompasses design and operational techniques to prevent harm, ensuring AI systems behave as intended, even under pressure or attack.\u003c/p\u003e\n"],["\u003cp\u003eAdversarial testing is a key safety technique where AI systems are intentionally challenged with malicious or harmful input to assess their robustness.\u003c/p\u003e\n"],["\u003cp\u003eGoogle's Responsible AI Practices provide recommendations for protecting AI systems, including methods for adversarial testing and safeguarding against attacks.\u003c/p\u003e\n"]]],[],null,["# Safety\n\n\u003cbr /\u003e\n\nAI **safety** includes a set of design and operational techniques to follow to\navoid and contain actions that can cause harm, intentionally or unintentionally.\nFor example, do AI systems behave as intended, even in the face of a security\nbreach or targeted attack? Is the AI system robust enough to operate safely\neven when perturbed? How do you plan ahead to prevent or avoid risks? Is the AI\nsystem reliable and stable under pressure?\n\nOne such safety technique is [adversarial testing](/machine-learning/guides/adv-testing),\nor the practice of trying to \"break\" your own application to learn how it\nbehaves when provided with malicious or inadvertently harmful input. The\n[Responsible Generative AI Toolkit](https://ai.google.dev/responsible/docs/evaluation)\nexplains more about safety evaluations, including adversarial testing. Learn\nmore about Google's work in this area and lessons\nlearned in the Keyword blog post, [Google's AI Red Team: the ethical hackers\nmaking AI\nsafer](https://blog.google/technology/safety-security/googles-ai-red-team-the-ethical-hackers-making-ai-safer/)\nor at [SAIF: Google's Guide to Secure AI](https://saif.google/)."]]