Cue word injection

  • 大模型安全:开源框架Guardrails安全护栏介绍与解析

    OpenGuardrails是首个完整开源的企业级大模型安全护栏平台,支持119种语言、统一LLM架构、可配置敏感度策略、多云部署。本报告深度解析其核心技术创新、应用场景、部署模式、性能对标与未来发展,为金融、医疗、法律等受管制行业的AI应用提供安全合规指引。通过分析OpenGuardrails的可配置策略、高效模型设计与生产级基础设施,揭示下一代AI安全护栏的发展方向。

    2026年1月6日
    06600
  • The MCP Governance Framework: How to build a next-generation security model that resists AI superpowers

    Focus on how MCP directly impacts the existing security system while empowering AI to actually "execute". On the one hand, MCP allows LLMs to access tools, databases, and business systems through a unified protocol, truly turning them into multi-agents that can cross systems rather than passive question-and-answer bots. On the other hand, this ability relies on "hybrid identity" and long-link authorization and authentication, so that the clear identity, minimal privileges and continuous verification required by zero trust are systematically weakened, and the context of poisoning, tool poisoning, supply chain attacks and other invisible threats are dramatically enlarged.
    Right now, governance must be rebuilt around MCP - with the gateway as the hub, unified identity, fine-grained authorization, and full-link auditing - in order to unlock the true value of agentic AI without sacrificing security.

    December 30, 2025
    08770
  • Artificial Intelligence Security Defense in Depth: Explanation of Google SAIF AI Security Framework

    With the widespread penetration of Large Language Models (LLM) and Generative Artificial Intelligence (GenAI) in enterprise applications, the traditional software security paradigm based on deterministic logic is struggling to cope with new stochastic threats such as model inversion, data poisoning, and cue word injection.Google's Secure AI Framework (SAIF), to be launched in 2023, proposes a systematic defense architecture that aims to combine traditional Cybersecurity best practices with the specificities of Artificial Intelligence (AI) systems. The Secure AI Framework (SAIF), launched by Google in 2023, proposes a systematic defense architecture that aims to combine the best practices of traditional cybersecurity with the specificities of AI systems. In this paper, we will analyze the six core pillars, ecological synergy mechanism and evolution path of SAIF from the perspective of architectural design, providing theoretical and practical references for the construction of enterprise-level AI security system.

    December 20, 2025
    01.8K0
  • Artificial Intelligence (AI) Big Model Security Risks and Defense In-Depth Report

    This report is based on the five core attack surfaces consisting of AI AI critical links from AI Assistants, Agents, Tools, Models, and Storage, with targeted security risks, defense architectures, and solutions.

    November 29, 2025
    09.6K0