Google has introduced that it is increasing its Vulnerability Rewards Program (VRP) to compensate researchers for locating assault eventualities tailor-made to generative synthetic intelligence (AI) programs in an effort to bolster AI security and safety.
“Generative AI raises new and totally different considerations than conventional digital safety, such because the potential for unfair bias, mannequin manipulation or misinterpretations of information (hallucinations),” Google’s Laurie Richardson and Royal Hansen stated.
Among the classes which are in scope embrace immediate injections, leakage of delicate knowledge from coaching datasets, mannequin manipulation, adversarial perturbation assaults that set off misclassification, and mannequin theft.
It is value noting that Google earlier this July instituted an AI Crimson Group to assist handle threats to AI programs as a part of its Safe AI Framework (SAIF).
Additionally introduced as a part of its dedication to safe AI are efforts to strengthen the AI provide chain through present open-source safety initiatives comparable to Provide Chain Ranges for Software program Artifacts (SLSA) and Sigstore.
“Digital signatures, comparable to these from Sigstore, which permit customers to confirm that the software program wasn’t tampered with or changed,” Google stated.
“Metadata comparable to SLSA provenance that inform us what’s in software program and the way it was constructed, permitting customers to make sure license compatibility, determine identified vulnerabilities, and detect extra superior threats.”
The event comes as OpenAI unveiled a brand new inner Preparedness crew to “monitor, consider, forecast, and shield” in opposition to catastrophic dangers to generative AI spanning cybersecurity, chemical, organic, radiological, and nuclear (CBRN) threats.
The 2 corporations, alongside Anthropic and Microsoft, have additionally introduced the creation of a $10 million AI Security Fund, centered on selling analysis within the discipline of AI security.



