DeepKeep, the main supplier of AI-Native Belief, Threat, and Safety Administration, declares the product launch of its GenAI Threat Evaluation module, designed to safe GenAI’s LLM and pc imaginative and prescient fashions, particularly specializing in penetration testing, figuring out potential vulnerabilities and threats to mannequin safety, trustworthiness and privateness.
Assessing and mitigating AI mannequin and software vulnerabilities ensures implementations are compliant, honest and moral. DeepKeep‘s Threat Evaluation module presents a complete ecosystem strategy by contemplating dangers related to mannequin deployment, and figuring out software weak spots.
DeepKeep’s evaluation gives an intensive examination of AI fashions, guaranteeing excessive requirements of accuracy, integrity, equity, and effectivity. The module helps safety groups streamline GenAI deployment processes, granting a spread of scoring metrics for analysis.
Core options embody:
- Penetration Testing
- Figuring out the mannequin’s tendency to hallucinate
- Figuring out the mannequin’s propensity to leak non-public information
- Assessing poisonous, offensive, dangerous, unfair, unethical, or discriminatory language
- Assessing biases and equity
- Weak spot evaluation
For instance, when making use of DeepKeep’s Threat Evaluation module to Meta’s LLM LlamaV2 7B to look at immediate manipulation sensitivity, findings pointed to a weak point in English-to-French translation as depicted within the chart under*:
“The market should be capable to belief its GenAI fashions, as increasingly more enterprises incorporate GenAI into day by day enterprise processes,” says Rony Ohayon, DeepKeep’s CEO and Founder. “Evaluating mannequin resilience is paramount, notably throughout its inference section with a view to present insights into the mannequin’s capacity to deal with varied eventualities successfully. DeepKeep’s aim is to empower companies with the arrogance to leverage GenAI applied sciences whereas sustaining excessive requirements of transparency and integrity.”
DeepKeep’s GenAI Threat Evaluation module secures AI alongside its AI Firewall, enabling reside safety in opposition to assaults on AI purposes. Detection capabilities cowl a variety of safety and security classes, leveraging DeepKeep’s proprietary expertise and cutting-edge analysis.
*ROUGE and METEOR are pure language processing (NLP) methods for evaluating machine studying outputs. Scores vary between 0-1, with 1 indicating perfection.