in

DeepKeep Launches Generative AI Risk Assessment Module


DeepKeep, the leading provider of AI-native trust, risk, and security management, announces the product launch of its Generative AI Risk Assessment module, designed to secure Large Language Models (LLMs) and computer vision models, specifically focusing on penetration testing, identifying potential vulnerabilities and threats to model security, trustworthiness and privacy.

Assessing and mitigating AI model and application vulnerabilities ensures implementations are compliant, fair and ethical. DeepKeep’s Risk Assessment module offers a comprehensive ecosystem approach by considering risks associated with model deployment, and identifying application weak spots.

DeepKeep’s assessment provides a thorough examination of AI models, ensuring high standards of accuracy, integrity, fairness, and efficiency. The module is helping security teams streamline Generative AI deployment processes, granting a range of scoring metrics for evaluation.

Core features include:

  • Penetration Testing
  • Identifying the model’s tendency to hallucinate
  • Identifying the model’s propensity to leak private data
  • Assessing toxic, offensive, harmful, unfair, unethical, or discriminatory language
  • Assessing biases and fairness
  • Weak spot analysis
LLM risk assessment measures. For example, when applying DeepKeep’s Risk Assessment module to Meta’s LLM LlamaV2 7B to examine prompt manipulation sensitivity, findings pointed to a weakness in English-to-French translation. Credit: DeepKeep.

“The market must be able to trust its GenAI models, as more and more enterprises incorporate GenAI into daily business processes,” says Rony Ohayon, DeepKeep’s CEO and Founder. “Evaluating model resilience is paramount, particularly during its inference phase in order to provide insights into the model’s ability to handle various scenarios effectively. DeepKeep’s goal is to empower businesses with the confidence to leverage GenAI technologies while maintaining high standards of transparency and integrity.”

DeepKeep’s Generative AI Risk Assessment module secures AI alongside its AI Firewall, enabling live protection against attacks on AI applications. Detection capabilities cover a wide range of security and safety categories, leveraging DeepKeep’s proprietary technology and cutting-edge research.

*ROUGE and METEOR are natural language processing (NLP) techniques for evaluating machine learning outputs. Scores range between 0-1, with 1 indicating perfection.


Bind AI Copilot (www.getbind.co)

Rajeev Sharma, CTO at Grid Dynamics — GenAI, Continuous Innovation, Rocket Propulsion Challenges, Leadership, Skills for Engineering Leaders, Ethical AI, Business Reimagination - AI Time Journal

Rajeev Sharma, CTO at Grid Dynamics — GenAI, Continuous Innovation, Rocket Propulsion Challenges, Leadership, Skills for Engineering Leaders, Ethical AI, Business Reimagination – AI Time Journal