Tuesday, November 5, 2024

Privacera Announces the General Availability of Its Generative AI Governance Solution Providing a Unified Platform for Data and AI Security

Related stories

Absci and Twist Bioscience Collaborate to Design Novel Antibody using Generative AI

Absci Corporation a data-first generative AI drug creation company, and...

GreyNoise Intelligence Discovers Zero-Day Vulnerabilities in Live Streaming Cameras with the Help of AI

GreyNoise Intelligence, the cybersecurity company providing real-time, verifiable threat...

Medidata Launches Bundled Solutions to Support Oncology and Vaccine Trials

Medidata, a Dassault Systèmes brand and leading provider of...

Blend Appoints Mike Mischel as SVP of AI Consulting

Blend, a leader in data science and AI-powered solutions,...

Patronus AI Launches Industry-First Self-Serve API for AI Evaluation and Guardrails

Patronus AI announced the launch of the Patronus API, the first...
spot_imgspot_img

Privacera, the AI and data security governance company founded by the creators of Apache Ranger™ and the industry’s first comprehensive generative AI governance solution, announced the General Availability (GA) of Privacera AI Governance (PAIG). PAIG allows organizations to securely innovate with generative AI (GenAI) technologies by securing the entire AI application lifecycle, from discovery and securing of sensitive fine-tuning data, Retrieval Augmented Generation (RAG) and user interactions feeding into AI-powered models, to model outputs as well as continuous monitoring of AI governance through comprehensive audit trails. Securing sensitive data and managing other risks with AI applications is crucial to enable organizations to accelerate their GenAI product strategies.

The emergence of Large Language Models (LLMs) is providing a vast range of opportunities to innovate and refine new experiences and products. Whether it’s content creation, developing new experiences around virtual assistance or improved productivity around code development, smaller and larger data-driven organizations are going to invest in diverse LLM-powered applications. With these opportunities, there is an increased need to secure and govern the use of LLMs within and outside of any enterprise, small or large. Such risks include sensitive and unauthorized data exposure, IP leakage, abuse of models, and regulatory compliance failures.

“With PAIG, Privacera is becoming the unified AI and data security platform for today’s modern data applications and products,” said Balaji Ganesan, co-founder and CEO of Privacera.  “Data-driven organizations need to think about how GenAI fits in their overall security and governance strategy. This will enable them to achieve enterprise-grade security in order to fully leverage GenAI to transform their businesses without exposing the business to unacceptable risk. Our new product capabilities allow enterprises to secure the end-to-end lifecycle for data and AI applications – from fine-tuning the LLMs, protecting VectorDB to validating and monitoring user prompts and replies to AI at scale.”

Also Read: Strobes Security Unveils Groundbreaking CTEM Platform: Elevating Offensive Cybersecurity to New Heights

PAIG enables organizations to responsibly leverage the power of GenAI by providing deep visibility into risks across use of any model and help enterprise teams to apply consistent controls to both AI applications and the underlying data required to train and fine-tune the AI applications. PAIG has been designed to be open and flexible to protect a range of GenAI applications, models and data – whether it’s structured, semi-structured or fully unstructured data sets. This design principle is particularly relevant as organizations are increasingly looking to apply GenAI techniques to a broad range of use cases to extract, organize and derive critical insights.

PAIG offers the following key capabilities:

  • Discover and classify sensitive data used to train, or fine-tune custom or generally available GenAI models and VectorDB
  • Protect models and VectorDB from being exposed to sensitive training or tuning data
  • Secure and continuously protect models from sensitive data prompt inputs and outputs with allow/deny, masking, or redaction of sensitive data in real-time
  • Comprehensive observability alongside built-in dashboards and user query analytics which provide enhanced transparency on who accessed what AI applications, what sensitive data was accessed or denied, what sensitive data assets are leveraged for each AI application, and what data protection policies are in place for each AI application
  • Ability to easily integrate with existing security monitoring and management tools
  • Open and extensible SDK to integrate seamlessly into your GenAI applications and LLM libraries

SOURCE: PRNewswire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img