Friday, November 22, 2024

Arkose Labs Becomes First Bot Management Company to Roll Out Protections for Enterprise GPT Applications

Related stories

Deep Instinct Expands Zero-Day Security to Amazon S3

Deep Instinct, the zero-day data security company built on...

Foxit Unveils AI Assistant in Admin Console

Foxit, a leading provider of innovative PDF and eSignature...

Instabase Names Junie Dinda CMO

Instabase, a leading applied artificial intelligence (AI) solution for...
spot_imgspot_img

Arkose Labs, the global leader in bot management and account security, announced the launch of its pioneering protection measures for GPT applications, addressing the urgent need for proactive defenses against new attack vectors, like GPT prompt compromise and LLM platform abuse.

Enterprises deploying GPT applications and providers pioneering LLM platforms are priority targets for bad actors, and the risks are substantial.

Before selecting Arkose Labs, a GPT platform was besieged by over 2 billion bot attacks. The attacks exhausted the platform’s processing capacity and cost tens of millions of dollars each month in compute resources. Genuine consumers had trouble accessing the service, as bots dominated the platform, employing proxies and doubling their efforts to scrape the platform’s insights, leveraging compromised account credentials. Within days of deploying Arkose Bot Manager, though, the GPT platform realized a 99.22% reduction in LLM platform abuse.

Also Read: Zscaler Unveils AI Innovations to Power the Industry’s Most Comprehensive Data Protection Platform

Arkose Labs’ new capabilities thwart emerging threat vectors, including:

  1. GPT prompt compromise: an attack type where bots are able to programmatically submit prompts and scrape the response with an intention to either train their own models, resell similar services or gain access to proprietary, confidential and personal information.
  2. LLM platform abuse: an attack type that creates unauthorized platform replicas and uses illegal reverse proxying that copies the platform’s insights. Those insights are used to create knock-off services that are increasingly used to generate phishing emails, create deepfake videos, and conduct other illicit acts. The insights are also used by bad actors to circumvent geographical restraints designated by China and other countries.

“Generative AI intensifies cybercrime not only by enhancing traditional attacks, like scraping, but also by introducing new threats like GPT prompt compromise and LLM platform abuse,” said Arkose Labs Chief Product Officer Ashish Jain. “The new protective measures we’re releasing today are battle tested and use AI to protect the AI that companies are deploying.”

“Our commitment is to stay ahead of cybercriminals, ensuring that our customers’ use of transformative AI technologies remains secure and productive,” added Vikas Shetty, vice president, product management, Arkose Labs. “Our proactive measures have proven effective, significantly reducing attack volumes and internal fraud costs while optimizing legitimate users’ experiences.”

Source: BusinessWire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img