Wednesday, April 2, 2025

Arkose Labs Becomes First Bot Management Company to Roll Out Protections for Enterprise GPT Applications

Related stories

Hacking the Hackers: How GenAI is Predicting and Preventing Cyber Attacks

In the high-stakes arena of cybersecurity, the rules of...

Veltris Acquires BPK to Boost AI & Digital in Healthcare

Veltris, a digital product engineering services provider backed by...

Accenture & Schaeffler Advance Industrial Humanoid Robots

Accenture has joined forces with Schaeffler AG to reshape...

EDGNEX Data Centers partners with Hyperco

The acquisition is expected to further strengthen the expansion...

Chef Robotics Raises $43M Series A to Scale AI Robotics

Chef Robotics, a leader in AI-powered robotic systems for...
spot_imgspot_img

Arkose Labs, the global leader in bot management and account security, announced the launch of its pioneering protection measures for GPT applications, addressing the urgent need for proactive defenses against new attack vectors, like GPT prompt compromise and LLM platform abuse.

Enterprises deploying GPT applications and providers pioneering LLM platforms are priority targets for bad actors, and the risks are substantial.

Before selecting Arkose Labs, a GPT platform was besieged by over 2 billion bot attacks. The attacks exhausted the platform’s processing capacity and cost tens of millions of dollars each month in compute resources. Genuine consumers had trouble accessing the service, as bots dominated the platform, employing proxies and doubling their efforts to scrape the platform’s insights, leveraging compromised account credentials. Within days of deploying Arkose Bot Manager, though, the GPT platform realized a 99.22% reduction in LLM platform abuse.

Also Read: Zscaler Unveils AI Innovations to Power the Industry’s Most Comprehensive Data Protection Platform

Arkose Labs’ new capabilities thwart emerging threat vectors, including:

  1. GPT prompt compromise: an attack type where bots are able to programmatically submit prompts and scrape the response with an intention to either train their own models, resell similar services or gain access to proprietary, confidential and personal information.
  2. LLM platform abuse: an attack type that creates unauthorized platform replicas and uses illegal reverse proxying that copies the platform’s insights. Those insights are used to create knock-off services that are increasingly used to generate phishing emails, create deepfake videos, and conduct other illicit acts. The insights are also used by bad actors to circumvent geographical restraints designated by China and other countries.

“Generative AI intensifies cybercrime not only by enhancing traditional attacks, like scraping, but also by introducing new threats like GPT prompt compromise and LLM platform abuse,” said Arkose Labs Chief Product Officer Ashish Jain. “The new protective measures we’re releasing today are battle tested and use AI to protect the AI that companies are deploying.”

“Our commitment is to stay ahead of cybercriminals, ensuring that our customers’ use of transformative AI technologies remains secure and productive,” added Vikas Shetty, vice president, product management, Arkose Labs. “Our proactive measures have proven effective, significantly reducing attack volumes and internal fraud costs while optimizing legitimate users’ experiences.”

Source: BusinessWire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img