Wednesday, May 14, 2025

John Snow Labs Launches Automated Testing for Responsible AI—the First No-Code Tool to Test and Evaluate Custom Language Models

Related stories

Anaconda Launches Unified AI Platform for Open Source

A Total Economic Impact Study Reveals Organizations Building AI...

Contentsquare Unveils Sense: AI Agent Acting as Analyst

Contentsquare brings together the power of AI, autonomous agents,...

Exabeam and AVANT Partner to Expand AI Security Access

New partnership empowers over 3,000 AVANT Trusted Advisors to...

HUMAIN & NVIDIA to Build AI Factories for Next-Gen Intelligence

NVIDIA and the Kingdom of Saudi Arabia (KSA) have...

Siemens Uses AI to Bridge IC Verification Gap with Questa One

Questa One redefines integrated circuit (IC) verification from a...
spot_imgspot_img

John Snow Labs, the AI for healthcare company, announced the release of Automated Responsible AI Testing Capabilities in the Generative AI Lab. This is a first-of-its-kind no-code tool to test and evaluate the safety and efficacy of custom language models. It enables non-technical domain experts to define, run, and share test suites for AI model bias, fairness, robustness, and accuracy.

This capability is based on John Snow Labs’ open-source LangTest library, which includes more than 100 test types for different aspects of Responsible AI, from bias and security, to toxicity and political leaning. LangTest uses Generative AI to automatically generate test cases, making it practical to produce a comprehensive set of tests in minutes instead of weeks. Created specifically for testing custom AI models, LangTest accounts for those not covered by general purpose benchmarks and leaderboards.

Also Read: eCential Robotics Appoints Clément Vidal as New Chief Executive Officer

Recent legislation in the US has made this kind of testing essential for companies looking to release new AI-based products and services, including:

  • The ACA Section 1557 Final Rule, which went into effect in June 2024, prohibiting discrimination in medical AI algorithms based on race, color, national origin, gender, age, or disability.
  • The HTI-1 Final Rule on transparency in medical decision support systems, which requires companies to show how they’ve trained and tested their models.
  • The American Bar Association Guidelines, requiring comprehensive internal and third-party audits prior to AI deployments in response to lawsuits against companies that provide models for automatically matching job descriptions with candidates’ resumes.

The need for a comprehensive testing solution for Large Language Models (LLMs) is urgent. Yet, many domain experts lack the technical expertise to do this. Similarly, many data scientists lack the domain expertise to build comprehensive, industry- and task-specific models. The Generative AI Lab enables domain experts to create, edit, and understand how a model is being tested without the need for a data scientist. The tool also embodies best practices such as versioning, sharing, and automated execution of tests for every new model.

“There has long been a gap between how AI models should be tested and how they often are. The new Generative AI Lab helps by making it far easier for teams to deliver AI models that are safe, effective, fair, and transparent,” said David Talby, CTO, John Snow Labs.

Source: GlobeNewsWire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img