John Snow Labs, the AI for healthcare company, announced the release of Automated Responsible AI Testing Capabilities in the Generative AI Lab. This is a first-of-its-kind no-code tool to test and evaluate the safety and efficacy of custom language models. It enables non-technical domain experts to define, run, and share test suites for AI model bias, fairness, robustness, and accuracy.
This capability is based on John Snow Labs’ open-source LangTest library, which includes more than 100 test types for different aspects of Responsible AI, from bias and security, to toxicity and political leaning. LangTest uses Generative AI to automatically generate test cases, making it practical to produce a comprehensive set of tests in minutes instead of weeks. Created specifically for testing custom AI models, LangTest accounts for those not covered by general purpose benchmarks and leaderboards.
Also Read: eCential Robotics Appoints Clément Vidal as New Chief Executive Officer
Recent legislation in the US has made this kind of testing essential for companies looking to release new AI-based products and services, including:
- The ACA Section 1557 Final Rule, which went into effect in June 2024, prohibiting discrimination in medical AI algorithms based on race, color, national origin, gender, age, or disability.
- The HTI-1 Final Rule on transparency in medical decision support systems, which requires companies to show how they’ve trained and tested their models.
- The American Bar Association Guidelines, requiring comprehensive internal and third-party audits prior to AI deployments in response to lawsuits against companies that provide models for automatically matching job descriptions with candidates’ resumes.
The need for a comprehensive testing solution for Large Language Models (LLMs) is urgent. Yet, many domain experts lack the technical expertise to do this. Similarly, many data scientists lack the domain expertise to build comprehensive, industry- and task-specific models. The Generative AI Lab enables domain experts to create, edit, and understand how a model is being tested without the need for a data scientist. The tool also embodies best practices such as versioning, sharing, and automated execution of tests for every new model.
“There has long been a gap between how AI models should be tested and how they often are. The new Generative AI Lab helps by making it far easier for teams to deliver AI models that are safe, effective, fair, and transparent,” said David Talby, CTO, John Snow Labs.
Source: GlobeNewsWire