Thursday, December 19, 2024

Forum Systems Releases GenAI Productivity-Risk Framework and two Fine-tuned Models for Enterprise Customers

Related stories

ePlus Launches Secure GenAI Accelerator

ePlus inc. announced its Secure GenAI Accelerator offering. Part of...

Lockheed Martin Launches Astris AI to Enable Secure AI Solution

Lockheed Martin has announced the formation of Astris AI,...

Boomi Expands Data Management with Rivery Acquisition

Boomi™, the intelligent integration and automation leader, announced a...

Tray.ai Launches Merlin Agent Builder to Break the Traps of Custom Code and SaaS Agents

Tray.ai, innovator of the AI-ready composable integration platform, announced...

Aily Labs and Mila Partner to Advance AI Agents and Decision Intelligence

Aily Labs GmbH, pioneer of an AI-powered decision intelligence...
spot_imgspot_img

Announcing a framework to balance productivity and risk for enterprise deployments of GenAI models and releasing two fine-tuned models— Forum Systems Mistral QS-Sentry and Llama 3 QS-Sentry—with optimized productivity-risk profiles

Forum Systems, a leader in LLM and API technologies, announced the public release of two language models fine-tuned to optimize their productivity-risk profile from the Gartner® Data & Analytics Summit 2024. The research is discussed at length in two recent articles, Framework for LLM Selection by Balancing Model Risk with Workforce Productivity and Improving Productivity-Risk Profile through LLM Fine-tuning. This groundbreaking work presents a framework for balancing productivity and risk in GenAI deployments, an urgent question among business leaders today.

“LLMs security has its tradeoffs. More restrictive models will be safer but may hamper productivity. If enterprises aren’t measuring the productivity-risk balance of their models, they are in the dark about whether they’ve achieved an optimal tradeoff,” remarked Mamoon Yunus, CEO of Forum Systems. He continued, “classic machine learning metrics like precision and recall can serve as proxies for productivity and risk. Fine-tuning on extensive manual multi-vote labeled data, our LLMs show superior performance compared to base models.”

The fine-tuned models—Mistral QS-Sentry and Llama 3 QS-Sentry—are based on Mistral-7B-Instruct-v0.2 and Meta-Llama-3-8B-Instruct.

Also Read: Reveald Launches Epiphany Validation Engine to Enhance AI-Driven Cyber Resilience

In the first article, Forum Systems developed a framework for balancing risk and productivity and assessed the productivity-risk profile of Mistral and Llama 3 before fine-tuning. It found that, when asked to classify prompts as either safe or unsafe, Mistral was more precise and thus aligned with higher productivity, while Llama 3 was more restrictive and thus aligned with lower risk.

The second article analyzed the models after they were fine-tuned on a hand-curated dataset of about 20,000 prompts. The study showed that the productivity-risk profile of both models can see meaningful improvements after fine-tuning. Forum Systems has released both fine-tuned models on Hugging Face to contribute to the broader community of researchers and those working in AI governance and AI alignment, believing its framework for analyzing the trade-offs between productivity will also prove valuable to business leaders deploying safe and effective GenAI offerings.

As Gartner analyst Arun Chandrasekaran recently stated, “Generative AI (GenAI) has the potential to transform businesses across industries. Most business and technology leaders believe that the benefits of GenAI far outweigh its risks.” He recommends, “Put responsible AI at the heart of your generative efforts. Promote harmonious interaction among humans and machines with design thinking and by incorporating human feedback into GenAI applications.” Forum Systems agrees with Chandrasekaran’s analysis and recommendation. Its work in optimizing productivity-risk profiles of small models demonstrates that enterprise-class responsible AI deployments are within reach through fine-tuning.

Source: PRNewsWire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img