Saturday, November 16, 2024

AI2, The Allen Institute for AI, Today Releases the OLMo: A State-of-the-Art, Truly Open-Source LLM and Framework

Related stories

Frost & Sullivan Launches FrostAI to Drive Growth Opportunities

Accelerating your Transformational Growth Journey: FrostAI’s elevated user experience...

Kyndryl & Microsoft Launch Services to Boost Cyber Resilience

New Kyndryl services, co-developed with Microsoft, provide augmented security...

Cognigy Launches Agentic AI for Enterprise Contact Centers

Cognigy, a global leader in AI-powered customer service solutions,...

Why Sustainable Data Centers Are Vital for the Future of Business

In today’s digital landscape, data centers are the critical...
spot_imgspot_img

The OLMo framework will drive a critical shift in AI development by providing the industry with a unique large, accurate, and open language model framework – creating an alternative to current models that are restrictive and closed.

As the world races to deploy AI models that are effective and safe, the demand for Open Large Language Models (LLMs) has exploded. The massive adoption of both open and closed AI models means that AI capabilities have leapfrogged our ability to understand how they are created. Releasing the OLMo framework will provide the industry with an opportunity to understand what is going on inside AI models.

The Allen Institute for AI (AI2) has released OLMo 7B, a truly open-source, state-of-the-art large language model released alongside the pre-training data and training code, something no open models of this scale offer today. This empowers researchers and developers to use the best and open models to advance the science of language models collectively.

“Open foundation models have been critical in driving a burst of innovation and development around generative AI,” said Yann LeCun, Chief AI Scientist at Meta. “The vibrant community that comes from open source is the fastest and most effective way to build the future of AI.”

OLMo and the framework is designed to aid researchers in training and experimenting with large language models. They are available for direct download on Hugging Face and in GitHub. This work was made possible, in part, via a collaboration with the Kempner Institute for the Study of Natural and Artificial Intelligence at Harvard University and partners including AMD, CSC – IT Center for Science (Finland), the Paul G. Allen School of Computer Science & Engineering at the University of Washington and Databricks.

The framework features a suite of completely open AI development tools, including:

  • Full pretraining data: The model is built on AI2’s Dolma set which features three trillion token open corpus for language model pretraining, including code that produces the training data.
  • Training code and model weights: The OLMo framework includes full model weights for four model variants at the 7B scale, each trained to at least 2T tokens. Inference code, training metrics and training logs are all provided.
  • Evaluation: We’ve released the evaluation suite used in development, complete with 500+ checkpoints per model, from every 1000 steps during the training process and evaluation code under the umbrella of the Catwalk project.

“I’m enthusiastic about getting OLMo into the hands of AI researchers,” said Eric Horvitz, Microsoft’s Chief Scientific Officer and a founding member of the AI2 Scientific Advisory Board. “The new offering continues Allen AI’s tradition of providing valuable open models, tools, and data, which have spurred numerous advancements in AI across the global community.”

Also Read: Copan Colibrí™ Receives Third FDA Clearance, Further Expanding Automated ID/AST Workup Capabilities

A truly open model

By making OLMo and its training data fully available to the public, AI2 has taken a big step towards collaboratively building the best open language model in the world. In the coming months, AI2 will continue to iterate on OLMo and will bring different model sizes, modalities, datasets, and capabilities into the OLMo family.

“Many language models today are published with limited transparency. Without having access to training data, researchers cannot scientifically understand how a model is working. It’s the equivalent of drug discovery without clinical trials or studying the solar system without a telescope,” said Hanna Hajishirzi, OLMo project lead, a senior director of NLP Research at AI2, and a professor in the UW’s Allen School. “With our new framework, researchers will finally be able to study the science of LLMs, which is critical to building the next generation of safe and trustworthy AI.”

With OLMo, AI researchers and developers will experience:

  • More Precision: With full insight into the training data behind the model, researchers will be able to work faster and no longer need to depend on qualitative assumptions of how we feel the model is performing but can test it scientifically.
  • Less Carbon: Currently one training run is equivalent to the emissions of nine US homes for one year. By opening the full training and evaluation ecosystem, it radically reduces developmental redundancies, which is critical in the decarbonization of AI
  • Lasting results: Keeping models and their datasets in the open and not behind APIs enables researchers to learn and build from previous models and work.

SOURCE: BusinessWire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img