Meta released Llama 3.1 today signaling a new era of open source AI. The Llama 3.1 models include 8b, 70B, LlamaGuard, and the new Llama 3.1 405B, the largest open source foundation model with over 400 billion parameters.  Llama 3.1 performance will rival OpenAI’s GPT-4 model, despite using less than half the parameters.

This is an important milestone for the AI development stack. For AI to flourish in the enterprise, businesses need to have choices. Enterprises want to adopt foundation models that best fit their use case and then specialize those models to achieve production quality performance against domain and business-specific criteria.

Image2

With Meta’s release of open source foundation models that match the performance of proprietary models at a lower cost, the market benefits from more options to build custom AI applications.

Day 1 availability: Llama 3.1 models available in Snorkel today

As of today, all Meta Llama 3.1 large language models are natively available as part of the Prompt Builder in Snorkel Flow:

Image3

Here’s a quick demo of how easy it is to activate Meta Llama 405B in Snorkel Flow. AI development teams can access Meta’s industry-leading Llama 3.1 models from their service of choice using Hugging Face, Together AI, Microsoft Azure ML, AWS SageMaker, and Google Vertex AI Model Garden.

Image1

The growing importance of OSS in the enterprise

With the launch of Llama 3.1 Meta has not been shy about its stance that AI will be driven by open source.  Mark Zuckerberg authored a blog today, Open Source AI Is the Path Forward, that outlines the parallels between early proprietary AI model development and early proprietary OS software development. He forecasts that closed, proprietary AI models will be replaced by open source models just as Unix was replaced by Linux as the dominant OS.

The reason for that is that, especially when it comes to infrastructure, developers and enterprises benefit from having more adaptable, affordable, and better foundations from which to build. Beyond the OS we have seen this play out in databases with the rise of MySQL and later MongoDB, as well as with key technologies that accelerated cloud adoption like Kafka and K8s.

We agree that open source models are likely to gain widespread enterprise adoption. This provides enterprises more flexibility in model size, data controls, and location, as well as affordability. The Llama 3.1 release shows that very soon open source models will also provide best-in-class performance.

Paired with Snorkel Flow’s data development capabilities, the LLama 3.1 model family will help enterprises cross the chasm from prototype to production:

  • LLM specialization via fine-tuning/alignment – With Snorkel Flow, enterprises can leverage a powerful combination of Llama 3.1 outputs with their own domain expertise to create fine-tuning datasets more quickly than ever.
  • SLM for real, practical use cases – The Wall Street Journal recently covered this trend, noting: “This category of AI software—called small or medium language models—is trained on less data and often designed for specific tasks.” Llama 3.1 can now be used to build SLMs via synthetic data & distillation approaches in Snorkel Flow.
  • LLM evals – Snorkel allows teams to combine powerful evaluators like Llama Guard 3 safety model with your own guardrails & acceptance criteria.

Meta’s continued leadership in open source model development is a critical step towards more efficient, lower cost AI systems that incorporate a variety of models and route queries based on best fit. The pattern here is not far off from the use of micro-service where orchestration enables a far more efficient use of compute resources with better outcomes. Fortunately, most of the major inference providers have also announced that they are supporting Llama 3.1 which means most enterprises are in a good position to quickly adopt the Llama models.

We believe Meta’s release today was a push towards flexibility—flexibility to choose the best foundation model for your task and specialize from there. The Llama 3.1 models provide another great option for our customers as they look for the right foundation model. We encourage enterprises to seriously consider the new Meta offerings.

Learn more about what Snorkel can do for your organization

Snorkel AI offers multiple ways for enterprises to uplevel their AI capabilities. Our Snorkel Flow data development platform empowers enterprise data scientists and subject matter experts to build and deploy high quality models end-to-end in-house. Our Snorkel Custom program puts our world-class engineers and researchers to work on your most promising challenges to deliver data sets or fully-built LLM or generative AI applications, fast.

See what Snorkel option is right for you. Book a demo today.