ImageImage

WEBINAR WITH LIVE DEMO

Combining structured and unstructured data to enhance RAG capabilities

August 22, 2024

10:00 AM PT / 1:00 PM ET

Image
Ines Chami

Co-founder & Chief Scientist
Numbers Station

Image
Shane Johnson

Senior Director of Product
Snorkel AI

The emergence of powerful text-to-SQL solutions and the increasing sophistication of RAG pipelines have created a novel opportunity to use both structured and unstructured data as context for LLM-powered AI assistants.

A modern RAG pipeline can combine information queried from databases with information extracted from unstructured documents (e.g., PDFs), enabling business users to gather meaningful insights from AI assistants and copilots.

In this webinar, presented by Snorkel and Numbers Station, we’ll explain how to elicit data-driven insights backed by domain-knowledge using a multi-agent architecture and a fine-tuned LLM and/or RAG pipeline.

Join us and learn how to:
  • Combine structured and unstructured data with RAG
  • Generate actionable insights derived from data and domain knowledge
  • Empower data-driven analysis without the need for SQL
  • Date: August 22, 2024
    Time: 10:00 AM PT / 1:00 PM ET

Register now

By submitting this form, I agree to the Terms of Use and acknowledge that my information will be used in accordance with the Privacy Policy.

Speakers

Image

Ines Chami

Co-founder & Chief Scientist
Numbers Station

Bio: Ines Chami is the Chief Scientist and Co-Founder of Numbers Station. She received her Ph.D. in the Institute for Computational and Mathematical Engineering from Stanford University where she was advised by Prof. Christopher Ré. Prior to attending Stanford, she studied Mathematics and Computer Science at Ecole Centrale Paris. Ines is particularly excited about building intelligent models to automate data-intensive work. Her work spans applications such as knowledge graph construction and data cleaning. For her work on graph representation learning, she won the 2021 Stanford Gene Golub Doctoral Dissertation Award. During her Ph.D. she interned at Microsoft AI and Research and Google Research where she co-authored the graph representation learning chapter of Kevin Murphy's "Probabilistic Machine Learning: An Introduction" book.

Image

Shane Johnson

Senior Director of Product
Snorkel AI

I started out as a developer and architect before pivoting to product/marketing. I'm still a developer at heart (and love coding for fun), but I love advocating for innovative products -- particularly to developers.

I've spent most of my time in the database space, but lately I've been going down the LLM rabbit hole.

ImageImage

Live Webinar

Speakers

Image

Ines Chami

Co-founder & Chief Scientist
Numbers Station

Ines Chami is the Chief Scientist and Co-Founder of Numbers Station. She received her Ph.D. in the Institute for Computational and Mathematical Engineering from Stanford University where she was advised by Prof. Christopher Ré. Prior to attending Stanford, she studied Mathematics and Computer Science at Ecole Centrale Paris. Ines is particularly excited about building intelligent models to automate data-intensive work. Her work spans applications such as knowledge graph construction and data cleaning. For her work on graph representation learning, she won the 2021 Stanford Gene Golub Doctoral Dissertation Award. During her Ph.D. she interned at Microsoft AI and Research and Google Research where she co-authored the graph representation learning chapter of Kevin Murphy's "Probabilistic Machine Learning: An Introduction" book.

Image

Skip McCormick

Managing Director, Governance & Risk Management | Artificial Intelligence Hub
BNY Mellon

Image

Shane Johnson

Senior Director of Product Marketing
Snorkel AI

I started out as a developer and architect before pivoting to product/marketing. I'm still a developer at heart (and love coding for fun), but I love advocating for innovative products -- particularly to developers.

I've spent most of my time in the database space, but lately I've been going down the LLM rabbit hole.

Image

Skip McCormick

Managing Director, Governance & Risk Management | Artificial Intelligence Hub
BNY Mellon

Register now

By submitting this form, I agree to the Terms of Use and acknowledge that my information will be used in accordance with the Privacy Policy.

Combining structured and unstructured data to enhance RAG capabilities

The reasoning capabilities of large, state-of-the-art open models such as Meta's Llama 3.1 405B rival those of popular closed models, but the compute resources required for low-latency inference often put them out of reach.

However, LLM distillation can be used to create small language models (SLMs) for specialized tasks that preserve the reasoning capabilities of large models while significantly reducing inference costs.

In this webinar, we'll provide an overview of LLM distillation, explain how it compares with fine-tuning, and introduce the latest techniques for training SLMs using larger models and knowledge transfer.

Watch this webinar on-demand and learn how to:

  • Train small-language models (SLMs) for specialized tasks
  • Choose between LLM fine-tuning and distillation
  • Reduce inference costs while preserving response quality

Schedule

Tuesday, March 12, 2024

7:45 PM to 8:30 PM

Arrive and mingle

8:30 PM to 10:45 PM

Dinner and conversation with data science leaders