A huge promise for LLMs is being able to answer questions and solve tasks of arbitrary complexity over an arbitrary number of data sources. The world has started to shift from simple RAG stacks, which are mostly good for answering pointed questions, to agents that can more autonomously reason over a diverse set of inputs, and interleave retrieval and tool use to produce sophisticated outputs.
Building a reliable multi-agent system is challenging. There’s a core question of developer ergonomics and production deployment – what makes sense outside a notebook setting. In this talk we outline some core building blocks for building advanced research assistants, including advanced RAG modules, event-driven workflow orchestration, and more.
Large Language Models (LLM’s) are starting to revolutionize how users can search for, interact with, and generate new content. Some recent stacks and toolkits around Retrieval Augmented Generation (RAG) have emerged where users are building applications such as chatbots using LLMs on their own private data. This opens the door to a vast array of applications. However while setting up a naive RAG stack is easy, there is a long-tail of data challenges that the user must tackle in order to make their application production-ready.
In this talk, we give practical tips on how to manage data for building a robust/reliable LLM software system, and how LlamaIndex provides the tools to do so.
Jerry is the co-founder and CEO of LlamaIndex, an open-source tool that provides a central data management/query interface for your LLM application.
Before this, he has spent his career at the intersection of ML, research, and startups. He led the ML monitoring team at Robust Intelligence, conducted self-driving AI research at Uber ATG, and worked on recommendation systems at Quora.
Jerry graduated from Princeton in 2017 with a degree in Computer Science.