We want to hear from you! Take our quick AI survey and share your insights on the current state of AI, how youโre implementing it, and what you expect to see in the future. Learn More
Retrieval augmented generation (RAG) is an important technique that pulls from external knowledge bases to help improve the quality of large language model (LLM) outputs. It also provides transparency into model sources that humans can cross-check.
However, according to Jerry Liu, co-founder and CEO of LlamaIndex, basic RAG systems can have primitive interfaces and poor quality understanding and planning, lack function calling or tool use and are stateless (with no memory). Data silos only exacerbate this problem. Liu spoke during VB Transform in San Francisco yesterday.
This can make it difficult to productionize LLM apps at scale, due to accuracy issues, difficulties with scaling and too many required parameters (requiring deep-tech expertise).
This means that there are many questions RAG simply canโt answer.
Register to access VB Transform On-Demand
In-person passes for VB Transform 2024 are now sold out! Don’t miss outรขยยregister now for exclusive on-demand access available after the conference. Learn More
โRAG was really just the beginning,โ Liu said onstage this week at VB Transform. Many core concepts of naive RAG are โkind of dumbโ and make โvery suboptimal decisions.โ
LlamaIndex aims to transcend these challenges by offering a platform that helps developers quickly and simply build next-generation LLM-powered apps. The framework offers data extraction that turns unstructured and semi-structured data into uniform, programmatically accessible formats; RAG that answers queries across internal data through question-answer systems and chatbots; and autonomous agents, Liu explained.
Synchronizing data so itโs always fresh
It is critical to tie together all the different types of data within an enterprise, whether unstructured or structured, Liu noted. Multi-agent systems can then โtap into the wealth of heterogeneous dataโ that companies contain.ย
โAny LLM application is only as good as your data,โ said Liu. โIf you donโt have good data quality, youโre not going to have good results.โ
LlamaCloud โ now available by waitlist โ features advanced extract, transform load (ETL) capabilities. This allows developers to โsynchronize data over time so itโs always fresh,โ Liu explained. โWhen you ask a question, youโre guaranteed to have the relevant context, no matter how complex or high level that question is.โ
LlamaIndexโs interface can handle questions both simple and complex, as well as high-level research tasks, and outputs could include short answers, structured outputs or even research reports, he said.ย
The companyโs LllamaParse is an advanced document parser specifically aimed at reducing LLM hallucinations. Liu said it has 500,000 monthly downloads and 14,000 unique users, and has processed more than 13 million pages.ย
โLlamaParse is currently the best technology I have seen for parsing complex document structures for enterprise RAG pipelines,โ said Dean Barr, applied AI lead at global investment firm The Carlyle Group. โIts ability to preserve nested tables, extract challenging spatial layouts and images is key to maintaining data integrity in advanced RAG and agentic model building.โ
Liu explained that LlamaIndexโs platform has been used in financial analyst assistance, centralized internet search, analytics dashboards for sensor data and internal LLM application development platforms, and in industries including technology, consulting, financial services and healthcare.ย
From simple agents to advanced, multi-agents
Importantly, LlamaIndex layers on agentic reasoning to help provide better query understanding, planning and tool use over different data interfaces, Liu explained. It also incorporates multiple agents that offer specialization and parallelization, and that help optimize cost and reduce latency.ย
The issue with single-agent systems is that โthe more stuff you try to cram into it, the more unreliable it becomes, even if the overall theoretical sophistication is higher,โ said Liu. Also, single agents canโt solve infinite sets of tasks. โIf you try to give an agent 10,000 tools, it doesnโt really do very well.โ
Multi-agents help each agent specialize in a given task, he explained. It has systems-level benefits such as parallelization costs and latency.
โThe idea is that by working together and communicating, you can solve even higher-level tasks,โ said Liu.ย
source: https://venturebeat.com/ai/how-llamaindex-is-ushering-in-the-future-of-rag-for-enterprises/

