A hidden release that makes LlamaParse way better at RAG over complex documents: we’ve made huge improvements to markdown-based table reconstruction - this allows you to parse very complex tables while making sure the rows/columns are well aligned! 🔨 Check out the results below (source image, before fix, after fix). Tutorial on how to build advanced RAG over your PDFs with lots of text and complex tables: https://lnkd.in/g5P2CMUA
LlamaIndex
Technology, Information and Internet
San Francisco, California 182,463 followers
The central interface between LLMs and your external data.
About us
The data framework for LLMs Python: Github: https://github.com/jerryjliu/llama_index Docs: https://docs.llamaindex.ai/ Typescript/Javascript: Github: https://github.com/run-llama/LlamaIndexTS Docs: https://ts.llamaindex.ai/ Other: Discord: discord.gg/dGcwcsnxhU LlamaHub: llamahub.ai Twitter: https://twitter.com/llama_index Blog: blog.llamaindex.ai #ai #llms #rag
- Website
-
https://www.llamaindex.ai/
External link for LlamaIndex
- Industry
- Technology, Information and Internet
- Company size
- 2-10 employees
- Headquarters
- San Francisco, California
- Type
- Public Company
Locations
-
Primary
San Francisco, California, US
Employees at LlamaIndex
Updates
-
Worked with AI but wondering why to use LlamaIndex? This new video is an introduction to LlamaIndex and in particular its agentic capabilities, covering: ➡️ What is LlamaIndex, including: ⭐️ Frameworks in Python and TypeScript ⭐️ LlamaParse service for parsing complex documents for LLMs ⭐️ LlamaCloud service for end-to-end enterprise RAG ⭐️ LlamaHub for free downloads of data connectors, LLM adapters, embedding models, vector stores, agent tools and more ➡️ Why you should use it, including: ⭐️Build faster ⭐️ Eliminate boilerplate ⭐️ Avoid early pitfalls ⭐️ Get into production ⭐️ Deliver business value ➡️ What can it do, including code examples of how to build: ⭐️ Retrieval augmented generation (RAG) ⭐️ World class parsing ⭐️ Agents and multi-agent systems https://lnkd.in/d_D3fenm
-
An underrated capability of sonnet-3.5 is that it’s really good at chart understanding 📊 - compared to gpt-4o it’s much better at inferring chart values into a structured table. Thanks to our brand-new LlamaParse release 💫 you can easily use SOTA multimodal models like sonnet-3.5 for document parsing and structuring, with added validation/scalability/reliability benefits behind our infrastructure. Check out our full tutorial below and example from the Llama 2 paper. Huge shoutout to Pierre-Loic Doulcet and Sacha Bron for the exciting new features. Additional releases: - Fast Mode: Run LlamaParse with our core text layout capabilities without OCR/models, for 0.1c a page. - Improved Table Reconstruction: Fewer hallucinations in reconstructing complex tables. Results coming soon. Notebook: https://lnkd.in/dBzRNYYc LlamaParse: https://lnkd.in/g3UmUkcD
-
New LlamaIndex Webinar 🚨We're excited to cohost a workshop with the cofounders of Deasie (Reece Griffiths, Leonard Platzer, Mikko Peiponen) on improving RAG with advanced parsing and metadata. The data processing layer is one of the most important pieces to get right for RAG. This means that AI engineers need to make careful decisions in terms of parsing and transformations - including metadata extraction and chunking - in order to make sure that their e2e QA system is surfacing relevant results. This is a nice two-part workshop that demonstrates the following: the value of good parsing itself over complex documents, with LlamaParse the value of additional value of adding in metadata through Deasie's powerful automated labeling platform We show overall experimental results over research papers validating the combination of parsing + metadata for good performance. This Thursday 9am PT, check it out! https://lu.ma/ufx77ve8 Deasie: https://deasie.com/ LlamaParse: https://lnkd.in/g3UmUkcD
-
Build-your-own graph query algorithm with LlamaIndex + Mistral AI 🔎🕸️ You’re not limited to text-to-cypher or pure vector search when you build GraphRAG - you can combine both techniques or define your own custom graph query algorithm, as long as you have access to the underlying property graph store. Ravi Theja Desetty has a full video tutorial and notebook showing you how to do this - combine text-to-cypher + vector retrieval with Cohere reranking, and synthesize the final results with mistral-large. Video: https://lnkd.in/g43-W-FJ Mistral AI cookbook: https://lnkd.in/gW6yEqdj
-
The future of document RAG is multimodal RAG 🖼️📑. In our brand-new cookbook we highlight a new multi-modal RAG architecture for processing a slide deck (heavy in text, diagrams, charts, tables) using LlamaParse, LlamaIndex, and gpt-4o. At the core is a hybrid text/image chunk. In contrast to standard RAG pipelines that only index/synthesize over text chunks, our multi-modal RAG setup does the following: 1. For each page, both a) parse out the text, but also b) screenshot each page as a separate image. You can use standard OCR techniques but also multimodal models for extraction. 2. Create a hybrid chunk that contains the parsed text, with a file link to the saved image 3. Use text embeddings to embed the chunk (note: you can use image embeddings as well, but we find gaps here) 4. Retrieve relevant chunks by text embeddings. 5. During synthesis, load in the text and image. This combines the benefits of text parsing (well studied), with the visual recognition benefits of multimodal models. We have a full cookbook showing how you can leverage LlamaParse to do both the text/image extraction, and then set it up with gpt-4o for a full pipeline. Check it out: https://lnkd.in/gjaFvwYE
-
We're co-hosting an online hackathon this weekend, starting Friday 19th! Use Llama 3 from Meta to build your AI app, and win thousands of dollars in cash and credits and other fun prizes from us and our co-hosts Together AI Milvus and lablab.ai! Sign up here: https://lnkd.in/dGN_kTgN
-
Deploy llama-agents running entirely self-hosted agents using Arcee.ai, Mistral AI and Ollama! In this video, AbdulMajedRaja RS walks you through ➡️ Setting up llama-agents ➡️ Getting local models running with Ollama ➡️ Building tools and running them with Arcee.ai https://lnkd.in/dAk4Su43
Local Llama Agents (Mistral + Arcee Agent)
https://www.youtube.com/
-
How exactly do knowledge graphs work? Before you dive deep into GraphRAG, learn the basics of property graphs - each node and relation can store a structured dictionary of properties. This underlying abstraction unlocks a lot of cool techniques during construction and retrieval! Don’t miss Ravi Theja Desetty's video explaining property graphs, different ways of constructing them with LLMs, and different ways of querying them complete with diagrams. We built these in collaboration with Neo4j. Introductory video: https://lnkd.in/gd_eRUT8 Property graphs with Neo4j: https://lnkd.in/gACUFB8Q
-
As a special weekend release, we’ve implemented an initial working implementation of GraphRAG concepts in LlamaIndex [in beta]. GraphRAG from Microsoft has a lot of cool ideas, and we’ve implemented the following main concepts: 1. Graph Generation: Build graph and community summaries 2. Community-based retrieval: use community summaries and aggregate intermediate answers. Huge shoutout to Ravi Theja Desetty for working on this, and this is just the start. Next up: advanced retrieval features, entity resolution, integration with Neo4j - stay tuned! Cookbook: https://lnkd.in/gGb8GJeq Image source from GraphRAG paper: GraphRAG: https://lnkd.in/gvRsrPn3 Paper: https://lnkd.in/emJ8w2fq