LlamaIndex

LlamaIndex

Technology, Information and Internet

San Francisco, California 182,463 followers

The central interface between LLMs and your external data.

About us

The data framework for LLMs Python: Github: https://github.com/jerryjliu/llama_index Docs: https://docs.llamaindex.ai/ Typescript/Javascript: Github: https://github.com/run-llama/LlamaIndexTS Docs: https://ts.llamaindex.ai/ Other: Discord: discord.gg/dGcwcsnxhU LlamaHub: llamahub.ai Twitter: https://twitter.com/llama_index Blog: blog.llamaindex.ai #ai #llms #rag

Website
https://www.llamaindex.ai/
Industry
Technology, Information and Internet
Company size
2-10 employees
Headquarters
San Francisco, California
Type
Public Company

Locations

Employees at LlamaIndex

Updates

  • View organization page for LlamaIndex, graphic

    182,463 followers

    A hidden release that makes LlamaParse way better at RAG over complex documents: we’ve made huge improvements to markdown-based table reconstruction - this allows you to parse very complex tables while making sure the rows/columns are well aligned! 🔨 Check out the results below (source image, before fix, after fix). Tutorial on how to build advanced RAG over your PDFs with lots of text and complex tables: https://lnkd.in/g5P2CMUA

    • No alternative text description for this image
  • View organization page for LlamaIndex, graphic

    182,463 followers

    Worked with AI but wondering why to use LlamaIndex? This new video is an introduction to LlamaIndex and in particular its agentic capabilities, covering: ➡️ What is LlamaIndex, including: ⭐️ Frameworks in Python and TypeScript ⭐️ LlamaParse service for parsing complex documents for LLMs ⭐️ LlamaCloud service for end-to-end enterprise RAG ⭐️ LlamaHub for free downloads of data connectors, LLM adapters, embedding models, vector stores, agent tools and more ➡️ Why you should use it, including: ⭐️Build faster ⭐️ Eliminate boilerplate ⭐️ Avoid early pitfalls ⭐️ Get into production ⭐️ Deliver business value ➡️ What can it do, including code examples of how to build: ⭐️ Retrieval augmented generation (RAG) ⭐️ World class parsing ⭐️ Agents and multi-agent systems https://lnkd.in/d_D3fenm

    • No alternative text description for this image
  • View organization page for LlamaIndex, graphic

    182,463 followers

    An underrated capability of sonnet-3.5 is that it’s really good at chart understanding 📊 - compared to gpt-4o it’s much better at inferring chart values into a structured table. Thanks to our brand-new LlamaParse release 💫 you can easily use SOTA multimodal models like sonnet-3.5 for document parsing and structuring, with added validation/scalability/reliability benefits behind our infrastructure. Check out our full tutorial below and example from the Llama 2 paper. Huge shoutout to Pierre-Loic Doulcet and Sacha Bron for the exciting new features. Additional releases: - Fast Mode: Run LlamaParse with our core text layout capabilities without OCR/models, for 0.1c a page.  - Improved Table Reconstruction: Fewer hallucinations in reconstructing complex tables. Results coming soon. Notebook: https://lnkd.in/dBzRNYYc LlamaParse: https://lnkd.in/g3UmUkcD

    • No alternative text description for this image
  • View organization page for LlamaIndex, graphic

    182,463 followers

    New LlamaIndex Webinar 🚨We're excited to cohost a workshop with the cofounders of Deasie (Reece Griffiths, Leonard Platzer, Mikko Peiponen) on improving RAG with advanced parsing and metadata. The data processing layer is one of the most important pieces to get right for RAG. This means that AI engineers need to make careful decisions in terms of parsing and transformations - including metadata extraction and chunking - in order to make sure that their e2e QA system is surfacing relevant results. This is a nice two-part workshop that demonstrates the following: the value of good parsing itself over complex documents, with LlamaParse the value of additional value of adding in metadata through Deasie's powerful automated labeling platform We show overall experimental results over research papers validating the combination of parsing + metadata for good performance. This Thursday 9am PT, check it out! https://lu.ma/ufx77ve8 Deasie: https://deasie.com/ LlamaParse: https://lnkd.in/g3UmUkcD

    • No alternative text description for this image
  • View organization page for LlamaIndex, graphic

    182,463 followers

    Build-your-own graph query algorithm with LlamaIndex + Mistral AI 🔎🕸️ You’re not limited to text-to-cypher or pure vector search when you build GraphRAG - you can combine both techniques or define your own custom graph query algorithm, as long as you have access to the underlying property graph store. Ravi Theja Desetty has a full video tutorial and notebook showing you how to do this - combine text-to-cypher + vector retrieval with Cohere reranking, and synthesize the final results with mistral-large. Video: https://lnkd.in/g43-W-FJ Mistral AI cookbook: https://lnkd.in/gW6yEqdj

    • No alternative text description for this image
  • View organization page for LlamaIndex, graphic

    182,463 followers

    The future of document RAG is multimodal RAG 🖼️📑. In our brand-new cookbook we highlight a new multi-modal RAG architecture for processing a slide deck (heavy in text, diagrams, charts, tables) using LlamaParse, LlamaIndex, and gpt-4o. At the core is a hybrid text/image chunk. In contrast to standard RAG pipelines that only index/synthesize over text chunks, our multi-modal RAG setup does the following:  1. For each page, both a) parse out the text, but also b) screenshot each page as a separate image. You can use standard OCR techniques but also multimodal models for extraction. 2. Create a hybrid chunk that contains the parsed text, with a file link to the saved image 3. Use text embeddings to embed the chunk (note: you can use image embeddings as well, but we find gaps here)  4. Retrieve relevant chunks by text embeddings. 5. During synthesis, load in the text and image. This combines the benefits of text parsing (well studied), with the visual recognition benefits of multimodal models. We have a full cookbook showing how you can leverage LlamaParse to do both the text/image extraction, and then set it up with gpt-4o for a full pipeline. Check it out: https://lnkd.in/gjaFvwYE

    • No alternative text description for this image
  • View organization page for LlamaIndex, graphic

    182,463 followers

    How exactly do knowledge graphs work? Before you dive deep into GraphRAG, learn the basics of property graphs - each node and relation can store a structured dictionary of properties. This underlying abstraction unlocks a lot of cool techniques during construction and retrieval! Don’t miss Ravi Theja Desetty's video explaining property graphs, different ways of constructing them with LLMs, and different ways of querying them complete with diagrams. We built these in collaboration with Neo4j. Introductory video: https://lnkd.in/gd_eRUT8 Property graphs with Neo4j: https://lnkd.in/gACUFB8Q

    • No alternative text description for this image
  • View organization page for LlamaIndex, graphic

    182,463 followers

    As a special weekend release, we’ve implemented an initial working implementation of GraphRAG concepts in LlamaIndex [in beta]. GraphRAG from Microsoft has a lot of cool ideas, and we’ve implemented the following main concepts:  1. Graph Generation: Build graph and community summaries 2. Community-based retrieval: use community summaries and aggregate intermediate answers. Huge shoutout to Ravi Theja Desetty for working on this, and this is just the start. Next up: advanced retrieval features, entity resolution, integration with Neo4j - stay tuned! Cookbook: https://lnkd.in/gGb8GJeq Image source from GraphRAG paper:  GraphRAG: https://lnkd.in/gvRsrPn3 Paper: https://lnkd.in/emJ8w2fq

    • No alternative text description for this image

Similar pages

Funding

LlamaIndex 1 total round

Last Round

Seed

US$ 8.5M

See more info on crunchbase