Skip to content

Welcome to LlamaIndex ๐Ÿฆ™ !#

LlamaIndex is a framework for building context-augmented LLM applications. Context augmentation refers to any use case that applies LLMs on top of your private or domain-specific data. Some popular use cases include the following:

  • Question-Answering Chatbots (commonly referred to as RAG systems, which stands for "Retrieval-Augmented Generation")
  • Document Understanding and Extraction
  • Autonomous Agents that can perform research and take actions

LlamaIndex provides the tools to build any of these above use cases from prototype to production. The tools allow you to both ingest/process this data and implement complex query workflows combining data access with LLM prompting.

LlamaIndex is available in Python (these docs) and Typescript.

Tip

Updating to LlamaIndex v0.10.0? Check out the migration guide.

๐Ÿš€ Why Context Augmentation?#

LLMs offer a natural language interface between humans and data. Widely available models come pre-trained on huge amounts of publicly available data. However, they are not trained on your data, which may be private or specific to the problem you're trying to solve. It's behind APIs, in SQL databases, or trapped in PDFs and slide decks.

LlamaIndex provides tooling to enable context augmentation. A popular example is Retrieval-Augmented Generation (RAG) which combines context with LLMs at inference time. Another is finetuning.

๐Ÿฆ™ LlamaIndex is the Data Framework for Context-Augmented LLM Apps#

LlamaIndex imposes no restriction on how you use LLMs. You can still use LLMs as auto-complete, chatbots, semi-autonomous agents, and more. It only makes LLMs more relevant to you.

LlamaIndex provides the following tools to help you quickly standup production-ready LLM applications:

  • Data connectors ingest your existing data from their native source and format. These could be APIs, PDFs, SQL, and (much) more.
  • Data indexes structure your data in intermediate representations that are easy and performant for LLMs to consume.
  • Engines provide natural language access to your data. For example:
    • Query engines are powerful interfaces for question-answering (e.g. a RAG pipeline).
    • Chat engines are conversational interfaces for multi-message, "back and forth" interactions with your data.
  • Agents are LLM-powered knowledge workers augmented by tools, from simple helper functions to API integrations and more.
  • Observability/Evaluation integrations that enable you to rigorously experiment, evaluate, and monitor your app in a virtuous cycle.

๐Ÿ‘จโ€๐Ÿ‘ฉโ€๐Ÿ‘งโ€๐Ÿ‘ฆ Who is LlamaIndex for?#

LlamaIndex provides tools for beginners, advanced users, and everyone in between.

Our high-level API allows beginner users to use LlamaIndex to ingest and query their data in 5 lines of code.

For more complex applications, our lower-level APIs allow advanced users to customize and extend any moduleโ€”data connectors, indices, retrievers, query engines, reranking modulesโ€”to fit their needs.

Getting Started#

To install the library:

pip install llama-index

We recommend starting at how to read these docs which will point you to the right place based on your experience level.

๐Ÿ—บ๏ธ Ecosystem#

To download or contribute, find LlamaIndex on:

LlamaCloud#

If you're an enterprise developer, check out LlamaCloud. It is a managed platform for data parsing and ingestion, allowing you to get production-quality data for your production LLM application.

Check out the following resources:

Community#

Need help? Have a feature suggestion? Join the LlamaIndex community:

Associated projects#

  • ๐Ÿก LlamaHub | A large (and growing!) collection of custom data connectors
  • SEC Insights | A LlamaIndex-powered application for financial research
  • create-llama | A CLI tool to quickly scaffold LlamaIndex projects