Open In Colab

Clarifai LLM#

Example notebook to call different LLM models using clarifai#

If you’re opening this Notebook on colab, you will probably need to install LlamaIndex 🦙.

%pip install llama-index-llms-clarifai
!pip install llama-index

Install clarifai

!pip install clarifai

Set clarifai PAT as environment variable.

import os


Import clarifai package

from llama_index.llms.clarifai import Clarifai

Explore various models according to your prefrence from Our Models page

# Example parameters
params = dict(

Initialize the LLM

# Method:1 using model_url parameter
llm_model = Clarifai(model_url=params["model_url"])
# Method:2 using model_name, app_id & user_id parameters
llm_model = Clarifai(

Call complete function

llm_reponse = llm_model.complete(
    prompt="write a 10 line rhyming poem about science"
Science is fun, it's true!
From atoms to galaxies, it's all new!
With experiments and tests, we learn so fast,
And discoveries come from the past.
It helps us understand the world around,
And makes our lives more profound.
So let's embrace this wondrous art,
And see where it takes us in the start!

Call chat function

from llama_index.core.llms import ChatMessage

messages = [
    ChatMessage(role="user", content="write about climate change in 50 lines")
Response =
user:  or less.
Climate change is a serious threat to our planet and its inhabitants. Rising temperatures are causing extreme weather events, such as hurricanes, droughts, and wildfires. Sea levels are rising, threatening coastal communities and ecosystems. The melting of polar ice caps is disrupting global navigation and commerce. Climate change is also exacerbating air pollution, which can lead to respiratory problems and other health issues. It's essential that we take action now to reduce greenhouse gas emissions and transition to renewable energy sources to mitigate the worst effects of climate change.

Using Inference parameters#

Alternatively you can call models with inference parameters.

# Here is an inference parameter example for GPT model.
inference_params = dict(temperature=str(0.3), max_tokens=20)
llm_reponse = llm_model.complete(
    prompt="What is nuclear fission and fusion?",
messages = [ChatMessage(role="user", content="Explain about the big bang")]
Response =, inference_params=params)