Skip to content

OctoAI

If you’re opening this Notebook on colab, you will probably need to install LlamaIndex 🦙.

%pip install llama-index-llms-octoai
%pip install llama-index
%pip install octoai-sdk

Include your OctoAI API key below. You can get yours at OctoAI.

Here are some instructions in case you need more guidance.

OCTOAI_API_KEY = ""

Initialize the Integration with the default model

Section titled “Initialize the Integration with the default model”
from llama_index.llms.octoai import OctoAI
octoai = OctoAI(token=OCTOAI_API_KEY)
response = octoai.complete("Paul Graham is ")
print(response)
from llama_index.core.llms import ChatMessage
messages = [
ChatMessage(
role="system",
content="Below is an instruction that describes a task. Write a response that appropriately completes the request.",
),
ChatMessage(role="user", content="Write a blog about Seattle"),
]
response = octoai.chat(messages)
print(response)

Using stream_complete endpoint

response = octoai.stream_complete("Paul Graham is ")
for r in response:
print(r.delta, end="")

Using stream_chat with a list of messages

from llama_index.core.llms import ChatMessage
messages = [
ChatMessage(
role="system",
content="Below is an instruction that describes a task. Write a response that appropriately completes the request.",
),
ChatMessage(role="user", content="Write a blog about Seattle"),
]
response = octoai.stream_chat(messages)
for r in response:
print(r.delta, end="")
# To customize your API token, do this
# otherwise it will lookup OCTOAI_TOKEN from your env variable
octoai = OctoAI(
model="mistral-7b-instruct", max_tokens=128, token=OCTOAI_API_KEY
)
response = octoai.complete("Paul Graham is ")
print(response)