Skip to content

Managed Data Sink

Use LlamaCloud managed index as data sink.

managed

Simply set data_sink_id to None when creating a pipeline

from llama_cloud import AsyncLlamaCloud, LlamaCloud
from llama_cloud.types.pipeline_create_params import (
EmbeddingConfigOpenAIEmbeddingConfig,
EmbeddingConfigOpenAIEmbeddingConfigComponent,
)
client = LlamaCloud(api_key=os.environ["LLAMA_CLOUD_API_KEY"])
pipeline = client.pipelines.create(
name="my-first-index",
project_id="my-project-id",
data_sink_id=None, # Use managed data sink
embedding_config=EmbeddingConfigOpenAIEmbeddingConfig(
component=EmbeddingConfigOpenAIEmbeddingConfigComponent(
api_key="sk-1234",
model_name="text-embedding-3-small",
),
type="OPENAI_EMBEDDING",
),
llama_parse_parameters={"parse_mode": "parse_document_with_agent", "model": "openai-gpt-4-1-mini"},
transform_config={"mode": "auto", "chunk_overlap": 128, "chunk_size": 1028},
)