Pipelines
Search Pipelines
Create Pipeline
Get Pipeline
Update Existing Pipeline
Delete Pipeline
Get Pipeline Status
Upsert Pipeline
Run Search
ModelsExpand Collapse
AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
chunking_config?: NoneChunkingConfig { mode } | CharacterChunkingConfig { chunk_overlap, chunk_size, mode } | TokenChunkingConfig { chunk_overlap, chunk_size, mode, separator } | 2 more
AzureOpenAIEmbedding { additional_kwargs, api_base, api_key, 12 more }
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
DataSinkCreate { component, name, sink_type }
Schema for creating a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
GeminiEmbedding { api_base, api_key, class_name, 7 more }
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbedding { token, class_name, cookies, 9 more }
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations?: Array<WebhookConfiguration> | null
Outbound webhook endpoints to notify on job status changes
webhook_events?: Array<"extract.pending" | "extract.success" | "extract.error" | 14 more> | null
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers?: Record<string, string> | null
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
MetadataFilters { filters, condition }
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
OpenAIEmbedding { additional_kwargs, api_base, api_key, 10 more }
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
Pipeline { id, embedding_config, name, 15 more }
Schema for a pipeline.
embedding_config: ManagedOpenAIEmbeddingConfig { component, type } | AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | 5 more
ManagedOpenAIEmbeddingConfig { component, type }
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
Schema for a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config?: EmbeddingModelConfig | null
Schema for an embedding model config.
embedding_config: AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | GeminiEmbeddingConfig { component, type } | 4 more
The embedding configuration for the embedding model config.
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id?: string | null
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters?: LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more } | null
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations?: Array<WebhookConfiguration> | null
Outbound webhook endpoints to notify on job status changes
webhook_events?: Array<"extract.pending" | "extract.success" | "extract.error" | 14 more> | null
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers?: Record<string, string> | null
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id?: string | null
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config?: PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys } | null
preset_retrieval_parameters?: PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha?: number | null
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff?: number | null
Minimum similarity score wrt query for retrieval
files_top_k?: number | null
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
transform_config?: AutoTransformConfig { chunk_overlap, chunk_size, mode } | AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
Configuration for the transformation.
AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
chunking_config?: NoneChunkingConfig { mode } | CharacterChunkingConfig { chunk_overlap, chunk_size, mode } | TokenChunkingConfig { chunk_overlap, chunk_size, mode, separator } | 2 more
PipelineCreate { name, data_sink, data_sink_id, 10 more }
Schema for creating a pipeline.
Schema for creating a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
data_sink_id?: string | null
Data sink ID. When provided instead of data_sink, the data sink will be looked up by ID.
embedding_config?: AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | GeminiEmbeddingConfig { component, type } | 4 more | null
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id?: string | null
Embedding model config ID. When provided instead of embedding_config, the embedding model config will be looked up by ID.
llama_parse_parameters?: LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations?: Array<WebhookConfiguration> | null
Outbound webhook endpoints to notify on job status changes
webhook_events?: Array<"extract.pending" | "extract.success" | "extract.error" | 14 more> | null
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers?: Record<string, string> | null
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id?: string | null
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config?: PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys } | null
preset_retrieval_parameters?: PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha?: number | null
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff?: number | null
Minimum similarity score wrt query for retrieval
files_top_k?: number | null
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
transform_config?: AutoTransformConfig { chunk_overlap, chunk_size, mode } | AdvancedModeTransformConfig { chunking_config, mode, segmentation_config } | null
Configuration for the transformation.
AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
chunking_config?: NoneChunkingConfig { mode } | CharacterChunkingConfig { chunk_overlap, chunk_size, mode } | TokenChunkingConfig { chunk_overlap, chunk_size, mode, separator } | 2 more
PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Schema for the search params for an retrieval execution that can be preset for a pipeline.
alpha?: number | null
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff?: number | null
Minimum similarity score wrt query for retrieval
files_top_k?: number | null
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
embedding_config: ManagedOpenAIEmbeddingConfig { component, type } | AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | 5 more
ManagedOpenAIEmbeddingConfig { component, type }
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
Schema for a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config?: EmbeddingModelConfig | null
Schema for an embedding model config.
embedding_config: AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | GeminiEmbeddingConfig { component, type } | 4 more
The embedding configuration for the embedding model config.
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id?: string | null
The ID of the EmbeddingModelConfig this pipeline is using.