Pipelines
Search Pipelines
Create Pipeline
Get Pipeline
Update Existing Pipeline
Delete Pipeline
Get Pipeline Status
Upsert Pipeline
Run Search
ModelsExpand Collapse
AdvancedModeTransformConfig = object { chunking_config, mode, segmentation_config }
chunking_config: optional object { mode } or object { chunk_overlap, chunk_size, mode } or object { chunk_overlap, chunk_size, mode, separator } or 2 more
AzureOpenAIEmbedding = object { additional_kwargs, api_base, api_key, 12 more }
AzureOpenAIEmbeddingConfig = object { component, type }
Configuration for the Azure OpenAI embedding model.
BedrockEmbeddingConfig = object { component, type }
component: optional BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
CohereEmbeddingConfig = object { component, type }
DataSinkCreate = object { component, name, sink_type }
Schema for creating a data sink.
component: map[unknown] or CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } or CloudPostgresVectorStore { database, embed_dim, host, 10 more } or 5 more
Component that implements the data sink
CloudPineconeVectorStore = object { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore = object { database, embed_dim, host, 10 more }
hnsw_settings: optional PgVectorHnswSettings { distance_method, ef_construction, ef_search, 2 more }
CloudQdrantVectorStore = object { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore = object { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch = object { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore = object { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
GeminiEmbedding = object { api_base, api_key, class_name, 7 more }
output_dimensionality: optional number
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
GeminiEmbeddingConfig = object { component, type }
Configuration for the Gemini embedding model.
output_dimensionality: optional number
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbedding = object { token, class_name, cookies, 9 more }
headers: optional map[string]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
HuggingFaceInferenceAPIEmbeddingConfig = object { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers: optional map[string]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
LlamaParseParameters = object { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: optional array of object { webhook_events, webhook_headers, webhook_output_format, webhook_url }
Outbound webhook endpoints to notify on job status changes
webhook_events: optional array of "extract.pending" or "extract.success" or "extract.error" or 14 more
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: optional map[string]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
MetadataFilters = object { filters, condition }
Metadata filters for vector stores.
MetadataFilter = object { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
OpenAIEmbedding = object { additional_kwargs, api_base, api_key, 10 more }
OpenAIEmbeddingConfig = object { component, type }
Configuration for the OpenAI embedding model.
Pipeline = object { id, embedding_config, name, 15 more }
Schema for a pipeline.
embedding_config: object { component, type } or AzureOpenAIEmbeddingConfig { component, type } or CohereEmbeddingConfig { component, type } or 5 more
ManagedOpenAIEmbedding = object { component, type }
AzureOpenAIEmbeddingConfig = object { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig = object { component, type }
GeminiEmbeddingConfig = object { component, type }
Configuration for the Gemini embedding model.
output_dimensionality: optional number
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig = object { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers: optional map[string]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig = object { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig = object { component, type }
BedrockEmbeddingConfig = object { component, type }
component: optional BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
Schema for a data sink.
component: map[unknown] or CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } or CloudPostgresVectorStore { database, embed_dim, host, 10 more } or 5 more
Component that implements the data sink
CloudPineconeVectorStore = object { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore = object { database, embed_dim, host, 10 more }
hnsw_settings: optional PgVectorHnswSettings { distance_method, ef_construction, ef_search, 2 more }
CloudQdrantVectorStore = object { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore = object { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch = object { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore = object { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config: optional object { id, embedding_config, name, 3 more }
Schema for an embedding model config.
embedding_config: AzureOpenAIEmbeddingConfig { component, type } or CohereEmbeddingConfig { component, type } or GeminiEmbeddingConfig { component, type } or 4 more
The embedding configuration for the embedding model config.
AzureOpenAIEmbeddingConfig = object { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig = object { component, type }
GeminiEmbeddingConfig = object { component, type }
Configuration for the Gemini embedding model.
output_dimensionality: optional number
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig = object { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers: optional map[string]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig = object { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig = object { component, type }
BedrockEmbeddingConfig = object { component, type }
component: optional BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id: optional string
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters: optional LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: optional array of object { webhook_events, webhook_headers, webhook_output_format, webhook_url }
Outbound webhook endpoints to notify on job status changes
webhook_events: optional array of "extract.pending" or "extract.success" or "extract.error" or 14 more
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: optional map[string]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: optional string
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config: optional PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys }
preset_retrieval_parameters: optional PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha: optional number
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: optional number
Minimum similarity score wrt query for retrieval
files_top_k: optional number
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter = object { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
transform_config: optional AutoTransformConfig { chunk_overlap, chunk_size, mode } or AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
Configuration for the transformation.
AdvancedModeTransformConfig = object { chunking_config, mode, segmentation_config }
chunking_config: optional object { mode } or object { chunk_overlap, chunk_size, mode } or object { chunk_overlap, chunk_size, mode, separator } or 2 more
PipelineCreate = object { name, data_sink, data_sink_id, 10 more }
Schema for creating a pipeline.
Schema for creating a data sink.
component: map[unknown] or CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } or CloudPostgresVectorStore { database, embed_dim, host, 10 more } or 5 more
Component that implements the data sink
CloudPineconeVectorStore = object { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore = object { database, embed_dim, host, 10 more }
hnsw_settings: optional PgVectorHnswSettings { distance_method, ef_construction, ef_search, 2 more }
CloudQdrantVectorStore = object { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore = object { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch = object { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore = object { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
data_sink_id: optional string
Data sink ID. When provided instead of data_sink, the data sink will be looked up by ID.
embedding_config: optional AzureOpenAIEmbeddingConfig { component, type } or CohereEmbeddingConfig { component, type } or GeminiEmbeddingConfig { component, type } or 4 more
AzureOpenAIEmbeddingConfig = object { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig = object { component, type }
GeminiEmbeddingConfig = object { component, type }
Configuration for the Gemini embedding model.
output_dimensionality: optional number
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig = object { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers: optional map[string]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig = object { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig = object { component, type }
BedrockEmbeddingConfig = object { component, type }
component: optional BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id: optional string
Embedding model config ID. When provided instead of embedding_config, the embedding model config will be looked up by ID.
llama_parse_parameters: optional LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: optional array of object { webhook_events, webhook_headers, webhook_output_format, webhook_url }
Outbound webhook endpoints to notify on job status changes
webhook_events: optional array of "extract.pending" or "extract.success" or "extract.error" or 14 more
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: optional map[string]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: optional string
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config: optional PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys }
preset_retrieval_parameters: optional PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha: optional number
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: optional number
Minimum similarity score wrt query for retrieval
files_top_k: optional number
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter = object { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
transform_config: optional AutoTransformConfig { chunk_overlap, chunk_size, mode } or AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
Configuration for the transformation.
AdvancedModeTransformConfig = object { chunking_config, mode, segmentation_config }
chunking_config: optional object { mode } or object { chunk_overlap, chunk_size, mode } or object { chunk_overlap, chunk_size, mode, separator } or 2 more
PresetRetrievalParams = object { alpha, class_name, dense_similarity_cutoff, 11 more }
Schema for the search params for an retrieval execution that can be preset for a pipeline.
alpha: optional number
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: optional number
Minimum similarity score wrt query for retrieval
files_top_k: optional number
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter = object { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
embedding_config: object { component, type } or AzureOpenAIEmbeddingConfig { component, type } or CohereEmbeddingConfig { component, type } or 5 more
ManagedOpenAIEmbedding = object { component, type }
AzureOpenAIEmbeddingConfig = object { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig = object { component, type }
GeminiEmbeddingConfig = object { component, type }
Configuration for the Gemini embedding model.
output_dimensionality: optional number
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig = object { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers: optional map[string]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig = object { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig = object { component, type }
BedrockEmbeddingConfig = object { component, type }
component: optional BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
Schema for a data sink.
component: map[unknown] or CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } or CloudPostgresVectorStore { database, embed_dim, host, 10 more } or 5 more
Component that implements the data sink
CloudPineconeVectorStore = object { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore = object { database, embed_dim, host, 10 more }
hnsw_settings: optional PgVectorHnswSettings { distance_method, ef_construction, ef_search, 2 more }
CloudQdrantVectorStore = object { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore = object { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch = object { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore = object { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config: optional object { id, embedding_config, name, 3 more }
Schema for an embedding model config.
embedding_config: AzureOpenAIEmbeddingConfig { component, type } or CohereEmbeddingConfig { component, type } or GeminiEmbeddingConfig { component, type } or 4 more
The embedding configuration for the embedding model config.
AzureOpenAIEmbeddingConfig = object { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig = object { component, type }
GeminiEmbeddingConfig = object { component, type }
Configuration for the Gemini embedding model.
output_dimensionality: optional number
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig = object { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers: optional map[string]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig = object { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig = object { component, type }
BedrockEmbeddingConfig = object { component, type }
component: optional BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id: optional string
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters: optional LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: optional array of object { webhook_events, webhook_headers, webhook_output_format, webhook_url }
Outbound webhook endpoints to notify on job status changes
webhook_events: optional array of "extract.pending" or "extract.success" or "extract.error" or 14 more
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: optional map[string]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: optional string
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config: optional PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys }
preset_retrieval_parameters: optional PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha: optional number
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: optional number
Minimum similarity score wrt query for retrieval
files_top_k: optional number
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter = object { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types