Pipelines
Search Pipelines
Create Pipeline
Get Pipeline
Update Existing Pipeline
Delete Pipeline
Get Pipeline Status
Upsert Pipeline
Run Search
ModelsExpand Collapse
AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
chunking_config?: NoneChunkingConfig { mode } | CharacterChunkingConfig { chunk_overlap, chunk_size, mode } | TokenChunkingConfig { chunk_overlap, chunk_size, mode, separator } | 2 more
AzureOpenAIEmbedding { additional_kwargs, api_base, api_key, 12 more }
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
DataSinkCreate { component, name, sink_type }
Schema for creating a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
GeminiEmbedding { api_base, api_key, class_name, 7 more }
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbedding { token, class_name, cookies, 9 more }
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
webhook_configurations?: Array<WebhookConfiguration> | null
Outbound webhook endpoints to notify on job status changes
webhook_events?: Array<"extract.pending" | "extract.success" | "extract.error" | 14 more> | null
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers?: Record<string, string> | null
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
MetadataFilters { filters, condition }
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
OpenAIEmbedding { additional_kwargs, api_base, api_key, 10 more }
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
Pipeline { id, embedding_config, name, 15 more }
Schema for a pipeline.
embedding_config: ManagedOpenAIEmbeddingConfig { component, type } | AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | 5 more
ManagedOpenAIEmbeddingConfig { component, type }
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
Schema for a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config?: EmbeddingModelConfig | null
Schema for an embedding model config.
embedding_config: AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | GeminiEmbeddingConfig { component, type } | 4 more
The embedding configuration for the embedding model config.
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id?: string | null
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters?: LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more } | null
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations?: Array<WebhookConfiguration> | null
Outbound webhook endpoints to notify on job status changes
webhook_events?: Array<"extract.pending" | "extract.success" | "extract.error" | 14 more> | null
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers?: Record<string, string> | null
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id?: string | null
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config?: PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys } | null
preset_retrieval_parameters?: PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha?: number | null
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff?: number | null
Minimum similarity score wrt query for retrieval
files_top_k?: number | null
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
transform_config?: AutoTransformConfig { chunk_overlap, chunk_size, mode } | AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
Configuration for the transformation.
AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
chunking_config?: NoneChunkingConfig { mode } | CharacterChunkingConfig { chunk_overlap, chunk_size, mode } | TokenChunkingConfig { chunk_overlap, chunk_size, mode, separator } | 2 more
PipelineCreate { name, data_sink, data_sink_id, 10 more }
Schema for creating a pipeline.
Schema for creating a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
data_sink_id?: string | null
Data sink ID. When provided instead of data_sink, the data sink will be looked up by ID.
embedding_config?: AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | GeminiEmbeddingConfig { component, type } | 4 more | null
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id?: string | null
Embedding model config ID. When provided instead of embedding_config, the embedding model config will be looked up by ID.
llama_parse_parameters?: LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more }
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations?: Array<WebhookConfiguration> | null
Outbound webhook endpoints to notify on job status changes
webhook_events?: Array<"extract.pending" | "extract.success" | "extract.error" | 14 more> | null
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers?: Record<string, string> | null
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id?: string | null
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config?: PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys } | null
preset_retrieval_parameters?: PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha?: number | null
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff?: number | null
Minimum similarity score wrt query for retrieval
files_top_k?: number | null
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
transform_config?: AutoTransformConfig { chunk_overlap, chunk_size, mode } | AdvancedModeTransformConfig { chunking_config, mode, segmentation_config } | null
Configuration for the transformation.
AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
chunking_config?: NoneChunkingConfig { mode } | CharacterChunkingConfig { chunk_overlap, chunk_size, mode } | TokenChunkingConfig { chunk_overlap, chunk_size, mode, separator } | 2 more
PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Schema for the search params for an retrieval execution that can be preset for a pipeline.
alpha?: number | null
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff?: number | null
Minimum similarity score wrt query for retrieval
files_top_k?: number | null
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
embedding_config: ManagedOpenAIEmbeddingConfig { component, type } | AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | 5 more
ManagedOpenAIEmbeddingConfig { component, type }
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
Schema for a data sink.
component: Record<string, unknown> | CloudPineconeVectorStore { api_key, index_name, class_name, 3 more } | CloudPostgresVectorStore { database, embed_dim, host, 10 more } | 5 more
Component that implements the data sink
CloudPineconeVectorStore { api_key, index_name, class_name, 3 more }
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
CloudPostgresVectorStore { database, embed_dim, host, 10 more }
CloudQdrantVectorStore { api_key, collection_name, url, 4 more }
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
CloudAzureAISearchVectorStore { search_service_api_key, search_service_endpoint, class_name, 8 more }
Cloud Azure AI Search Vector Store.
CloudMongoDBAtlasVectorSearch { collection_name, db_name, mongodb_uri, 5 more }
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
CloudAstraDBVectorStore { token, api_endpoint, collection_name, 4 more }
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config?: EmbeddingModelConfig | null
Schema for an embedding model config.
embedding_config: AzureOpenAIEmbeddingConfig { component, type } | CohereEmbeddingConfig { component, type } | GeminiEmbeddingConfig { component, type } | 4 more
The embedding configuration for the embedding model config.
AzureOpenAIEmbeddingConfig { component, type }
Configuration for the Azure OpenAI embedding model.
CohereEmbeddingConfig { component, type }
GeminiEmbeddingConfig { component, type }
Configuration for the Gemini embedding model.
output_dimensionality?: number | null
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
HuggingFaceInferenceAPIEmbeddingConfig { component, type }
Configuration for the HuggingFace Inference API embedding model.
headers?: Record<string, string> | null
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
OpenAIEmbeddingConfig { component, type }
Configuration for the OpenAI embedding model.
VertexAIEmbeddingConfig { component, type }
BedrockEmbeddingConfig { component, type }
component?: BedrockEmbedding { additional_kwargs, aws_access_key_id, aws_secret_access_key, 9 more }
embedding_model_config_id?: string | null
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters?: LlamaParseParameters { adaptive_long_table, aggressive_table_extraction, annotate_links, 116 more } | null
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations?: Array<WebhookConfiguration> | null
Outbound webhook endpoints to notify on job status changes
webhook_events?: Array<"extract.pending" | "extract.success" | "extract.error" | 14 more> | null
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers?: Record<string, string> | null
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id?: string | null
The ID of the ManagedPipeline this playground pipeline is linked to.
metadata_config?: PipelineMetadataConfig { excluded_embed_metadata_keys, excluded_llm_metadata_keys } | null
preset_retrieval_parameters?: PresetRetrievalParams { alpha, class_name, dense_similarity_cutoff, 11 more }
Preset retrieval parameters for the pipeline.
alpha?: number | null
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff?: number | null
Minimum similarity score wrt query for retrieval
files_top_k?: number | null
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
transform_config?: AutoTransformConfig { chunk_overlap, chunk_size, mode } | AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
Configuration for the transformation.
AdvancedModeTransformConfig { chunking_config, mode, segmentation_config }
chunking_config?: NoneChunkingConfig { mode } | CharacterChunkingConfig { chunk_overlap, chunk_size, mode } | TokenChunkingConfig { chunk_overlap, chunk_size, mode, separator } | 2 more
PipelineRetrieveResponse { pipeline_id, retrieval_nodes, class_name, 5 more }
Schema for the result of an retrieval execution.
retrieval_nodes: Array<RetrievalNode>
The nodes retrieved by the pipeline for the given query.
Provided for backward compatibility.
excluded_embed_metadata_keys?: Array<string>
Metadata keys that are excluded from text for the embed model.
metadata_template?: string
Template for how metadata is formatted, with {key} and {value} placeholders.
Metadata filters for vector stores.
MetadataFilter { key, value, operator }
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
PipelinesSync
Sync Pipeline
Cancel Pipeline Sync
PipelinesData Sources
List Pipeline Data Sources
Add Data Sources To Pipeline
Update Pipeline Data Source
Get Pipeline Data Source Status
Sync Pipeline Data Source
ModelsExpand Collapse
PipelineDataSource { id, component, data_source_id, 13 more }
Schema for a data source in a pipeline.
component: Record<string, unknown> | CloudS3DataSource { bucket, aws_access_id, aws_access_secret, 5 more } | CloudAzStorageBlobDataSource { account_url, container_name, account_key, 8 more } | 9 more
Component that implements the data source
CloudAzStorageBlobDataSource { account_url, container_name, account_key, 8 more }
CloudSharepointDataSource { client_id, client_secret, tenant_id, 11 more }
exclude_path_patterns?: Array<string> | null
List of regex patterns for file paths to exclude. Files whose paths (including filename) match any pattern will be excluded. Example: [‘/temp/’, ‘/backup/’, ‘.git/’, ‘.tmp$’, ’^~’]
CloudConfluenceDataSource { authentication_mechanism, server_url, api_token, 10 more }
CloudJiraDataSource { authentication_mechanism, query, api_token, 5 more }
CloudJiraDataSourceV2 { authentication_mechanism, query, server_url, 10 more }
CloudBoxDataSource { authentication_mechanism, class_name, client_id, 6 more }
custom_metadata?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
DataSourceGetDataSourcesResponse = Array<PipelineDataSource { id, component, data_source_id, 13 more } >
component: Record<string, unknown> | CloudS3DataSource { bucket, aws_access_id, aws_access_secret, 5 more } | CloudAzStorageBlobDataSource { account_url, container_name, account_key, 8 more } | 9 more
Component that implements the data source
CloudAzStorageBlobDataSource { account_url, container_name, account_key, 8 more }
CloudSharepointDataSource { client_id, client_secret, tenant_id, 11 more }
exclude_path_patterns?: Array<string> | null
List of regex patterns for file paths to exclude. Files whose paths (including filename) match any pattern will be excluded. Example: [‘/temp/’, ‘/backup/’, ‘.git/’, ‘.tmp$’, ’^~’]
CloudConfluenceDataSource { authentication_mechanism, server_url, api_token, 10 more }
CloudJiraDataSource { authentication_mechanism, query, api_token, 5 more }
CloudJiraDataSourceV2 { authentication_mechanism, query, server_url, 10 more }
CloudBoxDataSource { authentication_mechanism, class_name, client_id, 6 more }
custom_metadata?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
DataSourceUpdateDataSourcesResponse = Array<PipelineDataSource { id, component, data_source_id, 13 more } >
component: Record<string, unknown> | CloudS3DataSource { bucket, aws_access_id, aws_access_secret, 5 more } | CloudAzStorageBlobDataSource { account_url, container_name, account_key, 8 more } | 9 more
Component that implements the data source
CloudAzStorageBlobDataSource { account_url, container_name, account_key, 8 more }
CloudSharepointDataSource { client_id, client_secret, tenant_id, 11 more }
exclude_path_patterns?: Array<string> | null
List of regex patterns for file paths to exclude. Files whose paths (including filename) match any pattern will be excluded. Example: [‘/temp/’, ‘/backup/’, ‘.git/’, ‘.tmp$’, ’^~’]
CloudConfluenceDataSource { authentication_mechanism, server_url, api_token, 10 more }
CloudJiraDataSource { authentication_mechanism, query, api_token, 5 more }
CloudJiraDataSourceV2 { authentication_mechanism, query, server_url, 10 more }
CloudBoxDataSource { authentication_mechanism, class_name, client_id, 6 more }
custom_metadata?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
PipelinesImages
List File Page Screenshots
Get File Page Screenshot
Get File Page Figure
List File Pages Figures
PipelinesFiles
Get Pipeline File Status Counts
Get Pipeline File Status
Add Files To Pipeline Api
Update Pipeline File
Delete Pipeline File
List Pipeline Files2
ModelsExpand Collapse
PipelineFile { id, pipeline_id, config_hash, 16 more }
A file associated with a pipeline.
config_hash?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
custom_metadata?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
permission_info?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
resource_info?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
config_hash?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
custom_metadata?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
permission_info?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
resource_info?: Record<string, Record<string, unknown> | Array<unknown> | string | 2 more | null> | null
PipelinesMetadata
Import Pipeline Metadata
Delete Pipeline Files Metadata
PipelinesDocuments
Create Batch Pipeline Documents
Paginated List Pipeline Documents
Get Pipeline Document
Delete Pipeline Document
Get Pipeline Document Status
Sync Pipeline Document
List Pipeline Document Chunks
Upsert Batch Pipeline Documents
ModelsExpand Collapse
TextNode { class_name, embedding, end_char_idx, 11 more }
Provided for backward compatibility.
excluded_embed_metadata_keys?: Array<string>
Metadata keys that are excluded from text for the embed model.
metadata_template?: string
Template for how metadata is formatted, with {key} and {value} placeholders.
excluded_embed_metadata_keys?: Array<string>
Metadata keys that are excluded from text for the embed model.
metadata_template?: string
Template for how metadata is formatted, with {key} and {value} placeholders.