Pipelines
Search Pipelines
Update Existing Pipeline
Get Pipeline Status
Run Search
ModelsExpand Collapse
class AdvancedModeTransformConfig: …
chunking_config: Optional[ChunkingConfig]
class AzureOpenAIEmbedding: …
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class DataSinkCreate: …
Schema for creating a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
class GeminiEmbedding: …
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbedding: …
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class LlamaParseParameters: …
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
class MetadataFilters: …
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
class OpenAIEmbedding: …
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class Pipeline: …
Schema for a pipeline.
embedding_config: EmbeddingConfig
class EmbeddingConfigManagedOpenAIEmbeddingConfig: …
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
data_sink: Optional[DataSink]
Schema for a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config: Optional[EmbeddingModelConfig]
Schema for an embedding model config.
embedding_config: EmbeddingModelConfigEmbeddingConfig
The embedding configuration for the embedding model config.
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
embedding_model_config_id: Optional[str]
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters: Optional[LlamaParseParameters]
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: Optional[str]
The ID of the ManagedPipeline this playground pipeline is linked to.
preset_retrieval_parameters: Optional[PresetRetrievalParams]
Preset retrieval parameters for the pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
sparse_model_config: Optional[SparseModelConfig]
transform_config: Optional[TransformConfig]
Configuration for the transformation.
class AdvancedModeTransformConfig: …
chunking_config: Optional[ChunkingConfig]
class PipelineCreate: …
Schema for creating a pipeline.
data_sink: Optional[DataSinkCreate]
Schema for creating a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
data_sink_id: Optional[str]
Data sink ID. When provided instead of data_sink, the data sink will be looked up by ID.
embedding_config: Optional[EmbeddingConfig]
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
embedding_model_config_id: Optional[str]
Embedding model config ID. When provided instead of embedding_config, the embedding model config will be looked up by ID.
llama_parse_parameters: Optional[LlamaParseParameters]
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: Optional[str]
The ID of the ManagedPipeline this playground pipeline is linked to.
preset_retrieval_parameters: Optional[PresetRetrievalParams]
Preset retrieval parameters for the pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
sparse_model_config: Optional[SparseModelConfig]
transform_config: Optional[TransformConfig]
Configuration for the transformation.
class AdvancedModeTransformConfig: …
chunking_config: Optional[ChunkingConfig]
class PresetRetrievalParams: …
Schema for the search params for an retrieval execution that can be preset for a pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
List[Pipeline]
embedding_config: EmbeddingConfig
class EmbeddingConfigManagedOpenAIEmbeddingConfig: …
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
data_sink: Optional[DataSink]
Schema for a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config: Optional[EmbeddingModelConfig]
Schema for an embedding model config.
embedding_config: EmbeddingModelConfigEmbeddingConfig
The embedding configuration for the embedding model config.
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
embedding_model_config_id: Optional[str]
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters: Optional[LlamaParseParameters]
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: Optional[str]
The ID of the ManagedPipeline this playground pipeline is linked to.
preset_retrieval_parameters: Optional[PresetRetrievalParams]
Preset retrieval parameters for the pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
search_filters_inference_schema: Optional[Dict[str, Union[Dict[str, object], List[object], str, 3 more]]]
sparse_similarity_top_k: Optional[int]
Number of nodes for sparse retrieval.