Pipelines
Search Pipelines
Update Existing Pipeline
Get Pipeline Status
Run Search
ModelsExpand Collapse
class AdvancedModeTransformConfig: …
chunking_config: Optional[ChunkingConfig]
class AzureOpenAIEmbedding: …
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class DataSinkCreate: …
Schema for creating a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
class GeminiEmbedding: …
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbedding: …
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class LlamaParseParameters: …
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
class MetadataFilters: …
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
class OpenAIEmbedding: …
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class Pipeline: …
Schema for a pipeline.
embedding_config: EmbeddingConfig
class EmbeddingConfigManagedOpenAIEmbeddingConfig: …
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
data_sink: Optional[DataSink]
Schema for a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config: Optional[EmbeddingModelConfig]
Schema for an embedding model config.
embedding_config: EmbeddingModelConfigEmbeddingConfig
The embedding configuration for the embedding model config.
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
embedding_model_config_id: Optional[str]
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters: Optional[LlamaParseParameters]
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: Optional[str]
The ID of the ManagedPipeline this playground pipeline is linked to.
preset_retrieval_parameters: Optional[PresetRetrievalParams]
Preset retrieval parameters for the pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
sparse_model_config: Optional[SparseModelConfig]
transform_config: Optional[TransformConfig]
Configuration for the transformation.
class AdvancedModeTransformConfig: …
chunking_config: Optional[ChunkingConfig]
class PipelineCreate: …
Schema for creating a pipeline.
data_sink: Optional[DataSinkCreate]
Schema for creating a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
data_sink_id: Optional[str]
Data sink ID. When provided instead of data_sink, the data sink will be looked up by ID.
embedding_config: Optional[EmbeddingConfig]
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
embedding_model_config_id: Optional[str]
Embedding model config ID. When provided instead of embedding_config, the embedding model config will be looked up by ID.
llama_parse_parameters: Optional[LlamaParseParameters]
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: Optional[str]
The ID of the ManagedPipeline this playground pipeline is linked to.
preset_retrieval_parameters: Optional[PresetRetrievalParams]
Preset retrieval parameters for the pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
sparse_model_config: Optional[SparseModelConfig]
transform_config: Optional[TransformConfig]
Configuration for the transformation.
class AdvancedModeTransformConfig: …
chunking_config: Optional[ChunkingConfig]
class PresetRetrievalParams: …
Schema for the search params for an retrieval execution that can be preset for a pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
List[Pipeline]
embedding_config: EmbeddingConfig
class EmbeddingConfigManagedOpenAIEmbeddingConfig: …
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
data_sink: Optional[DataSink]
Schema for a data sink.
component: Component
Component that implements the data sink
class CloudPineconeVectorStore: …
Cloud Pinecone Vector Store.
This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.
Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion
class CloudPostgresVectorStore: …
class CloudQdrantVectorStore: …
Cloud Qdrant Vector Store.
This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.
Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client
class CloudMongoDBAtlasVectorSearch: …
Cloud MongoDB Atlas Vector Store.
This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.
Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index
class CloudAstraDBVectorStore: …
Cloud AstraDB Vector Store.
This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.
Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’
embedding_model_config: Optional[EmbeddingModelConfig]
Schema for an embedding model config.
embedding_config: EmbeddingModelConfigEmbeddingConfig
The embedding configuration for the embedding model config.
class AzureOpenAIEmbeddingConfig: …
component: Optional[AzureOpenAIEmbedding]
Configuration for the Azure OpenAI embedding model.
class CohereEmbeddingConfig: …
class GeminiEmbeddingConfig: …
component: Optional[GeminiEmbedding]
Configuration for the Gemini embedding model.
output_dimensionality: Optional[int]
Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.
class HuggingFaceInferenceAPIEmbeddingConfig: …
component: Optional[HuggingFaceInferenceAPIEmbedding]
Configuration for the HuggingFace Inference API embedding model.
headers: Optional[Dict[str, str]]
Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.
class OpenAIEmbeddingConfig: …
component: Optional[OpenAIEmbedding]
Configuration for the OpenAI embedding model.
class VertexAIEmbeddingConfig: …
embedding_model_config_id: Optional[str]
The ID of the EmbeddingModelConfig this pipeline is using.
llama_parse_parameters: Optional[LlamaParseParameters]
Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.
webhook_configurations: Optional[List[WebhookConfiguration]]
Outbound webhook endpoints to notify on job status changes
webhook_events: Optional[List[Literal["extract.pending", "extract.success", "extract.error", 14 more]]]
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: Optional[Dict[str, str]]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
managed_pipeline_id: Optional[str]
The ID of the ManagedPipeline this playground pipeline is linked to.
preset_retrieval_parameters: Optional[PresetRetrievalParams]
Preset retrieval parameters for the pipeline.
alpha: Optional[float]
Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.
dense_similarity_cutoff: Optional[float]
Minimum similarity score wrt query for retrieval
files_top_k: Optional[int]
Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).
search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
sparse_model_config: Optional[SparseModelConfig]
transform_config: Optional[TransformConfig]
Configuration for the transformation.
class AdvancedModeTransformConfig: …
chunking_config: Optional[ChunkingConfig]
class PipelineRetrieveResponse: …
Schema for the result of an retrieval execution.
retrieval_nodes: List[RetrievalNode]
The nodes retrieved by the pipeline for the given query.
Provided for backward compatibility.
excluded_embed_metadata_keys: Optional[List[str]]
Metadata keys that are excluded from text for the embed model.
excluded_llm_metadata_keys: Optional[List[str]]
Metadata keys that are excluded from text for the LLM.
metadata_template: Optional[str]
Template for how metadata is formatted, with {key} and {value} placeholders.
inferred_search_filters: Optional[MetadataFilters]
Metadata filters for vector stores.
filters: List[Filter]
class FilterMetadataFilter: …
Comprehensive metadata filter for vector stores to support more operators.
Value uses Strict types, as int, float and str are compatible types and were all converted to string before.
See: https://docs.pydantic.dev/latest/usage/types/#strict-types
PipelinesSync
Sync Pipeline
Cancel Pipeline Sync
PipelinesData Sources
List Pipeline Data Sources
Add Data Sources To Pipeline
Update Pipeline Data Source
Get Pipeline Data Source Status
Sync Pipeline Data Source
ModelsExpand Collapse
class PipelineDataSource: …
Schema for a data source in a pipeline.
component: Component
Component that implements the data source
class CloudAzStorageBlobDataSource: …
class CloudSharepointDataSource: …
exclude_path_patterns: Optional[List[str]]
List of regex patterns for file paths to exclude. Files whose paths (including filename) match any pattern will be excluded. Example: [‘/temp/’, ‘/backup/’, ‘.git/’, ‘.tmp$’, ’^~’]
class CloudConfluenceDataSource: …
class CloudJiraDataSource: …
class CloudJiraDataSourceV2: …
class CloudBoxDataSource: …
List[PipelineDataSource]
component: Component
Component that implements the data source
class CloudAzStorageBlobDataSource: …
class CloudSharepointDataSource: …
exclude_path_patterns: Optional[List[str]]
List of regex patterns for file paths to exclude. Files whose paths (including filename) match any pattern will be excluded. Example: [‘/temp/’, ‘/backup/’, ‘.git/’, ‘.tmp$’, ’^~’]
class CloudConfluenceDataSource: …
class CloudJiraDataSource: …
class CloudJiraDataSourceV2: …
class CloudBoxDataSource: …
List[PipelineDataSource]
component: Component
Component that implements the data source
class CloudAzStorageBlobDataSource: …
class CloudSharepointDataSource: …
exclude_path_patterns: Optional[List[str]]
List of regex patterns for file paths to exclude. Files whose paths (including filename) match any pattern will be excluded. Example: [‘/temp/’, ‘/backup/’, ‘.git/’, ‘.tmp$’, ’^~’]
class CloudConfluenceDataSource: …
class CloudJiraDataSource: …
class CloudJiraDataSourceV2: …
class CloudBoxDataSource: …
PipelinesImages
List File Page Screenshots
Get File Page Screenshot
Get File Page Figure
List File Pages Figures
PipelinesFiles
Get Pipeline File Status Counts
Get Pipeline File Status
Add Files To Pipeline Api
Update Pipeline File
Delete Pipeline File
List Pipeline Files2
ModelsExpand Collapse
class PipelineFile: …
A file associated with a pipeline.
List[PipelineFile]
PipelinesMetadata
Import Pipeline Metadata
Delete Pipeline Files Metadata
PipelinesDocuments
Create Batch Pipeline Documents
Paginated List Pipeline Documents
Get Pipeline Document
Delete Pipeline Document
Get Pipeline Document Status
Sync Pipeline Document
List Pipeline Document Chunks
Upsert Batch Pipeline Documents
ModelsExpand Collapse
class TextNode: …
Provided for backward compatibility.
excluded_embed_metadata_keys: Optional[List[str]]
Metadata keys that are excluded from text for the embed model.
excluded_llm_metadata_keys: Optional[List[str]]
Metadata keys that are excluded from text for the LLM.
metadata_template: Optional[str]
Template for how metadata is formatted, with {key} and {value} placeholders.
List[CloudDocument]
List[TextNode]
excluded_embed_metadata_keys: Optional[List[str]]
Metadata keys that are excluded from text for the embed model.
excluded_llm_metadata_keys: Optional[List[str]]
Metadata keys that are excluded from text for the LLM.
metadata_template: Optional[str]
Template for how metadata is formatted, with {key} and {value} placeholders.