Skip to content

Pipelines

Search Pipelines
client.Pipelines.List(ctx, query) (*[]Pipeline, error)
GET/api/v1/pipelines
Create Pipeline
client.Pipelines.New(ctx, params) (*Pipeline, error)
POST/api/v1/pipelines
Get Pipeline
client.Pipelines.Get(ctx, pipelineID) (*Pipeline, error)
GET/api/v1/pipelines/{pipeline_id}
Update Existing Pipeline
client.Pipelines.Update(ctx, pipelineID, body) (*Pipeline, error)
PUT/api/v1/pipelines/{pipeline_id}
Delete Pipeline
client.Pipelines.Delete(ctx, pipelineID) error
DELETE/api/v1/pipelines/{pipeline_id}
Get Pipeline Status
client.Pipelines.GetStatus(ctx, pipelineID, query) (*ManagedIngestionStatusResponse, error)
GET/api/v1/pipelines/{pipeline_id}/status
Upsert Pipeline
client.Pipelines.Upsert(ctx, params) (*Pipeline, error)
PUT/api/v1/pipelines
Run Search
client.Pipelines.Get(ctx, pipelineID, params) (*PipelineGetResponse, error)
POST/api/v1/pipelines/{pipeline_id}/retrieve
ModelsExpand Collapse
type AdvancedModeTransformConfig struct{…}
ChunkingConfig AdvancedModeTransformConfigChunkingConfigUnionoptional

Configuration for the chunking.

One of the following:
type AdvancedModeTransformConfigChunkingConfigNoneChunkingConfig struct{…}
Mode stringoptional
type AdvancedModeTransformConfigChunkingConfigCharacterChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
type AdvancedModeTransformConfigChunkingConfigTokenChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
Separator stringoptional
type AdvancedModeTransformConfigChunkingConfigSentenceChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
ParagraphSeparator stringoptional
Separator stringoptional
type AdvancedModeTransformConfigChunkingConfigSemanticChunkingConfig struct{…}
BreakpointPercentileThreshold int64optional
BufferSize int64optional
Mode stringoptional
Mode AdvancedModeTransformConfigModeoptional
SegmentationConfig AdvancedModeTransformConfigSegmentationConfigUnionoptional

Configuration for the segmentation.

One of the following:
type AdvancedModeTransformConfigSegmentationConfigNoneSegmentationConfig struct{…}
Mode stringoptional
type AdvancedModeTransformConfigSegmentationConfigPageSegmentationConfig struct{…}
Mode stringoptional
PageSeparator stringoptional
type AdvancedModeTransformConfigSegmentationConfigElementSegmentationConfig struct{…}
Mode stringoptional
type AutoTransformConfig struct{…}
ChunkOverlap int64optional

Chunk overlap for the transformation.

ChunkSize int64optional

Chunk size for the transformation.

exclusiveMinimum0
Mode AutoTransformConfigModeoptional
type AzureOpenAIEmbedding struct{…}
AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for Azure deployment.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for Azure OpenAI API.

AzureDeployment stringoptional

The Azure deployment to use.

AzureEndpoint stringoptional

The Azure endpoint to use.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
type AzureOpenAIEmbeddingConfig struct{…}
Component AzureOpenAIEmbeddingoptional

Configuration for the Azure OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for Azure deployment.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for Azure OpenAI API.

AzureDeployment stringoptional

The Azure deployment to use.

AzureEndpoint stringoptional

The Azure endpoint to use.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type AzureOpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type BedrockEmbedding struct{…}
AdditionalKwargs map[string, any]optional

Additional kwargs for the bedrock client.

AwsAccessKeyID stringoptional

AWS Access Key ID to use

AwsSecretAccessKey stringoptional

AWS Secret Access Key to use

AwsSessionToken stringoptional

AWS Session Token to use

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

The maximum number of API retries.

exclusiveMinimum0
ModelName stringoptional

The modelId of the Bedrock model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ProfileName stringoptional

The name of aws profile to use. If not given, then the default profile is used.

RegionName stringoptional

AWS region name to use. Uses region configured in AWS CLI if not passed

Timeout float64optional

The timeout for the Bedrock API request in seconds. It will be used for both connect and read timeouts.

type BedrockEmbeddingConfig struct{…}
Component BedrockEmbeddingoptional

Configuration for the Bedrock embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the bedrock client.

AwsAccessKeyID stringoptional

AWS Access Key ID to use

AwsSecretAccessKey stringoptional

AWS Secret Access Key to use

AwsSessionToken stringoptional

AWS Session Token to use

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

The maximum number of API retries.

exclusiveMinimum0
ModelName stringoptional

The modelId of the Bedrock model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ProfileName stringoptional

The name of aws profile to use. If not given, then the default profile is used.

RegionName stringoptional

AWS region name to use. Uses region configured in AWS CLI if not passed

Timeout float64optional

The timeout for the Bedrock API request in seconds. It will be used for both connect and read timeouts.

Type BedrockEmbeddingConfigTypeoptional

Type of the embedding model.

type CohereEmbedding struct{…}
APIKey string

The Cohere API key.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbeddingType stringoptional

Embedding type. If not provided float embedding_type is used when needed.

InputType stringoptional

Model Input type. If not provided, search_document and search_query are used when needed.

ModelName stringoptional

The modelId of the Cohere model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Truncate stringoptional

Truncation type - START/ END/ NONE

type CohereEmbeddingConfig struct{…}
Component CohereEmbeddingoptional

Configuration for the Cohere embedding model.

APIKey string

The Cohere API key.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbeddingType stringoptional

Embedding type. If not provided float embedding_type is used when needed.

InputType stringoptional

Model Input type. If not provided, search_document and search_query are used when needed.

ModelName stringoptional

The modelId of the Cohere model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Truncate stringoptional

Truncation type - START/ END/ NONE

Type CohereEmbeddingConfigTypeoptional

Type of the embedding model.

type DataSinkCreate struct{…}

Schema for creating a data sink.

Component DataSinkCreateComponentUnion

Component that implements the data sink

One of the following:
map[string, any]
type CloudPineconeVectorStore struct{…}

Cloud Pinecone Vector Store.

This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.

Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion

APIKey string

The API key for authenticating with Pinecone

formatpassword
IndexName string
ClassName stringoptional
InsertKwargs map[string, any]optional
Namespace stringoptional
SupportsNestedMetadataFilters booloptional
type CloudPostgresVectorStore struct{…}
Database string
EmbedDim int64
Host string
Password string
Port int64
SchemaName string
TableName string
User string
ClassName stringoptional
HnswSettings PgVectorHnswSettingsoptional

HNSW settings for PGVector.

DistanceMethod PgVectorHnswSettingsDistanceMethodoptional

The distance method to use.

One of the following:
const PgVectorHnswSettingsDistanceMethodL2 PgVectorHnswSettingsDistanceMethod = "l2"
const PgVectorHnswSettingsDistanceMethodIP PgVectorHnswSettingsDistanceMethod = "ip"
const PgVectorHnswSettingsDistanceMethodCosine PgVectorHnswSettingsDistanceMethod = "cosine"
const PgVectorHnswSettingsDistanceMethodL1 PgVectorHnswSettingsDistanceMethod = "l1"
const PgVectorHnswSettingsDistanceMethodHamming PgVectorHnswSettingsDistanceMethod = "hamming"
const PgVectorHnswSettingsDistanceMethodJaccard PgVectorHnswSettingsDistanceMethod = "jaccard"
EfConstruction int64optional

The number of edges to use during the construction phase.

minimum1

The number of edges to use during the search phase.

minimum1
M int64optional

The number of bi-directional links created for each new element.

minimum1
VectorType PgVectorHnswSettingsVectorTypeoptional

The type of vector to use.

One of the following:
const PgVectorHnswSettingsVectorTypeVector PgVectorHnswSettingsVectorType = "vector"
const PgVectorHnswSettingsVectorTypeHalfVec PgVectorHnswSettingsVectorType = "half_vec"
const PgVectorHnswSettingsVectorTypeBit PgVectorHnswSettingsVectorType = "bit"
const PgVectorHnswSettingsVectorTypeSparseVec PgVectorHnswSettingsVectorType = "sparse_vec"
PerformSetup booloptional
SupportsNestedMetadataFilters booloptional
type CloudQdrantVectorStore struct{…}

Cloud Qdrant Vector Store.

This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.

Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client

APIKey string
CollectionName string
URL string
ClassName stringoptional
ClientKwargs map[string, any]optional
MaxRetries int64optional
SupportsNestedMetadataFilters booloptional
type CloudAzureAISearchVectorStore struct{…}

Cloud Azure AI Search Vector Store.

SearchServiceAPIKey string
SearchServiceEndpoint string
ClassName stringoptional
ClientID stringoptional
ClientSecret stringoptional
EmbeddingDimension int64optional
FilterableMetadataFieldKeys map[string, any]optional
IndexName stringoptional
SearchServiceAPIVersion stringoptional
SupportsNestedMetadataFilters booloptional
TenantID stringoptional

Cloud MongoDB Atlas Vector Store.

This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.

Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index

type CloudMilvusVectorStore struct{…}

Cloud Milvus Vector Store.

Uri string
Token stringoptional
ClassName stringoptional
CollectionName stringoptional
EmbeddingDimension int64optional
SupportsNestedMetadataFilters booloptional
type CloudAstraDBVectorStore struct{…}

Cloud AstraDB Vector Store.

This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.

Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’

Token string

The Astra DB Application Token to use

formatpassword
APIEndpoint string

The Astra DB JSON API endpoint for your database

CollectionName string

Collection name to use. If not existing, it will be created

EmbeddingDimension int64

Length of the embedding vectors in use

ClassName stringoptional
Keyspace stringoptional

The keyspace to use. If not provided, ‘default_keyspace’

SupportsNestedMetadataFilters booloptional
Name string

The name of the data sink.

SinkType DataSinkCreateSinkType
One of the following:
const DataSinkCreateSinkTypePinecone DataSinkCreateSinkType = "PINECONE"
const DataSinkCreateSinkTypePostgres DataSinkCreateSinkType = "POSTGRES"
const DataSinkCreateSinkTypeQdrant DataSinkCreateSinkType = "QDRANT"
const DataSinkCreateSinkTypeAzureaiSearch DataSinkCreateSinkType = "AZUREAI_SEARCH"
const DataSinkCreateSinkTypeMongoDBAtlas DataSinkCreateSinkType = "MONGODB_ATLAS"
const DataSinkCreateSinkTypeMilvus DataSinkCreateSinkType = "MILVUS"
const DataSinkCreateSinkTypeAstraDB DataSinkCreateSinkType = "ASTRA_DB"
type GeminiEmbedding struct{…}
APIBase stringoptional

API base to access the model. Defaults to None.

APIKey stringoptional

API key to access the model. Defaults to None.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
ModelName stringoptional

The modelId of the Gemini model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

OutputDimensionality int64optional

Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.

TaskType stringoptional

The task for embedding model.

Title stringoptional

Title is only applicable for retrieval_document tasks, and is used to represent a document title. For other tasks, title is invalid.

Transport stringoptional

Transport to access the model. Defaults to None.

type GeminiEmbeddingConfig struct{…}
Component GeminiEmbeddingoptional

Configuration for the Gemini embedding model.

APIBase stringoptional

API base to access the model. Defaults to None.

APIKey stringoptional

API key to access the model. Defaults to None.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
ModelName stringoptional

The modelId of the Gemini model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

OutputDimensionality int64optional

Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.

TaskType stringoptional

The task for embedding model.

Title stringoptional

Title is only applicable for retrieval_document tasks, and is used to represent a document title. For other tasks, title is invalid.

Transport stringoptional

Transport to access the model. Defaults to None.

Type GeminiEmbeddingConfigTypeoptional

Type of the embedding model.

type HuggingFaceInferenceAPIEmbedding struct{…}
Token HuggingFaceInferenceAPIEmbeddingTokenUnionoptional

Hugging Face token. Will default to the locally saved token. Pass token=False if you don’t want to send your token to the server.

One of the following:
string
bool
ClassName stringoptional
Cookies map[string, string]optional

Additional cookies to send to the server.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
Headers map[string, string]optional

Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.

ModelName stringoptional

Hugging Face model name. If None, the task will be used.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Pooling HuggingFaceInferenceAPIEmbeddingPoolingoptional

Enum of possible pooling choices with pooling behaviors.

One of the following:
const HuggingFaceInferenceAPIEmbeddingPoolingCls HuggingFaceInferenceAPIEmbeddingPooling = "cls"
const HuggingFaceInferenceAPIEmbeddingPoolingMean HuggingFaceInferenceAPIEmbeddingPooling = "mean"
const HuggingFaceInferenceAPIEmbeddingPoolingLast HuggingFaceInferenceAPIEmbeddingPooling = "last"
QueryInstruction stringoptional

Instruction to prepend during query embedding.

Task stringoptional

Optional task to pick Hugging Face’s recommended model, used when model_name is left as default of None.

TextInstruction stringoptional

Instruction to prepend during text embedding.

Timeout float64optional

The maximum number of seconds to wait for a response from the server. Loading a new model in Inference API can take up to several minutes. Defaults to None, meaning it will loop until the server is available.

type HuggingFaceInferenceAPIEmbeddingConfig struct{…}

Configuration for the HuggingFace Inference API embedding model.

Token HuggingFaceInferenceAPIEmbeddingTokenUnionoptional

Hugging Face token. Will default to the locally saved token. Pass token=False if you don’t want to send your token to the server.

One of the following:
string
bool
ClassName stringoptional
Cookies map[string, string]optional

Additional cookies to send to the server.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
Headers map[string, string]optional

Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.

ModelName stringoptional

Hugging Face model name. If None, the task will be used.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Pooling HuggingFaceInferenceAPIEmbeddingPoolingoptional

Enum of possible pooling choices with pooling behaviors.

One of the following:
const HuggingFaceInferenceAPIEmbeddingPoolingCls HuggingFaceInferenceAPIEmbeddingPooling = "cls"
const HuggingFaceInferenceAPIEmbeddingPoolingMean HuggingFaceInferenceAPIEmbeddingPooling = "mean"
const HuggingFaceInferenceAPIEmbeddingPoolingLast HuggingFaceInferenceAPIEmbeddingPooling = "last"
QueryInstruction stringoptional

Instruction to prepend during query embedding.

Task stringoptional

Optional task to pick Hugging Face’s recommended model, used when model_name is left as default of None.

TextInstruction stringoptional

Instruction to prepend during text embedding.

Timeout float64optional

The maximum number of seconds to wait for a response from the server. Loading a new model in Inference API can take up to several minutes. Defaults to None, meaning it will loop until the server is available.

Type HuggingFaceInferenceAPIEmbeddingConfigTypeoptional

Type of the embedding model.

type LlamaParseParametersResp struct{…}
AdaptiveLongTable booloptional
AggressiveTableExtraction booloptional
AutoMode booloptional
AutoModeConfigurationJson stringoptional
AutoModeTriggerOnImageInPage booloptional
AutoModeTriggerOnRegexpInPage stringoptional
AutoModeTriggerOnTableInPage booloptional
AutoModeTriggerOnTextInPage stringoptional
AzureOpenAIAPIVersion stringoptional
AzureOpenAIDeploymentName stringoptional
AzureOpenAIEndpoint stringoptional
AzureOpenAIKey stringoptional
BboxBottom float64optional
BboxLeft float64optional
BboxRight float64optional
BboxTop float64optional
BoundingBox stringoptional
CompactMarkdownTable booloptional
ComplementalFormattingInstruction stringoptional
ContentGuidelineInstruction stringoptional
ContinuousMode booloptional
DisableImageExtraction booloptional
DisableOcr booloptional
DisableReconstruction booloptional
DoNotCache booloptional
DoNotUnrollColumns booloptional
EnableCostOptimizer booloptional
ExtractCharts booloptional
ExtractLayout booloptional
ExtractPrintedPageNumber booloptional
FastMode booloptional
FormattingInstruction stringoptional
Gpt4oAPIKey stringoptional
Gpt4oMode booloptional
GuessXlsxSheetName booloptional
HideFooters booloptional
HideHeaders booloptional
HighResOcr booloptional
HTMLMakeAllElementsVisible booloptional
HTMLRemoveFixedElements booloptional
HTMLRemoveNavigationElements booloptional
HTTPProxy stringoptional
IgnoreDocumentElementsForLayoutDetection booloptional
ImagesToSave []stringoptional
One of the following:
const LlamaParseParametersImagesToSaveScreenshot LlamaParseParametersImagesToSave = "screenshot"
const LlamaParseParametersImagesToSaveEmbedded LlamaParseParametersImagesToSave = "embedded"
const LlamaParseParametersImagesToSaveLayout LlamaParseParametersImagesToSave = "layout"
InlineImagesInMarkdown booloptional
InputS3Path stringoptional
InputS3Region stringoptional
InputURL stringoptional
InternalIsScreenshotJob booloptional
InvalidateCache booloptional
IsFormattingInstruction booloptional
JobTimeoutExtraTimePerPageInSeconds float64optional
JobTimeoutInSeconds float64optional
KeepPageSeparatorWhenMergingTables booloptional
Languages []ParsingLanguagesoptional
One of the following:
const ParsingLanguagesAf ParsingLanguages = "af"
const ParsingLanguagesAz ParsingLanguages = "az"
const ParsingLanguagesBs ParsingLanguages = "bs"
const ParsingLanguagesCs ParsingLanguages = "cs"
const ParsingLanguagesCy ParsingLanguages = "cy"
const ParsingLanguagesDa ParsingLanguages = "da"
const ParsingLanguagesDe ParsingLanguages = "de"
const ParsingLanguagesEn ParsingLanguages = "en"
const ParsingLanguagesEs ParsingLanguages = "es"
const ParsingLanguagesEt ParsingLanguages = "et"
const ParsingLanguagesFr ParsingLanguages = "fr"
const ParsingLanguagesGa ParsingLanguages = "ga"
const ParsingLanguagesHr ParsingLanguages = "hr"
const ParsingLanguagesHu ParsingLanguages = "hu"
const ParsingLanguagesID ParsingLanguages = "id"
const ParsingLanguagesIs ParsingLanguages = "is"
const ParsingLanguagesIt ParsingLanguages = "it"
const ParsingLanguagesKu ParsingLanguages = "ku"
const ParsingLanguagesLa ParsingLanguages = "la"
const ParsingLanguagesLt ParsingLanguages = "lt"
const ParsingLanguagesLv ParsingLanguages = "lv"
const ParsingLanguagesMi ParsingLanguages = "mi"
const ParsingLanguagesMs ParsingLanguages = "ms"
const ParsingLanguagesMt ParsingLanguages = "mt"
const ParsingLanguagesNl ParsingLanguages = "nl"
const ParsingLanguagesNo ParsingLanguages = "no"
const ParsingLanguagesOc ParsingLanguages = "oc"
const ParsingLanguagesPi ParsingLanguages = "pi"
const ParsingLanguagesPl ParsingLanguages = "pl"
const ParsingLanguagesPt ParsingLanguages = "pt"
const ParsingLanguagesRo ParsingLanguages = "ro"
const ParsingLanguagesRsLatin ParsingLanguages = "rs_latin"
const ParsingLanguagesSk ParsingLanguages = "sk"
const ParsingLanguagesSl ParsingLanguages = "sl"
const ParsingLanguagesSq ParsingLanguages = "sq"
const ParsingLanguagesSv ParsingLanguages = "sv"
const ParsingLanguagesSw ParsingLanguages = "sw"
const ParsingLanguagesTl ParsingLanguages = "tl"
const ParsingLanguagesTr ParsingLanguages = "tr"
const ParsingLanguagesUz ParsingLanguages = "uz"
const ParsingLanguagesVi ParsingLanguages = "vi"
const ParsingLanguagesAr ParsingLanguages = "ar"
const ParsingLanguagesFa ParsingLanguages = "fa"
const ParsingLanguagesUg ParsingLanguages = "ug"
const ParsingLanguagesUr ParsingLanguages = "ur"
const ParsingLanguagesBn ParsingLanguages = "bn"
const ParsingLanguagesAs ParsingLanguages = "as"
const ParsingLanguagesMni ParsingLanguages = "mni"
const ParsingLanguagesRu ParsingLanguages = "ru"
const ParsingLanguagesRsCyrillic ParsingLanguages = "rs_cyrillic"
const ParsingLanguagesBe ParsingLanguages = "be"
const ParsingLanguagesBg ParsingLanguages = "bg"
const ParsingLanguagesUk ParsingLanguages = "uk"
const ParsingLanguagesMn ParsingLanguages = "mn"
const ParsingLanguagesAbq ParsingLanguages = "abq"
const ParsingLanguagesAdy ParsingLanguages = "ady"
const ParsingLanguagesKbd ParsingLanguages = "kbd"
const ParsingLanguagesAva ParsingLanguages = "ava"
const ParsingLanguagesDar ParsingLanguages = "dar"
const ParsingLanguagesInh ParsingLanguages = "inh"
const ParsingLanguagesChe ParsingLanguages = "che"
const ParsingLanguagesLbe ParsingLanguages = "lbe"
const ParsingLanguagesLez ParsingLanguages = "lez"
const ParsingLanguagesTab ParsingLanguages = "tab"
const ParsingLanguagesTjk ParsingLanguages = "tjk"
const ParsingLanguagesHi ParsingLanguages = "hi"
const ParsingLanguagesMr ParsingLanguages = "mr"
const ParsingLanguagesNe ParsingLanguages = "ne"
const ParsingLanguagesBh ParsingLanguages = "bh"
const ParsingLanguagesMai ParsingLanguages = "mai"
const ParsingLanguagesAng ParsingLanguages = "ang"
const ParsingLanguagesBho ParsingLanguages = "bho"
const ParsingLanguagesMah ParsingLanguages = "mah"
const ParsingLanguagesSck ParsingLanguages = "sck"
const ParsingLanguagesNew ParsingLanguages = "new"
const ParsingLanguagesGom ParsingLanguages = "gom"
const ParsingLanguagesSa ParsingLanguages = "sa"
const ParsingLanguagesBgc ParsingLanguages = "bgc"
const ParsingLanguagesTh ParsingLanguages = "th"
const ParsingLanguagesChSim ParsingLanguages = "ch_sim"
const ParsingLanguagesChTra ParsingLanguages = "ch_tra"
const ParsingLanguagesJa ParsingLanguages = "ja"
const ParsingLanguagesKo ParsingLanguages = "ko"
const ParsingLanguagesTa ParsingLanguages = "ta"
const ParsingLanguagesTe ParsingLanguages = "te"
const ParsingLanguagesKn ParsingLanguages = "kn"
LayoutAware booloptional
LineLevelBoundingBox booloptional
MarkdownTableMultilineHeaderSeparator stringoptional
MaxPages int64optional
MaxPagesEnforced int64optional
MergeTablesAcrossPagesInMarkdown booloptional
Model stringoptional
OutlinedTableExtraction booloptional
OutputPdfOfDocument booloptional
OutputS3PathPrefix stringoptional
OutputS3Region stringoptional
OutputTablesAsHTML booloptional
PageErrorTolerance float64optional
PageHeaderPrefix stringoptional
PageHeaderSuffix stringoptional
PagePrefix stringoptional
PageSeparator stringoptional
PageSuffix stringoptional
ParseMode ParsingModeoptional

Enum for representing the mode of parsing to be used.

One of the following:
const ParsingModeParsePageWithoutLlm ParsingMode = "parse_page_without_llm"
const ParsingModeParsePageWithLlm ParsingMode = "parse_page_with_llm"
const ParsingModeParsePageWithLvm ParsingMode = "parse_page_with_lvm"
const ParsingModeParsePageWithAgent ParsingMode = "parse_page_with_agent"
const ParsingModeParsePageWithLayoutAgent ParsingMode = "parse_page_with_layout_agent"
const ParsingModeParseDocumentWithLlm ParsingMode = "parse_document_with_llm"
const ParsingModeParseDocumentWithLvm ParsingMode = "parse_document_with_lvm"
const ParsingModeParseDocumentWithAgent ParsingMode = "parse_document_with_agent"
ParsingInstruction stringoptional
PreciseBoundingBox booloptional
PremiumMode booloptional
PresentationOutOfBoundsContent booloptional
PresentationSkipEmbeddedData booloptional
PreserveLayoutAlignmentAcrossPages booloptional
PreserveVerySmallText booloptional
Preset stringoptional
Priority LlamaParseParametersPriorityoptional

The priority for the request. This field may be ignored or overwritten depending on the organization tier.

One of the following:
const LlamaParseParametersPriorityLow LlamaParseParametersPriority = "low"
const LlamaParseParametersPriorityMedium LlamaParseParametersPriority = "medium"
const LlamaParseParametersPriorityHigh LlamaParseParametersPriority = "high"
const LlamaParseParametersPriorityCritical LlamaParseParametersPriority = "critical"
ProjectID stringoptional
RemoveHiddenText booloptional
ReplaceFailedPageMode FailPageModeoptional

Enum for representing the different available page error handling modes.

One of the following:
const FailPageModeRawText FailPageMode = "raw_text"
const FailPageModeBlankPage FailPageMode = "blank_page"
const FailPageModeErrorMessage FailPageMode = "error_message"
ReplaceFailedPageWithErrorMessagePrefix stringoptional
ReplaceFailedPageWithErrorMessageSuffix stringoptional
SaveImages booloptional
SkipDiagonalText booloptional
SpecializedChartParsingAgentic booloptional
SpecializedChartParsingEfficient booloptional
SpecializedChartParsingPlus booloptional
SpecializedImageParsing booloptional
SpreadsheetExtractSubTables booloptional
SpreadsheetForceFormulaComputation booloptional
SpreadsheetIncludeHiddenSheets booloptional
StrictModeBuggyFont booloptional
StrictModeImageExtraction booloptional
StrictModeImageOcr booloptional
StrictModeReconstruction booloptional
StructuredOutput booloptional
StructuredOutputJsonSchema stringoptional
StructuredOutputJsonSchemaName stringoptional
SystemPrompt stringoptional
SystemPromptAppend stringoptional
TakeScreenshot booloptional
TargetPages stringoptional
Tier stringoptional
UseVendorMultimodalModel booloptional
UserPrompt stringoptional
VendorMultimodalAPIKey stringoptional
VendorMultimodalModelName stringoptional
Version stringoptional
WebhookConfigurations []LlamaParseParametersWebhookConfigurationRespoptional

Outbound webhook endpoints to notify on job status changes

WebhookEvents []stringoptional

Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.

One of the following:
const LlamaParseParametersWebhookConfigurationWebhookEventExtractPending LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.success"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractError LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.error"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractPartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventParsePending LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventParseRunning LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.running"
const LlamaParseParametersWebhookConfigurationWebhookEventParseSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.success"
const LlamaParseParametersWebhookConfigurationWebhookEventParseError LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.error"
const LlamaParseParametersWebhookConfigurationWebhookEventParsePartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventParseCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyPending LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifySuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.success"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyError LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.error"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyPartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventUnmappedEvent LlamaParseParametersWebhookConfigurationWebhookEvent = "unmapped_event"
WebhookHeaders map[string, string]optional

Custom HTTP headers sent with each webhook request (e.g. auth tokens)

WebhookOutputFormat stringoptional

Response format sent to the webhook: ‘string’ (default) or ‘json’

WebhookURL stringoptional

URL to receive webhook POST notifications

WebhookURL stringoptional
type LlmParametersResp struct{…}
ClassName stringoptional
ModelName LlmParametersModelNameoptional

The name of the model to use for LLM completions.

One of the following:
const LlmParametersModelNameGpt4O LlmParametersModelName = "GPT_4O"
const LlmParametersModelNameGpt4OMini LlmParametersModelName = "GPT_4O_MINI"
const LlmParametersModelNameGpt4_1 LlmParametersModelName = "GPT_4_1"
const LlmParametersModelNameGpt4_1Nano LlmParametersModelName = "GPT_4_1_NANO"
const LlmParametersModelNameGpt4_1Mini LlmParametersModelName = "GPT_4_1_MINI"
const LlmParametersModelNameAzureOpenAIGpt4O LlmParametersModelName = "AZURE_OPENAI_GPT_4O"
const LlmParametersModelNameAzureOpenAIGpt4OMini LlmParametersModelName = "AZURE_OPENAI_GPT_4O_MINI"
const LlmParametersModelNameAzureOpenAIGpt4_1 LlmParametersModelName = "AZURE_OPENAI_GPT_4_1"
const LlmParametersModelNameAzureOpenAIGpt4_1Mini LlmParametersModelName = "AZURE_OPENAI_GPT_4_1_MINI"
const LlmParametersModelNameAzureOpenAIGpt4_1Nano LlmParametersModelName = "AZURE_OPENAI_GPT_4_1_NANO"
const LlmParametersModelNameClaude4_5Sonnet LlmParametersModelName = "CLAUDE_4_5_SONNET"
const LlmParametersModelNameBedrockClaude3_5SonnetV1 LlmParametersModelName = "BEDROCK_CLAUDE_3_5_SONNET_V1"
const LlmParametersModelNameBedrockClaude3_5SonnetV2 LlmParametersModelName = "BEDROCK_CLAUDE_3_5_SONNET_V2"
SystemPrompt stringoptional

The system prompt to use for the completion.

maxLength3000
Temperature float64optional

The temperature value for the model.

UseChainOfThoughtReasoning booloptional

Whether to use chain of thought reasoning.

UseCitation booloptional

Whether to show citations in the response.

type ManagedIngestionStatusResponse struct{…}
Status ManagedIngestionStatusResponseStatus

Status of the ingestion.

One of the following:
const ManagedIngestionStatusResponseStatusNotStarted ManagedIngestionStatusResponseStatus = "NOT_STARTED"
const ManagedIngestionStatusResponseStatusInProgress ManagedIngestionStatusResponseStatus = "IN_PROGRESS"
const ManagedIngestionStatusResponseStatusSuccess ManagedIngestionStatusResponseStatus = "SUCCESS"
const ManagedIngestionStatusResponseStatusError ManagedIngestionStatusResponseStatus = "ERROR"
const ManagedIngestionStatusResponseStatusPartialSuccess ManagedIngestionStatusResponseStatus = "PARTIAL_SUCCESS"
const ManagedIngestionStatusResponseStatusCancelled ManagedIngestionStatusResponseStatus = "CANCELLED"
DeploymentDate Timeoptional

Date of the deployment.

formatdate-time
EffectiveAt Timeoptional

When the status is effective

formatdate-time
Error []ManagedIngestionStatusResponseErroroptional

List of errors that occurred during ingestion.

JobID string

ID of the job that failed.

formatuuid
Message string

List of errors that occurred during ingestion.

Step string

Name of the job that failed.

One of the following:
const ManagedIngestionStatusResponseErrorStepManagedIngestion ManagedIngestionStatusResponseErrorStep = "MANAGED_INGESTION"
const ManagedIngestionStatusResponseErrorStepDataSource ManagedIngestionStatusResponseErrorStep = "DATA_SOURCE"
const ManagedIngestionStatusResponseErrorStepFileUpdater ManagedIngestionStatusResponseErrorStep = "FILE_UPDATER"
const ManagedIngestionStatusResponseErrorStepParse ManagedIngestionStatusResponseErrorStep = "PARSE"
const ManagedIngestionStatusResponseErrorStepTransform ManagedIngestionStatusResponseErrorStep = "TRANSFORM"
const ManagedIngestionStatusResponseErrorStepIngestion ManagedIngestionStatusResponseErrorStep = "INGESTION"
const ManagedIngestionStatusResponseErrorStepMetadataUpdate ManagedIngestionStatusResponseErrorStep = "METADATA_UPDATE"
JobID stringoptional

ID of the latest job.

formatuuid
type MessageRole string

Message role.

One of the following:
const MessageRoleSystem MessageRole = "system"
const MessageRoleDeveloper MessageRole = "developer"
const MessageRoleUser MessageRole = "user"
const MessageRoleAssistant MessageRole = "assistant"
const MessageRoleFunction MessageRole = "function"
const MessageRoleTool MessageRole = "tool"
const MessageRoleChatbot MessageRole = "chatbot"
const MessageRoleModel MessageRole = "model"
type MetadataFilters struct{…}

Metadata filters for vector stores.

Filters []MetadataFiltersFilterUnion
One of the following:
type MetadataFiltersFilterMetadataFilter struct{…}

Comprehensive metadata filter for vector stores to support more operators.

Value uses Strict types, as int, float and str are compatible types and were all converted to string before.

See: https://docs.pydantic.dev/latest/usage/types/#strict-types

Key string
Value MetadataFiltersFilterMetadataFilterValueUnion
One of the following:
float64
string
type MetadataFiltersFilterMetadataFilterValueArray []string
type MetadataFiltersFilterMetadataFilterValueArray []float64
type MetadataFiltersFilterMetadataFilterValueArray []int64
Operator stringoptional

Vector store filter operator.

One of the following:
const MetadataFiltersFilterMetadataFilterOperatorEquals MetadataFiltersFilterMetadataFilterOperator = "=="
const MetadataFiltersFilterMetadataFilterOperatorGreater MetadataFiltersFilterMetadataFilterOperator = ">"
const MetadataFiltersFilterMetadataFilterOperatorLess MetadataFiltersFilterMetadataFilterOperator = "<"
const MetadataFiltersFilterMetadataFilterOperatorNotEquals MetadataFiltersFilterMetadataFilterOperator = "!="
const MetadataFiltersFilterMetadataFilterOperatorGreaterOrEquals MetadataFiltersFilterMetadataFilterOperator = ">="
const MetadataFiltersFilterMetadataFilterOperatorLessOrEquals MetadataFiltersFilterMetadataFilterOperator = "<="
const MetadataFiltersFilterMetadataFilterOperatorIn MetadataFiltersFilterMetadataFilterOperator = "in"
const MetadataFiltersFilterMetadataFilterOperatorNin MetadataFiltersFilterMetadataFilterOperator = "nin"
const MetadataFiltersFilterMetadataFilterOperatorAny MetadataFiltersFilterMetadataFilterOperator = "any"
const MetadataFiltersFilterMetadataFilterOperatorAll MetadataFiltersFilterMetadataFilterOperator = "all"
const MetadataFiltersFilterMetadataFilterOperatorTextMatch MetadataFiltersFilterMetadataFilterOperator = "text_match"
const MetadataFiltersFilterMetadataFilterOperatorTextMatchInsensitive MetadataFiltersFilterMetadataFilterOperator = "text_match_insensitive"
const MetadataFiltersFilterMetadataFilterOperatorContains MetadataFiltersFilterMetadataFilterOperator = "contains"
const MetadataFiltersFilterMetadataFilterOperatorIsEmpty MetadataFiltersFilterMetadataFilterOperator = "is_empty"
type MetadataFilters MetadataFilters

Metadata filters for vector stores.

Condition MetadataFiltersConditionoptional

Vector store filter conditions to combine different filters.

One of the following:
const MetadataFiltersConditionAnd MetadataFiltersCondition = "and"
const MetadataFiltersConditionOr MetadataFiltersCondition = "or"
const MetadataFiltersConditionNot MetadataFiltersCondition = "not"
type OpenAIEmbedding struct{…}
AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for OpenAI API.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for OpenAI API.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
type OpenAIEmbeddingConfig struct{…}
Component OpenAIEmbeddingoptional

Configuration for the OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for OpenAI API.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for OpenAI API.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type OpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type PageFigureNodeWithScore struct{…}

Page figure metadata with score

Node PageFigureNodeWithScoreNode
Confidence float64

The confidence of the figure

maximum1
minimum0
FigureName string

The name of the figure

FigureSize int64

The size of the figure in bytes

minimum0
FileID string

The ID of the file that the figure was taken from

formatuuid
PageIndex int64

The index of the page for which the figure is taken (0-indexed)

minimum0
IsLikelyNoise booloptional

Whether the figure is likely to be noise

Metadata map[string, any]optional

Metadata for the figure

Score float64

The score of the figure node

ClassName stringoptional
type PageScreenshotNodeWithScore struct{…}

Page screenshot metadata with score

Node PageScreenshotNodeWithScoreNode
FileID string

The ID of the file that the page screenshot was taken from

formatuuid
ImageSize int64

The size of the image in bytes

minimum0
PageIndex int64

The index of the page for which the screenshot is taken (0-indexed)

minimum0
Metadata map[string, any]optional

Metadata for the screenshot

Score float64

The score of the screenshot node

ClassName stringoptional
type Pipeline struct{…}

Schema for a pipeline.

ID string

Unique identifier

formatuuid
EmbeddingConfig PipelineEmbeddingConfigUnion
One of the following:
type PipelineEmbeddingConfigManagedOpenAIEmbedding struct{…}
Component PipelineEmbeddingConfigManagedOpenAIEmbeddingComponentoptional

Configuration for the Managed OpenAI embedding model.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Type stringoptional

Type of the embedding model.

type AzureOpenAIEmbeddingConfig struct{…}
Component AzureOpenAIEmbeddingoptional

Configuration for the Azure OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for Azure deployment.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for Azure OpenAI API.

AzureDeployment stringoptional

The Azure deployment to use.

AzureEndpoint stringoptional

The Azure endpoint to use.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type AzureOpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type CohereEmbeddingConfig struct{…}
Component CohereEmbeddingoptional

Configuration for the Cohere embedding model.

APIKey string

The Cohere API key.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbeddingType stringoptional

Embedding type. If not provided float embedding_type is used when needed.

InputType stringoptional

Model Input type. If not provided, search_document and search_query are used when needed.

ModelName stringoptional

The modelId of the Cohere model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Truncate stringoptional

Truncation type - START/ END/ NONE

Type CohereEmbeddingConfigTypeoptional

Type of the embedding model.

type GeminiEmbeddingConfig struct{…}
Component GeminiEmbeddingoptional

Configuration for the Gemini embedding model.

APIBase stringoptional

API base to access the model. Defaults to None.

APIKey stringoptional

API key to access the model. Defaults to None.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
ModelName stringoptional

The modelId of the Gemini model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

OutputDimensionality int64optional

Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.

TaskType stringoptional

The task for embedding model.

Title stringoptional

Title is only applicable for retrieval_document tasks, and is used to represent a document title. For other tasks, title is invalid.

Transport stringoptional

Transport to access the model. Defaults to None.

Type GeminiEmbeddingConfigTypeoptional

Type of the embedding model.

type HuggingFaceInferenceAPIEmbeddingConfig struct{…}

Configuration for the HuggingFace Inference API embedding model.

Token HuggingFaceInferenceAPIEmbeddingTokenUnionoptional

Hugging Face token. Will default to the locally saved token. Pass token=False if you don’t want to send your token to the server.

One of the following:
string
bool
ClassName stringoptional
Cookies map[string, string]optional

Additional cookies to send to the server.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
Headers map[string, string]optional

Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.

ModelName stringoptional

Hugging Face model name. If None, the task will be used.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Pooling HuggingFaceInferenceAPIEmbeddingPoolingoptional

Enum of possible pooling choices with pooling behaviors.

One of the following:
const HuggingFaceInferenceAPIEmbeddingPoolingCls HuggingFaceInferenceAPIEmbeddingPooling = "cls"
const HuggingFaceInferenceAPIEmbeddingPoolingMean HuggingFaceInferenceAPIEmbeddingPooling = "mean"
const HuggingFaceInferenceAPIEmbeddingPoolingLast HuggingFaceInferenceAPIEmbeddingPooling = "last"
QueryInstruction stringoptional

Instruction to prepend during query embedding.

Task stringoptional

Optional task to pick Hugging Face’s recommended model, used when model_name is left as default of None.

TextInstruction stringoptional

Instruction to prepend during text embedding.

Timeout float64optional

The maximum number of seconds to wait for a response from the server. Loading a new model in Inference API can take up to several minutes. Defaults to None, meaning it will loop until the server is available.

Type HuggingFaceInferenceAPIEmbeddingConfigTypeoptional

Type of the embedding model.

type OpenAIEmbeddingConfig struct{…}
Component OpenAIEmbeddingoptional

Configuration for the OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for OpenAI API.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for OpenAI API.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type OpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type VertexAIEmbeddingConfig struct{…}
Component VertexTextEmbeddingoptional

Configuration for the VertexAI embedding model.

ClientEmail string

The client email for the VertexAI credentials.

Location string

The default location to use when making API calls.

PrivateKey string

The private key for the VertexAI credentials.

PrivateKeyID string

The private key ID for the VertexAI credentials.

Project string

The default GCP project to use when making Vertex API calls.

TokenUri string

The token URI for the VertexAI credentials.

AdditionalKwargs map[string, any]optional

Additional kwargs for the Vertex.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbedMode VertexTextEmbeddingEmbedModeoptional

The embedding mode to use.

One of the following:
const VertexTextEmbeddingEmbedModeDefault VertexTextEmbeddingEmbedMode = "default"
const VertexTextEmbeddingEmbedModeClassification VertexTextEmbeddingEmbedMode = "classification"
const VertexTextEmbeddingEmbedModeClustering VertexTextEmbeddingEmbedMode = "clustering"
const VertexTextEmbeddingEmbedModeSimilarity VertexTextEmbeddingEmbedMode = "similarity"
const VertexTextEmbeddingEmbedModeRetrieval VertexTextEmbeddingEmbedMode = "retrieval"
ModelName stringoptional

The modelId of the VertexAI model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Type VertexAIEmbeddingConfigTypeoptional

Type of the embedding model.

type BedrockEmbeddingConfig struct{…}
Component BedrockEmbeddingoptional

Configuration for the Bedrock embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the bedrock client.

AwsAccessKeyID stringoptional

AWS Access Key ID to use

AwsSecretAccessKey stringoptional

AWS Secret Access Key to use

AwsSessionToken stringoptional

AWS Session Token to use

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

The maximum number of API retries.

exclusiveMinimum0
ModelName stringoptional

The modelId of the Bedrock model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ProfileName stringoptional

The name of aws profile to use. If not given, then the default profile is used.

RegionName stringoptional

AWS region name to use. Uses region configured in AWS CLI if not passed

Timeout float64optional

The timeout for the Bedrock API request in seconds. It will be used for both connect and read timeouts.

Type BedrockEmbeddingConfigTypeoptional

Type of the embedding model.

Name string
ProjectID string
ConfigHash PipelineConfigHashoptional

Hashes for the configuration of a pipeline.

EmbeddingConfigHash stringoptional

Hash of the embedding config.

ParsingConfigHash stringoptional

Hash of the llama parse parameters.

TransformConfigHash stringoptional

Hash of the transform config.

CreatedAt Timeoptional

Creation datetime

formatdate-time
DataSink DataSinkoptional

Schema for a data sink.

ID string

Unique identifier

formatuuid
Component DataSinkComponentUnion

Component that implements the data sink

One of the following:
type DataSinkComponentMap map[string, any]
type CloudPineconeVectorStore struct{…}

Cloud Pinecone Vector Store.

This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.

Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion

APIKey string

The API key for authenticating with Pinecone

formatpassword
IndexName string
ClassName stringoptional
InsertKwargs map[string, any]optional
Namespace stringoptional
SupportsNestedMetadataFilters booloptional
type CloudPostgresVectorStore struct{…}
Database string
EmbedDim int64
Host string
Password string
Port int64
SchemaName string
TableName string
User string
ClassName stringoptional
HnswSettings PgVectorHnswSettingsoptional

HNSW settings for PGVector.

DistanceMethod PgVectorHnswSettingsDistanceMethodoptional

The distance method to use.

One of the following:
const PgVectorHnswSettingsDistanceMethodL2 PgVectorHnswSettingsDistanceMethod = "l2"
const PgVectorHnswSettingsDistanceMethodIP PgVectorHnswSettingsDistanceMethod = "ip"
const PgVectorHnswSettingsDistanceMethodCosine PgVectorHnswSettingsDistanceMethod = "cosine"
const PgVectorHnswSettingsDistanceMethodL1 PgVectorHnswSettingsDistanceMethod = "l1"
const PgVectorHnswSettingsDistanceMethodHamming PgVectorHnswSettingsDistanceMethod = "hamming"
const PgVectorHnswSettingsDistanceMethodJaccard PgVectorHnswSettingsDistanceMethod = "jaccard"
EfConstruction int64optional

The number of edges to use during the construction phase.

minimum1

The number of edges to use during the search phase.

minimum1
M int64optional

The number of bi-directional links created for each new element.

minimum1
VectorType PgVectorHnswSettingsVectorTypeoptional

The type of vector to use.

One of the following:
const PgVectorHnswSettingsVectorTypeVector PgVectorHnswSettingsVectorType = "vector"
const PgVectorHnswSettingsVectorTypeHalfVec PgVectorHnswSettingsVectorType = "half_vec"
const PgVectorHnswSettingsVectorTypeBit PgVectorHnswSettingsVectorType = "bit"
const PgVectorHnswSettingsVectorTypeSparseVec PgVectorHnswSettingsVectorType = "sparse_vec"
PerformSetup booloptional
SupportsNestedMetadataFilters booloptional
type CloudQdrantVectorStore struct{…}

Cloud Qdrant Vector Store.

This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.

Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client

APIKey string
CollectionName string
URL string
ClassName stringoptional
ClientKwargs map[string, any]optional
MaxRetries int64optional
SupportsNestedMetadataFilters booloptional
type CloudAzureAISearchVectorStore struct{…}

Cloud Azure AI Search Vector Store.

SearchServiceAPIKey string
SearchServiceEndpoint string
ClassName stringoptional
ClientID stringoptional
ClientSecret stringoptional
EmbeddingDimension int64optional
FilterableMetadataFieldKeys map[string, any]optional
IndexName stringoptional
SearchServiceAPIVersion stringoptional
SupportsNestedMetadataFilters booloptional
TenantID stringoptional

Cloud MongoDB Atlas Vector Store.

This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.

Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index

type CloudMilvusVectorStore struct{…}

Cloud Milvus Vector Store.

Uri string
Token stringoptional
ClassName stringoptional
CollectionName stringoptional
EmbeddingDimension int64optional
SupportsNestedMetadataFilters booloptional
type CloudAstraDBVectorStore struct{…}

Cloud AstraDB Vector Store.

This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.

Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’

Token string

The Astra DB Application Token to use

formatpassword
APIEndpoint string

The Astra DB JSON API endpoint for your database

CollectionName string

Collection name to use. If not existing, it will be created

EmbeddingDimension int64

Length of the embedding vectors in use

ClassName stringoptional
Keyspace stringoptional

The keyspace to use. If not provided, ‘default_keyspace’

SupportsNestedMetadataFilters booloptional
Name string

The name of the data sink.

ProjectID string
SinkType DataSinkSinkType
One of the following:
const DataSinkSinkTypePinecone DataSinkSinkType = "PINECONE"
const DataSinkSinkTypePostgres DataSinkSinkType = "POSTGRES"
const DataSinkSinkTypeQdrant DataSinkSinkType = "QDRANT"
const DataSinkSinkTypeAzureaiSearch DataSinkSinkType = "AZUREAI_SEARCH"
const DataSinkSinkTypeMongoDBAtlas DataSinkSinkType = "MONGODB_ATLAS"
const DataSinkSinkTypeMilvus DataSinkSinkType = "MILVUS"
const DataSinkSinkTypeAstraDB DataSinkSinkType = "ASTRA_DB"
CreatedAt Timeoptional

Creation datetime

formatdate-time
UpdatedAt Timeoptional

Update datetime

formatdate-time
EmbeddingModelConfig PipelineEmbeddingModelConfigoptional

Schema for an embedding model config.

ID string

Unique identifier

formatuuid
EmbeddingConfig PipelineEmbeddingModelConfigEmbeddingConfigUnion

The embedding configuration for the embedding model config.

One of the following:
type AzureOpenAIEmbeddingConfig struct{…}
Component AzureOpenAIEmbeddingoptional

Configuration for the Azure OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for Azure deployment.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for Azure OpenAI API.

AzureDeployment stringoptional

The Azure deployment to use.

AzureEndpoint stringoptional

The Azure endpoint to use.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type AzureOpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type CohereEmbeddingConfig struct{…}
Component CohereEmbeddingoptional

Configuration for the Cohere embedding model.

APIKey string

The Cohere API key.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbeddingType stringoptional

Embedding type. If not provided float embedding_type is used when needed.

InputType stringoptional

Model Input type. If not provided, search_document and search_query are used when needed.

ModelName stringoptional

The modelId of the Cohere model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Truncate stringoptional

Truncation type - START/ END/ NONE

Type CohereEmbeddingConfigTypeoptional

Type of the embedding model.

type GeminiEmbeddingConfig struct{…}
Component GeminiEmbeddingoptional

Configuration for the Gemini embedding model.

APIBase stringoptional

API base to access the model. Defaults to None.

APIKey stringoptional

API key to access the model. Defaults to None.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
ModelName stringoptional

The modelId of the Gemini model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

OutputDimensionality int64optional

Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.

TaskType stringoptional

The task for embedding model.

Title stringoptional

Title is only applicable for retrieval_document tasks, and is used to represent a document title. For other tasks, title is invalid.

Transport stringoptional

Transport to access the model. Defaults to None.

Type GeminiEmbeddingConfigTypeoptional

Type of the embedding model.

type HuggingFaceInferenceAPIEmbeddingConfig struct{…}

Configuration for the HuggingFace Inference API embedding model.

Token HuggingFaceInferenceAPIEmbeddingTokenUnionoptional

Hugging Face token. Will default to the locally saved token. Pass token=False if you don’t want to send your token to the server.

One of the following:
string
bool
ClassName stringoptional
Cookies map[string, string]optional

Additional cookies to send to the server.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
Headers map[string, string]optional

Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.

ModelName stringoptional

Hugging Face model name. If None, the task will be used.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Pooling HuggingFaceInferenceAPIEmbeddingPoolingoptional

Enum of possible pooling choices with pooling behaviors.

One of the following:
const HuggingFaceInferenceAPIEmbeddingPoolingCls HuggingFaceInferenceAPIEmbeddingPooling = "cls"
const HuggingFaceInferenceAPIEmbeddingPoolingMean HuggingFaceInferenceAPIEmbeddingPooling = "mean"
const HuggingFaceInferenceAPIEmbeddingPoolingLast HuggingFaceInferenceAPIEmbeddingPooling = "last"
QueryInstruction stringoptional

Instruction to prepend during query embedding.

Task stringoptional

Optional task to pick Hugging Face’s recommended model, used when model_name is left as default of None.

TextInstruction stringoptional

Instruction to prepend during text embedding.

Timeout float64optional

The maximum number of seconds to wait for a response from the server. Loading a new model in Inference API can take up to several minutes. Defaults to None, meaning it will loop until the server is available.

Type HuggingFaceInferenceAPIEmbeddingConfigTypeoptional

Type of the embedding model.

type OpenAIEmbeddingConfig struct{…}
Component OpenAIEmbeddingoptional

Configuration for the OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for OpenAI API.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for OpenAI API.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type OpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type VertexAIEmbeddingConfig struct{…}
Component VertexTextEmbeddingoptional

Configuration for the VertexAI embedding model.

ClientEmail string

The client email for the VertexAI credentials.

Location string

The default location to use when making API calls.

PrivateKey string

The private key for the VertexAI credentials.

PrivateKeyID string

The private key ID for the VertexAI credentials.

Project string

The default GCP project to use when making Vertex API calls.

TokenUri string

The token URI for the VertexAI credentials.

AdditionalKwargs map[string, any]optional

Additional kwargs for the Vertex.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbedMode VertexTextEmbeddingEmbedModeoptional

The embedding mode to use.

One of the following:
const VertexTextEmbeddingEmbedModeDefault VertexTextEmbeddingEmbedMode = "default"
const VertexTextEmbeddingEmbedModeClassification VertexTextEmbeddingEmbedMode = "classification"
const VertexTextEmbeddingEmbedModeClustering VertexTextEmbeddingEmbedMode = "clustering"
const VertexTextEmbeddingEmbedModeSimilarity VertexTextEmbeddingEmbedMode = "similarity"
const VertexTextEmbeddingEmbedModeRetrieval VertexTextEmbeddingEmbedMode = "retrieval"
ModelName stringoptional

The modelId of the VertexAI model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Type VertexAIEmbeddingConfigTypeoptional

Type of the embedding model.

type BedrockEmbeddingConfig struct{…}
Component BedrockEmbeddingoptional

Configuration for the Bedrock embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the bedrock client.

AwsAccessKeyID stringoptional

AWS Access Key ID to use

AwsSecretAccessKey stringoptional

AWS Secret Access Key to use

AwsSessionToken stringoptional

AWS Session Token to use

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

The maximum number of API retries.

exclusiveMinimum0
ModelName stringoptional

The modelId of the Bedrock model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ProfileName stringoptional

The name of aws profile to use. If not given, then the default profile is used.

RegionName stringoptional

AWS region name to use. Uses region configured in AWS CLI if not passed

Timeout float64optional

The timeout for the Bedrock API request in seconds. It will be used for both connect and read timeouts.

Type BedrockEmbeddingConfigTypeoptional

Type of the embedding model.

Name string

The name of the embedding model config.

ProjectID string
CreatedAt Timeoptional

Creation datetime

formatdate-time
UpdatedAt Timeoptional

Update datetime

formatdate-time
EmbeddingModelConfigID stringoptional

The ID of the EmbeddingModelConfig this pipeline is using.

formatuuid
LlamaParseParameters LlamaParseParametersRespoptional

Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.

AdaptiveLongTable booloptional
AggressiveTableExtraction booloptional
AutoMode booloptional
AutoModeConfigurationJson stringoptional
AutoModeTriggerOnImageInPage booloptional
AutoModeTriggerOnRegexpInPage stringoptional
AutoModeTriggerOnTableInPage booloptional
AutoModeTriggerOnTextInPage stringoptional
AzureOpenAIAPIVersion stringoptional
AzureOpenAIDeploymentName stringoptional
AzureOpenAIEndpoint stringoptional
AzureOpenAIKey stringoptional
BboxBottom float64optional
BboxLeft float64optional
BboxRight float64optional
BboxTop float64optional
BoundingBox stringoptional
CompactMarkdownTable booloptional
ComplementalFormattingInstruction stringoptional
ContentGuidelineInstruction stringoptional
ContinuousMode booloptional
DisableImageExtraction booloptional
DisableOcr booloptional
DisableReconstruction booloptional
DoNotCache booloptional
DoNotUnrollColumns booloptional
EnableCostOptimizer booloptional
ExtractCharts booloptional
ExtractLayout booloptional
ExtractPrintedPageNumber booloptional
FastMode booloptional
FormattingInstruction stringoptional
Gpt4oAPIKey stringoptional
Gpt4oMode booloptional
GuessXlsxSheetName booloptional
HideFooters booloptional
HideHeaders booloptional
HighResOcr booloptional
HTMLMakeAllElementsVisible booloptional
HTMLRemoveFixedElements booloptional
HTMLRemoveNavigationElements booloptional
HTTPProxy stringoptional
IgnoreDocumentElementsForLayoutDetection booloptional
ImagesToSave []stringoptional
One of the following:
const LlamaParseParametersImagesToSaveScreenshot LlamaParseParametersImagesToSave = "screenshot"
const LlamaParseParametersImagesToSaveEmbedded LlamaParseParametersImagesToSave = "embedded"
const LlamaParseParametersImagesToSaveLayout LlamaParseParametersImagesToSave = "layout"
InlineImagesInMarkdown booloptional
InputS3Path stringoptional
InputS3Region stringoptional
InputURL stringoptional
InternalIsScreenshotJob booloptional
InvalidateCache booloptional
IsFormattingInstruction booloptional
JobTimeoutExtraTimePerPageInSeconds float64optional
JobTimeoutInSeconds float64optional
KeepPageSeparatorWhenMergingTables booloptional
Languages []ParsingLanguagesoptional
One of the following:
const ParsingLanguagesAf ParsingLanguages = "af"
const ParsingLanguagesAz ParsingLanguages = "az"
const ParsingLanguagesBs ParsingLanguages = "bs"
const ParsingLanguagesCs ParsingLanguages = "cs"
const ParsingLanguagesCy ParsingLanguages = "cy"
const ParsingLanguagesDa ParsingLanguages = "da"
const ParsingLanguagesDe ParsingLanguages = "de"
const ParsingLanguagesEn ParsingLanguages = "en"
const ParsingLanguagesEs ParsingLanguages = "es"
const ParsingLanguagesEt ParsingLanguages = "et"
const ParsingLanguagesFr ParsingLanguages = "fr"
const ParsingLanguagesGa ParsingLanguages = "ga"
const ParsingLanguagesHr ParsingLanguages = "hr"
const ParsingLanguagesHu ParsingLanguages = "hu"
const ParsingLanguagesID ParsingLanguages = "id"
const ParsingLanguagesIs ParsingLanguages = "is"
const ParsingLanguagesIt ParsingLanguages = "it"
const ParsingLanguagesKu ParsingLanguages = "ku"
const ParsingLanguagesLa ParsingLanguages = "la"
const ParsingLanguagesLt ParsingLanguages = "lt"
const ParsingLanguagesLv ParsingLanguages = "lv"
const ParsingLanguagesMi ParsingLanguages = "mi"
const ParsingLanguagesMs ParsingLanguages = "ms"
const ParsingLanguagesMt ParsingLanguages = "mt"
const ParsingLanguagesNl ParsingLanguages = "nl"
const ParsingLanguagesNo ParsingLanguages = "no"
const ParsingLanguagesOc ParsingLanguages = "oc"
const ParsingLanguagesPi ParsingLanguages = "pi"
const ParsingLanguagesPl ParsingLanguages = "pl"
const ParsingLanguagesPt ParsingLanguages = "pt"
const ParsingLanguagesRo ParsingLanguages = "ro"
const ParsingLanguagesRsLatin ParsingLanguages = "rs_latin"
const ParsingLanguagesSk ParsingLanguages = "sk"
const ParsingLanguagesSl ParsingLanguages = "sl"
const ParsingLanguagesSq ParsingLanguages = "sq"
const ParsingLanguagesSv ParsingLanguages = "sv"
const ParsingLanguagesSw ParsingLanguages = "sw"
const ParsingLanguagesTl ParsingLanguages = "tl"
const ParsingLanguagesTr ParsingLanguages = "tr"
const ParsingLanguagesUz ParsingLanguages = "uz"
const ParsingLanguagesVi ParsingLanguages = "vi"
const ParsingLanguagesAr ParsingLanguages = "ar"
const ParsingLanguagesFa ParsingLanguages = "fa"
const ParsingLanguagesUg ParsingLanguages = "ug"
const ParsingLanguagesUr ParsingLanguages = "ur"
const ParsingLanguagesBn ParsingLanguages = "bn"
const ParsingLanguagesAs ParsingLanguages = "as"
const ParsingLanguagesMni ParsingLanguages = "mni"
const ParsingLanguagesRu ParsingLanguages = "ru"
const ParsingLanguagesRsCyrillic ParsingLanguages = "rs_cyrillic"
const ParsingLanguagesBe ParsingLanguages = "be"
const ParsingLanguagesBg ParsingLanguages = "bg"
const ParsingLanguagesUk ParsingLanguages = "uk"
const ParsingLanguagesMn ParsingLanguages = "mn"
const ParsingLanguagesAbq ParsingLanguages = "abq"
const ParsingLanguagesAdy ParsingLanguages = "ady"
const ParsingLanguagesKbd ParsingLanguages = "kbd"
const ParsingLanguagesAva ParsingLanguages = "ava"
const ParsingLanguagesDar ParsingLanguages = "dar"
const ParsingLanguagesInh ParsingLanguages = "inh"
const ParsingLanguagesChe ParsingLanguages = "che"
const ParsingLanguagesLbe ParsingLanguages = "lbe"
const ParsingLanguagesLez ParsingLanguages = "lez"
const ParsingLanguagesTab ParsingLanguages = "tab"
const ParsingLanguagesTjk ParsingLanguages = "tjk"
const ParsingLanguagesHi ParsingLanguages = "hi"
const ParsingLanguagesMr ParsingLanguages = "mr"
const ParsingLanguagesNe ParsingLanguages = "ne"
const ParsingLanguagesBh ParsingLanguages = "bh"
const ParsingLanguagesMai ParsingLanguages = "mai"
const ParsingLanguagesAng ParsingLanguages = "ang"
const ParsingLanguagesBho ParsingLanguages = "bho"
const ParsingLanguagesMah ParsingLanguages = "mah"
const ParsingLanguagesSck ParsingLanguages = "sck"
const ParsingLanguagesNew ParsingLanguages = "new"
const ParsingLanguagesGom ParsingLanguages = "gom"
const ParsingLanguagesSa ParsingLanguages = "sa"
const ParsingLanguagesBgc ParsingLanguages = "bgc"
const ParsingLanguagesTh ParsingLanguages = "th"
const ParsingLanguagesChSim ParsingLanguages = "ch_sim"
const ParsingLanguagesChTra ParsingLanguages = "ch_tra"
const ParsingLanguagesJa ParsingLanguages = "ja"
const ParsingLanguagesKo ParsingLanguages = "ko"
const ParsingLanguagesTa ParsingLanguages = "ta"
const ParsingLanguagesTe ParsingLanguages = "te"
const ParsingLanguagesKn ParsingLanguages = "kn"
LayoutAware booloptional
LineLevelBoundingBox booloptional
MarkdownTableMultilineHeaderSeparator stringoptional
MaxPages int64optional
MaxPagesEnforced int64optional
MergeTablesAcrossPagesInMarkdown booloptional
Model stringoptional
OutlinedTableExtraction booloptional
OutputPdfOfDocument booloptional
OutputS3PathPrefix stringoptional
OutputS3Region stringoptional
OutputTablesAsHTML booloptional
PageErrorTolerance float64optional
PageHeaderPrefix stringoptional
PageHeaderSuffix stringoptional
PagePrefix stringoptional
PageSeparator stringoptional
PageSuffix stringoptional
ParseMode ParsingModeoptional

Enum for representing the mode of parsing to be used.

One of the following:
const ParsingModeParsePageWithoutLlm ParsingMode = "parse_page_without_llm"
const ParsingModeParsePageWithLlm ParsingMode = "parse_page_with_llm"
const ParsingModeParsePageWithLvm ParsingMode = "parse_page_with_lvm"
const ParsingModeParsePageWithAgent ParsingMode = "parse_page_with_agent"
const ParsingModeParsePageWithLayoutAgent ParsingMode = "parse_page_with_layout_agent"
const ParsingModeParseDocumentWithLlm ParsingMode = "parse_document_with_llm"
const ParsingModeParseDocumentWithLvm ParsingMode = "parse_document_with_lvm"
const ParsingModeParseDocumentWithAgent ParsingMode = "parse_document_with_agent"
ParsingInstruction stringoptional
PreciseBoundingBox booloptional
PremiumMode booloptional
PresentationOutOfBoundsContent booloptional
PresentationSkipEmbeddedData booloptional
PreserveLayoutAlignmentAcrossPages booloptional
PreserveVerySmallText booloptional
Preset stringoptional
Priority LlamaParseParametersPriorityoptional

The priority for the request. This field may be ignored or overwritten depending on the organization tier.

One of the following:
const LlamaParseParametersPriorityLow LlamaParseParametersPriority = "low"
const LlamaParseParametersPriorityMedium LlamaParseParametersPriority = "medium"
const LlamaParseParametersPriorityHigh LlamaParseParametersPriority = "high"
const LlamaParseParametersPriorityCritical LlamaParseParametersPriority = "critical"
ProjectID stringoptional
RemoveHiddenText booloptional
ReplaceFailedPageMode FailPageModeoptional

Enum for representing the different available page error handling modes.

One of the following:
const FailPageModeRawText FailPageMode = "raw_text"
const FailPageModeBlankPage FailPageMode = "blank_page"
const FailPageModeErrorMessage FailPageMode = "error_message"
ReplaceFailedPageWithErrorMessagePrefix stringoptional
ReplaceFailedPageWithErrorMessageSuffix stringoptional
SaveImages booloptional
SkipDiagonalText booloptional
SpecializedChartParsingAgentic booloptional
SpecializedChartParsingEfficient booloptional
SpecializedChartParsingPlus booloptional
SpecializedImageParsing booloptional
SpreadsheetExtractSubTables booloptional
SpreadsheetForceFormulaComputation booloptional
SpreadsheetIncludeHiddenSheets booloptional
StrictModeBuggyFont booloptional
StrictModeImageExtraction booloptional
StrictModeImageOcr booloptional
StrictModeReconstruction booloptional
StructuredOutput booloptional
StructuredOutputJsonSchema stringoptional
StructuredOutputJsonSchemaName stringoptional
SystemPrompt stringoptional
SystemPromptAppend stringoptional
TakeScreenshot booloptional
TargetPages stringoptional
Tier stringoptional
UseVendorMultimodalModel booloptional
UserPrompt stringoptional
VendorMultimodalAPIKey stringoptional
VendorMultimodalModelName stringoptional
Version stringoptional
WebhookConfigurations []LlamaParseParametersWebhookConfigurationRespoptional

Outbound webhook endpoints to notify on job status changes

WebhookEvents []stringoptional

Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.

One of the following:
const LlamaParseParametersWebhookConfigurationWebhookEventExtractPending LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.success"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractError LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.error"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractPartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventParsePending LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventParseRunning LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.running"
const LlamaParseParametersWebhookConfigurationWebhookEventParseSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.success"
const LlamaParseParametersWebhookConfigurationWebhookEventParseError LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.error"
const LlamaParseParametersWebhookConfigurationWebhookEventParsePartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventParseCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyPending LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifySuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.success"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyError LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.error"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyPartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventUnmappedEvent LlamaParseParametersWebhookConfigurationWebhookEvent = "unmapped_event"
WebhookHeaders map[string, string]optional

Custom HTTP headers sent with each webhook request (e.g. auth tokens)

WebhookOutputFormat stringoptional

Response format sent to the webhook: ‘string’ (default) or ‘json’

WebhookURL stringoptional

URL to receive webhook POST notifications

WebhookURL stringoptional
ManagedPipelineID stringoptional

The ID of the ManagedPipeline this playground pipeline is linked to.

formatuuid
MetadataConfig PipelineMetadataConfigoptional

Metadata configuration for the pipeline.

ExcludedEmbedMetadataKeys []stringoptional

List of metadata keys to exclude from embeddings

ExcludedLlmMetadataKeys []stringoptional

List of metadata keys to exclude from LLM during retrieval

PipelineType PipelineTypeoptional

Type of pipeline. Either PLAYGROUND or MANAGED.

One of the following:
const PipelineTypePlayground PipelineType = "PLAYGROUND"
const PipelineTypeManaged PipelineType = "MANAGED"
PresetRetrievalParameters PresetRetrievalParamsRespoptional

Preset retrieval parameters for the pipeline.

Alpha float64optional

Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.

maximum1
minimum0
ClassName stringoptional
DenseSimilarityCutoff float64optional

Minimum similarity score wrt query for retrieval

maximum1
minimum0
DenseSimilarityTopK int64optional

Number of nodes for dense retrieval.

maximum100
minimum1
EnableReranking booloptional

Enable reranking for retrieval

FilesTopK int64optional

Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).

maximum5
minimum1
RerankTopN int64optional

Number of reranked nodes for returning.

maximum100
minimum1
RetrievalMode RetrievalModeoptional

The retrieval mode for the query.

One of the following:
const RetrievalModeChunks RetrievalMode = "chunks"
const RetrievalModeFilesViaMetadata RetrievalMode = "files_via_metadata"
const RetrievalModeFilesViaContent RetrievalMode = "files_via_content"
const RetrievalModeAutoRouted RetrievalMode = "auto_routed"
DeprecatedRetrieveImageNodes booloptional

Whether to retrieve image nodes.

RetrievePageFigureNodes booloptional

Whether to retrieve page figure nodes.

RetrievePageScreenshotNodes booloptional

Whether to retrieve page screenshot nodes.

SearchFilters MetadataFiltersoptional

Metadata filters for vector stores.

Filters []MetadataFiltersFilterUnion
One of the following:
type MetadataFiltersFilterMetadataFilter struct{…}

Comprehensive metadata filter for vector stores to support more operators.

Value uses Strict types, as int, float and str are compatible types and were all converted to string before.

See: https://docs.pydantic.dev/latest/usage/types/#strict-types

Key string
Value MetadataFiltersFilterMetadataFilterValueUnion
One of the following:
float64
string
type MetadataFiltersFilterMetadataFilterValueArray []string
type MetadataFiltersFilterMetadataFilterValueArray []float64
type MetadataFiltersFilterMetadataFilterValueArray []int64
Operator stringoptional

Vector store filter operator.

One of the following:
const MetadataFiltersFilterMetadataFilterOperatorEquals MetadataFiltersFilterMetadataFilterOperator = "=="
const MetadataFiltersFilterMetadataFilterOperatorGreater MetadataFiltersFilterMetadataFilterOperator = ">"
const MetadataFiltersFilterMetadataFilterOperatorLess MetadataFiltersFilterMetadataFilterOperator = "<"
const MetadataFiltersFilterMetadataFilterOperatorNotEquals MetadataFiltersFilterMetadataFilterOperator = "!="
const MetadataFiltersFilterMetadataFilterOperatorGreaterOrEquals MetadataFiltersFilterMetadataFilterOperator = ">="
const MetadataFiltersFilterMetadataFilterOperatorLessOrEquals MetadataFiltersFilterMetadataFilterOperator = "<="
const MetadataFiltersFilterMetadataFilterOperatorIn MetadataFiltersFilterMetadataFilterOperator = "in"
const MetadataFiltersFilterMetadataFilterOperatorNin MetadataFiltersFilterMetadataFilterOperator = "nin"
const MetadataFiltersFilterMetadataFilterOperatorAny MetadataFiltersFilterMetadataFilterOperator = "any"
const MetadataFiltersFilterMetadataFilterOperatorAll MetadataFiltersFilterMetadataFilterOperator = "all"
const MetadataFiltersFilterMetadataFilterOperatorTextMatch MetadataFiltersFilterMetadataFilterOperator = "text_match"
const MetadataFiltersFilterMetadataFilterOperatorTextMatchInsensitive MetadataFiltersFilterMetadataFilterOperator = "text_match_insensitive"
const MetadataFiltersFilterMetadataFilterOperatorContains MetadataFiltersFilterMetadataFilterOperator = "contains"
const MetadataFiltersFilterMetadataFilterOperatorIsEmpty MetadataFiltersFilterMetadataFilterOperator = "is_empty"
type MetadataFilters MetadataFilters

Metadata filters for vector stores.

Condition MetadataFiltersConditionoptional

Vector store filter conditions to combine different filters.

One of the following:
const MetadataFiltersConditionAnd MetadataFiltersCondition = "and"
const MetadataFiltersConditionOr MetadataFiltersCondition = "or"
const MetadataFiltersConditionNot MetadataFiltersCondition = "not"
SearchFiltersInferenceSchema map[string, PresetRetrievalParamsSearchFiltersInferenceSchemaUnionResp]optional

JSON Schema that will be used to infer search_filters. Omit or leave as null to skip inference.

One of the following:
type PresetRetrievalParamsSearchFiltersInferenceSchemaMap map[string, any]
type PresetRetrievalParamsSearchFiltersInferenceSchemaArray []any
string
float64
bool
SparseSimilarityTopK int64optional

Number of nodes for sparse retrieval.

maximum100
minimum1
SparseModelConfig SparseModelConfigoptional

Configuration for sparse embedding models used in hybrid search.

This allows users to choose between Splade and BM25 models for sparse retrieval in managed data sinks.

ClassName stringoptional
ModelType SparseModelConfigModelTypeoptional

The sparse model type to use. ‘bm25’ uses Qdrant’s FastEmbed BM25 model (default for new pipelines), ‘splade’ uses HuggingFace Splade model, ‘auto’ selects based on deployment mode (BYOC uses term frequency, Cloud uses Splade).

One of the following:
const SparseModelConfigModelTypeSplade SparseModelConfigModelType = "splade"
const SparseModelConfigModelTypeBm25 SparseModelConfigModelType = "bm25"
const SparseModelConfigModelTypeAuto SparseModelConfigModelType = "auto"
Status PipelineStatusoptional

Status of the pipeline.

One of the following:
const PipelineStatusCreated PipelineStatus = "CREATED"
const PipelineStatusDeleting PipelineStatus = "DELETING"
TransformConfig PipelineTransformConfigUnionoptional

Configuration for the transformation.

One of the following:
type AutoTransformConfig struct{…}
ChunkOverlap int64optional

Chunk overlap for the transformation.

ChunkSize int64optional

Chunk size for the transformation.

exclusiveMinimum0
Mode AutoTransformConfigModeoptional
type AdvancedModeTransformConfig struct{…}
ChunkingConfig AdvancedModeTransformConfigChunkingConfigUnionoptional

Configuration for the chunking.

One of the following:
type AdvancedModeTransformConfigChunkingConfigNoneChunkingConfig struct{…}
Mode stringoptional
type AdvancedModeTransformConfigChunkingConfigCharacterChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
type AdvancedModeTransformConfigChunkingConfigTokenChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
Separator stringoptional
type AdvancedModeTransformConfigChunkingConfigSentenceChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
ParagraphSeparator stringoptional
Separator stringoptional
type AdvancedModeTransformConfigChunkingConfigSemanticChunkingConfig struct{…}
BreakpointPercentileThreshold int64optional
BufferSize int64optional
Mode stringoptional
Mode AdvancedModeTransformConfigModeoptional
SegmentationConfig AdvancedModeTransformConfigSegmentationConfigUnionoptional

Configuration for the segmentation.

One of the following:
type AdvancedModeTransformConfigSegmentationConfigNoneSegmentationConfig struct{…}
Mode stringoptional
type AdvancedModeTransformConfigSegmentationConfigPageSegmentationConfig struct{…}
Mode stringoptional
PageSeparator stringoptional
type AdvancedModeTransformConfigSegmentationConfigElementSegmentationConfig struct{…}
Mode stringoptional
UpdatedAt Timeoptional

Update datetime

formatdate-time
type PipelineCreate struct{…}

Schema for creating a pipeline.

Name string
DataSink DataSinkCreateoptional

Schema for creating a data sink.

Component DataSinkCreateComponentUnion

Component that implements the data sink

One of the following:
map[string, any]
type CloudPineconeVectorStore struct{…}

Cloud Pinecone Vector Store.

This class is used to store the configuration for a Pinecone vector store, so that it can be created and used in LlamaCloud.

Args: api_key (str): API key for authenticating with Pinecone index_name (str): name of the Pinecone index namespace (optional[str]): namespace to use in the Pinecone index insert_kwargs (optional[dict]): additional kwargs to pass during insertion

APIKey string

The API key for authenticating with Pinecone

formatpassword
IndexName string
ClassName stringoptional
InsertKwargs map[string, any]optional
Namespace stringoptional
SupportsNestedMetadataFilters booloptional
type CloudPostgresVectorStore struct{…}
Database string
EmbedDim int64
Host string
Password string
Port int64
SchemaName string
TableName string
User string
ClassName stringoptional
HnswSettings PgVectorHnswSettingsoptional

HNSW settings for PGVector.

DistanceMethod PgVectorHnswSettingsDistanceMethodoptional

The distance method to use.

One of the following:
const PgVectorHnswSettingsDistanceMethodL2 PgVectorHnswSettingsDistanceMethod = "l2"
const PgVectorHnswSettingsDistanceMethodIP PgVectorHnswSettingsDistanceMethod = "ip"
const PgVectorHnswSettingsDistanceMethodCosine PgVectorHnswSettingsDistanceMethod = "cosine"
const PgVectorHnswSettingsDistanceMethodL1 PgVectorHnswSettingsDistanceMethod = "l1"
const PgVectorHnswSettingsDistanceMethodHamming PgVectorHnswSettingsDistanceMethod = "hamming"
const PgVectorHnswSettingsDistanceMethodJaccard PgVectorHnswSettingsDistanceMethod = "jaccard"
EfConstruction int64optional

The number of edges to use during the construction phase.

minimum1

The number of edges to use during the search phase.

minimum1
M int64optional

The number of bi-directional links created for each new element.

minimum1
VectorType PgVectorHnswSettingsVectorTypeoptional

The type of vector to use.

One of the following:
const PgVectorHnswSettingsVectorTypeVector PgVectorHnswSettingsVectorType = "vector"
const PgVectorHnswSettingsVectorTypeHalfVec PgVectorHnswSettingsVectorType = "half_vec"
const PgVectorHnswSettingsVectorTypeBit PgVectorHnswSettingsVectorType = "bit"
const PgVectorHnswSettingsVectorTypeSparseVec PgVectorHnswSettingsVectorType = "sparse_vec"
PerformSetup booloptional
SupportsNestedMetadataFilters booloptional
type CloudQdrantVectorStore struct{…}

Cloud Qdrant Vector Store.

This class is used to store the configuration for a Qdrant vector store, so that it can be created and used in LlamaCloud.

Args: collection_name (str): name of the Qdrant collection url (str): url of the Qdrant instance api_key (str): API key for authenticating with Qdrant max_retries (int): maximum number of retries in case of a failure. Defaults to 3 client_kwargs (dict): additional kwargs to pass to the Qdrant client

APIKey string
CollectionName string
URL string
ClassName stringoptional
ClientKwargs map[string, any]optional
MaxRetries int64optional
SupportsNestedMetadataFilters booloptional
type CloudAzureAISearchVectorStore struct{…}

Cloud Azure AI Search Vector Store.

SearchServiceAPIKey string
SearchServiceEndpoint string
ClassName stringoptional
ClientID stringoptional
ClientSecret stringoptional
EmbeddingDimension int64optional
FilterableMetadataFieldKeys map[string, any]optional
IndexName stringoptional
SearchServiceAPIVersion stringoptional
SupportsNestedMetadataFilters booloptional
TenantID stringoptional

Cloud MongoDB Atlas Vector Store.

This class is used to store the configuration for a MongoDB Atlas vector store, so that it can be created and used in LlamaCloud.

Args: mongodb_uri (str): URI for connecting to MongoDB Atlas db_name (str): name of the MongoDB database collection_name (str): name of the MongoDB collection vector_index_name (str): name of the MongoDB Atlas vector index fulltext_index_name (str): name of the MongoDB Atlas full-text index

type CloudMilvusVectorStore struct{…}

Cloud Milvus Vector Store.

Uri string
Token stringoptional
ClassName stringoptional
CollectionName stringoptional
EmbeddingDimension int64optional
SupportsNestedMetadataFilters booloptional
type CloudAstraDBVectorStore struct{…}

Cloud AstraDB Vector Store.

This class is used to store the configuration for an AstraDB vector store, so that it can be created and used in LlamaCloud.

Args: token (str): The Astra DB Application Token to use. api_endpoint (str): The Astra DB JSON API endpoint for your database. collection_name (str): Collection name to use. If not existing, it will be created. embedding_dimension (int): Length of the embedding vectors in use. keyspace (optional[str]): The keyspace to use. If not provided, ‘default_keyspace’

Token string

The Astra DB Application Token to use

formatpassword
APIEndpoint string

The Astra DB JSON API endpoint for your database

CollectionName string

Collection name to use. If not existing, it will be created

EmbeddingDimension int64

Length of the embedding vectors in use

ClassName stringoptional
Keyspace stringoptional

The keyspace to use. If not provided, ‘default_keyspace’

SupportsNestedMetadataFilters booloptional
Name string

The name of the data sink.

SinkType DataSinkCreateSinkType
One of the following:
const DataSinkCreateSinkTypePinecone DataSinkCreateSinkType = "PINECONE"
const DataSinkCreateSinkTypePostgres DataSinkCreateSinkType = "POSTGRES"
const DataSinkCreateSinkTypeQdrant DataSinkCreateSinkType = "QDRANT"
const DataSinkCreateSinkTypeAzureaiSearch DataSinkCreateSinkType = "AZUREAI_SEARCH"
const DataSinkCreateSinkTypeMongoDBAtlas DataSinkCreateSinkType = "MONGODB_ATLAS"
const DataSinkCreateSinkTypeMilvus DataSinkCreateSinkType = "MILVUS"
const DataSinkCreateSinkTypeAstraDB DataSinkCreateSinkType = "ASTRA_DB"
DataSinkID stringoptional

Data sink ID. When provided instead of data_sink, the data sink will be looked up by ID.

formatuuid
EmbeddingConfig PipelineCreateEmbeddingConfigUnionoptional
One of the following:
type AzureOpenAIEmbeddingConfig struct{…}
Component AzureOpenAIEmbeddingoptional

Configuration for the Azure OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for Azure deployment.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for Azure OpenAI API.

AzureDeployment stringoptional

The Azure deployment to use.

AzureEndpoint stringoptional

The Azure endpoint to use.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type AzureOpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type CohereEmbeddingConfig struct{…}
Component CohereEmbeddingoptional

Configuration for the Cohere embedding model.

APIKey string

The Cohere API key.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbeddingType stringoptional

Embedding type. If not provided float embedding_type is used when needed.

InputType stringoptional

Model Input type. If not provided, search_document and search_query are used when needed.

ModelName stringoptional

The modelId of the Cohere model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Truncate stringoptional

Truncation type - START/ END/ NONE

Type CohereEmbeddingConfigTypeoptional

Type of the embedding model.

type GeminiEmbeddingConfig struct{…}
Component GeminiEmbeddingoptional

Configuration for the Gemini embedding model.

APIBase stringoptional

API base to access the model. Defaults to None.

APIKey stringoptional

API key to access the model. Defaults to None.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
ModelName stringoptional

The modelId of the Gemini model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

OutputDimensionality int64optional

Optional reduced dimension for output embeddings. Supported by models/text-embedding-004 and newer (e.g. gemini-embedding-001). Not supported by models/embedding-001.

TaskType stringoptional

The task for embedding model.

Title stringoptional

Title is only applicable for retrieval_document tasks, and is used to represent a document title. For other tasks, title is invalid.

Transport stringoptional

Transport to access the model. Defaults to None.

Type GeminiEmbeddingConfigTypeoptional

Type of the embedding model.

type HuggingFaceInferenceAPIEmbeddingConfig struct{…}

Configuration for the HuggingFace Inference API embedding model.

Token HuggingFaceInferenceAPIEmbeddingTokenUnionoptional

Hugging Face token. Will default to the locally saved token. Pass token=False if you don’t want to send your token to the server.

One of the following:
string
bool
ClassName stringoptional
Cookies map[string, string]optional

Additional cookies to send to the server.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
Headers map[string, string]optional

Additional headers to send to the server. By default only the authorization and user-agent headers are sent. Values in this dictionary will override the default values.

ModelName stringoptional

Hugging Face model name. If None, the task will be used.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Pooling HuggingFaceInferenceAPIEmbeddingPoolingoptional

Enum of possible pooling choices with pooling behaviors.

One of the following:
const HuggingFaceInferenceAPIEmbeddingPoolingCls HuggingFaceInferenceAPIEmbeddingPooling = "cls"
const HuggingFaceInferenceAPIEmbeddingPoolingMean HuggingFaceInferenceAPIEmbeddingPooling = "mean"
const HuggingFaceInferenceAPIEmbeddingPoolingLast HuggingFaceInferenceAPIEmbeddingPooling = "last"
QueryInstruction stringoptional

Instruction to prepend during query embedding.

Task stringoptional

Optional task to pick Hugging Face’s recommended model, used when model_name is left as default of None.

TextInstruction stringoptional

Instruction to prepend during text embedding.

Timeout float64optional

The maximum number of seconds to wait for a response from the server. Loading a new model in Inference API can take up to several minutes. Defaults to None, meaning it will loop until the server is available.

Type HuggingFaceInferenceAPIEmbeddingConfigTypeoptional

Type of the embedding model.

type OpenAIEmbeddingConfig struct{…}
Component OpenAIEmbeddingoptional

Configuration for the OpenAI embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the OpenAI API.

APIBase stringoptional

The base URL for OpenAI API.

APIKey stringoptional

The OpenAI API key.

APIVersion stringoptional

The version for OpenAI API.

ClassName stringoptional
DefaultHeaders map[string, string]optional

The default headers for API requests.

Dimensions int64optional

The number of dimensions on the output embedding vectors. Works only with v3 embedding models.

EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

Maximum number of retries.

minimum0
ModelName stringoptional

The name of the OpenAI embedding model.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ReuseClient booloptional

Reuse the OpenAI client between requests. When doing anything with large volumes of async API calls, setting this to false can improve stability.

Timeout float64optional

Timeout for each request.

minimum0
Type OpenAIEmbeddingConfigTypeoptional

Type of the embedding model.

type VertexAIEmbeddingConfig struct{…}
Component VertexTextEmbeddingoptional

Configuration for the VertexAI embedding model.

ClientEmail string

The client email for the VertexAI credentials.

Location string

The default location to use when making API calls.

PrivateKey string

The private key for the VertexAI credentials.

PrivateKeyID string

The private key ID for the VertexAI credentials.

Project string

The default GCP project to use when making Vertex API calls.

TokenUri string

The token URI for the VertexAI credentials.

AdditionalKwargs map[string, any]optional

Additional kwargs for the Vertex.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbedMode VertexTextEmbeddingEmbedModeoptional

The embedding mode to use.

One of the following:
const VertexTextEmbeddingEmbedModeDefault VertexTextEmbeddingEmbedMode = "default"
const VertexTextEmbeddingEmbedModeClassification VertexTextEmbeddingEmbedMode = "classification"
const VertexTextEmbeddingEmbedModeClustering VertexTextEmbeddingEmbedMode = "clustering"
const VertexTextEmbeddingEmbedModeSimilarity VertexTextEmbeddingEmbedMode = "similarity"
const VertexTextEmbeddingEmbedModeRetrieval VertexTextEmbeddingEmbedMode = "retrieval"
ModelName stringoptional

The modelId of the VertexAI model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Type VertexAIEmbeddingConfigTypeoptional

Type of the embedding model.

type BedrockEmbeddingConfig struct{…}
Component BedrockEmbeddingoptional

Configuration for the Bedrock embedding model.

AdditionalKwargs map[string, any]optional

Additional kwargs for the bedrock client.

AwsAccessKeyID stringoptional

AWS Access Key ID to use

AwsSecretAccessKey stringoptional

AWS Secret Access Key to use

AwsSessionToken stringoptional

AWS Session Token to use

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
MaxRetries int64optional

The maximum number of API retries.

exclusiveMinimum0
ModelName stringoptional

The modelId of the Bedrock model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

ProfileName stringoptional

The name of aws profile to use. If not given, then the default profile is used.

RegionName stringoptional

AWS region name to use. Uses region configured in AWS CLI if not passed

Timeout float64optional

The timeout for the Bedrock API request in seconds. It will be used for both connect and read timeouts.

Type BedrockEmbeddingConfigTypeoptional

Type of the embedding model.

EmbeddingModelConfigID stringoptional

Embedding model config ID. When provided instead of embedding_config, the embedding model config will be looked up by ID.

formatuuid
LlamaParseParameters LlamaParseParametersRespoptional

Settings that can be configured for how to use LlamaParse to parse files within a LlamaCloud pipeline.

AdaptiveLongTable booloptional
AggressiveTableExtraction booloptional
AutoMode booloptional
AutoModeConfigurationJson stringoptional
AutoModeTriggerOnImageInPage booloptional
AutoModeTriggerOnRegexpInPage stringoptional
AutoModeTriggerOnTableInPage booloptional
AutoModeTriggerOnTextInPage stringoptional
AzureOpenAIAPIVersion stringoptional
AzureOpenAIDeploymentName stringoptional
AzureOpenAIEndpoint stringoptional
AzureOpenAIKey stringoptional
BboxBottom float64optional
BboxLeft float64optional
BboxRight float64optional
BboxTop float64optional
BoundingBox stringoptional
CompactMarkdownTable booloptional
ComplementalFormattingInstruction stringoptional
ContentGuidelineInstruction stringoptional
ContinuousMode booloptional
DisableImageExtraction booloptional
DisableOcr booloptional
DisableReconstruction booloptional
DoNotCache booloptional
DoNotUnrollColumns booloptional
EnableCostOptimizer booloptional
ExtractCharts booloptional
ExtractLayout booloptional
ExtractPrintedPageNumber booloptional
FastMode booloptional
FormattingInstruction stringoptional
Gpt4oAPIKey stringoptional
Gpt4oMode booloptional
GuessXlsxSheetName booloptional
HideFooters booloptional
HideHeaders booloptional
HighResOcr booloptional
HTMLMakeAllElementsVisible booloptional
HTMLRemoveFixedElements booloptional
HTMLRemoveNavigationElements booloptional
HTTPProxy stringoptional
IgnoreDocumentElementsForLayoutDetection booloptional
ImagesToSave []stringoptional
One of the following:
const LlamaParseParametersImagesToSaveScreenshot LlamaParseParametersImagesToSave = "screenshot"
const LlamaParseParametersImagesToSaveEmbedded LlamaParseParametersImagesToSave = "embedded"
const LlamaParseParametersImagesToSaveLayout LlamaParseParametersImagesToSave = "layout"
InlineImagesInMarkdown booloptional
InputS3Path stringoptional
InputS3Region stringoptional
InputURL stringoptional
InternalIsScreenshotJob booloptional
InvalidateCache booloptional
IsFormattingInstruction booloptional
JobTimeoutExtraTimePerPageInSeconds float64optional
JobTimeoutInSeconds float64optional
KeepPageSeparatorWhenMergingTables booloptional
Languages []ParsingLanguagesoptional
One of the following:
const ParsingLanguagesAf ParsingLanguages = "af"
const ParsingLanguagesAz ParsingLanguages = "az"
const ParsingLanguagesBs ParsingLanguages = "bs"
const ParsingLanguagesCs ParsingLanguages = "cs"
const ParsingLanguagesCy ParsingLanguages = "cy"
const ParsingLanguagesDa ParsingLanguages = "da"
const ParsingLanguagesDe ParsingLanguages = "de"
const ParsingLanguagesEn ParsingLanguages = "en"
const ParsingLanguagesEs ParsingLanguages = "es"
const ParsingLanguagesEt ParsingLanguages = "et"
const ParsingLanguagesFr ParsingLanguages = "fr"
const ParsingLanguagesGa ParsingLanguages = "ga"
const ParsingLanguagesHr ParsingLanguages = "hr"
const ParsingLanguagesHu ParsingLanguages = "hu"
const ParsingLanguagesID ParsingLanguages = "id"
const ParsingLanguagesIs ParsingLanguages = "is"
const ParsingLanguagesIt ParsingLanguages = "it"
const ParsingLanguagesKu ParsingLanguages = "ku"
const ParsingLanguagesLa ParsingLanguages = "la"
const ParsingLanguagesLt ParsingLanguages = "lt"
const ParsingLanguagesLv ParsingLanguages = "lv"
const ParsingLanguagesMi ParsingLanguages = "mi"
const ParsingLanguagesMs ParsingLanguages = "ms"
const ParsingLanguagesMt ParsingLanguages = "mt"
const ParsingLanguagesNl ParsingLanguages = "nl"
const ParsingLanguagesNo ParsingLanguages = "no"
const ParsingLanguagesOc ParsingLanguages = "oc"
const ParsingLanguagesPi ParsingLanguages = "pi"
const ParsingLanguagesPl ParsingLanguages = "pl"
const ParsingLanguagesPt ParsingLanguages = "pt"
const ParsingLanguagesRo ParsingLanguages = "ro"
const ParsingLanguagesRsLatin ParsingLanguages = "rs_latin"
const ParsingLanguagesSk ParsingLanguages = "sk"
const ParsingLanguagesSl ParsingLanguages = "sl"
const ParsingLanguagesSq ParsingLanguages = "sq"
const ParsingLanguagesSv ParsingLanguages = "sv"
const ParsingLanguagesSw ParsingLanguages = "sw"
const ParsingLanguagesTl ParsingLanguages = "tl"
const ParsingLanguagesTr ParsingLanguages = "tr"
const ParsingLanguagesUz ParsingLanguages = "uz"
const ParsingLanguagesVi ParsingLanguages = "vi"
const ParsingLanguagesAr ParsingLanguages = "ar"
const ParsingLanguagesFa ParsingLanguages = "fa"
const ParsingLanguagesUg ParsingLanguages = "ug"
const ParsingLanguagesUr ParsingLanguages = "ur"
const ParsingLanguagesBn ParsingLanguages = "bn"
const ParsingLanguagesAs ParsingLanguages = "as"
const ParsingLanguagesMni ParsingLanguages = "mni"
const ParsingLanguagesRu ParsingLanguages = "ru"
const ParsingLanguagesRsCyrillic ParsingLanguages = "rs_cyrillic"
const ParsingLanguagesBe ParsingLanguages = "be"
const ParsingLanguagesBg ParsingLanguages = "bg"
const ParsingLanguagesUk ParsingLanguages = "uk"
const ParsingLanguagesMn ParsingLanguages = "mn"
const ParsingLanguagesAbq ParsingLanguages = "abq"
const ParsingLanguagesAdy ParsingLanguages = "ady"
const ParsingLanguagesKbd ParsingLanguages = "kbd"
const ParsingLanguagesAva ParsingLanguages = "ava"
const ParsingLanguagesDar ParsingLanguages = "dar"
const ParsingLanguagesInh ParsingLanguages = "inh"
const ParsingLanguagesChe ParsingLanguages = "che"
const ParsingLanguagesLbe ParsingLanguages = "lbe"
const ParsingLanguagesLez ParsingLanguages = "lez"
const ParsingLanguagesTab ParsingLanguages = "tab"
const ParsingLanguagesTjk ParsingLanguages = "tjk"
const ParsingLanguagesHi ParsingLanguages = "hi"
const ParsingLanguagesMr ParsingLanguages = "mr"
const ParsingLanguagesNe ParsingLanguages = "ne"
const ParsingLanguagesBh ParsingLanguages = "bh"
const ParsingLanguagesMai ParsingLanguages = "mai"
const ParsingLanguagesAng ParsingLanguages = "ang"
const ParsingLanguagesBho ParsingLanguages = "bho"
const ParsingLanguagesMah ParsingLanguages = "mah"
const ParsingLanguagesSck ParsingLanguages = "sck"
const ParsingLanguagesNew ParsingLanguages = "new"
const ParsingLanguagesGom ParsingLanguages = "gom"
const ParsingLanguagesSa ParsingLanguages = "sa"
const ParsingLanguagesBgc ParsingLanguages = "bgc"
const ParsingLanguagesTh ParsingLanguages = "th"
const ParsingLanguagesChSim ParsingLanguages = "ch_sim"
const ParsingLanguagesChTra ParsingLanguages = "ch_tra"
const ParsingLanguagesJa ParsingLanguages = "ja"
const ParsingLanguagesKo ParsingLanguages = "ko"
const ParsingLanguagesTa ParsingLanguages = "ta"
const ParsingLanguagesTe ParsingLanguages = "te"
const ParsingLanguagesKn ParsingLanguages = "kn"
LayoutAware booloptional
LineLevelBoundingBox booloptional
MarkdownTableMultilineHeaderSeparator stringoptional
MaxPages int64optional
MaxPagesEnforced int64optional
MergeTablesAcrossPagesInMarkdown booloptional
Model stringoptional
OutlinedTableExtraction booloptional
OutputPdfOfDocument booloptional
OutputS3PathPrefix stringoptional
OutputS3Region stringoptional
OutputTablesAsHTML booloptional
PageErrorTolerance float64optional
PageHeaderPrefix stringoptional
PageHeaderSuffix stringoptional
PagePrefix stringoptional
PageSeparator stringoptional
PageSuffix stringoptional
ParseMode ParsingModeoptional

Enum for representing the mode of parsing to be used.

One of the following:
const ParsingModeParsePageWithoutLlm ParsingMode = "parse_page_without_llm"
const ParsingModeParsePageWithLlm ParsingMode = "parse_page_with_llm"
const ParsingModeParsePageWithLvm ParsingMode = "parse_page_with_lvm"
const ParsingModeParsePageWithAgent ParsingMode = "parse_page_with_agent"
const ParsingModeParsePageWithLayoutAgent ParsingMode = "parse_page_with_layout_agent"
const ParsingModeParseDocumentWithLlm ParsingMode = "parse_document_with_llm"
const ParsingModeParseDocumentWithLvm ParsingMode = "parse_document_with_lvm"
const ParsingModeParseDocumentWithAgent ParsingMode = "parse_document_with_agent"
ParsingInstruction stringoptional
PreciseBoundingBox booloptional
PremiumMode booloptional
PresentationOutOfBoundsContent booloptional
PresentationSkipEmbeddedData booloptional
PreserveLayoutAlignmentAcrossPages booloptional
PreserveVerySmallText booloptional
Preset stringoptional
Priority LlamaParseParametersPriorityoptional

The priority for the request. This field may be ignored or overwritten depending on the organization tier.

One of the following:
const LlamaParseParametersPriorityLow LlamaParseParametersPriority = "low"
const LlamaParseParametersPriorityMedium LlamaParseParametersPriority = "medium"
const LlamaParseParametersPriorityHigh LlamaParseParametersPriority = "high"
const LlamaParseParametersPriorityCritical LlamaParseParametersPriority = "critical"
ProjectID stringoptional
RemoveHiddenText booloptional
ReplaceFailedPageMode FailPageModeoptional

Enum for representing the different available page error handling modes.

One of the following:
const FailPageModeRawText FailPageMode = "raw_text"
const FailPageModeBlankPage FailPageMode = "blank_page"
const FailPageModeErrorMessage FailPageMode = "error_message"
ReplaceFailedPageWithErrorMessagePrefix stringoptional
ReplaceFailedPageWithErrorMessageSuffix stringoptional
SaveImages booloptional
SkipDiagonalText booloptional
SpecializedChartParsingAgentic booloptional
SpecializedChartParsingEfficient booloptional
SpecializedChartParsingPlus booloptional
SpecializedImageParsing booloptional
SpreadsheetExtractSubTables booloptional
SpreadsheetForceFormulaComputation booloptional
SpreadsheetIncludeHiddenSheets booloptional
StrictModeBuggyFont booloptional
StrictModeImageExtraction booloptional
StrictModeImageOcr booloptional
StrictModeReconstruction booloptional
StructuredOutput booloptional
StructuredOutputJsonSchema stringoptional
StructuredOutputJsonSchemaName stringoptional
SystemPrompt stringoptional
SystemPromptAppend stringoptional
TakeScreenshot booloptional
TargetPages stringoptional
Tier stringoptional
UseVendorMultimodalModel booloptional
UserPrompt stringoptional
VendorMultimodalAPIKey stringoptional
VendorMultimodalModelName stringoptional
Version stringoptional
WebhookConfigurations []LlamaParseParametersWebhookConfigurationRespoptional

Outbound webhook endpoints to notify on job status changes

WebhookEvents []stringoptional

Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.

One of the following:
const LlamaParseParametersWebhookConfigurationWebhookEventExtractPending LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.success"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractError LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.error"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractPartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventExtractCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "extract.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventParsePending LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventParseRunning LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.running"
const LlamaParseParametersWebhookConfigurationWebhookEventParseSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.success"
const LlamaParseParametersWebhookConfigurationWebhookEventParseError LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.error"
const LlamaParseParametersWebhookConfigurationWebhookEventParsePartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventParseCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "parse.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyPending LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.pending"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifySuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.success"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyError LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.error"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyPartialSuccess LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.partial_success"
const LlamaParseParametersWebhookConfigurationWebhookEventClassifyCancelled LlamaParseParametersWebhookConfigurationWebhookEvent = "classify.cancelled"
const LlamaParseParametersWebhookConfigurationWebhookEventUnmappedEvent LlamaParseParametersWebhookConfigurationWebhookEvent = "unmapped_event"
WebhookHeaders map[string, string]optional

Custom HTTP headers sent with each webhook request (e.g. auth tokens)

WebhookOutputFormat stringoptional

Response format sent to the webhook: ‘string’ (default) or ‘json’

WebhookURL stringoptional

URL to receive webhook POST notifications

WebhookURL stringoptional
ManagedPipelineID stringoptional

The ID of the ManagedPipeline this playground pipeline is linked to.

formatuuid
MetadataConfig PipelineMetadataConfigoptional

Metadata configuration for the pipeline.

ExcludedEmbedMetadataKeys []stringoptional

List of metadata keys to exclude from embeddings

ExcludedLlmMetadataKeys []stringoptional

List of metadata keys to exclude from LLM during retrieval

PipelineType PipelineTypeoptional

Type of pipeline. Either PLAYGROUND or MANAGED.

One of the following:
const PipelineTypePlayground PipelineType = "PLAYGROUND"
const PipelineTypeManaged PipelineType = "MANAGED"
PresetRetrievalParameters PresetRetrievalParamsRespoptional

Preset retrieval parameters for the pipeline.

Alpha float64optional

Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.

maximum1
minimum0
ClassName stringoptional
DenseSimilarityCutoff float64optional

Minimum similarity score wrt query for retrieval

maximum1
minimum0
DenseSimilarityTopK int64optional

Number of nodes for dense retrieval.

maximum100
minimum1
EnableReranking booloptional

Enable reranking for retrieval

FilesTopK int64optional

Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).

maximum5
minimum1
RerankTopN int64optional

Number of reranked nodes for returning.

maximum100
minimum1
RetrievalMode RetrievalModeoptional

The retrieval mode for the query.

One of the following:
const RetrievalModeChunks RetrievalMode = "chunks"
const RetrievalModeFilesViaMetadata RetrievalMode = "files_via_metadata"
const RetrievalModeFilesViaContent RetrievalMode = "files_via_content"
const RetrievalModeAutoRouted RetrievalMode = "auto_routed"
DeprecatedRetrieveImageNodes booloptional

Whether to retrieve image nodes.

RetrievePageFigureNodes booloptional

Whether to retrieve page figure nodes.

RetrievePageScreenshotNodes booloptional

Whether to retrieve page screenshot nodes.

SearchFilters MetadataFiltersoptional

Metadata filters for vector stores.

Filters []MetadataFiltersFilterUnion
One of the following:
type MetadataFiltersFilterMetadataFilter struct{…}

Comprehensive metadata filter for vector stores to support more operators.

Value uses Strict types, as int, float and str are compatible types and were all converted to string before.

See: https://docs.pydantic.dev/latest/usage/types/#strict-types

Key string
Value MetadataFiltersFilterMetadataFilterValueUnion
One of the following:
float64
string
type MetadataFiltersFilterMetadataFilterValueArray []string
type MetadataFiltersFilterMetadataFilterValueArray []float64
type MetadataFiltersFilterMetadataFilterValueArray []int64
Operator stringoptional

Vector store filter operator.

One of the following:
const MetadataFiltersFilterMetadataFilterOperatorEquals MetadataFiltersFilterMetadataFilterOperator = "=="
const MetadataFiltersFilterMetadataFilterOperatorGreater MetadataFiltersFilterMetadataFilterOperator = ">"
const MetadataFiltersFilterMetadataFilterOperatorLess MetadataFiltersFilterMetadataFilterOperator = "<"
const MetadataFiltersFilterMetadataFilterOperatorNotEquals MetadataFiltersFilterMetadataFilterOperator = "!="
const MetadataFiltersFilterMetadataFilterOperatorGreaterOrEquals MetadataFiltersFilterMetadataFilterOperator = ">="
const MetadataFiltersFilterMetadataFilterOperatorLessOrEquals MetadataFiltersFilterMetadataFilterOperator = "<="
const MetadataFiltersFilterMetadataFilterOperatorIn MetadataFiltersFilterMetadataFilterOperator = "in"
const MetadataFiltersFilterMetadataFilterOperatorNin MetadataFiltersFilterMetadataFilterOperator = "nin"
const MetadataFiltersFilterMetadataFilterOperatorAny MetadataFiltersFilterMetadataFilterOperator = "any"
const MetadataFiltersFilterMetadataFilterOperatorAll MetadataFiltersFilterMetadataFilterOperator = "all"
const MetadataFiltersFilterMetadataFilterOperatorTextMatch MetadataFiltersFilterMetadataFilterOperator = "text_match"
const MetadataFiltersFilterMetadataFilterOperatorTextMatchInsensitive MetadataFiltersFilterMetadataFilterOperator = "text_match_insensitive"
const MetadataFiltersFilterMetadataFilterOperatorContains MetadataFiltersFilterMetadataFilterOperator = "contains"
const MetadataFiltersFilterMetadataFilterOperatorIsEmpty MetadataFiltersFilterMetadataFilterOperator = "is_empty"
type MetadataFilters MetadataFilters

Metadata filters for vector stores.

Condition MetadataFiltersConditionoptional

Vector store filter conditions to combine different filters.

One of the following:
const MetadataFiltersConditionAnd MetadataFiltersCondition = "and"
const MetadataFiltersConditionOr MetadataFiltersCondition = "or"
const MetadataFiltersConditionNot MetadataFiltersCondition = "not"
SearchFiltersInferenceSchema map[string, PresetRetrievalParamsSearchFiltersInferenceSchemaUnionResp]optional

JSON Schema that will be used to infer search_filters. Omit or leave as null to skip inference.

One of the following:
type PresetRetrievalParamsSearchFiltersInferenceSchemaMap map[string, any]
type PresetRetrievalParamsSearchFiltersInferenceSchemaArray []any
string
float64
bool
SparseSimilarityTopK int64optional

Number of nodes for sparse retrieval.

maximum100
minimum1
SparseModelConfig SparseModelConfigoptional

Configuration for sparse embedding models used in hybrid search.

This allows users to choose between Splade and BM25 models for sparse retrieval in managed data sinks.

ClassName stringoptional
ModelType SparseModelConfigModelTypeoptional

The sparse model type to use. ‘bm25’ uses Qdrant’s FastEmbed BM25 model (default for new pipelines), ‘splade’ uses HuggingFace Splade model, ‘auto’ selects based on deployment mode (BYOC uses term frequency, Cloud uses Splade).

One of the following:
const SparseModelConfigModelTypeSplade SparseModelConfigModelType = "splade"
const SparseModelConfigModelTypeBm25 SparseModelConfigModelType = "bm25"
const SparseModelConfigModelTypeAuto SparseModelConfigModelType = "auto"
Status stringoptional

Status of the pipeline deployment.

TransformConfig PipelineCreateTransformConfigUnionoptional

Configuration for the transformation.

One of the following:
type AutoTransformConfig struct{…}
ChunkOverlap int64optional

Chunk overlap for the transformation.

ChunkSize int64optional

Chunk size for the transformation.

exclusiveMinimum0
Mode AutoTransformConfigModeoptional
type AdvancedModeTransformConfig struct{…}
ChunkingConfig AdvancedModeTransformConfigChunkingConfigUnionoptional

Configuration for the chunking.

One of the following:
type AdvancedModeTransformConfigChunkingConfigNoneChunkingConfig struct{…}
Mode stringoptional
type AdvancedModeTransformConfigChunkingConfigCharacterChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
type AdvancedModeTransformConfigChunkingConfigTokenChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
Separator stringoptional
type AdvancedModeTransformConfigChunkingConfigSentenceChunkingConfig struct{…}
ChunkOverlap int64optional
ChunkSize int64optional
Mode stringoptional
ParagraphSeparator stringoptional
Separator stringoptional
type AdvancedModeTransformConfigChunkingConfigSemanticChunkingConfig struct{…}
BreakpointPercentileThreshold int64optional
BufferSize int64optional
Mode stringoptional
Mode AdvancedModeTransformConfigModeoptional
SegmentationConfig AdvancedModeTransformConfigSegmentationConfigUnionoptional

Configuration for the segmentation.

One of the following:
type AdvancedModeTransformConfigSegmentationConfigNoneSegmentationConfig struct{…}
Mode stringoptional
type AdvancedModeTransformConfigSegmentationConfigPageSegmentationConfig struct{…}
Mode stringoptional
PageSeparator stringoptional
type AdvancedModeTransformConfigSegmentationConfigElementSegmentationConfig struct{…}
Mode stringoptional
type PipelineMetadataConfig struct{…}
ExcludedEmbedMetadataKeys []stringoptional

List of metadata keys to exclude from embeddings

ExcludedLlmMetadataKeys []stringoptional

List of metadata keys to exclude from LLM during retrieval

type PipelineType string

Enum for representing the type of a pipeline

One of the following:
const PipelineTypePlayground PipelineType = "PLAYGROUND"
const PipelineTypeManaged PipelineType = "MANAGED"
type PresetRetrievalParamsResp struct{…}

Schema for the search params for an retrieval execution that can be preset for a pipeline.

Alpha float64optional

Alpha value for hybrid retrieval to determine the weights between dense and sparse retrieval. 0 is sparse retrieval and 1 is dense retrieval.

maximum1
minimum0
ClassName stringoptional
DenseSimilarityCutoff float64optional

Minimum similarity score wrt query for retrieval

maximum1
minimum0
DenseSimilarityTopK int64optional

Number of nodes for dense retrieval.

maximum100
minimum1
EnableReranking booloptional

Enable reranking for retrieval

FilesTopK int64optional

Number of files to retrieve (only for retrieval mode files_via_metadata and files_via_content).

maximum5
minimum1
RerankTopN int64optional

Number of reranked nodes for returning.

maximum100
minimum1
RetrievalMode RetrievalModeoptional

The retrieval mode for the query.

One of the following:
const RetrievalModeChunks RetrievalMode = "chunks"
const RetrievalModeFilesViaMetadata RetrievalMode = "files_via_metadata"
const RetrievalModeFilesViaContent RetrievalMode = "files_via_content"
const RetrievalModeAutoRouted RetrievalMode = "auto_routed"
DeprecatedRetrieveImageNodes booloptional

Whether to retrieve image nodes.

RetrievePageFigureNodes booloptional

Whether to retrieve page figure nodes.

RetrievePageScreenshotNodes booloptional

Whether to retrieve page screenshot nodes.

SearchFilters MetadataFiltersoptional

Metadata filters for vector stores.

Filters []MetadataFiltersFilterUnion
One of the following:
type MetadataFiltersFilterMetadataFilter struct{…}

Comprehensive metadata filter for vector stores to support more operators.

Value uses Strict types, as int, float and str are compatible types and were all converted to string before.

See: https://docs.pydantic.dev/latest/usage/types/#strict-types

Key string
Value MetadataFiltersFilterMetadataFilterValueUnion
One of the following:
float64
string
type MetadataFiltersFilterMetadataFilterValueArray []string
type MetadataFiltersFilterMetadataFilterValueArray []float64
type MetadataFiltersFilterMetadataFilterValueArray []int64
Operator stringoptional

Vector store filter operator.

One of the following:
const MetadataFiltersFilterMetadataFilterOperatorEquals MetadataFiltersFilterMetadataFilterOperator = "=="
const MetadataFiltersFilterMetadataFilterOperatorGreater MetadataFiltersFilterMetadataFilterOperator = ">"
const MetadataFiltersFilterMetadataFilterOperatorLess MetadataFiltersFilterMetadataFilterOperator = "<"
const MetadataFiltersFilterMetadataFilterOperatorNotEquals MetadataFiltersFilterMetadataFilterOperator = "!="
const MetadataFiltersFilterMetadataFilterOperatorGreaterOrEquals MetadataFiltersFilterMetadataFilterOperator = ">="
const MetadataFiltersFilterMetadataFilterOperatorLessOrEquals MetadataFiltersFilterMetadataFilterOperator = "<="
const MetadataFiltersFilterMetadataFilterOperatorIn MetadataFiltersFilterMetadataFilterOperator = "in"
const MetadataFiltersFilterMetadataFilterOperatorNin MetadataFiltersFilterMetadataFilterOperator = "nin"
const MetadataFiltersFilterMetadataFilterOperatorAny MetadataFiltersFilterMetadataFilterOperator = "any"
const MetadataFiltersFilterMetadataFilterOperatorAll MetadataFiltersFilterMetadataFilterOperator = "all"
const MetadataFiltersFilterMetadataFilterOperatorTextMatch MetadataFiltersFilterMetadataFilterOperator = "text_match"
const MetadataFiltersFilterMetadataFilterOperatorTextMatchInsensitive MetadataFiltersFilterMetadataFilterOperator = "text_match_insensitive"
const MetadataFiltersFilterMetadataFilterOperatorContains MetadataFiltersFilterMetadataFilterOperator = "contains"
const MetadataFiltersFilterMetadataFilterOperatorIsEmpty MetadataFiltersFilterMetadataFilterOperator = "is_empty"
type MetadataFilters MetadataFilters

Metadata filters for vector stores.

Condition MetadataFiltersConditionoptional

Vector store filter conditions to combine different filters.

One of the following:
const MetadataFiltersConditionAnd MetadataFiltersCondition = "and"
const MetadataFiltersConditionOr MetadataFiltersCondition = "or"
const MetadataFiltersConditionNot MetadataFiltersCondition = "not"
SearchFiltersInferenceSchema map[string, PresetRetrievalParamsSearchFiltersInferenceSchemaUnionResp]optional

JSON Schema that will be used to infer search_filters. Omit or leave as null to skip inference.

One of the following:
type PresetRetrievalParamsSearchFiltersInferenceSchemaMap map[string, any]
type PresetRetrievalParamsSearchFiltersInferenceSchemaArray []any
string
float64
bool
SparseSimilarityTopK int64optional

Number of nodes for sparse retrieval.

maximum100
minimum1
type RetrievalMode string
One of the following:
const RetrievalModeChunks RetrievalMode = "chunks"
const RetrievalModeFilesViaMetadata RetrievalMode = "files_via_metadata"
const RetrievalModeFilesViaContent RetrievalMode = "files_via_content"
const RetrievalModeAutoRouted RetrievalMode = "auto_routed"
type SparseModelConfig struct{…}

Configuration for sparse embedding models used in hybrid search.

This allows users to choose between Splade and BM25 models for sparse retrieval in managed data sinks.

ClassName stringoptional
ModelType SparseModelConfigModelTypeoptional

The sparse model type to use. ‘bm25’ uses Qdrant’s FastEmbed BM25 model (default for new pipelines), ‘splade’ uses HuggingFace Splade model, ‘auto’ selects based on deployment mode (BYOC uses term frequency, Cloud uses Splade).

One of the following:
const SparseModelConfigModelTypeSplade SparseModelConfigModelType = "splade"
const SparseModelConfigModelTypeBm25 SparseModelConfigModelType = "bm25"
const SparseModelConfigModelTypeAuto SparseModelConfigModelType = "auto"
type VertexAIEmbeddingConfig struct{…}
Component VertexTextEmbeddingoptional

Configuration for the VertexAI embedding model.

ClientEmail string

The client email for the VertexAI credentials.

Location string

The default location to use when making API calls.

PrivateKey string

The private key for the VertexAI credentials.

PrivateKeyID string

The private key ID for the VertexAI credentials.

Project string

The default GCP project to use when making Vertex API calls.

TokenUri string

The token URI for the VertexAI credentials.

AdditionalKwargs map[string, any]optional

Additional kwargs for the Vertex.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbedMode VertexTextEmbeddingEmbedModeoptional

The embedding mode to use.

One of the following:
const VertexTextEmbeddingEmbedModeDefault VertexTextEmbeddingEmbedMode = "default"
const VertexTextEmbeddingEmbedModeClassification VertexTextEmbeddingEmbedMode = "classification"
const VertexTextEmbeddingEmbedModeClustering VertexTextEmbeddingEmbedMode = "clustering"
const VertexTextEmbeddingEmbedModeSimilarity VertexTextEmbeddingEmbedMode = "similarity"
const VertexTextEmbeddingEmbedModeRetrieval VertexTextEmbeddingEmbedMode = "retrieval"
ModelName stringoptional

The modelId of the VertexAI model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

Type VertexAIEmbeddingConfigTypeoptional

Type of the embedding model.

type VertexTextEmbedding struct{…}
ClientEmail string

The client email for the VertexAI credentials.

Location string

The default location to use when making API calls.

PrivateKey string

The private key for the VertexAI credentials.

PrivateKeyID string

The private key ID for the VertexAI credentials.

Project string

The default GCP project to use when making Vertex API calls.

TokenUri string

The token URI for the VertexAI credentials.

AdditionalKwargs map[string, any]optional

Additional kwargs for the Vertex.

ClassName stringoptional
EmbedBatchSize int64optional

The batch size for embedding calls.

maximum2048
exclusiveMinimum0
EmbedMode VertexTextEmbeddingEmbedModeoptional

The embedding mode to use.

One of the following:
const VertexTextEmbeddingEmbedModeDefault VertexTextEmbeddingEmbedMode = "default"
const VertexTextEmbeddingEmbedModeClassification VertexTextEmbeddingEmbedMode = "classification"
const VertexTextEmbeddingEmbedModeClustering VertexTextEmbeddingEmbedMode = "clustering"
const VertexTextEmbeddingEmbedModeSimilarity VertexTextEmbeddingEmbedMode = "similarity"
const VertexTextEmbeddingEmbedModeRetrieval VertexTextEmbeddingEmbedMode = "retrieval"
ModelName stringoptional

The modelId of the VertexAI model to use.

NumWorkers int64optional

The number of workers to use for async embedding calls.

PipelinesSync

Sync Pipeline
client.Pipelines.Sync.New(ctx, pipelineID) (*Pipeline, error)
POST/api/v1/pipelines/{pipeline_id}/sync
Cancel Pipeline Sync
client.Pipelines.Sync.Cancel(ctx, pipelineID) (*Pipeline, error)
POST/api/v1/pipelines/{pipeline_id}/sync/cancel

PipelinesData Sources

List Pipeline Data Sources
client.Pipelines.DataSources.GetDataSources(ctx, pipelineID) (*[]PipelineDataSource, error)
GET/api/v1/pipelines/{pipeline_id}/data-sources
Add Data Sources To Pipeline
client.Pipelines.DataSources.UpdateDataSources(ctx, pipelineID, body) (*[]PipelineDataSource, error)
PUT/api/v1/pipelines/{pipeline_id}/data-sources
Update Pipeline Data Source
client.Pipelines.DataSources.Update(ctx, dataSourceID, params) (*PipelineDataSource, error)
PUT/api/v1/pipelines/{pipeline_id}/data-sources/{data_source_id}
Get Pipeline Data Source Status
client.Pipelines.DataSources.GetStatus(ctx, dataSourceID, query) (*ManagedIngestionStatusResponse, error)
GET/api/v1/pipelines/{pipeline_id}/data-sources/{data_source_id}/status
Sync Pipeline Data Source
client.Pipelines.DataSources.Sync(ctx, dataSourceID, params) (*Pipeline, error)
POST/api/v1/pipelines/{pipeline_id}/data-sources/{data_source_id}/sync
ModelsExpand Collapse
type PipelineDataSource struct{…}

Schema for a data source in a pipeline.

ID string

Unique identifier

formatuuid
Component PipelineDataSourceComponentUnion

Component that implements the data source

One of the following:
type PipelineDataSourceComponentMap map[string, any]
type CloudS3DataSource struct{…}
Bucket string

The name of the S3 bucket to read from.

AwsAccessID stringoptional

The AWS access ID to use for authentication.

AwsAccessSecret stringoptional

The AWS access secret to use for authentication.

formatpassword
ClassName stringoptional
Prefix stringoptional

The prefix of the S3 objects to read from.

RegexPattern stringoptional

The regex pattern to filter S3 objects. Must be a valid regex pattern.

S3EndpointURL stringoptional

The S3 endpoint URL to use for authentication.

SupportsAccessControl booloptional
type CloudAzStorageBlobDataSource struct{…}
AccountURL string

The Azure Storage Blob account URL to use for authentication.

ContainerName string

The name of the Azure Storage Blob container to read from.

AccountKey stringoptional

The Azure Storage Blob account key to use for authentication.

formatpassword
AccountName stringoptional

The Azure Storage Blob account name to use for authentication.

Blob stringoptional

The blob name to read from.

ClassName stringoptional
ClientID stringoptional

The Azure AD client ID to use for authentication.

ClientSecret stringoptional

The Azure AD client secret to use for authentication.

formatpassword
Prefix stringoptional

The prefix of the Azure Storage Blob objects to read from.

SupportsAccessControl booloptional
TenantID stringoptional

The Azure AD tenant ID to use for authentication.

type CloudGoogleDriveDataSource struct{…}
FolderID string

The ID of the Google Drive folder to read from.

ClassName stringoptional
ServiceAccountKey map[string, string]optional

A dictionary containing secret values

SupportsAccessControl booloptional
type CloudOneDriveDataSource struct{…}
ClientID string

The client ID to use for authentication.

ClientSecret string

The client secret to use for authentication.

formatpassword
TenantID string

The tenant ID to use for authentication.

UserPrincipalName string

The user principal name to use for authentication.

ClassName stringoptional
FolderID stringoptional

The ID of the OneDrive folder to read from.

FolderPath stringoptional

The path of the OneDrive folder to read from.

RequiredExts []stringoptional

The list of required file extensions.

SupportsAccessControl booloptional
type CloudSharepointDataSource struct{…}
ClientID string

The client ID to use for authentication.

ClientSecret string

The client secret to use for authentication.

formatpassword
TenantID string

The tenant ID to use for authentication.

ClassName stringoptional
DriveName stringoptional

The name of the Sharepoint drive to read from.

ExcludePathPatterns []stringoptional

List of regex patterns for file paths to exclude. Files whose paths (including filename) match any pattern will be excluded. Example: [‘/temp/’, ‘/backup/’, ‘.git/’, ‘.tmp$’, ’^~’]

FolderID stringoptional

The ID of the Sharepoint folder to read from.

FolderPath stringoptional

The path of the Sharepoint folder to read from.

GetPermissions booloptional

Whether to get permissions for the sharepoint site.

IncludePathPatterns []stringoptional

List of regex patterns for file paths to include. Full paths (including filename) must match at least one pattern to be included. Example: [‘/reports/’, ‘/docs/..pdf$’, ‘^Report..pdf$’]

RequiredExts []stringoptional

The list of required file extensions.

SiteID stringoptional

The ID of the SharePoint site to download from.

SiteName stringoptional

The name of the SharePoint site to download from.

SupportsAccessControl booloptional
type CloudSlackDataSource struct{…}
SlackToken string

Slack Bot Token.

formatpassword
ChannelIDs stringoptional

Slack Channel.

ChannelPatterns stringoptional

Slack Channel name pattern.

ClassName stringoptional
EarliestDate stringoptional

Earliest date.

EarliestDateTimestamp float64optional

Earliest date timestamp.

LatestDate stringoptional

Latest date.

LatestDateTimestamp float64optional

Latest date timestamp.

SupportsAccessControl booloptional
type CloudNotionPageDataSource struct{…}
IntegrationToken string

The integration token to use for authentication.

formatpassword
ClassName stringoptional
DatabaseIDs stringoptional

The Notion Database Id to read content from.

PageIDs stringoptional

The Page ID’s of the Notion to read from.

SupportsAccessControl booloptional
type CloudConfluenceDataSource struct{…}
AuthenticationMechanism string

Type of Authentication for connecting to Confluence APIs.

ServerURL string

The server URL of the Confluence instance.

APIToken stringoptional

The API token to use for authentication.

formatpassword
ClassName stringoptional
Cql stringoptional

The CQL query to use for fetching pages.

FailureHandling FailureHandlingConfigoptional

Configuration for handling failures during processing. Key-value object controlling failure handling behaviors.

Example: { “skip_list_failures”: true }

Currently supports:

  • skip_list_failures: Skip failed batches/lists and continue processing
SkipListFailures booloptional

Whether to skip failed batches/lists and continue processing

IndexRestrictedPages booloptional

Whether to index restricted pages.

KeepMarkdownFormat booloptional

Whether to keep the markdown format.

Label stringoptional

The label to use for fetching pages.

PageIDs stringoptional

The page IDs of the Confluence to read from.

SpaceKey stringoptional

The space key to read from.

SupportsAccessControl booloptional
UserName stringoptional

The username to use for authentication.

type CloudJiraDataSource struct{…}

Cloud Jira Data Source integrating JiraReader.

AuthenticationMechanism string

Type of Authentication for connecting to Jira APIs.

Query string

JQL (Jira Query Language) query to search.

APIToken stringoptional

The API/ Access Token used for Basic, PAT and OAuth2 authentication.

formatpassword
ClassName stringoptional
CloudID stringoptional

The cloud ID, used in case of OAuth2.

Email stringoptional

The email address to use for authentication.

ServerURL stringoptional

The server url for Jira Cloud.

SupportsAccessControl booloptional
type CloudJiraDataSourceV2 struct{…}

Cloud Jira Data Source integrating JiraReaderV2.

AuthenticationMechanism string

Type of Authentication for connecting to Jira APIs.

Query string

JQL (Jira Query Language) query to search.

ServerURL string

The server url for Jira Cloud.

APIToken stringoptional

The API Access Token used for Basic, PAT and OAuth2 authentication.

formatpassword
APIVersion CloudJiraDataSourceV2APIVersionoptional

Jira REST API version to use (2 or 3). 3 supports Atlassian Document Format (ADF).

One of the following:
const CloudJiraDataSourceV2APIVersion2 CloudJiraDataSourceV2APIVersion = "2"
const CloudJiraDataSourceV2APIVersion3 CloudJiraDataSourceV2APIVersion = "3"
ClassName stringoptional
CloudID stringoptional

The cloud ID, used in case of OAuth2.

Email stringoptional

The email address to use for authentication.

Expand stringoptional

Fields to expand in the response.

Fields []stringoptional

List of fields to retrieve from Jira. If None, retrieves all fields.

GetPermissions booloptional

Whether to fetch project role permissions and issue-level security

RequestsPerMinute int64optional

Rate limit for Jira API requests per minute.

SupportsAccessControl booloptional
type CloudBoxDataSource struct{…}
AuthenticationMechanism CloudBoxDataSourceAuthenticationMechanism

The type of authentication to use (Developer Token or CCG)

One of the following:
const CloudBoxDataSourceAuthenticationMechanismDeveloperToken CloudBoxDataSourceAuthenticationMechanism = "developer_token"
const CloudBoxDataSourceAuthenticationMechanismCcg CloudBoxDataSourceAuthenticationMechanism = "ccg"
ClassName stringoptional
ClientID stringoptional

Box API key used for identifying the application the user is authenticating with

ClientSecret stringoptional

Box API secret used for making auth requests.

formatpassword
DeveloperToken stringoptional

Developer token for authentication if authentication_mechanism is ‘developer_token’.

formatpassword
EnterpriseID stringoptional

Box Enterprise ID, if provided authenticates as service.

FolderID stringoptional

The ID of the Box folder to read from.

SupportsAccessControl booloptional
UserID stringoptional

Box User ID, if provided authenticates as user.

DataSourceID string

The ID of the data source.

formatuuid
LastSyncedAt Time

The last time the data source was automatically synced.

formatdate-time
Name string

The name of the data source.

PipelineID string

The ID of the pipeline.

formatuuid
ProjectID string
SourceType PipelineDataSourceSourceType
One of the following:
const PipelineDataSourceSourceTypeS3 PipelineDataSourceSourceType = "S3"
const PipelineDataSourceSourceTypeAzureStorageBlob PipelineDataSourceSourceType = "AZURE_STORAGE_BLOB"
const PipelineDataSourceSourceTypeGoogleDrive PipelineDataSourceSourceType = "GOOGLE_DRIVE"
const PipelineDataSourceSourceTypeMicrosoftOnedrive PipelineDataSourceSourceType = "MICROSOFT_ONEDRIVE"
const PipelineDataSourceSourceTypeMicrosoftSharepoint PipelineDataSourceSourceType = "MICROSOFT_SHAREPOINT"
const PipelineDataSourceSourceTypeSlack PipelineDataSourceSourceType = "SLACK"
const PipelineDataSourceSourceTypeNotionPage PipelineDataSourceSourceType = "NOTION_PAGE"
const PipelineDataSourceSourceTypeConfluence PipelineDataSourceSourceType = "CONFLUENCE"
const PipelineDataSourceSourceTypeJira PipelineDataSourceSourceType = "JIRA"
const PipelineDataSourceSourceTypeJiraV2 PipelineDataSourceSourceType = "JIRA_V2"
const PipelineDataSourceSourceTypeBox PipelineDataSourceSourceType = "BOX"
CreatedAt Timeoptional

Creation datetime

formatdate-time
CustomMetadata map[string, PipelineDataSourceCustomMetadataUnion]optional

Custom metadata that will be present on all data loaded from the data source

One of the following:
type PipelineDataSourceCustomMetadataMap map[string, any]
type PipelineDataSourceCustomMetadataArray []any
string
float64
bool
Status PipelineDataSourceStatusoptional

The status of the data source in the pipeline.

One of the following:
const PipelineDataSourceStatusNotStarted PipelineDataSourceStatus = "NOT_STARTED"
const PipelineDataSourceStatusInProgress PipelineDataSourceStatus = "IN_PROGRESS"
const PipelineDataSourceStatusSuccess PipelineDataSourceStatus = "SUCCESS"
const PipelineDataSourceStatusError PipelineDataSourceStatus = "ERROR"
const PipelineDataSourceStatusCancelled PipelineDataSourceStatus = "CANCELLED"
StatusUpdatedAt Timeoptional

The last time the status was updated.

formatdate-time
SyncInterval float64optional

The interval at which the data source should be synced.

SyncScheduleSetBy stringoptional

The id of the user who set the sync schedule.

UpdatedAt Timeoptional

Update datetime

formatdate-time
VersionMetadata DataSourceReaderVersionMetadataoptional

Version metadata for the data source

ReaderVersion DataSourceReaderVersionMetadataReaderVersionoptional

The version of the reader to use for this data source.

One of the following:
const DataSourceReaderVersionMetadataReaderVersion1_0 DataSourceReaderVersionMetadataReaderVersion = "1.0"
const DataSourceReaderVersionMetadataReaderVersion2_0 DataSourceReaderVersionMetadataReaderVersion = "2.0"
const DataSourceReaderVersionMetadataReaderVersion2_1 DataSourceReaderVersionMetadataReaderVersion = "2.1"

PipelinesImages

List File Page Screenshots
client.Pipelines.Images.ListPageScreenshots(ctx, id, query) (*[]PipelineImageListPageScreenshotsResponse, error)
GET/api/v1/files/{id}/page_screenshots
Get File Page Screenshot
client.Pipelines.Images.GetPageScreenshot(ctx, pageIndex, params) (*PipelineImageGetPageScreenshotResponse, error)
GET/api/v1/files/{id}/page_screenshots/{page_index}
Get File Page Figure
client.Pipelines.Images.GetPageFigure(ctx, figureName, params) (*PipelineImageGetPageFigureResponse, error)
GET/api/v1/files/{id}/page-figures/{page_index}/{figure_name}
List File Pages Figures
client.Pipelines.Images.ListPageFigures(ctx, id, query) (*[]PipelineImageListPageFiguresResponse, error)
GET/api/v1/files/{id}/page-figures

PipelinesFiles

Get Pipeline File Status Counts
client.Pipelines.Files.GetStatusCounts(ctx, pipelineID, query) (*PipelineFileGetStatusCountsResponse, error)
GET/api/v1/pipelines/{pipeline_id}/files/status-counts
Get Pipeline File Status
client.Pipelines.Files.GetStatus(ctx, fileID, query) (*ManagedIngestionStatusResponse, error)
GET/api/v1/pipelines/{pipeline_id}/files/{file_id}/status
Add Files To Pipeline Api
client.Pipelines.Files.New(ctx, pipelineID, body) (*[]PipelineFile, error)
PUT/api/v1/pipelines/{pipeline_id}/files
Update Pipeline File
client.Pipelines.Files.Update(ctx, fileID, params) (*PipelineFile, error)
PUT/api/v1/pipelines/{pipeline_id}/files/{file_id}
Delete Pipeline File
client.Pipelines.Files.Delete(ctx, fileID, body) error
DELETE/api/v1/pipelines/{pipeline_id}/files/{file_id}
List Pipeline Files2
Deprecated
client.Pipelines.Files.List(ctx, pipelineID, query) (*PaginatedPipelineFiles[PipelineFile], error)
GET/api/v1/pipelines/{pipeline_id}/files2
ModelsExpand Collapse
type PipelineFile struct{…}

A file associated with a pipeline.

ID string

Unique identifier for the pipeline file.

formatuuid
PipelineID string

The ID of the pipeline that the file is associated with.

formatuuid
ConfigHash map[string, PipelineFileConfigHashUnion]optional

Hashes for the configuration of the pipeline.

One of the following:
type PipelineFileConfigHashMap map[string, any]
type PipelineFileConfigHashArray []any
string
float64
bool
CreatedAt Timeoptional

When the pipeline file was created.

formatdate-time
CustomMetadata map[string, PipelineFileCustomMetadataUnion]optional

Custom metadata for the file.

One of the following:
type PipelineFileCustomMetadataMap map[string, any]
type PipelineFileCustomMetadataArray []any
string
float64
bool
DataSourceID stringoptional

The ID of the data source that the file belongs to.

formatuuid
ExternalFileID stringoptional

The ID of the file in the external system.

FileID stringoptional

The ID of the file.

formatuuid
FileSize int64optional

Size of the file in bytes.

FileType stringoptional

File type (e.g. pdf, docx, etc.).

IndexedPageCount int64optional

The number of pages that have been indexed for this file.

LastModifiedAt Timeoptional

The last modified time of the file.

formatdate-time
Name stringoptional

Name of the file.

PermissionInfo map[string, PipelineFilePermissionInfoUnion]optional

Permission information for the file.

One of the following:
type PipelineFilePermissionInfoMap map[string, any]
type PipelineFilePermissionInfoArray []any
string
float64
bool
ProjectID stringoptional

The ID of the project that the file belongs to.

formatuuid
ResourceInfo map[string, PipelineFileResourceInfoUnion]optional

Resource information for the file.

One of the following:
type PipelineFileResourceInfoMap map[string, any]
type PipelineFileResourceInfoArray []any
string
float64
bool
Status PipelineFileStatusoptional

Status of the pipeline file.

One of the following:
const PipelineFileStatusNotStarted PipelineFileStatus = "NOT_STARTED"
const PipelineFileStatusInProgress PipelineFileStatus = "IN_PROGRESS"
const PipelineFileStatusSuccess PipelineFileStatus = "SUCCESS"
const PipelineFileStatusError PipelineFileStatus = "ERROR"
const PipelineFileStatusCancelled PipelineFileStatus = "CANCELLED"
StatusUpdatedAt Timeoptional

The last time the status was updated.

formatdate-time
UpdatedAt Timeoptional

When the pipeline file was last updated.

formatdate-time

PipelinesMetadata

Import Pipeline Metadata
client.Pipelines.Metadata.New(ctx, pipelineID, body) (*PipelineMetadataNewResponse, error)
PUT/api/v1/pipelines/{pipeline_id}/metadata
Delete Pipeline Files Metadata
client.Pipelines.Metadata.DeleteAll(ctx, pipelineID) error
DELETE/api/v1/pipelines/{pipeline_id}/metadata

PipelinesDocuments

Create Batch Pipeline Documents
client.Pipelines.Documents.New(ctx, pipelineID, body) (*[]CloudDocument, error)
POST/api/v1/pipelines/{pipeline_id}/documents
Paginated List Pipeline Documents
client.Pipelines.Documents.List(ctx, pipelineID, query) (*PaginatedCloudDocuments[CloudDocument], error)
GET/api/v1/pipelines/{pipeline_id}/documents/paginated
Get Pipeline Document
client.Pipelines.Documents.Get(ctx, documentID, query) (*CloudDocument, error)
GET/api/v1/pipelines/{pipeline_id}/documents/{document_id}
Delete Pipeline Document
client.Pipelines.Documents.Delete(ctx, documentID, body) error
DELETE/api/v1/pipelines/{pipeline_id}/documents/{document_id}
Get Pipeline Document Status
client.Pipelines.Documents.GetStatus(ctx, documentID, query) (*ManagedIngestionStatusResponse, error)
GET/api/v1/pipelines/{pipeline_id}/documents/{document_id}/status
Sync Pipeline Document
client.Pipelines.Documents.Sync(ctx, documentID, body) (*PipelineDocumentSyncResponse, error)
POST/api/v1/pipelines/{pipeline_id}/documents/{document_id}/sync
List Pipeline Document Chunks
client.Pipelines.Documents.GetChunks(ctx, documentID, query) (*[]TextNode, error)
GET/api/v1/pipelines/{pipeline_id}/documents/{document_id}/chunks
Upsert Batch Pipeline Documents
client.Pipelines.Documents.Upsert(ctx, pipelineID, body) (*[]CloudDocument, error)
PUT/api/v1/pipelines/{pipeline_id}/documents
ModelsExpand Collapse
type CloudDocument struct{…}

Cloud document stored in S3.

ID string
Metadata map[string, any]
Text string
ExcludedEmbedMetadataKeys []stringoptional
ExcludedLlmMetadataKeys []stringoptional
PagePositions []int64optional

indices in the CloudDocument.text where a new page begins. e.g. Second page starts at index specified by page_positions[1].

StatusMetadata map[string, any]optional
type CloudDocumentCreate struct{…}

Create a new cloud document.

Metadata map[string, any]
Text string
ID stringoptional
ExcludedEmbedMetadataKeys []stringoptional
ExcludedLlmMetadataKeys []stringoptional
PagePositions []int64optional

indices in the CloudDocument.text where a new page begins. e.g. Second page starts at index specified by page_positions[1].

type TextNode struct{…}

Provided for backward compatibility.

ClassName stringoptional
Embedding []float64optional

Embedding of the node.

EndCharIdx int64optional

End char index of the node.

ExcludedEmbedMetadataKeys []stringoptional

Metadata keys that are excluded from text for the embed model.

ExcludedLlmMetadataKeys []stringoptional

Metadata keys that are excluded from text for the LLM.

ExtraInfo map[string, any]optional

A flat dictionary of metadata fields

ID stringoptional

Unique ID of the node.

MetadataSeperator stringoptional

Separator between metadata fields when converting to string.

MetadataTemplate stringoptional

Template for how metadata is formatted, with {key} and {value} placeholders.

Mimetype stringoptional

MIME type of the node content.

Relationships map[string, TextNodeRelationshipUnion]optional

A mapping of relationships to other node information.

One of the following:
type TextNodeRelationshipRelatedNodeInfo struct{…}
NodeID string
ClassName stringoptional
Hash stringoptional
Metadata map[string, any]optional
NodeType stringoptional
One of the following:
string
One of the following:
const TextNodeRelationshipRelatedNodeInfoNodeType1 TextNodeRelationshipRelatedNodeInfoNodeType = "1"
const TextNodeRelationshipRelatedNodeInfoNodeType2 TextNodeRelationshipRelatedNodeInfoNodeType = "2"
const TextNodeRelationshipRelatedNodeInfoNodeType3 TextNodeRelationshipRelatedNodeInfoNodeType = "3"
const TextNodeRelationshipRelatedNodeInfoNodeType4 TextNodeRelationshipRelatedNodeInfoNodeType = "4"
const TextNodeRelationshipRelatedNodeInfoNodeType5 TextNodeRelationshipRelatedNodeInfoNodeType = "5"
string
type TextNodeRelationshipArray []TextNodeRelationshipArrayItem
NodeID string
ClassName stringoptional
Hash stringoptional
Metadata map[string, any]optional
NodeType stringoptional
One of the following:
string
One of the following:
const TextNodeRelationshipArrayItemNodeType1 TextNodeRelationshipArrayItemNodeType = "1"
const TextNodeRelationshipArrayItemNodeType2 TextNodeRelationshipArrayItemNodeType = "2"
const TextNodeRelationshipArrayItemNodeType3 TextNodeRelationshipArrayItemNodeType = "3"
const TextNodeRelationshipArrayItemNodeType4 TextNodeRelationshipArrayItemNodeType = "4"
const TextNodeRelationshipArrayItemNodeType5 TextNodeRelationshipArrayItemNodeType = "5"
string
StartCharIdx int64optional

Start char index of the node.

Text stringoptional

Text content of the node.

TextTemplate stringoptional

Template for how text is formatted, with {content} and {metadata_str} placeholders.