Beta
BetaAgent Data
Get Agent Data
Update Agent Data
Delete Agent Data
Create Agent Data
Search Agent Data
Aggregate Agent Data
Delete Agent Data By Query
ModelsExpand Collapse
BetaSheets
Create Spreadsheet Job
List Spreadsheet Jobs
Get Spreadsheet Job
Get Result Region
Delete Spreadsheet Job
ModelsExpand Collapse
SheetsJob = object { id, config, created_at, 10 more }
A spreadsheet parsing job
config: SheetsParsingConfig { extraction_range, flatten_hierarchical_tables, generate_additional_metadata, 5 more }
Configuration for the parsing job
extraction_range: optional string
A1 notation of the range to extract a single region from. If None, the entire sheet is used.
flatten_hierarchical_tables: optional boolean
Return a flattened dataframe when a detected table is recognized as hierarchical.
generate_additional_metadata: optional boolean
Whether to generate additional metadata (title, description) for each extracted region.
include_hidden_cells: optional boolean
Whether to include hidden cells when extracting regions from the spreadsheet.
sheet_names: optional array of string
The names of the sheets to extract regions from. If empty, all sheets will be processed.
specialization: optional string
Optional specialization mode for domain-specific extraction. Supported values: ‘financial-standard’, ‘financial-enhanced’, ‘financial-precise’. Default None uses the general-purpose pipeline.
Schema for a file.
expires_at: optional string
The expiration date for the file. Files past this date can be deleted.
purpose: optional string
The intended purpose of the file (e.g., ‘user_data’, ‘parse’, ‘extract’, ‘split’, ‘classify’)
regions: optional array of object { location, region_type, sheet_name, 3 more }
All extracted regions (populated when job is complete)
SheetsParsingConfig = object { extraction_range, flatten_hierarchical_tables, generate_additional_metadata, 5 more }
Configuration for spreadsheet parsing and region extraction
extraction_range: optional string
A1 notation of the range to extract a single region from. If None, the entire sheet is used.
flatten_hierarchical_tables: optional boolean
Return a flattened dataframe when a detected table is recognized as hierarchical.
generate_additional_metadata: optional boolean
Whether to generate additional metadata (title, description) for each extracted region.
include_hidden_cells: optional boolean
Whether to include hidden cells when extracting regions from the spreadsheet.
sheet_names: optional array of string
The names of the sheets to extract regions from. If empty, all sheets will be processed.
specialization: optional string
Optional specialization mode for domain-specific extraction. Supported values: ‘financial-standard’, ‘financial-enhanced’, ‘financial-precise’. Default None uses the general-purpose pipeline.
BetaDirectories
Create Directory
List Directories
Get Directory
Update Directory
Delete Directory
ModelsExpand Collapse
DirectoryCreateResponse = object { id, name, project_id, 5 more }
DirectoryListResponse = object { id, name, project_id, 5 more }
DirectoryGetResponse = object { id, name, project_id, 5 more }
DirectoryUpdateResponse = object { id, name, project_id, 5 more }
BetaDirectoriesFiles
Add Directory File
List Directory Files
Get Directory File
Update Directory File
Delete Directory File
Upload File To Directory
ModelsExpand Collapse
BetaBatch
Create Batch Job
List Batch Jobs
Get Batch Job Status
Cancel Batch Job
ModelsExpand Collapse
BatchGetStatusResponse = object { job, progress_percentage }
BetaBatchJob Items
List Batch Job Items
Get Item Processing Results
ModelsExpand Collapse
JobItemListResponse = object { item_id, item_name, status, 7 more }
Detailed information about an item in a batch job.
JobItemGetProcessingResultsResponse = object { item_id, item_name, processing_results }
Response containing all processing results for an item.
processing_results: optional array of object { item_id, job_config, job_type, 5 more }
List of all processing operations performed on this item
job_config: object { correlation_id, job_name, parameters, 6 more } or ClassifyJob { id, project_id, rules, 9 more }
Job configuration used for processing
BatchParseJobRecordCreate = object { correlation_id, job_name, parameters, 6 more }
Batch-specific parse job record for batch processing.
This model contains the metadata and configuration for a batch parse job, but excludes file-specific information. It’s used as input to the batch parent workflow and combined with DirectoryFile data to create full ParseJobRecordCreate instances for each file.
Attributes: job_name: Must be PARSE_RAW_FILE partitions: Partitions for job output location parameters: Generic parse configuration (BatchParseJobConfig) session_id: Upstream request ID for tracking correlation_id: Correlation ID for cross-service tracking parent_job_execution_id: Parent job execution ID if nested user_id: User who created the job project_id: Project this job belongs to webhook_url: Optional webhook URL for job completion notifications
correlation_id: optional string
The correlation ID for this job. Used for tracking the job across services.
parameters: optional object { adaptive_long_table, aggressive_table_extraction, annotate_links, 122 more }
Generic parse job configuration for batch processing.
This model contains the parsing configuration that applies to all files in a batch, but excludes file-specific fields like file_name, file_id, etc. Those file-specific fields are populated from DirectoryFile data when creating individual ParseJobRecordCreate instances for each file.
The fields in this model should be generic settings that apply uniformly to all files being processed in the batch.
output_s3_path_prefix: optional string
If specified, llamaParse will save the output to the specified path. All output file will use this ‘prefix’ should be a valid s3:// url
webhook_configurations: optional array of object { webhook_events, webhook_headers, webhook_output_format, webhook_url }
Outbound webhook endpoints to notify on job status changes
webhook_events: optional array of "extract.pending" or "extract.success" or "extract.error" or 14 more
Events to subscribe to (e.g. ‘parse.success’, ‘extract.error’). If null, all events are delivered.
webhook_headers: optional map[string]
Custom HTTP headers sent with each webhook request (e.g. auth tokens)
partitions: optional map[string]
The partitions for this execution. Used for determining where to save job output.