Get Batch Job Status
beta.batch.get_status(strjob_id, BatchGetStatusParams**kwargs) -> BatchGetStatusResponse
GET/api/v1/beta/batch-processing/{job_id}
Get detailed status of a batch processing job.
Returns current progress, file counts, and estimated completion time.
Parameters
job_id: str
organization_id: Optional[str]
project_id: Optional[str]
Returns
Get Batch Job Status
import os
from llama_cloud import LlamaCloud
client = LlamaCloud(
api_key=os.environ.get("LLAMA_CLOUD_API_KEY"), # This is the default and can be omitted
)
response = client.beta.batch.get_status(
job_id="job_id",
)
print(response.job){
"job": {
"id": "id",
"job_type": "parse",
"project_id": "project_id",
"status": "pending",
"total_items": 0,
"completed_at": "2019-12-27T18:11:19.117Z",
"created_at": "2019-12-27T18:11:19.117Z",
"directory_id": "directory_id",
"effective_at": "2019-12-27T18:11:19.117Z",
"error_message": "error_message",
"failed_items": 0,
"job_record_id": "job_record_id",
"processed_items": 0,
"skipped_items": 0,
"started_at": "2019-12-27T18:11:19.117Z",
"updated_at": "2019-12-27T18:11:19.117Z",
"workflow_id": "workflow_id"
},
"progress_percentage": 0
}Returns Examples
{
"job": {
"id": "id",
"job_type": "parse",
"project_id": "project_id",
"status": "pending",
"total_items": 0,
"completed_at": "2019-12-27T18:11:19.117Z",
"created_at": "2019-12-27T18:11:19.117Z",
"directory_id": "directory_id",
"effective_at": "2019-12-27T18:11:19.117Z",
"error_message": "error_message",
"failed_items": 0,
"job_record_id": "job_record_id",
"processed_items": 0,
"skipped_items": 0,
"started_at": "2019-12-27T18:11:19.117Z",
"updated_at": "2019-12-27T18:11:19.117Z",
"workflow_id": "workflow_id"
},
"progress_percentage": 0
}