/v1/chat/completions
and /v1/completions
.
endpoint
parameter to /v1/chat/completions
or /v1/completions
for OpenAI API compatibility.custom_id
to locate inference results in the output file after batch completion. Parameters in the body
field of each line are sent as actual inference request parameters to the endpoint.
Below is an example input file containing 2 requests:
batch
.
48h
and is currently non-adjustable.
Code Example
Python
Status | Description |
---|---|
VALIDATING | The input file is being validated before the batch can begin |
PROGRESS | Batch is in progress |
COMPLETED | Batch processing completed successfully |
FAILED | Batch processing failed |
EXPIRED | Batch exceeded deadline |
CANCELLING | Batch is being cancelled |
CANCELLED | Batch was cancelled |
output_file_id
field from the Batch object.
The result output file will be deleted 30 days after the batch inference concludes, so please retrieve it promptly via the interface.
Code Example
Python
deepseek/deepseek-r1-0528
Error Code | Description | Solution |
---|---|---|
400 | Invalid request format | Check JSONL syntax and required fields |
401 | Authentication failed | Verify API key |
404 | Batch not found | Check batch ID |
429 | Rate limit exceeded | Reduce request frequency |
500 | Server error | Contact us |