For the following example to work, save your API key to the NEBIUS_API_KEY environment variable.

Create a batch: POST

Request:
client.batches.create(
    input_file_id=batch_requests.id,
    endpoint="/v1/chat/completions",
    completion_window="24h",
    metadata={
        "description": "Asynchronous job"
    }
)

Get batch info: GET

Request:
import os
from openai import OpenAI

client = OpenAI(
    base_url="https://api.studio.nebius.com/v1/",
    api_key=os.environ.get("NEBIUS_API_KEY"),
)

client.batches.retrieve("batch_123")

List batches: GET

Request:
import os
from openai import OpenAI

client = OpenAI(
    base_url="https://api.studio.nebius.com/v1/",
    api_key=os.environ.get("NEBIUS_API_KEY"),
)

client.batches.list()

Cancel a batch: POST

Request:
client.batches.cancel("batch_123")
To learn more about batch inference, see Batch inference. For detailed field descriptions, see the API reference.