Create a AI21 inference endpoint
PUT
/_inference/{task_type}/{ai21_inference_id} Create an inference endpoint to perform an inference task with the ai21 service.
Required authorization
- Cluster privileges:
manage_inference
Parameters
path Path Parameters
| Name | Type |
|---|---|
task_type
required
The type of the inference task that the model will perform. | type InferenceTypesAi21TaskType = "completion" | "chat_completion" |
ai21_inference_id
required
The unique identifier of the inference endpoint. | type TypesId = string |
query Query Parameters
| Name | Type |
|---|---|
timeout Specifies the amount of time to wait for the inference endpoint to be created. | type TypesDuration = string | "-1" | "0" |
Request Body
application/json
required
{ service: InferenceTypesAi21ServiceType ;service_settings: InferenceTypesAi21ServiceSettings ; }
Responses
200 application/json
type InferenceTypesInferenceEndpointInfoAi21 = interface InferenceTypesInferenceEndpoint {
chunking_settings?:InferenceTypesInferenceChunkingSettings ;
service: string;
service_settings:InferenceTypesServiceSettings ;
task_settings?:InferenceTypesTaskSettings ;
} & { inference_id: string;task_type:InferenceTypesTaskTypeAi21 ; }
chunking_settings?:
service: string;
service_settings:
task_settings?:
} & { inference_id: string;task_type: