Create an Amazon Bedrock inference endpoint
PUT
/_inference/{task_type}/{amazonbedrock_inference_id} Create an inference endpoint to perform an inference task with the amazonbedrock service.
info You need to provide the access and secret keys only once, during the inference model creation. The get inference API does not retrieve your access or secret keys. After creating the inference model, you cannot change the associated key pairs. If you want to use a different access and secret key pair, delete the inference model and recreate it with the same name and the updated keys.
Required authorization
- Cluster privileges:
manage_inference
Parameters
path Path Parameters
| Name | Type |
|---|---|
task_type
required
The type of the inference task that the model will perform. | type InferenceTypesAmazonBedrockTaskType = "chat_completion" | "completion" | "text_embedding" |
amazonbedrock_inference_id
required
The unique identifier of the inference endpoint. | type TypesId = string |
query Query Parameters
| Name | Type |
|---|---|
timeout Specifies the amount of time to wait for the inference endpoint to be created. | type TypesDuration = string | "-1" | "0" |
Request Body
application/json
required
{
chunking_settings?:InferenceTypesInferenceChunkingSettings ;
service:InferenceTypesAmazonBedrockServiceType ;
service_settings:InferenceTypesAmazonBedrockServiceSettings ;
task_settings?:InferenceTypesAmazonBedrockTaskSettings ;
}
chunking_settings?:
service:
service_settings:
task_settings?:
}
Responses
200 application/json
type InferenceTypesInferenceEndpointInfoAmazonBedrock = interface InferenceTypesInferenceEndpoint {
chunking_settings?:InferenceTypesInferenceChunkingSettings ;
service: string;
service_settings:InferenceTypesServiceSettings ;
task_settings?:InferenceTypesTaskSettings ;
} & { inference_id: string;task_type:InferenceTypesTaskTypeAmazonBedrock ; }
chunking_settings?:
service: string;
service_settings:
task_settings?:
} & { inference_id: string;task_type: