Create an Elastic Inference Service (EIS) inference endpoint
Added in 8.12.0
Create an inference endpoint to perform an inference task through the Elastic Inference Service (EIS).
Path parameters
-
task_type
string Required The type of the inference task that the model will perform. NOTE: The
chat_completion
task type only supports streaming and only through the _stream API.Value is
chat_completion
. -
eis_inference_id
string Required The unique identifier of the inference endpoint.
Body
-
service
string Required Value is
elastic
. -
service_settings
object Required
PUT
/_inference/{task_type}/{eis_inference_id}
curl \
--request PUT 'http://api.example.com/_inference/{task_type}/{eis_inference_id}' \
--header "Authorization: $API_KEY" \
--header "Content-Type: application/json" \
--data '{"service":"elastic","service_settings":{"model_id":"string","rate_limit":{"requests_per_minute":42.0}}}'