Elasticsearch billing dimensions

edit

Elasticsearch billing dimensions

edit

Elasticsearch is priced based on consumption of the underlying infrastructure that supports your use case, with the performance characteristics you need. Measurements are in Virtual Compute Units (VCUs). Each VCU represents a fraction of RAM, CPU, and local disk for caching.

The number of VCUs you need is determined by:

  • Volume and ingestion rate of your data
  • Data retention requirements
  • Search query volume
  • Search Power setting
  • Machine learning usage

For detailed Elasticsearch Serverless project rates, see the Elasticsearch Serverless pricing page.

VCU types: Search, Indexing, and ML

edit

Elasticsearch uses three VCU types:

  • Indexing: The VCUs used to index incoming documents.
  • Search: The VCUs used to return search results, with the latency and queries per second (QPS) you require.
  • Machine learning: The VCUs used to perform inference, NLP tasks, and other ML activities.

Data storage and billing

edit

Elasticsearch Serverless projects store data in the Search AI Lake. You are charged per GB of stored data at rest. Note that if you perform operations at ingest such as vectorization or enrichment, the size of your stored data will differ from the size of the original source data.

Managing Elasticsearch costs

edit

You can control costs using the following strategies:

  • Search Power setting: Search Power controls the speed of searches against your data. With Search Power, you can improve search performance by adding more resources for querying, or you can reduce provisioned resources to cut costs.
  • Time series data retention: By limiting the number of days of time series data that are available for caching, you can reduce the number of search VCUs required.
  • Machine learning trained model autoscaling: Configure your trained model deployment to allow it to scale down to zero allocations when there are no active inference requests: