elasticsearch/docs/reference/ml
Dimitris Athanasiou 68c51f3ada
[ML] Rename threading params in _start trained model deployment API (#86597)
When starting a trained model deployment the user can tweak performance
by setting the `model_threads` and `inference_threads` parameters.
These parameters are hard to understand and cause confusion.

This commit renames these as well as the fields where their values are
reported in the stats API.

- `model_threads` => `number_of_allocations`
- `inference_threads` => `threads_per_allocation`

Now the terminology is as follows.

A model deployment starts with a requested `number_of_allocations`.
Each allocation means the model gets another thread for executing
parallel inference requests. Thus, more allocations should increase
throughput. In its turn, each allocation is may be using a number
of threads to parallelize each individual inference request.
This is the `threads_per_allocation` setting and increases inference
speed (which might also result in improved throughput).
2022-05-10 17:41:00 +03:00
..
anomaly-detection [DOCS] Move ML info and upgrade APIs (#84005) 2022-02-16 11:23:00 -08:00
common/apis [ML] Add ML memory stats API (#83802) 2022-02-17 09:19:14 +00:00
df-analytics/apis [DOCS] Adds size param to evaluate DFA API docs (#85735) 2022-04-07 10:03:09 +02:00
images [DOCS] Refresh automated screenshots (#84543) 2022-03-02 09:30:07 -08:00
trained-models/apis [ML] Rename threading params in _start trained model deployment API (#86597) 2022-05-10 17:41:00 +03:00
ml-shared.asciidoc [ML] add roberta/bart docs (#85001) 2022-03-17 12:14:57 -04:00