@@ -534,20 +534,24 @@ shards to become available. Both work exactly how they work in the
534534<<docs-bulk,Bulk API>>.
535535
536536`requests_per_second` can be set to any positive decimal number (`1.4`, `6`,
537- `1000`, etc) and throttles the number of batches that the reindex issues by
538- padding each batch with a wait time. The throttling can be disabled by
539- setting `requests_per_second` to `-1`.
537+ `1000`, etc) and throttles rate at which reindex issues batches of index
538+ operations by padding each batch with a wait time. The throttling can be
539+ disabled by setting `requests_per_second` to `-1`.
540540
541- The throttling is done waiting between bulk batches so that it can manipulate the
542- scroll timeout. The wait time is the difference between the request scroll search
543- size divided by the `requests_per_second` and the `batch_write_time`. By default
544- the scroll batch size is `1000`, so if the `requests_per_second` is set to `500`:
541+ The throttling is done by waiting between batches so that scroll that reindex
542+ uses internally can be given a timeout that takes into account the padding.
543+ The padding time is the difference between the batch size divided by the
544+ `requests_per_second` and the time spent writing. By default the batch size is
545+ `1000`, so if the `requests_per_second` is set to `500`:
545546
546- `target_total_time` = `1000` / `500 per second` = `2 seconds` +
547- `wait_time` = `target_total_time` - `batch_write_time` = `2 seconds` - `.5 seconds` = `1.5 seconds`
547+ [source,txt]
548+ --------------------------------------------------
549+ target_time = 1000 / 500 per second = 2 seconds
550+ wait_time = target_time - write_time = 2 seconds - .5 seconds = 1.5 seconds
551+ --------------------------------------------------
548552
549- Since the batch isn't broken into multiple bulk requests large batch sizes will
550- cause Elasticsearch to create many requests and then wait for a while before
553+ Since the batch is issued as a single `_bulk` request large batch sizes will
554+ cause Elasticsearch to create many requests and then wait for a while before
551555starting the next set. This is "bursty" instead of "smooth". The default is `-1`.
552556
553557[float]
0 commit comments