Skip to content

Commit

Permalink
Merge remote-tracking branch 'upstream/main' into further-chunked-xco…
Browse files Browse the repository at this point in the history
…ntent-builder
  • Loading branch information
thecoop committed Oct 8, 2024
2 parents eb11f5d + 4af241b commit 2674bca
Show file tree
Hide file tree
Showing 119 changed files with 4,952 additions and 1,900 deletions.
5 changes: 5 additions & 0 deletions docs/changelog/111336.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
pr: 111336
summary: Use the same chunking configurations for models in the Elasticsearch service
area: Machine Learning
type: enhancement
issues: []
5 changes: 5 additions & 0 deletions docs/changelog/113812.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
pr: 113812
summary: Add Streaming Inference spec
area: Machine Learning
type: enhancement
issues: []
5 changes: 5 additions & 0 deletions docs/changelog/114080.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
pr: 114080
summary: Stream Cohere Completion
area: Machine Learning
type: enhancement
issues: []
17 changes: 17 additions & 0 deletions docs/changelog/114231.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
pr: 114231
summary: Remove cluster state from `/_cluster/reroute` response
area: Allocation
type: breaking
issues:
- 88978
breaking:
title: Remove cluster state from `/_cluster/reroute` response
area: REST API
details: >-
The `POST /_cluster/reroute` API no longer returns the cluster state in its
response. The `?metric` query parameter to this API now has no effect and
its use will be forbidden in a future version.
impact: >-
Cease usage of the `?metric` query parameter when calling the
`POST /_cluster/reroute` API.
notable: false
4 changes: 2 additions & 2 deletions docs/reference/cluster/reroute.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ Changes the allocation of shards in a cluster.
[[cluster-reroute-api-request]]
==== {api-request-title}

`POST /_cluster/reroute?metric=none`
`POST /_cluster/reroute`

[[cluster-reroute-api-prereqs]]
==== {api-prereq-title}
Expand Down Expand Up @@ -193,7 +193,7 @@ This is a short example of a simple reroute API call:

[source,console]
--------------------------------------------------
POST /_cluster/reroute?metric=none
POST /_cluster/reroute
{
"commands": [
{
Expand Down
2 changes: 1 addition & 1 deletion docs/reference/commands/shard-tool.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -95,7 +95,7 @@ Changing allocation id V8QXk-QXSZinZMT-NvEq4w to tjm9Ve6uTBewVFAlfUMWjA
You should run the following command to allocate this shard:
POST /_cluster/reroute?metric=none
POST /_cluster/reroute
{
"commands" : [
{
Expand Down
89 changes: 89 additions & 0 deletions docs/reference/ml/trained-models/apis/infer-trained-model.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -225,6 +225,17 @@ include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizatio
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
=======
`deberta_v2`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-deberta-v2]
+
.Properties of deberta_v2
[%collapsible%open]
=======
`truncate`::::
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate-deberta-v2]
=======
`roberta`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
Expand Down Expand Up @@ -301,6 +312,17 @@ include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizatio
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
=======
`deberta_v2`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-deberta-v2]
+
.Properties of deberta_v2
[%collapsible%open]
=======
`truncate`::::
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate-deberta-v2]
=======
`roberta`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
Expand Down Expand Up @@ -397,6 +419,21 @@ include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizatio
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
=======
`deberta_v2`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-deberta-v2]
+
.Properties of deberta_v2
[%collapsible%open]
=======
`span`::::
(Optional, integer)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-span]

`truncate`::::
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate-deberta-v2]
=======
`roberta`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
Expand Down Expand Up @@ -517,6 +554,21 @@ include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizatio
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
=======
`deberta_v2`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-deberta-v2]
+
.Properties of deberta_v2
[%collapsible%open]
=======
`span`::::
(Optional, integer)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-span]

`truncate`::::
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate-deberta-v2]
=======
`roberta`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
Expand Down Expand Up @@ -608,6 +660,17 @@ include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizatio
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
=======
`deberta_v2`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-deberta-v2]
+
.Properties of deberta_v2
[%collapsible%open]
=======
`truncate`::::
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate-deberta-v2]
=======
`roberta`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
Expand Down Expand Up @@ -687,6 +750,21 @@ include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizatio
(Optional, integer)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-span]

`with_special_tokens`::::
(Optional, boolean)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
=======
`deberta_v2`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-deberta-v2]
+
.Properties of deberta_v2
[%collapsible%open]
=======
`span`::::
(Optional, integer)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-span]

`with_special_tokens`::::
(Optional, boolean)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-bert-with-special-tokens]
Expand Down Expand Up @@ -790,6 +868,17 @@ include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenizatio
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate]
=======
`deberta_v2`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-deberta-v2]
+
.Properties of deberta_v2
[%collapsible%open]
=======
`truncate`::::
(Optional, string)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-truncate-deberta-v2]
=======
`roberta`::::
(Optional, object)
include::{es-ref-dir}/ml/ml-shared.asciidoc[tag=inference-config-nlp-tokenization-roberta]
Expand Down
9 changes: 5 additions & 4 deletions docs/reference/rest-api/common-parms.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -1298,10 +1298,11 @@ tag::wait_for_active_shards[]
`wait_for_active_shards`::
+
--
(Optional, string) The number of shard copies that must be active before
proceeding with the operation. Set to `all` or any positive integer up
to the total number of shards in the index (`number_of_replicas+1`).
Default: 1, the primary shard.
(Optional, string) The number of copies of each shard that must be active
before proceeding with the operation. Set to `all` or any non-negative integer
up to the total number of copies of each shard in the index
(`number_of_replicas+1`). Defaults to `1`, meaning to wait just for each
primary shard to be active.

See <<index-wait-for-active-shards>>.
--
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -89,6 +89,16 @@ PUT semantic-embeddings
It will be used to generate the embeddings based on the input text.
Every time you ingest data into the related `semantic_text` field, this endpoint will be used for creating the vector representation of the text.

[NOTE]
====
If you're using web crawlers or connectors to generate indices, you have to
<<indices-put-mapping,update the index mappings>> for these indices to
include the `semantic_text` field. Once the mapping is updated, you'll need to run
a full web crawl or a full connector sync. This ensures that all existing
documents are reprocessed and updated with the new semantic embeddings,
enabling semantic search on the updated data.
====


[discrete]
[[semantic-text-load-data]]
Expand Down Expand Up @@ -118,6 +128,13 @@ Create the embeddings from the text by reindexing the data from the `test-data`
The data in the `content` field will be reindexed into the `content` semantic text field of the destination index.
The reindexed data will be processed by the {infer} endpoint associated with the `content` semantic text field.

[NOTE]
====
This step uses the reindex API to simulate data ingestion. If you are working with data that has already been indexed,
rather than using the test-data set, reindexing is required to ensure that the data is processed by the {infer} endpoint
and the necessary embeddings are generated.
====

[source,console]
------------------------------------------------------------
POST _reindex?wait_for_completion=false
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,12 +2,12 @@
=== Red or yellow cluster health status

A red or yellow cluster health status indicates one or more shards are not assigned to
a node.
a node.

* **Red health status**: The cluster has some unassigned primary shards, which
means that some operations such as searches and indexing may fail.
* **Yellow health status**: The cluster has no unassigned primary shards but some
unassigned replica shards. This increases your risk of data loss and can degrade
means that some operations such as searches and indexing may fail.
* **Yellow health status**: The cluster has no unassigned primary shards but some
unassigned replica shards. This increases your risk of data loss and can degrade
cluster performance.

When your cluster has a red or yellow health status, it will continue to process
Expand All @@ -16,8 +16,8 @@ cleanup activities until the cluster returns to green health status. For instanc
some <<index-lifecycle-management,{ilm-init}>> actions require the index on which they
operate to have a green health status.

In many cases, your cluster will recover to green health status automatically.
If the cluster doesn't automatically recover, then you must <<fix-red-yellow-cluster-status,manually address>>
In many cases, your cluster will recover to green health status automatically.
If the cluster doesn't automatically recover, then you must <<fix-red-yellow-cluster-status,manually address>>
the remaining problems so management and cleanup activities can proceed.

[discrete]
Expand Down Expand Up @@ -107,7 +107,7 @@ asynchronously in the background.

[source,console]
----
POST _cluster/reroute?metric=none
POST _cluster/reroute
----

[discrete]
Expand Down Expand Up @@ -231,10 +231,10 @@ unassigned. See <<high-jvm-memory-pressure>>.

If a node containing a primary shard is lost, {es} can typically replace it
using a replica on another node. If you can't recover the node and replicas
don't exist or are irrecoverable, <<cluster-allocation-explain,Allocation
Explain>> will report `no_valid_shard_copy` and you'll need to do one of the following:
don't exist or are irrecoverable, <<cluster-allocation-explain,Allocation
Explain>> will report `no_valid_shard_copy` and you'll need to do one of the following:

* restore the missing data from <<snapshot-restore,snapshot>>
* restore the missing data from <<snapshot-restore,snapshot>>
* index the missing data from its original data source
* accept data loss on the index-level by running <<indices-delete-index,Delete Index>>
* accept data loss on the shard-level by executing <<cluster-reroute,Cluster Reroute>> allocate_stale_primary or allocate_empty_primary command with `accept_data_loss: true`
Expand All @@ -246,7 +246,7 @@ resulting in data loss.
+
[source,console]
----
POST _cluster/reroute?metric=none
POST _cluster/reroute
{
"commands": [
{
Expand Down
5 changes: 3 additions & 2 deletions libs/simdvec/build.gradle
Original file line number Diff line number Diff line change
Expand Up @@ -23,14 +23,15 @@ dependencies {
}
}

tasks.named("compileMain21Java").configure {
// compileMain21Java does not exist within idea (see MrJarPlugin) so we cannot reference directly by name
tasks.matching { it.name == "compileMain21Java" }.configureEach {
options.compilerArgs << '--add-modules=jdk.incubator.vector'
// we remove Werror, since incubating suppression (-Xlint:-incubating)
// is only support since JDK 22
options.compilerArgs -= '-Werror'
}

test {
tasks.named('test').configure {
if (JavaVersion.current().majorVersion.toInteger() >= 21) {
jvmArgs '--add-modules=jdk.incubator.vector'
}
Expand Down
Loading

0 comments on commit 2674bca

Please sign in to comment.