10 KiB
layout | title | nav_order | has_children | has_toc | redirect_from | |
---|---|---|---|---|---|---|
default | Neural search | 200 | false | false |
|
Neural search
Neural search transforms text into vectors and facilitates vector search both at ingestion time and at search time. During ingestion, neural search transforms document text into vector embeddings and indexes both the text and its vector embeddings in a k-NN index. When you use a neural query during search, neural search converts the query text into vector embeddings, uses vector search to compare the query and document embeddings, and returns the closest results.
The Neural Search plugin comes bundled with OpenSearch and is generally available as of OpenSearch 2.9. For more information, see Managing plugins.
Using neural search
To use neural search, follow these steps:
- Create an ingest pipeline.
- Create an index for ingestion.
- Ingest documents into the index.
- Search the index using neural search.
Step 1: Create an ingest pipeline
To generate vector embeddings for text fields, you need to create a neural search ingest pipeline. An ingest pipeline consists of a series of processors that manipulate documents during ingestion, allowing the documents to be vectorized.
Path and HTTP method
The following API operation creates a neural search ingest pipeline:
PUT _ingest/pipeline/<pipeline_name>
Path parameter
Use pipeline_name
to create a name for your neural search ingest pipeline.
Request fields
In the pipeline request body, you must set up a text_embedding
processor (the only processor supported by neural search), which will convert the text in a document field to vector embeddings. The processor's field_map
determines the input fields from which to generate vector embeddings and the output fields in which to store the embeddings:
"text_embedding": {
"model_id": "<model_id>",
"field_map": {
"<input_field>": "<vector_field>"
}
}
The following table lists the text_embedding
processor request fields.
Field | Data type | Description |
---|---|---|
model_id |
String | The ID of the model that will be used to generate the embeddings. The model must be indexed in OpenSearch before it can be used in neural search. For more information, see ML Framework and Semantic search. |
field_map.<input_field> |
String | The name of the field from which to obtain text for generating text embeddings. |
field_map.<vector_field> |
String | The name of the vector field in which to store the generated text embeddings. |
Example request
The following example request creates an ingest pipeline where the text from passage_text
will be converted into text embeddings and the embeddings will be stored in passage_embedding
:
PUT /_ingest/pipeline/nlp-ingest-pipeline
{
"description": "An NLP ingest pipeline",
"processors": [
{
"text_embedding": {
"model_id": "bQ1J8ooBpBj3wT4HVUsb",
"field_map": {
"passage_text": "passage_embedding"
}
}
}
]
}
{% include copy-curl.html %}
Step 2: Create an index for ingestion
In order to use the text embedding processor defined in your pipelines, create a k-NN index with mapping data that aligns with the maps specified in your pipeline. For example, the <vector_field>
defined in the field_map
of your processor must be mapped as a k-NN vector field with a dimension that matches the model dimension. Similarly, the <input_field>
defined in your processor should be mapped as text
in your index.
Example request
The following example request creates a k-NN index that is set up with a default ingest pipeline:
PUT /my-nlp-index
{
"settings": {
"index.knn": true,
"default_pipeline": "nlp-ingest-pipeline"
},
"mappings": {
"properties": {
"id": {
"type": "text"
},
"passage_embedding": {
"type": "knn_vector",
"dimension": 768,
"method": {
"engine": "lucene",
"space_type": "l2",
"name": "hnsw",
"parameters": {}
}
},
"passage_text": {
"type": "text"
}
}
}
}
{% include copy-curl.html %}
For more information about creating a k-NN index and the methods it supports, see k-NN index.
Step 3: Ingest documents into the index
To ingest documents into the index created in the previous step, send a POST request for each document:
PUT /my-nlp-index/_doc/1
{
"passage_text": "Hello world",
"id": "s1"
}
{% include copy-curl.html %}
PUT /my-nlp-index/_doc/2
{
"passage_text": "Hi planet",
"id": "s2"
}
{% include copy-curl.html %}
Before the document is ingested into the index, the ingest pipeline runs the text_embedding
processor on the document, generating text embeddings for the passage_text
field. The indexed document contains the passage_text
field that has the original text and the passage_embedding
field that has the vector embeddings.
Step 4: Search the index using neural search
To perform vector search on your index, use the neural
query clause either in the k-NN plugin API or Query DSL queries. You can refine the results by using a k-NN search filter.
Neural query request fields
Include the following request fields under the neural
query clause:
"neural": {
"<vector_field>": {
"query_text": "<query_text>",
"model_id": "<model_id>",
"k": 100
}
}
The top-level vector_field
specifies the vector field against which to run a search query. The following table lists the other neural query fields.
Field | Data type | Description |
---|---|---|
query_text |
String | The query text from which to generate text embeddings. |
model_id |
String | The ID of the model that will be used to generate text embeddings from the query text. The model must be indexed in OpenSearch before it can be used in neural search. |
k |
Integer | The number of results returned by the k-NN search. |
Example request
The following example request uses a Boolean query to combine a filter clause and two query clauses---a neural query and a match
query. The script_score
query assigns custom weights to the query clauses:
GET /my-nlp-index/_search
{
"_source": {
"excludes": [
"passage_embedding"
]
},
"query": {
"bool": {
"filter": {
"wildcard": { "id": "*1" }
},
"should": [
{
"script_score": {
"query": {
"neural": {
"passage_embedding": {
"query_text": "Hi world",
"model_id": "bQ1J8ooBpBj3wT4HVUsb",
"k": 100
}
}
},
"script": {
"source": "_score * 1.5"
}
}
},
{
"script_score": {
"query": {
"match": {
"passage_text": "Hi world"
}
},
"script": {
"source": "_score * 1.7"
}
}
}
]
}
}
}
{% include copy-curl.html %}
The response contains the matching document:
{
"took" : 36,
"timed_out" : false,
"_shards" : {
"total" : 1,
"successful" : 1,
"skipped" : 0,
"failed" : 0
},
"hits" : {
"total" : {
"value" : 1,
"relation" : "eq"
},
"max_score" : 1.2251667,
"hits" : [
{
"_index" : "my-nlp-index",
"_id" : "1",
"_score" : 1.2251667,
"_source" : {
"passage_text" : "Hello world",
"id" : "s1"
}
}
]
}
}
Setting a default model on an index or field
To eliminate passing the model ID with each neural query request, you can set a default model on a k-NN index or a field.
First, create a search pipeline with a neural_query_enricher
request processor. To set a default model for an index, provide the model ID in the default_model_id
parameter. To set a default model for a specific field, provide the field name and the corresponding model ID in the neural_field_default_id
map. If you provide both default_model_id
and neural_field_default_id
, neural_field_default_id
takes precedence:
PUT /_search/pipeline/default_model_pipeline
{
"request_processors": [
{
"neural_query_enricher" : {
"default_model_id": "bQ1J8ooBpBj3wT4HVUsb",
"neural_field_default_id": {
"my_field_1": "uZj0qYoBMtvQlfhaYeud",
"my_field_2": "upj0qYoBMtvQlfhaZOuM"
}
}
}
]
}
{% include copy-curl.html %}
Then set the default model for your index:
PUT /my-nlp-index/_settings
{
"index.search.default_pipeline" : "default_model_pipeline"
}
{% include copy-curl.html %}
You can now omit the model ID when searching:
GET /my-nlp-index/_search
{
"_source": {
"excludes": [
"passage_embedding"
]
},
"query": {
"neural": {
"passage_embedding": {
"query_text": "Hi world",
"k": 100
}
}
}
}
{% include copy-curl.html %}
The response contains both documents:
{
"took" : 41,
"timed_out" : false,
"_shards" : {
"total" : 1,
"successful" : 1,
"skipped" : 0,
"failed" : 0
},
"hits" : {
"total" : {
"value" : 2,
"relation" : "eq"
},
"max_score" : 1.22762,
"hits" : [
{
"_index" : "my-nlp-index",
"_id" : "2",
"_score" : 1.22762,
"_source" : {
"passage_text" : "Hi planet",
"id" : "s2"
}
},
{
"_index" : "my-nlp-index",
"_id" : "1",
"_score" : 1.2251667,
"_source" : {
"passage_text" : "Hello world",
"id" : "s1"
}
}
]
}
}