You're viewing version 2.14 of the OpenSearch documentation. This version is no longer maintained. For the latest version, see the current documentation. For information about OpenSearch version maintenance, see Release Schedule and Maintenance Policy.
ML inference processor
The ml_inference
processor is used to generate inferences from machine learning (ML) models connected to the OpenSearch ML Commons plugin. The inferences are added as new fields to the ingested documents.
PREREQUISITE
Before using the ml_inference
processor, you must have an ML model connected to your OpenSearch cluster through the ML Commons plugin. For more information, see Connecting to externally hosted models.
Syntax
The following is the syntax for the ml-inference
processor:
{
"ml_inference": {
"model_id": "<model_id>",
"input_map": [
{
"<model_input>": "<document_field>"
}
],
"output_map": [
{
"<new_document_field>": "<model_output>"
}
],
"model_config":{
"<model_config_field>": "<config_value>"
}
}
}
Configuration parameters
The following table lists the required and optional parameters for the ml-inference
processor.
Parameter | Data type | Required/Optional | Description |
---|---|---|---|
model_id | String | Required | The ID of the ML model connected to the OpenSearch cluster through the ML Commons plugin. |
input_map | Array | Optional | An array specifying how to map fields from the ingested document to the model’s input. If no input mapping is specified, then all fields from the document are used as the model input by default. The input_map size indicates the number of prediction tasks. In each map, the key represents the model input field name, and the value specifies the document field used to construct the model input. |
input_map.<input_key> | String | Optional | The model input field name. |
input_map.<input_field> | String | Optional | The name of the field from the ingested document to be used as the model’s input. |
output_map | Array | Optional | An array specifying how to map the model’s output to new fields in the ingested document. |
output_map.<output_field> | String | Optional | The name of the new field in the ingested document in which the model’s output (specified by the output_key ) is stored. If no output mapping is specified, then all fields from the model prediction output are added to the new document field by default. |
output_map.<output_key> | String | Optional | The key representing the model output to be stored in the output_field . |
model_config | Object | Optional | Custom configuration options for the ML model. |
max_prediction_tasks | Integer | Optional | The maximum number of concurrent prediction tasks that can run during document ingestion. Default is 10 . |
description | String | Optional | A brief description of the processor. |
tag | String | Optional | An identifier tag for the processor. Useful for debugging to distinguish between processors of the same type. |
ignore_missing | Boolean | Optional | If true and any of the input fields defined in input_map or output_map are missing, then the missing fields are ignored. Otherwise, a missing field causes a failure. Default is false . |
ignore_failure | Boolean | Optional | If true and any exceptions occur, then they are ignored and ingestion continues. Otherwise, an exception occurs and ingestion is canceled. Default is false . |
Using the processor
Follow these steps to use the processor in a pipeline. You must provide a model ID when creating the processor. Only remote models are currently supported. For more information about creating remote models, see Connecting to externally hosted models.
Step 1: Create a pipeline
The following example creates an ingest pipeline where the model requires an input
field, produces a data
field for prediction, and converts the passage_text
field into text embeddings stored in the passage_embedding
field. The input_map
and output_map
mappings support standard JSON path notation for complex data structures.
PUT /_ingest/pipeline/ml_inference_pipeline
{
"description": "Generate passage_embedding for ingested documents",
"processors": [
{
"ml_inference": {
"model_id": "<your model id>",
"input_map": [
{
"input": "passage_text"
}
],
"output_map": [
{
"passage_embedding": "data"
}
]
}
}
]
}
Step 2 (Optional): Test the pipeline
It is recommended that you test your pipeline before you ingest documents.
To test the pipeline, run the following query:
POST _ingest/pipeline/ml_inference_pipeline/_simulate
{
"docs": [
{
"_index": "testindex1",
"_id": "1",
"_source":{
"passage_text": "hello world"
}
}
]
}
Response
The response confirms that, in addition to the passage_text
field, the processor has generated text embeddings in the passage_embedding
field:
{
"docs" : [
{
"doc" : {
"_index" : "testindex1",
"_id" : "1",
"_source" : {
"passage_embedding" : [
0.017304314,
-0.021530833,
0.050184276,
0.08962978,
...
],
"passage_text" : "hello world"
},
"_ingest" : {
"timestamp" : "2023-10-11T22:35:53.654650086Z"
}
}
}
]
}
Once you have created an ingest pipeline, you need to create an index for ingestion and ingest documents into the index.
Limitation
The ml_inference
processor currently supports only remote models connected through a connector. Local models uploaded to an OpenSearch cluster are not yet supported. Check the OpenSearch Project Roadmap for updates on upcoming features.