Versions (relevant - OpenSearch/Dashboard/Server OS/Browser):
opensearch version : 2.13
Describe the issue:
Text chunking is in progress by referring to the link below.
(Text chunking - OpenSearch Documentation)
An error occurs during Step 3.
Step 1: Create a pipeline
PUT _ingest/pipeline/text-chunking-embedding-ingest-pipeline
{
"description": "A text chunking and embedding ingest pipeline",
"processors": [
{
"text_chunking": {
"algorithm": {
"fixed_token_length": {
"token_limit": 10,
"overlap_rate": 0.2,
"tokenizer": "standard"
}
},
"field_map": {
"passage_text": "passage_chunk"
}
}
},
{
"text_embedding": {
"model_id": "GkyElI8BKGaMwVo6PeZn",
"field_map": {
"passage_chunk": "passage_chunk_embedding"
}
}
}
]
}
Step 2: Create an index for ingestion
PUT testindex
{
"settings": {
"index": {
"knn": true,
"default_pipeline": "text-chunking-embedding-ingest-pipeline"
}
},
"mappings": {
"properties": {
"passage_text": {
"type": "text"
},
"passage_chunk_embedding": {
"type": "nested",
"properties": {
"knn": {
"type": "knn_vector",
"dimension": 768
}
}
}
}
}
}
Step 3: Ingest documents into the index
POST testindex/_doc?pipeline=text-chunking-embedding-ingest-pipeline
{
"passage_text": "This is an example document to be chunked. The document contains a single paragraph, two sentences and 24 tokens by standard tokenizer in OpenSearch."
}
The following error occurs:
{
"error": {
"root_cause": [
{
"type": "index_not_found_exception",
"reason": "no such index [testindex]",
"index": "testindex",
"index_uuid": "lE1DNT22ShW-eD_0gSkxUg"
}
],
"type": "index_not_found_exception",
"reason": "no such index [testindex]",
"index": "testindex",
"index_uuid": "lE1DNT22ShW-eD_0gSkxUg"
},
"status": 404
}
The model used “GkyElI8BKGaMwVo6PeZn” is a TEXT_EMBEDDING model
Is there something I did wrong?