You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Coming from flaky integration test failures due to the memory limitations of Github action runners which facilitate integration tests (Reference), it has been determined that mitigation would require using a smaller model to test out local model registration within an integration test cluster.
The amazon/neural-sparse/opensearch-neural-sparse-tokenizer-v1 has been determined to be the smallest model we can use for testing (Documentation) , however the required fields for registering a sparse encoding model is different than the required fields for registering a custom text embedding model.
In order to replace the model used in integration testing with this sparse encoding model, it is required to separate out the RegisterLocalModelStep into a RegisterCustomLocalModelStep and RegisterSparseEncodingLocalModelStep. Additionally, we require support for registering an OpenSearch -provided pretained model, which does not require a URL.
What solution would you like?
The RegisterCustomLocalModelStep relates to the following documentation and will have the following required and optional fields :
Required keys :
name
version
model_format
function_name
model_content_hash_value
url
model_type
embedding_dimension
framework_type
Optional keys :
description
model_group_id
all_config
deploy
The RegisterSparseEncodingModelStep relates to the following documentation and will have the following required and optional fields
Required keys :
name
version
model_format
function_name
model_content_hash_value
url
Optional keys :
description
model_group_id
deploy
The RegisterLocalPretrainedModelStep relates to the following documentation and will have the following required and optional fields
Required keys :
name
version
model_format
Optional keys :
description
model_group_id
deploy
The text was updated successfully, but these errors were encountered:
joshpalis
changed the title
[FEATURE] Separate Local Model Registration to support Custom and Sparse Encoding Models
[FEATURE] Separate Local Model Registration to support Custom, Pretrained and Sparse Encoding Models
Jan 8, 2024
Is your feature request related to a problem?
Coming from flaky integration test failures due to the memory limitations of Github action runners which facilitate integration tests (Reference), it has been determined that mitigation would require using a smaller model to test out local model registration within an integration test cluster.
The
amazon/neural-sparse/opensearch-neural-sparse-tokenizer-v1
has been determined to be the smallest model we can use for testing (Documentation) , however the required fields for registering asparse encoding model
is different than the required fields for registering a customtext embedding model
.In order to replace the model used in integration testing with this sparse encoding model, it is required to separate out the
RegisterLocalModelStep
into aRegisterCustomLocalModelStep
andRegisterSparseEncodingLocalModelStep
. Additionally, we require support for registering anOpenSearch -provided pretained model
, which does not require a URL.What solution would you like?
The
RegisterCustomLocalModelStep
relates to the following documentation and will have the following required and optional fields :Required keys :
name
version
model_format
function_name
model_content_hash_value
url
model_type
embedding_dimension
framework_type
Optional keys :
description
model_group_id
all_config
deploy
The
RegisterSparseEncodingModelStep
relates to the following documentation and will have the following required and optional fieldsRequired keys :
name
version
model_format
function_name
model_content_hash_value
url
Optional keys :
description
model_group_id
deploy
The
RegisterLocalPretrainedModelStep
relates to the following documentation and will have the following required and optional fieldsRequired keys :
name
version
model_format
Optional keys :
description
model_group_id
deploy
The text was updated successfully, but these errors were encountered: