Lucidworks AI Prediction API
embedding
use case tokens before being sent to any pre-trained embedding model or custom embedding model.
This API is used to help debug embedding
model tokens to ensure the input to the pre-trained or custom embedding model is valid, and within the model’s processing limits.
embedding
use cases.
APPLICATION_ID
for your Lucidworks AI application. For more information, see credentials to use APIs.machinelearning.predict
. For more information, see Authentication API.MODEL_ID
field for the request. The path is: /ai/tokenization/MODEL_ID
. For more information about supported models, see Embedding use cases./ai/tokenization/MODEL_ID
request are common to all of the Async Chunking API requests, such as the modelConfig
parameter.
Also referred to as hyperparameters, these fields set certain controls on the response.
Refer to the API spec for more information.
modelConfig
parameter, vectorQuantizationMethod
. The methods are named min-max
and max-scale
.
min-max
method creates tensors of embeddings and converts them to uint8 by normalizing them to the range [0, 255].max-scale
method finds the maximum absolute value along each embedding, normalizes the embeddings by scaling them to a range of -127 to 127, and returns the quantized embeddings as an 8-bit integer tensor.max-scale
method has no loss at the ten-thousandths place during evaluation against non-quantized vectors.
However, other methods lose precision when evaluated against non-quantized vectors, with min-max
losing the most precision.
The syntax example is:
APPLICATION_ID
, MODEL_ID
, and ACCESS_TOKEN
fields with your information.