Skip to main content

LLM Generate Embeddings

A system task to generate embeddings from the input data provided. Embeddings are the processed input text converted into a sequence of vectors, which can then be stored in a vector database for retrieval later. You can use a model that was previously integrated to generate these embeddings.

Definitions

{
"name": "llm_generate_embeddings_task",
"taskReferenceName": "llm_generate_embeddings_task_ref",
"inputParameters": {
"llmProvider": "azure_openai",
"model": "text-davinci-003",
"text": "${workflow.input.text}"
},
"type": "LLM_GENERATE_EMBEDDINGS"
}
],
"inputParameters": [
"text"
]

Input Parameters

ParameterDescription
llmProviderSelect the required LLM provider. You can only choose providers to which you have access for at least one model from that provider.

Note:If you haven’t configured your AI / LLM provider on your Orkes console, navigate to the Integrations tab and set it up. Refer to the documentation for integrating LLM providers with Orkes console and providing access to required groups.
modelChoose from the available language models provided by the selected LLM provider. You can only choose models for which you have access.

For example, If your LLM provider is Azure Open AI and you’ve configured text-davinci-003 as the language model, you can select it here.
textProvide the text to be converted and stored as a vector. The text can also be passed as parameters.

Output Parameters

ParameterDescription
resultA JSON array containing the vectors of the indexed data.

Examples



  1. Add task type LLM Generate Embeddings.
  2. Choose the LLM provider and language model.
  3. Provide the text to be embedded.

LLM Generate Embeddings Task