Skip to content

Infer

Infer (infer)

Execute inference workloads (LLM, embeddings, anomaly detection).

AI & ML json

Minimal example

actions:
- infer: {}
JSON
{
"actions": [
{
"infer": {}
}
]
}

Contents

Fields

FieldTypeRequiredDescription
descriptionstringdescribe this step.
conditionlua-expression (string)Only run this action if the condition is met.
Examples: 2 * count()
workloadWorkloadMode-specific configuration (llm, embedding, or anomaly).
timeout-msduration (integer)Maximum time to wait under infer backpressure (milliseconds, 0 uses the built-in default).
cacheCacheOptional cache configuration.
rate-limitRate LimitOptional rate limiting configuration.
concurrencynumber (integer)Maximum concurrent in-flight requests (overrides provider defaults).
Examples: 42, 1.2e-10
streamingboolean (bool)Enable streaming responses when supported (LLM mode only).
on-errorstringError handling strategy (e.g. fail, skip, dlq:name).

Schema

Workload Options

OptionNameTypeDescription
llm-completionLlm CompletionobjectLarge language model completion.
embeddingEmbeddingobjectText or document embeddings.
anomaly-detectAnomaly DetectobjectAnomaly detection scoring.

Workload - Llm Completion - Llm - Prompt Fields

FieldTypeRequiredDescription
systemstringSystem message supplied to the model.
templatestringPrompt template (may include {{json record}} style placeholders).
schemastringOptional JSON schema or tool definitions for structured output.

Workload - Llm Completion - Llm - Field Map Fields

FieldTypeRequiredDescription
inputfield (string)Field containing the prompt input payload.
Examples: data_field
outputfield (string)Field to write model response into.
Examples: data_field
usagefield (string)Optional field to capture token usage metadata.
Examples: data_field

Workload - Llm Completion - Llm Fields

FieldTypeRequiredDescription
providerProviderProvider selection (llama-server or openai-compat).
Allowed values: llama-server, openai-compat
modelstringModel identifier or path (provider specific).
endpointurl (string)Optional HTTP endpoint for remote providers (OpenAI-compatible / llama-server).
Examples: https://example.com/path
api-keystringSecret or variable reference for the API key / bearer token (use `${dyn
promptPromptPrompt template configuration.
field-mapField MapField mapping for input/output projection.
input-fieldfield (string)Field containing the prompt input payload.
Examples: data_field
response-fieldfield (string)Field to write model response into (required).
Examples: data_field
usage-fieldfield (string)Optional field to capture token usage metadata.
Examples: data_field
response-formatResponse FormatDesired response format emitted by the provider.
Allowed values: str, json
concurrencynumber (integer)Maximum concurrent requests for this provider (overrides action-level concurrency).
Examples: 42, 1.2e-10
streamingboolean (bool)Enable streaming token responses when supported.
temperaturenumber (integer)Temperature parameter for text generation.
Examples: 42, 1.2e-10
top-pnumber (integer)Top-p parameter for nucleus sampling.
Examples: 42, 1.2e-10
max-tokensnumber (integer)Maximum tokens to generate (remote providers only).
Examples: 42, 1.2e-10

Workload - Llm Completion Fields

FieldTypeRequiredDescription
llmLlmLLM configuration (required when mode = llm-completion).

Workload - Embedding - Embedding Fields

FieldTypeRequiredDescription
providerProviderEmbedding provider selection.
Allowed values: llama-server, openai-compat
modelstringModel identifier or path (provider specific).
endpointurl (string)HTTP endpoint for remote embedding providers.
Examples: https://example.com/path
api-keystringSecret or variable reference for the API key / bearer token (use `${dyn
input-fieldfield (string)Field containing the text payload to embed.
Examples: data_field
response-fieldfield (string)Field to write the embedding vector to (required).
Examples: data_field
additional-response-fieldsstring[]Additional fields to mirror the embedding vector into.
normalizeboolean (bool)Normalize embedding vectors before emitting.
concurrencynumber (integer)Maximum concurrent embedding requests.
Examples: 42, 1.2e-10

Workload - Embedding Fields

FieldTypeRequiredDescription
embeddingEmbeddingEmbedding configuration (required when mode = embedding).

Workload - Anomaly Detect - Anomaly - Training Fields

FieldTypeRequiredDescription
modeModeTraining mode (online or offline).
Allowed values: online, offline
windownumber (integer)Sliding window size for online training.
Examples: 42, 1.2e-10
paramsmap (string)Algorithm-specific parameters.
model-variablestringSerialized model payload reference (use `${dyn

Workload - Anomaly Detect - Anomaly Fields

FieldTypeRequiredDescription
algoAlgoAlgorithm selection.
Allowed values: zscore, isolation-forest, one-class-svm
fieldsstring[]Numeric fields to monitor for anomalies.
output-fieldfield (string)Field to write anomaly score to.
Examples: data_field
flag-fieldfield (string)Optional field to emit boolean anomaly flag.
Examples: data_field
input-fieldfield (string)Field containing the anomaly input payload.
Examples: data_field
response-fieldfield (string)Field to write the anomaly response into (defaults to output_field).
Examples: data_field
score-thresholdnumber (integer)Score threshold to mark anomalies (optional, algorithm specific).
Examples: 42, 1.2e-10
trainingTrainingTraining configuration (online/offline).

Workload - Anomaly Detect Fields

FieldTypeRequiredDescription
anomalyAnomalyAnomaly configuration (required when mode = anomaly-detect).

Cache Fields

FieldTypeRequiredDescription
namespacestringCache namespace identifier.
max-entriesnumber (integer)Maximum cache entries.
Examples: 42, 1.2e-10
ttlduration (string)Cache entry TTL (e.g. “5m”).

Rate Limit Fields

FieldTypeRequiredDescription
requests-per-secondnumber (integer)Maximum requests per second.
Examples: 42, 1.2e-10
tokens-per-minutenumber (integer)Maximum tokens per minute (LLM providers).
Examples: 42, 1.2e-10
max-concurrencynumber (integer)Maximum concurrent requests.
Examples: 42, 1.2e-10

Workload - Anomaly Detect - Anomaly - Training - Params Table

ParameterValue
paramvalue

Value format: templated-text.

Workload - Llm Completion - Llm - Provider Options

ValueDescription
llama-serverLlama Server
openai-compatOpenai Compat

Workload - Llm Completion - Llm - Response Format Options

ValueDescription
strEmit text/string output (default)
jsonRequest JSON formatted responses via provider-specific hinting

Workload - Embedding - Embedding - Provider Options

ValueDescription
llama-serverLlama Server
openai-compatOpenai Compat

Workload - Anomaly Detect - Anomaly - Algo Options

ValueDescription
zscoreZscore
isolation-forestIsolation Forest
one-class-svmOne Class Svm

Workload - Anomaly Detect - Anomaly - Training - Mode Options

ValueDescription
onlineOnline
offlineOffline