10.0+41.10 Artificial Intelligence

The following options serve to integrate LLMs ("Large Language Models") with QF-Test.

LLM integration options
Figure 41.38:  AI options
LLM Configurations (System)
Server script name: OPT_LLM_PROVIDERS

In this table you can make third-party LLMs usable in QF-Test, for example to use them with the Check text with AI or ai.ask() (see The ai module). Each line represents a specific language model.

The following table columns are available:

Type
The name of the API provider or type. If the API type you need is not yet available, please get in touch with our support team.
Name
The name of this configuration. Used to reference it in other places.
URL
The URL at which the LLM can be accessed. Please get this information from the documentation of the LLM provider. Usually ends with /v1 or similar.
API key
The API key serves to authenticate with the respective LLM provider. Get it from your LLM provider.
Model name
Many providers offer multiple LLM models like "gpt-4o" or "gemini-2.0-flash". You can learn about the actually available model names from your LLM provider.
Prompt templates (System)
Server script name: OPT_LLM_PROMPT_TEMPLATES

In this table you can configure templates for LLM requests, to use with Check text with AI, for example. Each line represents a specific template.

The following table columns are available:

Name
The name of this template with which it can be referenced elsewhere.
Template

The text content of the template which would be sent to an LLM. The following placeholders are available:

  • {{context}}: Additional application-specific information, see Check context.
  • {{goodCompletions}}: A list of example results which should be interpreted as positive.
  • {{badCompletions}}: A list of example results which should be interpreted as negative.
  • {{completion}}: The actual result, e.g. from a Check text with AI.
  • {{positiveResponse}}: The string configured in "Positive response", to teach the model how to reply.
  • {{negativeResponse}}: The string configured in "Negative response", to teach the model how to reply.
Positive response
The response text that should be considered positive, e.g. "yes". Case-insensitive.
Negative response
The response text that should be considered negative, e.g. "no". Case-insensitive.
Default timeout for LLM requests (ms) (System)
Server script name: OPT_LLM_REQUEST_TIMEOUT

The default timeout for LLM requests in milliseconds. If a request, like in the context of a Check text with AI, takes longer than the configured duration, an exception is thrown.