Request Headers
The authorization token (optional).
Request Query
Whether to include metadata in the response. If false, all other query fields are ignored.
Whether to only return the user's own metadata statistics. If true, authorization is required.
An RFC 3339 start date to filter metadata statistics by.
An RFC 3339 end date to filter metadata statistics by.
Request Body
Model ID used to generate the response.
The mode of the model, which determines whether it generates a response or selects from the generated options.
Variants
The model generates a response.
The model selects a Generate ID. The model will output reasoning, even if the LLM is not a reasoning model. Best for non-reasoning models.
The model selects a Generate ID.
The model selects one or more Generate IDs as a probability distribution. The model will output reasoning, even if the LLM is not a reasoning model. Best for non-reasoning models.
The model selects one or more Generate IDs as a probability distribution.
If the mode is one of the select logprobs modes, this controls how many of the top options are returned with their probabilities.
This setting aims to control the repetition of tokens based on how often they appear in the input. It tries to use less frequently those tokens that appear more in the input, proportional to how frequently they occur. Token penalty scales with the number of occurrences. Negative values will encourage token reuse.
Accepts a JSON object that maps tokens (specified by their token ID in the tokenizer) to an associated bias value from -100 to 100. Mathematically, the bias is added to the logits generated by the model prior to sampling. The exact effect will vary per model, but values between -1 and 1 should decrease or increase likelihood of selection; values like -100 or 100 should result in a ban or exclusive selection of the relevant token.
An upper bound for the number of tokens that can be generated for a completion, including visible output tokens and reasoning tokens.
This setting aims to control the presence of tokens in the output. It tries to encourage the model to use tokens that are less present in the input, proportional to their presence in the input. Token presence scales with the number of occurrences. Negative values will encourage more diverse token usage.
Constrains effort on reasoning for some reasoning models.
Variants
Stop generation immediately if the model encounters any token specified in the stop array.
Variants
Items
This setting influences the variety in the model’s responses. Lower values lead to more predictable and typical responses, while higher values encourage more diverse and less common responses. At 0, the model always gives the same response for a given input.
This setting limits the model’s choices to a percentage of likely tokens: only the top tokens whose probabilities add up to P. A lower value makes the model’s responses more predictable, while the default setting allows for a full range of token choices. Think of it like a dynamic Top-K.
This sets the upper limit for the number of tokens the model can generate in response. It won’t produce more than this limit. The maximum value is the context length minus the prompt length.
Represents the minimum probability for a token to be considered, relative to the probability of the most likely token. (The value changes depending on the confidence level of the most probable token.) If your Min-P is set to 0.1, that means it will only allow for tokens that are at least 1/10th as probable as the best possible option.
OpenRouter provider preferences.
Properties
List of provider slugs to try in order.
Items
Whether to allow backup providers when the primary is unavailable.
Only use providers that support all parameters in your request.
Control whether to use providers that may store data.
Variants
List of provider slugs to allow for this request.
Items
List of provider slugs to skip for this request.
Items
List of quantization levels to filter by.
Items
Sort providers by price or throughput.
OpenRouter reasoning configuration.
Properties
An upper bound for the number of tokens that can be generated for reasoning.
Constrains effort on reasoning for some reasoning models.
Variants
Whether reasoning is enabled for this request.
Helps to reduce the repetition of tokens from the input. A higher value makes the model less likely to repeat tokens, but too high a value can make the output less coherent (often with run-on sentences that lack small words). Token penalty scales based on original token’s probability.
Consider only the top tokens with “sufficiently high” probabilities based on the probability of the most likely token. Think of it like a dynamic Top-P. A lower Top-A value focuses the choices based on the highest probability token but with a narrower scope. A higher Top-A value does not necessarily affect the creativity of the output, but rather refines the filtering process based on the maximum probability.
This limits the model’s choice of tokens at each step, making it choose from a smaller set. A value of 1 means the model will always pick the most likely next token, leading to predictable results. By default this setting is disabled, making the model to consider all choices.
Controls the verbosity and length of the model response. Lower values produce more concise responses, while higher values produce more detailed and comprehensive responses.
Variants
Fallback models. Will be tried in order if the first one fails.
Items
The weight of the model, which determines its influence on the Confidence Score. Must match the weight strategy of the parent Model.
Variants
A static weight value.
Properties
The static weight value.
A dynamic weight value based on training table data.
Properties
The base weight value, uninfluenced by training table data.
The minimum weight value. A model that never matches the correct answer will have this weight.
The maximum weight value. A model that always matches the correct answer will have this weight.
Response Body
Model ID used to generate the response.
The mode of the model, which determines whether it generates a response or selects from the generated options.
Variants
The model generates a response.
The model selects a Generate ID. The model will output reasoning, even if the LLM is not a reasoning model. Best for non-reasoning models.
The model selects a Generate ID.
The model selects one or more Generate IDs as a probability distribution. The model will output reasoning, even if the LLM is not a reasoning model. Best for non-reasoning models.
The model selects one or more Generate IDs as a probability distribution.
If the mode is one of the select logprobs modes, this controls how many of the top options are returned with their probabilities.
This setting aims to control the repetition of tokens based on how often they appear in the input. It tries to use less frequently those tokens that appear more in the input, proportional to how frequently they occur. Token penalty scales with the number of occurrences. Negative values will encourage token reuse.
Accepts a JSON object that maps tokens (specified by their token ID in the tokenizer) to an associated bias value from -100 to 100. Mathematically, the bias is added to the logits generated by the model prior to sampling. The exact effect will vary per model, but values between -1 and 1 should decrease or increase likelihood of selection; values like -100 or 100 should result in a ban or exclusive selection of the relevant token.
An upper bound for the number of tokens that can be generated for a completion, including visible output tokens and reasoning tokens.
This setting aims to control the presence of tokens in the output. It tries to encourage the model to use tokens that are less present in the input, proportional to their presence in the input. Token presence scales with the number of occurrences. Negative values will encourage more diverse token usage.
Constrains effort on reasoning for some reasoning models.
Variants
Stop generation immediately if the model encounters any token specified in the stop array.
Variants
Items
This setting influences the variety in the model’s responses. Lower values lead to more predictable and typical responses, while higher values encourage more diverse and less common responses. At 0, the model always gives the same response for a given input.
This setting limits the model’s choices to a percentage of likely tokens: only the top tokens whose probabilities add up to P. A lower value makes the model’s responses more predictable, while the default setting allows for a full range of token choices. Think of it like a dynamic Top-K.
This sets the upper limit for the number of tokens the model can generate in response. It won’t produce more than this limit. The maximum value is the context length minus the prompt length.
Represents the minimum probability for a token to be considered, relative to the probability of the most likely token. (The value changes depending on the confidence level of the most probable token.) If your Min-P is set to 0.1, that means it will only allow for tokens that are at least 1/10th as probable as the best possible option.
OpenRouter provider preferences.
Properties
List of provider slugs to try in order.
Items
Whether to allow backup providers when the primary is unavailable.
Only use providers that support all parameters in your request.
Control whether to use providers that may store data.
Variants
List of provider slugs to allow for this request.
Items
List of provider slugs to skip for this request.
Items
List of quantization levels to filter by.
Items
Sort providers by price or throughput.
OpenRouter reasoning configuration.
Properties
An upper bound for the number of tokens that can be generated for reasoning.
Constrains effort on reasoning for some reasoning models.
Variants
Whether reasoning is enabled for this request.
Helps to reduce the repetition of tokens from the input. A higher value makes the model less likely to repeat tokens, but too high a value can make the output less coherent (often with run-on sentences that lack small words). Token penalty scales based on original token’s probability.
Consider only the top tokens with “sufficiently high” probabilities based on the probability of the most likely token. Think of it like a dynamic Top-P. A lower Top-A value focuses the choices based on the highest probability token but with a narrower scope. A higher Top-A value does not necessarily affect the creativity of the output, but rather refines the filtering process based on the maximum probability.
This limits the model’s choice of tokens at each step, making it choose from a smaller set. A value of 1 means the model will always pick the most likely next token, leading to predictable results. By default this setting is disabled, making the model to consider all choices.
Controls the verbosity and length of the model response. Lower values produce more concise responses, while higher values produce more detailed and comprehensive responses.
Variants
Fallback models. Will be tried in order if the first one fails.
Items
The weight of the model, which determines its influence on the Confidence Score. Must match the weight strategy of the parent Model.
Variants
A static weight value.
Properties
The static weight value.
A dynamic weight value based on training table data.
Properties
The base weight value, uninfluenced by training table data.
The minimum weight value. A model that never matches the correct answer will have this weight.
The maximum weight value. A model that always matches the correct answer will have this weight.
A base62 22-character unique identifier for the Query LLM. This is a hash of all parameters.
A base62 22-character unique identifier for the Query LLM. This is a hash of some parameters. Only present with Training Table Weight.
The ID of the user who created the Query LLM
The RFC 3339 timestamp when the Query LLM was created
The number of requests made with the Query LLM
The number of chat completion tokens generated by the Query LLM
The number of chat prompt tokens processed by the Query LLM
The total cost of chat completions generated by the Query LLM, in Credits
The number of embedding completion tokens generated by the Query LLM
The number of embedding prompt tokens processed by the Query LLM
The total cost of embedding completions generated by the Query LLM, in Credits