This function sends a message history to the OpenAI Chat Completions API and returns the assistant's reply.
Usage
openai_chat(
.llm,
.model = "gpt-4o",
.max_completion_tokens = NULL,
.frequency_penalty = NULL,
.logit_bias = NULL,
.presence_penalty = NULL,
.seed = NULL,
.stop = NULL,
.stream = FALSE,
.temperature = NULL,
.top_p = NULL,
.api_url = "https://api.openai.com/",
.timeout = 60,
.verbose = FALSE,
.json = FALSE,
.json_schema = NULL,
.max_tries = 3,
.dry_run = FALSE,
.compatible = FALSE,
.api_path = "/v1/chat/completions"
)
Arguments
- .llm
An
LLMMessage
object containing the conversation history.- .model
The identifier of the model to use (default: "gpt-4o").
- .max_completion_tokens
An upper bound for the number of tokens that can be generated for a completion, including visible output tokens and reasoning tokens.
- .frequency_penalty
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far.
- .logit_bias
A named list modifying the likelihood of specified tokens appearing in the completion.
- .presence_penalty
Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far.
- .seed
If specified, the system will make a best effort to sample deterministically.
- .stop
Up to 4 sequences where the API will stop generating further tokens.
- .stream
If set to TRUE, the answer will be streamed to console as it comes (default: FALSE).
- .temperature
What sampling temperature to use, between 0 and 2. Higher values make the output more random.
- .top_p
An alternative to sampling with temperature, called nucleus sampling.
- .api_url
Base URL for the API (default: "https://api.openai.com/").
- .timeout
Request timeout in seconds (default: 60).
- .verbose
Should additional information be shown after the API call (default: FALSE).
- .json
Should output be in JSON mode (default: FALSE).
- .json_schema
A JSON schema object as R list to enforce the output structure (If defined has precedence over JSON mode).
- .max_tries
Maximum retries to perform request
- .dry_run
If TRUE, perform a dry run and return the request object (default: FALSE).
- .compatible
If TRUE, skip API and rate-limit checks for OpenAI compatible APIs (default: FALSE).
- .api_path
The path relative to the base
.api_url
for the API (default: "/v1/chat/completions").