Skip to contents

Typically passed to ql_generate().

Usage

ql_prompt(
  prompt,
  system = NULL,
  format = NULL,
  model = NULL,
  images = NULL,
  temperature = NULL,
  seed = NULL,
  host = NULL,
  hash = TRUE
)

Arguments

prompt

A prompt for the LLM.

system

System message to pass to the model. See official documentation for details. For example: "You are a helpful assistant."

model

The name of the model, e.g. llama3.2 or phi3.5:3.8b. Run ollama list from the command line to see a list of locally available models.

temperature

Numeric value comprised between 0 and 1 passed to the model. When set to 0 and with the same seed, the response to the same prompt is always exactly the same. When closer to one, the response is more variable and creative. Use 0 for consistent responses. Setting this to 0.7 is a common choice for creative or interactive tasks.

seed

An integer. When temperature is set to 0 and the seed is constant, the model consistently returns the same response to the same prompt.

host

The address where the Ollama API can be reached, e.g. http://localhost:11434 for locally deployed Ollama.

hash

Defaults to TRUE. If TRUE, adds a column with the hash of all other components of the prompt. Used internally for caching. Can be added separately with ql_hash().

Value

A tibble with all main components of a query, to be passed to ql_generate().

Details

For more details and context about each parameter, see https://github.com/ollama/ollama/blob/main/docs/api.md.

Examples

ql_prompt("a haiku")
#> # A tibble: 1 × 7
#>   prompt  system                        seed temperature model    format hash   
#>   <chr>   <chr>                        <dbl>       <dbl> <chr>    <chr>  <chr>  
#> 1 a haiku You are a helpful assistant.     0           0 llama3.2 ""     e64e99…