Skip to content

Commit

Permalink
Implement chat_vllm()
Browse files Browse the repository at this point in the history
Fixes #140
  • Loading branch information
hadley committed Nov 7, 2024
1 parent dda34a3 commit 3bf0c51
Show file tree
Hide file tree
Showing 5 changed files with 146 additions and 0 deletions.
1 change: 1 addition & 0 deletions DESCRIPTION
Original file line number Diff line number Diff line change
Expand Up @@ -73,6 +73,7 @@ Collate:
'provider-groq.R'
'provider-ollama.R'
'provider-perplexity.R'
'provider-vllm.R'
'shiny.R'
'tokens.R'
'tools-def-auto.R'
Expand Down
1 change: 1 addition & 0 deletions NAMESPACE
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@ export(chat_groq)
export(chat_ollama)
export(chat_openai)
export(chat_perplexity)
export(chat_vllm)
export(content_image_file)
export(content_image_plot)
export(content_image_url)
Expand Down
2 changes: 2 additions & 0 deletions NEWS.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,7 @@
# elmer (development version)

* New `chat_vllm()` to chat with models served by vLLM (#140).

* New `chat_bedrock()` to chat with AWS bedrock models (#50).

* New `chat$extract_data()` uses the structured data API where available (and tool calling otherwise) to extract data structured according to a known type specification. You can create specs with functions `type_boolean()`, `type_integer()`, `type_number()`, `type_string()`, `type_enum()`, `type_array()`, and `type_object()` (#31).
Expand Down
79 changes: 79 additions & 0 deletions R/provider-vllm.R
Original file line number Diff line number Diff line change
@@ -0,0 +1,79 @@
#' @include provider-openai.R
#' @include content.R
NULL

#' Chat with a model hosted by vLLM
#'
#' @description
#' [vLLM](https://docs.vllm.ai/en/latest/) is an open source library that
#' provides an efficient and convenient LLMs model server. You can use
#' `chat_vllm()` to connect to endpoints powered by vLLM.
#'
#' @inheritParams chat_openai
#' @inherit chat_openai return
#' @export
chat_vllm <- function(base_url,
system_prompt = NULL,
turns = NULL,
model,
seed = NULL,
api_args = list(),
api_key = vllm_key(),
echo = NULL) {

check_string(base_url)
turns <- normalize_turns(turns, system_prompt)
check_string(api_key)
if (missing(model)) {
models <- vllm_models(base_url, api_key)
cli::cli_abort(c(
"Must specify {.arg model}.",
i = "Available models: {.str {models}}."
))
}
if (is_testing() && is.null(seed)) {
seed <- seed %||% 1014
}
echo <- check_echo(echo)

provider <- ProviderVllm(
base_url = base_url,
model = model,
seed = seed,
extra_args = api_args,
api_key = api_key
)
Chat$new(provider = provider, turns = turns, echo = echo)
}

ProviderVllm <- new_class(
"ProviderVllm",
parent = ProviderOpenAI,
package = "elmer",
)

# Just like OpenAI but no strict
method(as_json, list(ProviderVllm, ToolDef)) <- function(provider, x) {
list(
type = "function",
"function" = compact(list(
name = x@name,
description = x@description,
parameters = as_json(provider, x@arguments)
))
)
}

vllm_key <- function() {
key_get("VLLM_KEY")
}

vllm_models <- function(base_url, key = vllm_key()) {
req <- request(base_url)
req <- req_auth_bearer_token(req, key)
req <- req_url_path(req, "/v1/models")
resp <- req_perform(req)
json <- resp_body_json(resp)

map_chr(json$data, "[[", "id")
}
63 changes: 63 additions & 0 deletions man/chat_vllm.Rd

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

0 comments on commit 3bf0c51

Please sign in to comment.