Nothing
test_that("Ollama not found error", {
local_mocked_bindings(
test_connection = function() {
x <- list()
x$status_code <- 400
x
}
)
m_defaults_reset()
expect_error(llm_use())
})
test_that("Init code is covered", {
local_mocked_bindings(
test_connection = function() {
x <- list()
x$status_code <- 200
x
},
list_models = function() data.frame(name = c("model1", "model2")),
menu = function(...) 1
)
m_defaults_reset()
expect_snapshot(llm_use(.cache = ""))
})
test_that("Stops cache", {
expect_snapshot(
llm_use("simulate_llm", "echo", .force = TRUE, .cache = "")
)
})
test_that("Chat objects work", {
withr::with_envvar(
new = list(OPENAI_API_KEY = "test"),
{
m_defaults_reset()
chat <- ellmer::chat_openai()
expect_snapshot(
llm_use(chat, .cache = "")
)
}
)
})
test_that("Ensures empty llm_use works with Chat", {
withr::with_envvar(
new = list(OPENAI_API_KEY = "test"),
{
m_defaults_reset()
chat <- ellmer::chat_openai()
llm_use(chat, .cache = "", .silent = TRUE)
expect_snapshot(
llm_use()
)
}
)
})
test_that("Chat error message", {
withr::with_envvar(
new = list(OPENAI_API_KEY = "test"),
{
m_defaults_reset()
chat <- ellmer::chat_openai()
expect_error(
llm_use(chat, .cache = "", model = "test")
)
}
)
})
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.