2023-04-21 02:54:25 -04:00
|
|
|
# frozen_string_literal: true
|
|
|
|
require "rails_helper"
|
|
|
|
|
2023-05-05 14:28:31 -04:00
|
|
|
require_relative "../../support/openai_completions_inference_stubs"
|
|
|
|
|
2023-04-21 02:54:25 -04:00
|
|
|
describe DiscourseAi::Inference::OpenAiCompletions do
|
|
|
|
before { SiteSetting.ai_openai_api_key = "abc-123" }
|
|
|
|
|
2023-06-20 20:39:51 -04:00
|
|
|
context "when configured using Azure" do
|
2023-08-16 21:00:11 -04:00
|
|
|
it "Supports custom Azure endpoints for completions" do
|
|
|
|
gpt_url_base =
|
2023-06-20 20:39:51 -04:00
|
|
|
"https://company.openai.azure.com/openai/deployments/deployment/chat/completions?api-version=2023-03-15-preview"
|
2023-08-16 21:00:11 -04:00
|
|
|
key = "12345"
|
|
|
|
SiteSetting.ai_openai_api_key = key
|
2023-06-20 20:39:51 -04:00
|
|
|
|
2023-08-16 21:00:11 -04:00
|
|
|
[
|
|
|
|
{ setting_name: "ai_openai_gpt35_url", model: "gpt-35-turbo" },
|
|
|
|
{ setting_name: "ai_openai_gpt35_16k_url", model: "gpt-35-16k-turbo" },
|
|
|
|
{ setting_name: "ai_openai_gpt4_url", model: "gpt-4" },
|
|
|
|
{ setting_name: "ai_openai_gpt4_32k_url", model: "gpt-4-32k" },
|
|
|
|
].each do |config|
|
|
|
|
gpt_url = "#{gpt_url_base}/#{config[:model]}"
|
|
|
|
setting_name = config[:setting_name]
|
|
|
|
model = config[:model]
|
|
|
|
|
|
|
|
SiteSetting.public_send("#{setting_name}=".to_sym, gpt_url)
|
|
|
|
|
|
|
|
expected = {
|
|
|
|
id: "chatcmpl-7TfPzOyBGW5K6dyWp3NPU0mYLGZRQ",
|
|
|
|
object: "chat.completion",
|
|
|
|
created: 1_687_305_079,
|
|
|
|
model: model,
|
|
|
|
choices: [
|
|
|
|
{
|
|
|
|
index: 0,
|
|
|
|
finish_reason: "stop",
|
|
|
|
message: {
|
|
|
|
role: "assistant",
|
|
|
|
content: "Hi there! How can I assist you today?",
|
|
|
|
},
|
2023-06-20 20:39:51 -04:00
|
|
|
},
|
2023-08-16 21:00:11 -04:00
|
|
|
],
|
|
|
|
usage: {
|
|
|
|
completion_tokens: 10,
|
|
|
|
prompt_tokens: 9,
|
|
|
|
total_tokens: 19,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
stub_request(:post, gpt_url).with(
|
|
|
|
body: "{\"model\":\"#{model}\",\"messages\":[{\"role\":\"user\",\"content\":\"hello\"}]}",
|
|
|
|
headers: {
|
|
|
|
"Api-Key" => "12345",
|
|
|
|
"Content-Type" => "application/json",
|
|
|
|
"Host" => "company.openai.azure.com",
|
2023-06-20 20:39:51 -04:00
|
|
|
},
|
2023-08-16 21:00:11 -04:00
|
|
|
).to_return(status: 200, body: expected.to_json, headers: {})
|
|
|
|
|
|
|
|
result =
|
|
|
|
DiscourseAi::Inference::OpenAiCompletions.perform!(
|
|
|
|
[role: "user", content: "hello"],
|
|
|
|
model,
|
|
|
|
)
|
|
|
|
|
|
|
|
expect(result).to eq(expected)
|
|
|
|
end
|
2023-06-20 20:39:51 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2023-06-19 18:45:31 -04:00
|
|
|
it "supports function calling" do
|
|
|
|
prompt = [role: "system", content: "you are weatherbot"]
|
|
|
|
prompt << { role: "user", content: "what is the weather in sydney?" }
|
|
|
|
|
|
|
|
functions = []
|
|
|
|
|
|
|
|
function =
|
2023-08-22 17:49:36 -04:00
|
|
|
DiscourseAi::Inference::Function.new(
|
2023-06-19 18:45:31 -04:00
|
|
|
name: "get_weather",
|
|
|
|
description: "Get the weather in a city",
|
|
|
|
)
|
|
|
|
|
|
|
|
function.add_parameter(
|
|
|
|
name: "location",
|
|
|
|
type: "string",
|
|
|
|
description: "the city name",
|
|
|
|
required: true,
|
|
|
|
)
|
|
|
|
|
|
|
|
function.add_parameter(
|
|
|
|
name: "unit",
|
|
|
|
type: "string",
|
|
|
|
description: "the unit of measurement celcius c or fahrenheit f",
|
|
|
|
enum: %w[c f],
|
|
|
|
required: true,
|
|
|
|
)
|
|
|
|
|
|
|
|
functions << function
|
|
|
|
|
|
|
|
function_calls = []
|
|
|
|
current_function_call = nil
|
|
|
|
|
|
|
|
deltas = [
|
|
|
|
{ role: "assistant" },
|
|
|
|
{ function_call: { name: "get_weather", arguments: "" } },
|
|
|
|
{ function_call: { arguments: "{ \"location\": " } },
|
|
|
|
{ function_call: { arguments: "\"sydney\", \"unit\": \"c\" }" } },
|
|
|
|
]
|
|
|
|
|
|
|
|
OpenAiCompletionsInferenceStubs.stub_streamed_response(
|
|
|
|
prompt,
|
|
|
|
deltas,
|
|
|
|
model: "gpt-3.5-turbo-0613",
|
|
|
|
req_opts: {
|
|
|
|
functions: functions,
|
|
|
|
stream: true,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
DiscourseAi::Inference::OpenAiCompletions.perform!(
|
|
|
|
prompt,
|
|
|
|
"gpt-3.5-turbo-0613",
|
|
|
|
functions: functions,
|
|
|
|
) do |json, cancel|
|
|
|
|
fn = json.dig(:choices, 0, :delta, :function_call)
|
|
|
|
if fn && fn[:name]
|
|
|
|
current_function_call = { name: fn[:name], arguments: +fn[:arguments].to_s.dup }
|
|
|
|
function_calls << current_function_call
|
|
|
|
elsif fn && fn[:arguments] && current_function_call
|
|
|
|
current_function_call[:arguments] << fn[:arguments]
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
expect(function_calls.length).to eq(1)
|
|
|
|
expect(function_calls[0][:name]).to eq("get_weather")
|
|
|
|
expect(JSON.parse(function_calls[0][:arguments])).to eq(
|
|
|
|
{ "location" => "sydney", "unit" => "c" },
|
|
|
|
)
|
|
|
|
|
|
|
|
prompt << { role: "function", name: "get_weather", content: 22.to_json }
|
|
|
|
|
|
|
|
OpenAiCompletionsInferenceStubs.stub_response(
|
|
|
|
prompt,
|
|
|
|
"The current temperature in Sydney is 22 degrees Celsius.",
|
|
|
|
model: "gpt-3.5-turbo-0613",
|
|
|
|
req_opts: {
|
|
|
|
functions: functions,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
result =
|
|
|
|
DiscourseAi::Inference::OpenAiCompletions.perform!(
|
|
|
|
prompt,
|
|
|
|
"gpt-3.5-turbo-0613",
|
|
|
|
functions: functions,
|
|
|
|
)
|
|
|
|
|
|
|
|
expect(result.dig(:choices, 0, :message, :content)).to eq(
|
|
|
|
"The current temperature in Sydney is 22 degrees Celsius.",
|
|
|
|
)
|
|
|
|
end
|
|
|
|
|
2023-10-04 18:00:45 -04:00
|
|
|
it "supports rate limits" do
|
|
|
|
stub_request(:post, "https://api.openai.com/v1/chat/completions").to_return(
|
|
|
|
[
|
|
|
|
{ status: 429, body: "", headers: {} },
|
|
|
|
{ status: 429, body: "", headers: {} },
|
|
|
|
{ status: 200, body: { choices: [message: { content: "ok" }] }.to_json, headers: {} },
|
|
|
|
],
|
|
|
|
)
|
|
|
|
completions =
|
|
|
|
DiscourseAi::Inference::OpenAiCompletions.perform!(
|
|
|
|
[{ role: "user", content: "hello" }],
|
|
|
|
"gpt-3.5-turbo",
|
|
|
|
temperature: 0.5,
|
|
|
|
top_p: 0.8,
|
|
|
|
max_tokens: 700,
|
|
|
|
retries: 3,
|
|
|
|
retry_timeout: 0,
|
|
|
|
)
|
|
|
|
|
|
|
|
expect(completions.dig(:choices, 0, :message, :content)).to eq("ok")
|
|
|
|
end
|
|
|
|
|
|
|
|
it "supports will raise once rate limit is met" do
|
|
|
|
stub_request(:post, "https://api.openai.com/v1/chat/completions").to_return(
|
|
|
|
[
|
|
|
|
{ status: 429, body: "", headers: {} },
|
|
|
|
{ status: 429, body: "", headers: {} },
|
|
|
|
{ status: 429, body: "", headers: {} },
|
|
|
|
],
|
|
|
|
)
|
|
|
|
|
|
|
|
expect do
|
|
|
|
DiscourseAi::Inference::OpenAiCompletions.perform!(
|
|
|
|
[{ role: "user", content: "hello" }],
|
|
|
|
"gpt-3.5-turbo",
|
|
|
|
temperature: 0.5,
|
|
|
|
top_p: 0.8,
|
|
|
|
max_tokens: 700,
|
|
|
|
retries: 3,
|
|
|
|
retry_timeout: 0,
|
|
|
|
)
|
|
|
|
end.to raise_error(DiscourseAi::Inference::OpenAiCompletions::CompletionFailed)
|
|
|
|
end
|
|
|
|
|
2023-04-21 02:54:25 -04:00
|
|
|
it "can complete a trivial prompt" do
|
2023-05-05 14:28:31 -04:00
|
|
|
response_text = "1. Serenity\\n2. Laughter\\n3. Adventure"
|
|
|
|
prompt = [role: "user", content: "write 3 words"]
|
2023-04-25 21:44:29 -04:00
|
|
|
user_id = 183
|
2023-05-05 14:28:31 -04:00
|
|
|
req_opts = { temperature: 0.5, top_p: 0.8, max_tokens: 700 }
|
|
|
|
|
|
|
|
OpenAiCompletionsInferenceStubs.stub_response(prompt, response_text, req_opts: req_opts)
|
2023-04-25 21:44:29 -04:00
|
|
|
|
2023-04-21 02:54:25 -04:00
|
|
|
completions =
|
|
|
|
DiscourseAi::Inference::OpenAiCompletions.perform!(
|
|
|
|
prompt,
|
|
|
|
"gpt-3.5-turbo",
|
|
|
|
temperature: 0.5,
|
|
|
|
top_p: 0.8,
|
|
|
|
max_tokens: 700,
|
2023-04-25 21:44:29 -04:00
|
|
|
user_id: user_id,
|
2023-04-21 02:54:25 -04:00
|
|
|
)
|
2023-05-05 14:28:31 -04:00
|
|
|
|
|
|
|
expect(completions.dig(:choices, 0, :message, :content)).to eq(response_text)
|
2023-04-25 21:44:29 -04:00
|
|
|
|
|
|
|
expect(AiApiAuditLog.count).to eq(1)
|
|
|
|
log = AiApiAuditLog.first
|
|
|
|
|
2023-05-05 14:28:31 -04:00
|
|
|
body = { model: "gpt-3.5-turbo", messages: prompt }.merge(req_opts).to_json
|
|
|
|
request_body = OpenAiCompletionsInferenceStubs.response(response_text).to_json
|
2023-04-25 21:44:29 -04:00
|
|
|
|
|
|
|
expect(log.provider_id).to eq(AiApiAuditLog::Provider::OpenAI)
|
2023-05-05 14:28:31 -04:00
|
|
|
expect(log.request_tokens).to eq(337)
|
|
|
|
expect(log.response_tokens).to eq(162)
|
|
|
|
expect(log.raw_request_payload).to eq(body)
|
|
|
|
expect(log.raw_response_payload).to eq(request_body)
|
2023-04-21 02:54:25 -04:00
|
|
|
end
|
|
|
|
|
|
|
|
it "can operate in streaming mode" do
|
2023-05-05 14:28:31 -04:00
|
|
|
deltas = [
|
|
|
|
{ role: "assistant" },
|
|
|
|
{ content: "Mount" },
|
|
|
|
{ content: "ain" },
|
|
|
|
{ content: " " },
|
|
|
|
{ content: "Tree " },
|
|
|
|
{ content: "Frog" },
|
|
|
|
]
|
2023-04-21 02:54:25 -04:00
|
|
|
|
|
|
|
prompt = [role: "user", content: "write 3 words"]
|
|
|
|
content = +""
|
|
|
|
|
2023-05-05 14:28:31 -04:00
|
|
|
OpenAiCompletionsInferenceStubs.stub_streamed_response(
|
2023-04-21 02:54:25 -04:00
|
|
|
prompt,
|
2023-05-05 14:28:31 -04:00
|
|
|
deltas,
|
|
|
|
req_opts: {
|
|
|
|
stream: true,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
DiscourseAi::Inference::OpenAiCompletions.perform!(prompt, "gpt-3.5-turbo") do |partial, cancel|
|
|
|
|
data = partial.dig(:choices, 0, :delta, :content)
|
2023-04-21 02:54:25 -04:00
|
|
|
content << data if data
|
|
|
|
cancel.call if content.split(" ").length == 2
|
|
|
|
end
|
|
|
|
|
|
|
|
expect(content).to eq("Mountain Tree ")
|
2023-04-25 21:44:29 -04:00
|
|
|
|
|
|
|
expect(AiApiAuditLog.count).to eq(1)
|
|
|
|
log = AiApiAuditLog.first
|
|
|
|
|
2023-05-05 14:28:31 -04:00
|
|
|
request_body = { model: "gpt-3.5-turbo", messages: prompt, stream: true }.to_json
|
2023-04-25 21:44:29 -04:00
|
|
|
|
|
|
|
expect(log.provider_id).to eq(AiApiAuditLog::Provider::OpenAI)
|
2023-05-20 03:45:54 -04:00
|
|
|
expect(log.request_tokens).to eq(4)
|
|
|
|
expect(log.response_tokens).to eq(3)
|
2023-04-25 21:44:29 -04:00
|
|
|
expect(log.raw_request_payload).to eq(request_body)
|
2023-05-05 14:28:31 -04:00
|
|
|
expect(log.raw_response_payload).to be_present
|
2023-04-21 02:54:25 -04:00
|
|
|
end
|
|
|
|
end
|