51 lines
1.9 KiB
Ruby
51 lines
1.9 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
module DiscourseAi
|
|
module Automation
|
|
AVAILABLE_MODELS = [
|
|
{ id: "gpt-4-turbo", name: "discourse_automation.ai_models.gpt_4_turbo" },
|
|
{ id: "gpt-4", name: "discourse_automation.ai_models.gpt_4" },
|
|
{ id: "gpt-3.5-turbo", name: "discourse_automation.ai_models.gpt_3_5_turbo" },
|
|
{ id: "gemini-pro", name: "discourse_automation.ai_models.gemini_pro" },
|
|
{ id: "claude-2", name: "discourse_automation.ai_models.claude_2" },
|
|
{ id: "claude-3-sonnet", name: "discourse_automation.ai_models.claude_3_sonnet" },
|
|
{ id: "claude-3-opus", name: "discourse_automation.ai_models.claude_3_opus" },
|
|
{ id: "claude-3-haiku", name: "discourse_automation.ai_models.claude_3_haiku" },
|
|
{
|
|
id: "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
|
name: "discourse_automation.ai_models.mixtral_8x7b_instruct_v0_1",
|
|
},
|
|
{
|
|
id: "mistralai/Mistral-7B-Instruct-v0.2",
|
|
name: "discourse_automation.ai_models.mistral_7b_instruct_v0_2",
|
|
},
|
|
{ id: "command-r", name: "discourse_automation.ai_models.command_r" },
|
|
{ id: "command-r-plus", name: "discourse_automation.ai_models.command_r_plus" },
|
|
]
|
|
|
|
def self.translate_model(model)
|
|
return "google:gemini-pro" if model == "gemini-pro"
|
|
return "open_ai:#{model}" if model.start_with? "gpt"
|
|
return "cohere:#{model}" if model.start_with? "command"
|
|
|
|
if model.start_with? "claude"
|
|
if DiscourseAi::Completions::Endpoints::AwsBedrock.correctly_configured?(model)
|
|
return "aws_bedrock:#{model}"
|
|
else
|
|
return "anthropic:#{model}"
|
|
end
|
|
end
|
|
|
|
if model.start_with?("mistral")
|
|
if DiscourseAi::Completions::Endpoints::Vllm.correctly_configured?(model)
|
|
return "vllm:#{model}"
|
|
else
|
|
return "hugging_face:#{model}"
|
|
end
|
|
end
|
|
|
|
raise "Unknown model #{model}"
|
|
end
|
|
end
|
|
end
|