FEATURE: remove gpt-4-turbo-0125 preview swap with gpt-4-turbo (#568)
Open AI just released gpt-4-turbo (with vision) This change stops using the old preview model and swaps with the officially released gpt-4-turbo To come is an implementation of vision.
This commit is contained in:
parent
f20ada00db
commit
6de9c53a71
|
@ -11,7 +11,6 @@ module DiscourseAi
|
||||||
gpt-4
|
gpt-4
|
||||||
gpt-3.5-turbo-16k
|
gpt-3.5-turbo-16k
|
||||||
gpt-4-32k
|
gpt-4-32k
|
||||||
gpt-4-0125-preview
|
|
||||||
gpt-4-turbo
|
gpt-4-turbo
|
||||||
gpt-4-vision-preview
|
gpt-4-vision-preview
|
||||||
].include?(model_name)
|
].include?(model_name)
|
||||||
|
@ -124,8 +123,6 @@ module DiscourseAi
|
||||||
8192
|
8192
|
||||||
when "gpt-4-32k"
|
when "gpt-4-32k"
|
||||||
32_768
|
32_768
|
||||||
when "gpt-4-0125-preview"
|
|
||||||
131_072
|
|
||||||
when "gpt-4-turbo"
|
when "gpt-4-turbo"
|
||||||
131_072
|
131_072
|
||||||
else
|
else
|
||||||
|
|
|
@ -13,7 +13,6 @@ module DiscourseAi
|
||||||
gpt-4
|
gpt-4
|
||||||
gpt-3.5-turbo-16k
|
gpt-3.5-turbo-16k
|
||||||
gpt-4-32k
|
gpt-4-32k
|
||||||
gpt-4-0125-preview
|
|
||||||
gpt-4-turbo
|
gpt-4-turbo
|
||||||
gpt-4-vision-preview
|
gpt-4-vision-preview
|
||||||
].include?(model_name)
|
].include?(model_name)
|
||||||
|
@ -75,7 +74,7 @@ module DiscourseAi
|
||||||
end
|
end
|
||||||
|
|
||||||
def default_options
|
def default_options
|
||||||
{ model: model == "gpt-4-turbo" ? "gpt-4-0125-preview" : model }
|
{ model: model }
|
||||||
end
|
end
|
||||||
|
|
||||||
def provider_id
|
def provider_id
|
||||||
|
|
|
@ -7,7 +7,7 @@ module DiscourseAi
|
||||||
foldable_models = [
|
foldable_models = [
|
||||||
Models::OpenAi.new("open_ai:gpt-4", max_tokens: 8192),
|
Models::OpenAi.new("open_ai:gpt-4", max_tokens: 8192),
|
||||||
Models::OpenAi.new("open_ai:gpt-4-32k", max_tokens: 32_768),
|
Models::OpenAi.new("open_ai:gpt-4-32k", max_tokens: 32_768),
|
||||||
Models::OpenAi.new("open_ai:gpt-4-0125-preview", max_tokens: 100_000),
|
Models::OpenAi.new("open_ai:gpt-4-turbo", max_tokens: 100_000),
|
||||||
Models::OpenAi.new("open_ai:gpt-3.5-turbo", max_tokens: 4096),
|
Models::OpenAi.new("open_ai:gpt-3.5-turbo", max_tokens: 4096),
|
||||||
Models::OpenAi.new("open_ai:gpt-3.5-turbo-16k", max_tokens: 16_384),
|
Models::OpenAi.new("open_ai:gpt-3.5-turbo-16k", max_tokens: 16_384),
|
||||||
Models::Llama2.new(
|
Models::Llama2.new(
|
||||||
|
|
Loading…
Reference in New Issue