mirror of
https://github.com/discourse/discourse-ai.git
synced 2025-06-29 11:02:17 +00:00
FIX: Don't echo prompt back on HF/TGI (#338)
* FIX: Don't echo prompt back on HF/TGI * teeeeests
This commit is contained in:
parent
d8267d8da0
commit
252efdf142
@ -33,6 +33,7 @@ module DiscourseAi
|
|||||||
token_limit = SiteSetting.ai_hugging_face_token_limit || 4_000
|
token_limit = SiteSetting.ai_hugging_face_token_limit || 4_000
|
||||||
|
|
||||||
payload[:parameters][:max_new_tokens] = token_limit - prompt_size(prompt)
|
payload[:parameters][:max_new_tokens] = token_limit - prompt_size(prompt)
|
||||||
|
payload[:parameters][:return_full_text] = false
|
||||||
|
|
||||||
payload[:stream] = true if @streaming_mode
|
payload[:stream] = true if @streaming_mode
|
||||||
end
|
end
|
||||||
|
@ -40,6 +40,7 @@ module ::DiscourseAi
|
|||||||
parameters[:max_new_tokens] = token_limit - prompt_size
|
parameters[:max_new_tokens] = token_limit - prompt_size
|
||||||
parameters[:temperature] = temperature if temperature
|
parameters[:temperature] = temperature if temperature
|
||||||
parameters[:repetition_penalty] = repetition_penalty if repetition_penalty
|
parameters[:repetition_penalty] = repetition_penalty if repetition_penalty
|
||||||
|
parameters[:return_full_text] = false
|
||||||
|
|
||||||
payload[:stream] = true if block_given?
|
payload[:stream] = true if block_given?
|
||||||
|
|
||||||
|
@ -18,6 +18,7 @@ RSpec.describe DiscourseAi::Completions::Endpoints::HuggingFace do
|
|||||||
.tap do |payload|
|
.tap do |payload|
|
||||||
payload[:parameters][:max_new_tokens] = (SiteSetting.ai_hugging_face_token_limit || 4_000) -
|
payload[:parameters][:max_new_tokens] = (SiteSetting.ai_hugging_face_token_limit || 4_000) -
|
||||||
model.prompt_size(prompt)
|
model.prompt_size(prompt)
|
||||||
|
payload[:parameters][:return_full_text] = false
|
||||||
end
|
end
|
||||||
.to_json
|
.to_json
|
||||||
end
|
end
|
||||||
@ -29,6 +30,7 @@ RSpec.describe DiscourseAi::Completions::Endpoints::HuggingFace do
|
|||||||
payload[:parameters][:max_new_tokens] = (SiteSetting.ai_hugging_face_token_limit || 4_000) -
|
payload[:parameters][:max_new_tokens] = (SiteSetting.ai_hugging_face_token_limit || 4_000) -
|
||||||
model.prompt_size(prompt)
|
model.prompt_size(prompt)
|
||||||
payload[:stream] = true
|
payload[:stream] = true
|
||||||
|
payload[:parameters][:return_full_text] = false
|
||||||
end
|
end
|
||||||
.to_json
|
.to_json
|
||||||
end
|
end
|
||||||
|
Loading…
x
Reference in New Issue
Block a user