mirror of
https://github.com/discourse/discourse-ai.git
synced 2025-02-27 22:09:23 +00:00
A recent change meant that llm instance got cached internally, repeat calls to inference would cache data in Endpoint object leading model to failures. Both Gemini and Open AI expect a clean endpoint object cause they set data. This amends internals to make sure llm.generate will always operate on clean objects
Discourse AI Plugin
Plugin Summary
For more information, please see: https://meta.discourse.org/t/discourse-ai/259214?u=falco
Description
Languages
Ruby
79.2%
JavaScript
17.3%
SCSS
2.1%
CSS
0.6%
HTML
0.5%
Other
0.3%