6623928b95
A recent change meant that llm instance got cached internally, repeat calls to inference would cache data in Endpoint object leading model to failures. Both Gemini and Open AI expect a clean endpoint object cause they set data. This amends internals to make sure llm.generate will always operate on clean objects |
||
---|---|---|
.. | ||
anthropic_spec.rb | ||
aws_bedrock_spec.rb | ||
cohere_spec.rb | ||
endpoint_compliance.rb | ||
gemini_spec.rb | ||
hugging_face_spec.rb | ||
open_ai_spec.rb | ||
vllm_spec.rb |