discourse-ai/spec/shared
Roman Rizzi 9a79afcdbf
DEV: Better strategies for summarization (#88)
* DEV: Better strategies for summarization

The strategy responsibility needs to be "Given a collection of texts, I know how to summarize them most efficiently, using the minimum amount of requests and maximizing token usage".

There are different token limits for each model, so it all boils down to two different strategies:

Fold all these texts into a single one, doing the summarization in chunks, and then build a summary from those.
Build it by combining texts in a single prompt, and truncate it according to your token limits.

While the latter is less than ideal, we need it for "bart-large-cnn-samsum" and "flan-t5-base-samsum", both with low limits. The rest will rely on folding.

* Expose summarized chunks to users
2023-06-27 12:26:33 -03:00
..
inference DEV: Better strategies for summarization (#88) 2023-06-27 12:26:33 -03:00
chat_message_classificator_spec.rb DEV: Fix test (#43) 2023-04-18 15:43:10 -03:00
classificator_spec.rb DEV: DiscourseAI -> DiscourseAi rename to have consistent folders and files (#9) 2023-03-14 16:03:50 -03:00
post_classificator_spec.rb DEV: DiscourseAI -> DiscourseAi rename to have consistent folders and files (#9) 2023-03-14 16:03:50 -03:00
tokenizer.rb FIX: OpenAI Tokenizer was failing to truncate mid emojis (#91) 2023-06-16 15:15:36 -03:00