mirror of
https://github.com/discourse/discourse-ai.git
synced 2025-03-09 02:40:50 +00:00
* FIX/REFACTOR: FoldContent revamp We hit a snag with our hot topic gist strategy: the regex we used to split the content didn't work, so we cannot send the original post separately. This was important for letting the model focus on what's new in the topic. The algorithm doesn’t give us full control over how prompts are written, and figuring out how to format the content isn't straightforward. This means we're having to use more complicated workarounds, like regex. To tackle this, I'm suggesting we simplify the approach a bit. Let's focus on summarizing as much as we can upfront, then gradually add new content until there's nothing left to summarize. Also, the "extend" part is mostly for models with small context windows, which shouldn't pose a problem 99% of the time with the content volume we're dealing with. * Fix fold docs * Use #shift instead of #pop to get the first elem, not the last
144 lines
5.0 KiB
Ruby
144 lines
5.0 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
module DiscourseAi
|
|
module Summarization
|
|
# This class offers a generic way of summarizing content from multiple sources using different prompts.
|
|
#
|
|
# It summarizes large amounts of content by recursively summarizing it in smaller chunks that
|
|
# fit the given model context window, finally concatenating the disjoint summaries
|
|
# into a final version.
|
|
#
|
|
class FoldContent
|
|
def initialize(llm, strategy, persist_summaries: true)
|
|
@llm = llm
|
|
@strategy = strategy
|
|
@persist_summaries = persist_summaries
|
|
end
|
|
|
|
attr_reader :llm, :strategy
|
|
|
|
# @param user { User } - User object used for auditing usage.
|
|
# @param &on_partial_blk { Block - Optional } - The passed block will get called with the LLM partial response alongside a cancel function.
|
|
# Note: The block is only called with results of the final summary, not intermediate summaries.
|
|
#
|
|
# @returns { AiSummary } - Resulting summary.
|
|
def summarize(user, &on_partial_blk)
|
|
base_summary = ""
|
|
initial_pos = 0
|
|
folded_summary =
|
|
fold(content_to_summarize, base_summary, initial_pos, user, &on_partial_blk)
|
|
|
|
clean_summary =
|
|
Nokogiri::HTML5.fragment(folded_summary).css("ai")&.first&.text || folded_summary
|
|
|
|
if persist_summaries
|
|
AiSummary.store!(
|
|
strategy.target,
|
|
strategy.type,
|
|
llm_model.name,
|
|
clean_summary,
|
|
content_to_summarize.map { |c| c[:id] },
|
|
)
|
|
else
|
|
AiSummary.new(summarized_text: clean_summary)
|
|
end
|
|
end
|
|
|
|
# @returns { AiSummary } - Resulting summary.
|
|
#
|
|
# Finds a summary matching the target and strategy. Marks it as outdates if the strategy found newer content
|
|
def existing_summary
|
|
if !defined?(@existing_summary)
|
|
summary = AiSummary.find_by(target: strategy.target, summary_type: strategy.type)
|
|
|
|
if summary
|
|
@existing_summary = summary
|
|
|
|
if existing_summary.original_content_sha != latest_sha
|
|
@existing_summary.mark_as_outdated
|
|
end
|
|
end
|
|
end
|
|
@existing_summary
|
|
end
|
|
|
|
def delete_cached_summaries!
|
|
AiSummary.where(target: strategy.target, summary_type: strategy.type).destroy_all
|
|
end
|
|
|
|
private
|
|
|
|
attr_reader :persist_summaries
|
|
|
|
def llm_model
|
|
llm.llm_model
|
|
end
|
|
|
|
def content_to_summarize
|
|
@targets_data ||= strategy.targets_data
|
|
end
|
|
|
|
def latest_sha
|
|
@latest_sha ||= AiSummary.build_sha(content_to_summarize.map { |c| c[:id] }.join)
|
|
end
|
|
|
|
# @param items { Array<Hash> } - Content to summarize. Structure will be: { poster: who wrote the content, id: a way to order content, text: content }
|
|
# @param summary { String } - Intermediate summaries that we'll keep extending as part of our "folding" algorithm.
|
|
# @param cursor { Integer } - Idx to know how much we already summarized.
|
|
# @param user { User } - User object used for auditing usage.
|
|
# @param &on_partial_blk { Block - Optional } - The passed block will get called with the LLM partial response alongside a cancel function.
|
|
# Note: The block is only called with results of the final summary, not intermediate summaries.
|
|
#
|
|
# The summarization algorithm.
|
|
# The idea is to build an initial summary packing as much content as we can. Once we have the initial summary, we'll keep extending using the leftover
|
|
# content until there is nothing left.
|
|
#
|
|
# @returns { String } - Resulting summary.
|
|
def fold(items, summary, cursor, user, &on_partial_blk)
|
|
tokenizer = llm_model.tokenizer_class
|
|
tokens_left = available_tokens - tokenizer.size(summary)
|
|
iteration_content = []
|
|
|
|
items.each_with_index do |item, idx|
|
|
next if idx < cursor
|
|
|
|
as_text = "(#{item[:id]} #{item[:poster]} said: #{item[:text]} "
|
|
|
|
if tokenizer.below_limit?(as_text, tokens_left)
|
|
iteration_content << item
|
|
tokens_left -= tokenizer.size(as_text)
|
|
cursor += 1
|
|
else
|
|
break
|
|
end
|
|
end
|
|
|
|
prompt =
|
|
(
|
|
if summary.blank?
|
|
strategy.first_summary_prompt(iteration_content)
|
|
else
|
|
strategy.summary_extension_prompt(summary, iteration_content)
|
|
end
|
|
)
|
|
|
|
if cursor == items.length
|
|
llm.generate(prompt, user: user, feature_name: "summarize", &on_partial_blk)
|
|
else
|
|
latest_summary =
|
|
llm.generate(prompt, user: user, max_tokens: 600, feature_name: "summarize")
|
|
fold(items, latest_summary, cursor, user, &on_partial_blk)
|
|
end
|
|
end
|
|
|
|
def available_tokens
|
|
# Reserve tokens for the response and the base prompt
|
|
# ~500 words
|
|
reserved_tokens = 700
|
|
|
|
llm_model.max_prompt_tokens - reserved_tokens
|
|
end
|
|
end
|
|
end
|
|
end
|