# frozen_string_literal: true module DiscourseAi module Embeddings class SemanticSearch def self.clear_cache_for(query) digest = OpenSSL::Digest::SHA1.hexdigest(query) hyde_key = "semantic-search-#{digest}-#{SiteSetting.ai_embeddings_semantic_search_hyde_model}" Discourse.cache.delete(hyde_key) Discourse.cache.delete("#{hyde_key}-#{SiteSetting.ai_embeddings_model}") end def initialize(guardian) @guardian = guardian end def cached_query?(query) digest = OpenSSL::Digest::SHA1.hexdigest(query) embedding_key = build_embedding_key( digest, SiteSetting.ai_embeddings_semantic_search_hyde_model, SiteSetting.ai_embeddings_model, ) Discourse.cache.read(embedding_key).present? end def search_for_topics(query, page = 1) max_results_per_page = 100 limit = [Search.per_filter, max_results_per_page].min + 1 offset = (page - 1) * limit search = Search.new(query, { guardian: guardian }) search_term = search.term return [] if search_term.nil? || search_term.length < SiteSetting.min_search_term_length strategy = DiscourseAi::Embeddings::Strategies::Truncation.new vector_rep = DiscourseAi::Embeddings::VectorRepresentations::Base.current_representation(strategy) digest = OpenSSL::Digest::SHA1.hexdigest(search_term) hyde_key = build_hyde_key(digest, SiteSetting.ai_embeddings_semantic_search_hyde_model) embedding_key = build_embedding_key( digest, SiteSetting.ai_embeddings_semantic_search_hyde_model, SiteSetting.ai_embeddings_model, ) hypothetical_post = Discourse .cache .fetch(hyde_key, expires_in: 1.week) { hypothetical_post_from(search_term) } hypothetical_post_embedding = Discourse .cache .fetch(embedding_key, expires_in: 1.week) { vector_rep.vector_from(hypothetical_post) } candidate_topic_ids = vector_rep.asymmetric_topics_similarity_search( hypothetical_post_embedding, limit: limit, offset: offset, ) semantic_results = ::Post .where(post_type: ::Topic.visible_post_types(guardian.user)) .public_posts .where("topics.visible") .where(topic_id: candidate_topic_ids, post_number: 1) .order("array_position(ARRAY#{candidate_topic_ids}, posts.topic_id)") query_filter_results = search.apply_filters(semantic_results) guardian.filter_allowed_categories(query_filter_results) end def quick_search(query) max_semantic_results_per_page = 100 search = Search.new(query, { guardian: guardian }) search_term = search.term return [] if search_term.nil? || search_term.length < SiteSetting.min_search_term_length strategy = DiscourseAi::Embeddings::Strategies::Truncation.new vector_rep = DiscourseAi::Embeddings::VectorRepresentations::Base.current_representation(strategy) digest = OpenSSL::Digest::SHA1.hexdigest(search_term) embedding_key = build_embedding_key( digest, SiteSetting.ai_embeddings_semantic_search_hyde_model, SiteSetting.ai_embeddings_model, ) search_term_embedding = Discourse .cache .fetch(embedding_key, expires_in: 1.week) do vector_rep.vector_from(search_term, asymetric: true) end candidate_post_ids = vector_rep.asymmetric_posts_similarity_search( search_term_embedding, limit: max_semantic_results_per_page, offset: 0, ) semantic_results = ::Post .where(post_type: ::Topic.visible_post_types(guardian.user)) .public_posts .where("topics.visible") .where(id: candidate_post_ids) .order("array_position(ARRAY#{candidate_post_ids}, posts.id)") filtered_results = search.apply_filters(semantic_results) rerank_posts_payload = filtered_results .map(&:cooked) .map { Nokogiri::HTML5.fragment(_1).text } .map { _1.truncate(2000, omission: "") } reranked_results = DiscourseAi::Inference::HuggingFaceTextEmbeddings.rerank( search_term, rerank_posts_payload, ) reordered_ids = reranked_results.map { _1[:index] }.map { filtered_results[_1].id }.take(5) reranked_semantic_results = ::Post .where(post_type: ::Topic.visible_post_types(guardian.user)) .public_posts .where("topics.visible") .where(id: reordered_ids) .order("array_position(ARRAY#{reordered_ids}, posts.id)") guardian.filter_allowed_categories(reranked_semantic_results) end private attr_reader :guardian def build_hyde_key(digest, hyde_model) "semantic-search-#{digest}-#{hyde_model}" end def build_embedding_key(digest, hyde_model, embedding_model) "#{build_hyde_key(digest, hyde_model)}-#{embedding_model}" end def hypothetical_post_from(search_term) prompt = DiscourseAi::Completions::Prompt.new(<<~TEXT.strip) You are a content creator for a forum. The forum description is as follows: #{SiteSetting.title} #{SiteSetting.site_description} Put the forum post between tags. TEXT prompt.push(type: :user, content: <<~TEXT.strip) Using this description, write a forum post about the subject inside the XML tags: #{search_term} TEXT llm_response = DiscourseAi::Completions::Llm.proxy( SiteSetting.ai_embeddings_semantic_search_hyde_model, ).generate(prompt, user: @guardian.user) Nokogiri::HTML5.fragment(llm_response).at("ai")&.text&.presence || llm_response end end end end