mirror of
https://github.com/discourse/discourse-ai.git
synced 2025-07-08 23:32:45 +00:00
This PR introduces several enhancements and refactorings to the AI Persona and RAG (Retrieval-Augmented Generation) functionalities within the discourse-ai plugin. Here's a breakdown of the changes: **1. LLM Model Association for RAG and Personas:** - **New Database Columns:** Adds `rag_llm_model_id` to both `ai_personas` and `ai_tools` tables. This allows specifying a dedicated LLM for RAG indexing, separate from the persona's primary LLM. Adds `default_llm_id` and `question_consolidator_llm_id` to `ai_personas`. - **Migration:** Includes a migration (`20250210032345_migrate_persona_to_llm_model_id.rb`) to populate the new `default_llm_id` and `question_consolidator_llm_id` columns in `ai_personas` based on the existing `default_llm` and `question_consolidator_llm` string columns, and a post migration to remove the latter. - **Model Changes:** The `AiPersona` and `AiTool` models now `belong_to` an `LlmModel` via `rag_llm_model_id`. The `LlmModel.proxy` method now accepts an `LlmModel` instance instead of just an identifier. `AiPersona` now has `default_llm_id` and `question_consolidator_llm_id` attributes. - **UI Updates:** The AI Persona and AI Tool editors in the admin panel now allow selecting an LLM for RAG indexing (if PDF/image support is enabled). The RAG options component displays an LLM selector. - **Serialization:** The serializers (`AiCustomToolSerializer`, `AiCustomToolListSerializer`, `LocalizedAiPersonaSerializer`) have been updated to include the new `rag_llm_model_id`, `default_llm_id` and `question_consolidator_llm_id` attributes. **2. PDF and Image Support for RAG:** - **Site Setting:** Introduces a new hidden site setting, `ai_rag_pdf_images_enabled`, to control whether PDF and image files can be indexed for RAG. This defaults to `false`. - **File Upload Validation:** The `RagDocumentFragmentsController` now checks the `ai_rag_pdf_images_enabled` setting and allows PDF, PNG, JPG, and JPEG files if enabled. Error handling is included for cases where PDF/image indexing is attempted with the setting disabled. - **PDF Processing:** Adds a new utility class, `DiscourseAi::Utils::PdfToImages`, which uses ImageMagick (`magick`) to convert PDF pages into individual PNG images. A maximum PDF size and conversion timeout are enforced. - **Image Processing:** A new utility class, `DiscourseAi::Utils::ImageToText`, is included to handle OCR for the images and PDFs. - **RAG Digestion Job:** The `DigestRagUpload` job now handles PDF and image uploads. It uses `PdfToImages` and `ImageToText` to extract text and create document fragments. - **UI Updates:** The RAG uploader component now accepts PDF and image file types if `ai_rag_pdf_images_enabled` is true. The UI text is adjusted to indicate supported file types. **3. Refactoring and Improvements:** - **LLM Enumeration:** The `DiscourseAi::Configuration::LlmEnumerator` now provides a `values_for_serialization` method, which returns a simplified array of LLM data (id, name, vision_enabled) suitable for use in serializers. This avoids exposing unnecessary details to the frontend. - **AI Helper:** The `AiHelper::Assistant` now takes optional `helper_llm` and `image_caption_llm` parameters in its constructor, allowing for greater flexibility. - **Bot and Persona Updates:** Several updates were made across the codebase, changing the string based association to a LLM to the new model based. - **Audit Logs:** The `DiscourseAi::Completions::Endpoints::Base` now formats raw request payloads as pretty JSON for easier auditing. - **Eval Script:** An evaluation script is included. **4. Testing:** - The PR introduces a new eval system for LLMs, this allows us to test how functionality works across various LLM providers. This lives in `/evals`
291 lines
8.2 KiB
Ruby
291 lines
8.2 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
RSpec.describe AiPersona do
|
|
fab!(:llm_model)
|
|
fab!(:seeded_llm_model) { Fabricate(:llm_model, id: -1) }
|
|
|
|
it "validates context settings" do
|
|
persona =
|
|
AiPersona.new(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
tools: [],
|
|
allowed_group_ids: [],
|
|
)
|
|
|
|
expect(persona.valid?).to eq(true)
|
|
|
|
persona.max_context_posts = 0
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:max_context_posts]).to eq(["must be greater than 0"])
|
|
|
|
persona.max_context_posts = 1
|
|
expect(persona.valid?).to eq(true)
|
|
|
|
persona.max_context_posts = nil
|
|
expect(persona.valid?).to eq(true)
|
|
end
|
|
|
|
it "validates tools" do
|
|
persona =
|
|
AiPersona.new(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
tools: [],
|
|
allowed_group_ids: [],
|
|
)
|
|
|
|
Fabricate(:ai_tool, id: 1)
|
|
Fabricate(:ai_tool, id: 2, name: "Archie search", tool_name: "search")
|
|
|
|
expect(persona.valid?).to eq(true)
|
|
|
|
persona.tools = %w[search image_generation]
|
|
expect(persona.valid?).to eq(true)
|
|
|
|
persona.tools = %w[search image_generation search]
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:tools]).to eq(["Can not have duplicate tools"])
|
|
|
|
persona.tools = [["custom-1", { test: "test" }, false], ["custom-2", { test: "test" }, false]]
|
|
expect(persona.valid?).to eq(true)
|
|
expect(persona.errors[:tools]).to eq([])
|
|
|
|
persona.tools = [["custom-1", { test: "test" }, false], ["custom-1", { test: "test" }, false]]
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:tools]).to eq(["Can not have duplicate tools"])
|
|
|
|
persona.tools = [
|
|
["custom-1", { test: "test" }, false],
|
|
["custom-2", { test: "test" }, false],
|
|
"image_generation",
|
|
]
|
|
expect(persona.valid?).to eq(true)
|
|
expect(persona.errors[:tools]).to eq([])
|
|
|
|
persona.tools = [
|
|
["custom-1", { test: "test" }, false],
|
|
["custom-2", { test: "test" }, false],
|
|
"Search",
|
|
]
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:tools]).to eq(["Can not have duplicate tools"])
|
|
end
|
|
|
|
it "allows creation of user" do
|
|
persona =
|
|
AiPersona.create!(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
tools: [],
|
|
allowed_group_ids: [],
|
|
)
|
|
|
|
user = persona.create_user!
|
|
expect(user.username).to eq("test_bot")
|
|
expect(user.name).to eq("Test")
|
|
expect(user.bot?).to be(true)
|
|
expect(user.id).to be <= AiPersona::FIRST_PERSONA_USER_ID
|
|
end
|
|
|
|
it "removes all rag embeddings when rag params change" do
|
|
persona =
|
|
AiPersona.create!(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
tools: [],
|
|
allowed_group_ids: [],
|
|
rag_chunk_tokens: 10,
|
|
rag_chunk_overlap_tokens: 5,
|
|
)
|
|
|
|
id =
|
|
RagDocumentFragment.create!(
|
|
target: persona,
|
|
fragment: "test",
|
|
fragment_number: 1,
|
|
upload: Fabricate(:upload),
|
|
).id
|
|
|
|
persona.rag_chunk_tokens = 20
|
|
persona.save!
|
|
|
|
expect(RagDocumentFragment.exists?(id)).to eq(false)
|
|
end
|
|
|
|
it "defines singleton methods on system persona classes" do
|
|
forum_helper = AiPersona.find_by(name: "Forum Helper")
|
|
forum_helper.update!(
|
|
user_id: 1,
|
|
default_llm_id: llm_model.id,
|
|
max_context_posts: 3,
|
|
allow_topic_mentions: true,
|
|
allow_personal_messages: true,
|
|
allow_chat_channel_mentions: true,
|
|
allow_chat_direct_messages: true,
|
|
)
|
|
|
|
klass = forum_helper.class_instance
|
|
|
|
expect(klass.id).to eq(forum_helper.id)
|
|
expect(klass.system).to eq(true)
|
|
# tl 0 by default
|
|
expect(klass.allowed_group_ids).to eq([10])
|
|
expect(klass.user_id).to eq(1)
|
|
expect(klass.default_llm_id).to eq(llm_model.id)
|
|
expect(klass.max_context_posts).to eq(3)
|
|
expect(klass.allow_topic_mentions).to eq(true)
|
|
expect(klass.allow_personal_messages).to eq(true)
|
|
expect(klass.allow_chat_channel_mentions).to eq(true)
|
|
expect(klass.allow_chat_direct_messages).to eq(true)
|
|
end
|
|
|
|
it "defines singleton methods non persona classes" do
|
|
persona =
|
|
AiPersona.create!(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
tools: [],
|
|
allowed_group_ids: [],
|
|
default_llm_id: llm_model.id,
|
|
max_context_posts: 3,
|
|
allow_topic_mentions: true,
|
|
allow_personal_messages: true,
|
|
allow_chat_channel_mentions: true,
|
|
allow_chat_direct_messages: true,
|
|
user_id: 1,
|
|
)
|
|
|
|
klass = persona.class_instance
|
|
|
|
expect(klass.id).to eq(persona.id)
|
|
expect(klass.system).to eq(false)
|
|
expect(klass.allowed_group_ids).to eq([])
|
|
expect(klass.user_id).to eq(1)
|
|
expect(klass.default_llm_id).to eq(llm_model.id)
|
|
expect(klass.max_context_posts).to eq(3)
|
|
expect(klass.allow_topic_mentions).to eq(true)
|
|
expect(klass.allow_personal_messages).to eq(true)
|
|
expect(klass.allow_chat_channel_mentions).to eq(true)
|
|
expect(klass.allow_chat_direct_messages).to eq(true)
|
|
end
|
|
|
|
it "does not allow setting allowing chat without a default_llm" do
|
|
persona =
|
|
AiPersona.create(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
allowed_group_ids: [],
|
|
default_llm: nil,
|
|
allow_chat_channel_mentions: true,
|
|
)
|
|
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:default_llm].first).to eq(
|
|
I18n.t("discourse_ai.ai_bot.personas.default_llm_required"),
|
|
)
|
|
|
|
persona =
|
|
AiPersona.create(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
allowed_group_ids: [],
|
|
default_llm: nil,
|
|
allow_chat_direct_messages: true,
|
|
)
|
|
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:default_llm].first).to eq(
|
|
I18n.t("discourse_ai.ai_bot.personas.default_llm_required"),
|
|
)
|
|
|
|
persona =
|
|
AiPersona.create(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
allowed_group_ids: [],
|
|
default_llm: nil,
|
|
allow_topic_mentions: true,
|
|
)
|
|
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:default_llm].first).to eq(
|
|
I18n.t("discourse_ai.ai_bot.personas.default_llm_required"),
|
|
)
|
|
end
|
|
|
|
it "validates allowed seeded model" do
|
|
persona =
|
|
AiPersona.new(
|
|
name: "test",
|
|
description: "test",
|
|
system_prompt: "test",
|
|
tools: [],
|
|
allowed_group_ids: [],
|
|
default_llm_id: seeded_llm_model.id,
|
|
)
|
|
|
|
SiteSetting.ai_bot_allowed_seeded_models = ""
|
|
|
|
expect(persona.valid?).to eq(false)
|
|
expect(persona.errors[:default_llm]).to include(
|
|
I18n.t("discourse_ai.llm.configuration.invalid_seeded_model"),
|
|
)
|
|
|
|
SiteSetting.ai_bot_allowed_seeded_models = "-1"
|
|
expect(persona.valid?).to eq(true)
|
|
end
|
|
|
|
it "does not leak caches between sites" do
|
|
AiPersona.create!(
|
|
name: "pun_bot",
|
|
description: "you write puns",
|
|
system_prompt: "you are pun bot",
|
|
tools: ["ImageCommand"],
|
|
allowed_group_ids: [Group::AUTO_GROUPS[:trust_level_0]],
|
|
)
|
|
|
|
AiPersona.all_personas
|
|
|
|
expect(AiPersona.persona_cache[:value].length).to be > (0)
|
|
RailsMultisite::ConnectionManagement.stubs(:current_db) { "abc" }
|
|
expect(AiPersona.persona_cache[:value]).to eq(nil)
|
|
end
|
|
|
|
describe "system persona validations" do
|
|
let(:system_persona) do
|
|
AiPersona.create!(
|
|
name: "system_persona",
|
|
description: "system persona",
|
|
system_prompt: "system persona",
|
|
tools: %w[Search Time],
|
|
system: true,
|
|
)
|
|
end
|
|
|
|
context "when modifying a system persona" do
|
|
it "allows changing tool options without allowing tool additions/removals" do
|
|
tools = [["Search", { "base_query" => "abc" }], ["Time"]]
|
|
system_persona.update!(tools: tools)
|
|
|
|
system_persona.reload
|
|
expect(system_persona.tools).to eq(tools)
|
|
|
|
invalid_tools = ["Time"]
|
|
system_persona.update(tools: invalid_tools)
|
|
expect(system_persona.errors[:base]).to include(
|
|
I18n.t("discourse_ai.ai_bot.personas.cannot_edit_system_persona"),
|
|
)
|
|
end
|
|
end
|
|
end
|
|
end
|