# frozen_string_literal: true RSpec.describe DiscourseAi::AiBot::Playground do subject(:playground) { described_class.new(bot) } fab!(:claude_2) { Fabricate(:llm_model, name: "claude-2") } fab!(:opus_model) { Fabricate(:anthropic_model) } fab!(:bot_user) do toggle_enabled_bots(bots: [claude_2]) SiteSetting.ai_bot_enabled = true claude_2.reload.user end fab!(:bot) do persona = AiPersona .find( DiscourseAi::AiBot::Personas::Persona.system_personas[ DiscourseAi::AiBot::Personas::General ], ) .class_instance .new DiscourseAi::AiBot::Bot.as(bot_user, persona: persona) end fab!(:admin) { Fabricate(:admin, refresh_auto_groups: true) } fab!(:user) { Fabricate(:user, refresh_auto_groups: true) } fab!(:pm) do Fabricate( :private_message_topic, title: "This is my special PM", user: user, topic_allowed_users: [ Fabricate.build(:topic_allowed_user, user: user), Fabricate.build(:topic_allowed_user, user: bot_user), ], ) end fab!(:first_post) do Fabricate(:post, topic: pm, user: user, post_number: 1, raw: "This is a reply by the user") end fab!(:second_post) do Fabricate(:post, topic: pm, user: bot_user, post_number: 2, raw: "This is a bot reply") end fab!(:third_post) do Fabricate( :post, topic: pm, user: user, post_number: 3, raw: "This is a second reply by the user", ) end before { SiteSetting.ai_embeddings_enabled = false } after do # we must reset cache on persona cause data can be rolled back AiPersona.persona_cache.flush! end describe "is_bot_user_id?" do it "properly detects ALL bots as bot users" do persona = Fabricate(:ai_persona, enabled: false) persona.create_user! expect(DiscourseAi::AiBot::Playground.is_bot_user_id?(persona.user_id)).to eq(true) end end describe "custom tool integration" do let!(:custom_tool) do AiTool.create!( name: "search", summary: "searching for things", description: "A test custom tool", parameters: [{ name: "query", type: "string", description: "Input for the custom tool" }], script: "function invoke(params) { return 'Custom tool result: ' + params.query; }; function details() { return 'did stuff'; }", created_by: user, ) end let!(:ai_persona) { Fabricate(:ai_persona, tools: ["custom-#{custom_tool.id}"]) } let(:tool_call) do DiscourseAi::Completions::ToolCall.new( name: "search", id: "666", parameters: { query: "Can you use the custom tool", }, ) end let(:bot) { DiscourseAi::AiBot::Bot.as(bot_user, persona: ai_persona.class_instance.new) } let(:playground) { DiscourseAi::AiBot::Playground.new(bot) } it "can create uploads from a tool" do custom_tool.update!(script: <<~JS) let imageBase64 = "iVBORw0KGgoAAAANSUhEUgAAAAEAAAABCAQAAAC1HAwCAAAAC0lEQVR42mP8/wcAAgEB/awxUE0AAAAASUVORK5CYII=" function invoke(params) { let image = upload.create("image.png", imageBase64); chain.setCustomRaw(`![image](${image.short_url})`); return image.id; }; JS tool_name = "custom-#{custom_tool.id}" ai_persona.update!(tools: [[tool_name, nil, true]], tool_details: false) reply_post = nil prompts = nil responses = [tool_call] DiscourseAi::Completions::Llm.with_prepared_responses(responses) do |_, _, _prompts| new_post = Fabricate(:post, raw: "Can you use the custom tool?") reply_post = playground.reply_to(new_post) prompts = _prompts end expect(prompts.length).to eq(1) upload_id = prompts[0].messages[3][:content].to_i upload = Upload.find(upload_id) expect(reply_post.raw).to eq("![image](#{upload.short_url})") end it "can force usage of a tool" do tool_name = "custom-#{custom_tool.id}" ai_persona.update!(tools: [[tool_name, nil, true]], forced_tool_count: 1) responses = [tool_call, "custom tool did stuff (maybe)"] prompts = nil reply_post = nil DiscourseAi::Completions::Llm.with_prepared_responses(responses) do |_, _, _prompts| new_post = Fabricate(:post, raw: "Can you use the custom tool?") reply_post = playground.reply_to(new_post) prompts = _prompts end expect(prompts.length).to eq(2) expect(prompts[0].tool_choice).to eq("search") expect(prompts[1].tool_choice).to eq(nil) ai_persona.update!(forced_tool_count: 1) responses = ["no tool call here"] DiscourseAi::Completions::Llm.with_prepared_responses(responses) do |_, _, _prompts| new_post = Fabricate(:post, raw: "Will you use the custom tool?", topic: reply_post.topic) _reply_post = playground.reply_to(new_post) prompts = _prompts end expect(prompts.length).to eq(1) expect(prompts[0].tool_choice).to eq(nil) end it "uses custom tool in conversation" do persona_klass = AiPersona.all_personas.find { |p| p.name == ai_persona.name } bot = DiscourseAi::AiBot::Bot.as(bot_user, persona: persona_klass.new) playground = DiscourseAi::AiBot::Playground.new(bot) responses = [tool_call, "custom tool did stuff (maybe)"] reply_post = nil DiscourseAi::Completions::Llm.with_prepared_responses(responses) do |_, _, _prompt| new_post = Fabricate(:post, raw: "Can you use the custom tool?") reply_post = playground.reply_to(new_post) end expected = <<~TXT.strip
searching for things

did stuff

custom tool did stuff (maybe) TXT expect(reply_post.raw).to eq(expected) custom_prompt = PostCustomPrompt.find_by(post_id: reply_post.id).custom_prompt expected_prompt = [ [ "{\"arguments\":{\"query\":\"Can you use the custom tool\"}}", "666", "tool_call", "search", ], ["\"Custom tool result: Can you use the custom tool\"", "666", "tool", "search"], ["custom tool did stuff (maybe)", "claude-2"], ] expect(custom_prompt).to eq(expected_prompt) custom_tool.update!(enabled: false) # so we pick up new cache persona_klass = AiPersona.all_personas.find { |p| p.name == ai_persona.name } bot = DiscourseAi::AiBot::Bot.as(bot_user, persona: persona_klass.new) playground = DiscourseAi::AiBot::Playground.new(bot) responses = ["custom tool did stuff (maybe)", tool_call] # lets ensure tool does not run... DiscourseAi::Completions::Llm.with_prepared_responses(responses) do |_, _, _prompt| new_post = Fabricate(:post, raw: "Can you use the custom tool?") reply_post = playground.reply_to(new_post) end expect(reply_post.raw.strip).to eq("custom tool did stuff (maybe)") end end describe "image support" do before do Jobs.run_immediately! SiteSetting.ai_bot_allowed_groups = "#{Group::AUTO_GROUPS[:trust_level_0]}" end fab!(:persona) do AiPersona.create!( name: "Test Persona", description: "A test persona", allowed_group_ids: [Group::AUTO_GROUPS[:trust_level_0]], enabled: true, system_prompt: "You are a helpful bot", vision_enabled: true, vision_max_pixels: 1_000, default_llm: "custom:#{opus_model.id}", allow_topic_mentions: true, ) end fab!(:upload) it "sends images to llm" do post = nil persona.create_user! image = "![image](upload://#{upload.base62_sha1}.jpg)" body = "Hey @#{persona.user.username}, can you help me with this image? #{image}" prompts = nil DiscourseAi::Completions::Llm.with_prepared_responses( ["I understood image"], ) do |_, _, inner_prompts| post = create_post(title: "some new topic I created", raw: body) prompts = inner_prompts end expect(prompts[0].messages[1][:upload_ids]).to eq([upload.id]) expect(prompts[0].max_pixels).to eq(1000) post.topic.reload last_post = post.topic.posts.order(:post_number).last expect(last_post.raw).to eq("I understood image") end end describe "persona with user support" do before do Jobs.run_immediately! SiteSetting.ai_bot_allowed_groups = "#{Group::AUTO_GROUPS[:trust_level_0]}" end fab!(:persona) do persona = AiPersona.create!( name: "Test Persona", description: "A test persona", allowed_group_ids: [Group::AUTO_GROUPS[:trust_level_0]], enabled: true, system_prompt: "You are a helpful bot", ) persona.create_user! persona.update!( default_llm: "custom:#{claude_2.id}", allow_chat_channel_mentions: true, allow_topic_mentions: true, ) persona end context "with chat channels" do fab!(:channel) { Fabricate(:chat_channel) } fab!(:membership) do Fabricate(:user_chat_channel_membership, user: user, chat_channel: channel) end let(:guardian) { Guardian.new(user) } before do SiteSetting.ai_bot_enabled = true SiteSetting.chat_allowed_groups = "#{Group::AUTO_GROUPS[:trust_level_0]}" Group.refresh_automatic_groups! persona.update!(allow_chat_channel_mentions: true, default_llm: "custom:#{opus_model.id}") end it "should behave in a sane way when threading is enabled" do channel.update!(threading_enabled: true) message = ChatSDK::Message.create( channel_id: channel.id, raw: "thread 1 message 1", guardian: guardian, ) message = ChatSDK::Message.create( channel_id: channel.id, raw: "thread 1 message 2", in_reply_to_id: message.id, guardian: guardian, ) thread = message.thread thread.update!(title: "a magic thread") message = ChatSDK::Message.create( channel_id: channel.id, raw: "thread 2 message 1", guardian: guardian, ) message = ChatSDK::Message.create( channel_id: channel.id, raw: "thread 2 message 2", in_reply_to_id: message.id, guardian: guardian, ) prompts = nil DiscourseAi::Completions::Llm.with_prepared_responses(["world"]) do |_, _, _prompts| message = ChatSDK::Message.create( channel_id: channel.id, raw: "Hello @#{persona.user.username}", guardian: guardian, ) prompts = _prompts end # don't start a thread cause it will get confusing message.reload expect(message.thread_id).to be_nil prompt = prompts[0] content = prompt.messages[1][:content] # this is fragile by design, mainly so the example can be ultra clear expected = (<<~TEXT).strip You are replying inside a Discourse chat channel. Here is a summary of the conversation so far: {{{ #{user.username}: (a magic thread) thread 1 message 1 #{user.username}: thread 2 message 1 }}} Your instructions: #{user.username} said Hello TEXT expect(content.strip).to eq(expected) end it "should reply to a mention if properly enabled" do prompts = nil ChatSDK::Message.create( channel_id: channel.id, raw: "This is a story about stuff", guardian: guardian, ) DiscourseAi::Completions::Llm.with_prepared_responses(["world"]) do |_, _, _prompts| ChatSDK::Message.create( channel_id: channel.id, raw: "Hello @#{persona.user.username}", guardian: guardian, ) prompts = _prompts end expect(prompts.length).to eq(1) prompt = prompts[0] expect(prompt.messages.length).to eq(2) expect(prompt.messages[1][:content]).to include("story about stuff") expect(prompt.messages[1][:content]).to include("Hello") last_message = Chat::Message.where(chat_channel_id: channel.id).order("id desc").first expect(last_message.message).to eq("world") end end context "with chat dms" do fab!(:dm_channel) { Fabricate(:direct_message_channel, users: [user, persona.user]) } before do SiteSetting.chat_allowed_groups = "#{Group::AUTO_GROUPS[:trust_level_0]}" Group.refresh_automatic_groups! persona.update!( allow_chat_direct_messages: true, allow_topic_mentions: false, allow_chat_channel_mentions: false, default_llm: "custom:#{opus_model.id}", ) SiteSetting.ai_bot_enabled = true end let(:guardian) { Guardian.new(user) } it "can supply context" do post = Fabricate(:post, raw: "this is post content") prompts = nil message = DiscourseAi::Completions::Llm.with_prepared_responses(["World"]) do |_, _, _prompts| prompts = _prompts ChatSDK::Message.create( raw: "Hello", channel_id: dm_channel.id, context_post_ids: [post.id], guardian:, ) end expect(prompts[0].messages[1][:content]).to include("this is post content") message.reload reply = ChatSDK::Thread.messages(thread_id: message.thread_id, guardian: guardian).last expect(reply.message).to eq("World") expect(message.thread_id).to be_present end it "can run tools" do persona.update!(tools: ["Time"]) tool_call1 = DiscourseAi::Completions::ToolCall.new( name: "time", id: "time", parameters: { timezone: "Buenos Aires", }, ) tool_call2 = DiscourseAi::Completions::ToolCall.new( name: "time", id: "time", parameters: { timezone: "Sydney", }, ) responses = [[tool_call1, tool_call2], "The time is 2023-12-14 17:24:00 -0300"] message = DiscourseAi::Completions::Llm.with_prepared_responses(responses) do ChatSDK::Message.create(channel_id: dm_channel.id, raw: "Hello", guardian: guardian) end message.reload expect(message.thread_id).to be_present reply = ChatSDK::Thread.messages(thread_id: message.thread_id, guardian: guardian).last expect(reply.message).to eq("The time is 2023-12-14 17:24:00 -0300") # it also needs to have tool details now set on message prompt = ChatMessageCustomPrompt.find_by(message_id: reply.id) expect(prompt.custom_prompt.length).to eq(5) # TODO in chat I am mixed on including this in the context, but I guess maybe? # thinking about this end it "can reply to a chat message" do message = DiscourseAi::Completions::Llm.with_prepared_responses(["World"]) do ChatSDK::Message.create(channel_id: dm_channel.id, raw: "Hello", guardian: guardian) end message.reload expect(message.thread_id).to be_present thread_messages = ChatSDK::Thread.messages(thread_id: message.thread_id, guardian: guardian) expect(thread_messages.length).to eq(2) expect(thread_messages.last.message).to eq("World") # it also needs to include history per config - first feed some history persona.update!(enabled: false) persona_guardian = Guardian.new(persona.user) 4.times do |i| ChatSDK::Message.create( channel_id: dm_channel.id, thread_id: message.thread_id, raw: "request #{i}", guardian: guardian, ) ChatSDK::Message.create( channel_id: dm_channel.id, thread_id: message.thread_id, raw: "response #{i}", guardian: persona_guardian, ) end persona.update!(max_context_posts: 4, enabled: true) prompts = nil DiscourseAi::Completions::Llm.with_prepared_responses( ["World 2"], ) do |_response, _llm, _prompts| ChatSDK::Message.create( channel_id: dm_channel.id, thread_id: message.thread_id, raw: "Hello", guardian: guardian, ) prompts = _prompts end expect(prompts.length).to eq(1) mapped = prompts[0] .messages .map { |m| "#{m[:type]}: #{m[:content]}" if m[:type] != :system } .compact .join("\n") .strip # why? # 1. we set context to 4 # 2. however PromptMessagesBuilder will enforce rules of starting with :user and ending with it # so one of the model messages is dropped expected = (<<~TEXT).strip user: request 3 model: response 3 user: Hello TEXT expect(mapped).to eq(expected) end end it "replies to whispers with a whisper" do post = nil DiscourseAi::Completions::Llm.with_prepared_responses(["Yes I can"]) do post = create_post( title: "My public topic", raw: "Hey @#{persona.user.username}, can you help me?", post_type: Post.types[:whisper], ) end post.topic.reload last_post = post.topic.posts.order(:post_number).last expect(last_post.raw).to eq("Yes I can") expect(last_post.user_id).to eq(persona.user_id) expect(last_post.post_type).to eq(Post.types[:whisper]) end it "allows mentioning a persona" do # we still should be able to mention with no bots toggle_enabled_bots(bots: []) persona.update!(allow_topic_mentions: true) post = nil DiscourseAi::Completions::Llm.with_prepared_responses(["Yes I can"]) do post = create_post( title: "My public topic", raw: "Hey @#{persona.user.username}, can you help me?", ) end post.topic.reload last_post = post.topic.posts.order(:post_number).last expect(last_post.raw).to eq("Yes I can") expect(last_post.user_id).to eq(persona.user_id) persona.update!(allow_topic_mentions: false) post = create_post( title: "My public topic ABC", raw: "Hey @#{persona.user.username}, can you help me?", ) expect(post.topic.posts.last.post_number).to eq(1) end it "allows PMing a persona even when no particular bots are enabled" do SiteSetting.ai_bot_enabled = true toggle_enabled_bots(bots: []) post = nil DiscourseAi::Completions::Llm.with_prepared_responses( ["Yes I can", "Magic Title"], llm: "custom:#{claude_2.id}", ) do post = create_post( title: "I just made a PM", raw: "Hey there #{persona.user.username}, can you help me?", target_usernames: "#{user.username},#{persona.user.username}", archetype: Archetype.private_message, user: admin, ) end last_post = post.topic.posts.order(:post_number).last expect(last_post.raw).to eq("Yes I can") expect(last_post.user_id).to eq(persona.user_id) last_post.topic.reload expect(last_post.topic.allowed_users.pluck(:user_id)).to include(persona.user_id) expect(last_post.topic.participant_count).to eq(2) # ensure it can be disabled persona.update!(allow_personal_messages: false) post = create_post( raw: "Hey there #{persona.user.username}, can you help me please", topic_id: post.topic.id, user: admin, ) expect(post.post_number).to eq(3) end it "can tether a persona unconditionally to an llm" do gpt_35_turbo = Fabricate(:llm_model, name: "gpt-3.5-turbo") # If you start a PM with GPT 3.5 bot, replies should come from it, not from Claude SiteSetting.ai_bot_enabled = true toggle_enabled_bots(bots: [gpt_35_turbo, claude_2]) post = nil persona.update!(force_default_llm: true, default_llm: "custom:#{gpt_35_turbo.id}") DiscourseAi::Completions::Llm.with_prepared_responses( ["Yes I can", "Magic Title"], llm: "custom:#{gpt_35_turbo.id}", ) do post = create_post( title: "I just made a PM", raw: "hello world", target_usernames: "#{user.username},#{claude_2.user.username}", archetype: Archetype.private_message, user: admin, custom_fields: { "ai_persona_id" => persona.id, }, ) end last_post = post.topic.posts.order(:post_number).last expect(last_post.raw).to eq("Yes I can") expect(last_post.user_id).to eq(persona.user_id) end it "picks the correct llm for persona in PMs" do gpt_35_turbo = Fabricate(:llm_model, name: "gpt-3.5-turbo") # If you start a PM with GPT 3.5 bot, replies should come from it, not from Claude SiteSetting.ai_bot_enabled = true toggle_enabled_bots(bots: [gpt_35_turbo, claude_2]) post = nil gpt3_5_bot_user = gpt_35_turbo.reload.user messages = nil DiscourseAi::Completions::Llm.with_prepared_responses( ["Yes I can", "Magic Title"], llm: "custom:#{gpt_35_turbo.id}", ) do messages = MessageBus.track_publish do post = create_post( title: "I just made a PM", raw: "Hey @#{persona.user.username}, can you help me?", target_usernames: "#{user.username},#{gpt3_5_bot_user.username}", archetype: Archetype.private_message, user: admin, ) end end title_update_message = messages.find { |m| m.channel == "/discourse-ai/ai-bot/topic/#{post.topic.id}" } expect(title_update_message.data).to eq({ title: "Magic Title" }) last_post = post.topic.posts.order(:post_number).last expect(last_post.raw).to eq("Yes I can") expect(last_post.user_id).to eq(persona.user_id) last_post.topic.reload expect(last_post.topic.allowed_users.pluck(:user_id)).to include(persona.user_id) # does not reply if replying directly to a user # nothing is mocked, so this would result in HTTP error # if we were going to reply create_post( raw: "Please ignore this bot, I am replying to a user", topic: post.topic, user: admin, reply_to_post_number: post.post_number, ) # replies as correct persona if replying direct to persona DiscourseAi::Completions::Llm.with_prepared_responses( ["Another reply"], llm: "custom:#{gpt_35_turbo.id}", ) do create_post( raw: "Please ignore this bot, I am replying to a user", topic: post.topic, user: admin, reply_to_post_number: last_post.post_number, ) end last_post = post.topic.posts.order(:post_number).last expect(last_post.raw).to eq("Another reply") expect(last_post.user_id).to eq(persona.user_id) end end describe "#title_playground" do let(:expected_response) { "This is a suggested title" } before { SiteSetting.min_personal_message_post_length = 5 } it "updates the title using bot suggestions" do DiscourseAi::Completions::Llm.with_prepared_responses([expected_response]) do playground.title_playground(third_post) expect(pm.reload.title).to eq(expected_response) end end end describe "#reply_to" do it "streams the bot reply through MB and create a new post in the PM with a cooked responses" do expected_bot_response = "Hello this is a bot and what you just said is an interesting question" DiscourseAi::Completions::Llm.with_prepared_responses([expected_bot_response]) do messages = MessageBus.track_publish("discourse-ai/ai-bot/topic/#{pm.id}") do playground.reply_to(third_post) end reply = pm.reload.posts.last noop_signal = messages.pop expect(noop_signal.data[:noop]).to eq(true) done_signal = messages.pop expect(done_signal.data[:done]).to eq(true) expect(done_signal.data[:cooked]).to eq(reply.cooked) expect(messages.first.data[:raw]).to eq("") expect(reply.cooked).to eq(PrettyText.cook(expected_bot_response)) messages[1..-1].each do |m| expect(expected_bot_response.start_with?(m.data[:raw])).to eq(true) end end end it "supports multiple function calls" do tool_call1 = DiscourseAi::Completions::ToolCall.new( name: "search", id: "search", parameters: { search_query: "testing various things", }, ) tool_call2 = DiscourseAi::Completions::ToolCall.new( name: "search", id: "search", parameters: { search_query: "another search", }, ) response2 = "I found stuff" DiscourseAi::Completions::Llm.with_prepared_responses( [[tool_call1, tool_call2], response2], ) { playground.reply_to(third_post) } last_post = third_post.topic.reload.posts.order(:post_number).last expect(last_post.raw).to include("testing various things") expect(last_post.raw).to include("another search") expect(last_post.raw).to include("I found stuff") end it "supports disabling tool details" do persona = Fabricate(:ai_persona, tool_details: false, tools: ["Search"]) bot = DiscourseAi::AiBot::Bot.as(bot_user, persona: persona.class_instance.new) playground = described_class.new(bot) response1 = DiscourseAi::Completions::ToolCall.new( name: "search", id: "search", parameters: { search_query: "testing various things", }, ) response2 = "I found stuff" DiscourseAi::Completions::Llm.with_prepared_responses([response1, response2]) do playground.reply_to(third_post) end last_post = third_post.topic.reload.posts.order(:post_number).last expect(last_post.raw).to eq("I found stuff") end it "does not include placeholders in conversation context but includes all completions" do response1 = DiscourseAi::Completions::ToolCall.new( name: "search", id: "search", parameters: { search_query: "testing various things", }, ) response2 = "I found some really amazing stuff!" DiscourseAi::Completions::Llm.with_prepared_responses([response1, response2]) do playground.reply_to(third_post) end last_post = third_post.topic.reload.posts.order(:post_number).last custom_prompt = PostCustomPrompt.where(post_id: last_post.id).first.custom_prompt expect(custom_prompt.length).to eq(3) expect(custom_prompt.to_s).not_to include("
") expect(custom_prompt.last.first).to eq(response2) expect(custom_prompt.last.last).to eq(bot_user.username) end context "with Dall E bot" do before { SiteSetting.ai_openai_api_key = "123" } let(:persona) do AiPersona.find( DiscourseAi::AiBot::Personas::Persona.system_personas[ DiscourseAi::AiBot::Personas::DallE3 ], ) end let(:bot) { DiscourseAi::AiBot::Bot.as(bot_user, persona: persona.class_instance.new) } let(:data) do image = "iVBORw0KGgoAAAANSUhEUgAAAAEAAAABCAYAAAAfFcSJAAAADUlEQVR42mP8z8BQDwAEhQGAhKmMIQAAAABJRU5ErkJggg==" [{ b64_json: image, revised_prompt: "a pink cow 1" }] end let(:response) do DiscourseAi::Completions::ToolCall.new( name: "dall_e", id: "dall_e", parameters: { prompts: ["a pink cow"], }, ) end it "properly returns an image when skipping tool details" do persona.update!(tool_details: false) WebMock.stub_request(:post, SiteSetting.ai_openai_dall_e_3_url).to_return( status: 200, body: { data: data }.to_json, ) DiscourseAi::Completions::Llm.with_prepared_responses([response]) do playground.reply_to(third_post) end last_post = third_post.topic.reload.posts.order(:post_number).last expect(last_post.raw).to include("a pink cow") end it "does not include placeholders in conversation context (simulate DALL-E)" do WebMock.stub_request(:post, SiteSetting.ai_openai_dall_e_3_url).to_return( status: 200, body: { data: data }.to_json, ) DiscourseAi::Completions::Llm.with_prepared_responses([response]) do playground.reply_to(third_post) end last_post = third_post.topic.reload.posts.order(:post_number).last custom_prompt = PostCustomPrompt.where(post_id: last_post.id).first.custom_prompt # DALL E has custom_raw, we do not want to inject this into the prompt stream expect(custom_prompt.length).to eq(2) expect(custom_prompt.to_s).not_to include("
") end end end describe "#available_bot_usernames" do it "includes persona users" do persona = Fabricate(:ai_persona) persona.create_user! expect(playground.available_bot_usernames).to include(persona.user.username) end end describe "#conversation_context" do context "with limited context" do before do @old_persona = playground.bot.persona persona = Fabricate(:ai_persona, max_context_posts: 1) playground.bot.persona = persona.class_instance.new end after { playground.bot.persona = @old_persona } it "respects max_context_post" do context = playground.conversation_context(third_post) expect(context).to contain_exactly( *[{ type: :user, id: user.username, content: third_post.raw }], ) end end xit "includes previous posts ordered by post_number" do context = playground.conversation_context(third_post) expect(context).to contain_exactly( *[ { type: :user, id: user.username, content: third_post.raw }, { type: :model, content: second_post.raw }, { type: :user, id: user.username, content: first_post.raw }, ], ) end xit "only include regular posts" do first_post.update!(post_type: Post.types[:whisper]) context = playground.conversation_context(third_post) # skips leading model reply which makes no sense cause first post was whisper expect(context).to contain_exactly( *[{ type: :user, id: user.username, content: third_post.raw }], ) end context "with custom prompts" do it "When post custom prompt is present, we use that instead of the post content" do custom_prompt = [ [ { name: "time", arguments: { name: "time", timezone: "Buenos Aires" } }.to_json, "time", "tool_call", ], [ { args: { timezone: "Buenos Aires" }, time: "2023-12-14 17:24:00 -0300" }.to_json, "time", "tool", ], ["I replied to the time command", bot_user.username], ] PostCustomPrompt.create!(post: second_post, custom_prompt: custom_prompt) context = playground.conversation_context(third_post) expect(context).to contain_exactly( *[ { type: :user, id: user.username, content: first_post.raw }, { type: :tool_call, content: custom_prompt.first.first, id: "time" }, { type: :tool, id: "time", content: custom_prompt.second.first }, { type: :model, content: custom_prompt.third.first }, { type: :user, id: user.username, content: third_post.raw }, ], ) end end end end