discourse-ai/config/locales/client.en.yml

238 lines
9.9 KiB
YAML

en:
admin_js:
admin:
site_settings:
categories:
discourse_ai: "Discourse AI"
js:
discourse_automation:
ai_models:
gpt_4_turbo: GPT 4 Turbo
gpt_4: GPT 4
gpt_3_5_turbo: GPT 3.5 Turbo
claude_2: Claude 2
gemini_pro: Gemini Pro
scriptables:
llm_report:
fields:
sender:
label: "Sender"
description: "The user that will send the report"
receivers:
label: "Receivers"
description: "The users that will receive the report (emails will be sent direct emails, usernames will be sent a PM)"
topic_id:
label: "Topic ID"
description: "The topic id to post the report to"
title:
label: "Title"
description: "The title of the report"
days:
label: "Days"
description: "The timespan of the report"
offset:
label: "Offset"
description: "When testing you may want to run the report historically, use offset to start the report in an earlier date"
instructions:
label: "Instructions"
description: "The instructions provided to the large language model"
sample_size:
label: "Sample Size"
description: "The number of posts to sample for the report"
tokens_per_post:
label: "Tokens per post"
description: "The number of llm tokens to use per post"
model:
label: "Model"
description: "LLM to use for report generation"
categories:
label: "Categories"
description: "Filter topics only to these categories"
tags:
label: "Tags"
description: "Filter topics only to these tags"
exclude_tags:
label: "Exclude Tags"
description: "Exclude topics with these tags"
exclude_categories:
label: "Exclude Categories"
description: "Exclude topics with these categories"
allow_secure_categories:
label: "Allow secure categories"
description: "Allow the report to be generated for topics in secure categories"
debug_mode:
label: "Debug Mode"
description: "Enable debug mode to see the raw input and output of the LLM"
priority_group:
label: "Priority Group"
description: "Prioritize content from this group in the report"
temperature:
label: "Temperature"
description: "Temperature to use for the LLM, increase to increase randomness (0 to use model default)"
top_p:
label: "Top P"
description: "Top P to use for the LLM, increase to increase randomness (0 to use model default)"
llm_triage:
fields:
system_prompt:
label: "System Prompt"
description: "The prompt that will be used to triage, be sure for it to reply with a single word you can use to trigger the action"
search_for_text:
label: "Search for text"
description: "If the following text appears in the llm reply, apply this actions"
category:
label: "Category"
description: "Category to apply to the topic"
tags:
label: "Tags"
description: "Tags to apply to the topic"
canned_reply:
label: "Reply"
description: "Raw text of canned reply to post on the topic"
canned_reply_user:
label: "Reply User"
description: "Username of the user to post the canned reply"
hide_topic:
label: "Hide topic"
description: "Make topic non visible to the public if triggered"
model:
label: "Model"
description: "Either gpt-4 or gpt-3-5-turbo or claude-2"
discourse_ai:
title: "AI"
modals:
select_option: "Select an option..."
ai_persona:
name: Name
description: Description
no_llm_selected: "No language model selected"
max_context_posts: "Max Context Posts"
max_context_posts_help: "The maximum number of posts to use as context for the AI when responding to a user. (empty for default)"
mentionable: Mentionable
mentionable_help: If enabled, users in allowed groups can mention this user in posts and messages, the AI will respond as this persona.
user: User
create_user: Create User
create_user_help: You can optionally attach a user to this persona. If you do, the AI will use this user to respond to requests.
default_llm: Default Language Model
default_llm_help: The default language model to use for this persona. Required if you wish to mention persona on public posts.
system_prompt: System Prompt
save: Save
saved: AI Persona Saved
enabled: Enabled
commands: Enabled Commands
allowed_groups: Allowed Groups
confirm_delete: Are you sure you want to delete this persona?
new: New
title: "AI Personas"
delete: Delete
temperature: Temperature
temperature_help: Temperature to use for the LLM, increase to increase creativity (leave empty to use model default, generally a value from 0.0 to 2.0)
top_p: Top P
top_p_help: Top P to use for the LLM, increase to increase randomness (leave empty to use model default, generally a value from 0.0 to 1.0)
priority: Priority
priority_help: Priority personas are displayed to users at the top of the persona list. If multiple personas have priority, they will be sorted alphabetically.
command_options: "Command Options"
no_persona_selected: |
## What are AI Personas?
AI Personas are a powerful feature that allows you to customize the behavior of the AI engine in your Discourse forum. They act as a 'system message' that guides the AI's responses and interactions, helping to create a more personalized and engaging user experience.
## Why use AI Personas?
With AI Personas, you can tailor the AI's behavior to better fit the context and tone of your forum. Whether you want the AI to be more formal for a professional setting, more casual for a community forum, or even embody a specific character for a role-playing game, AI Personas give you the flexibility to do so.
## Group-Specific Access to AI Personas
Moreover, you can set it up so that certain user groups have access to specific personas. This means you can have different AI behaviors for different sections of your forum, further enhancing the diversity and richness of your community's interactions.
related_topics:
title: "Related Topics"
pill: "Related"
ai_helper:
title: "Suggest changes using AI"
description: "Choose one of the options below, and the AI will suggest you a new version of the text."
selection_hint: "Hint: You can also select a portion of the text before opening the helper to rewrite only that."
suggest: "Suggest with AI"
missing_content: "Please enter some content to generate suggestions."
context_menu:
back: "Back"
trigger: "AI"
undo: "Undo"
loading: "AI is generating"
cancel: "Cancel"
regen: "Try Again"
view_changes: "View Changes"
confirm: "Confirm"
revert: "Revert"
changes: "Changes"
custom_prompt:
title: "Custom Prompt"
placeholder: "Enter a custom prompt..."
submit: "Send Prompt"
post_options_menu:
trigger: "Ask AI"
title: "Ask AI"
loading: "AI is generating"
close: "Close"
copy: "Copy"
copied: "Copied!"
cancel: "Cancel"
fast_edit:
suggest_button: "Suggest Edit"
thumbnail_suggestions:
title: "Suggested Thumbnails"
select: "Select"
selected: "Selected"
reviewables:
model_used: "Model used:"
accuracy: "Accuracy:"
embeddings:
semantic_search: "Topics (Semantic)"
semantic_search_loading: "Searching for more results using AI"
semantic_search_disabled_sort: "AI Search disabled for this sort order, sort by Relevance to enable."
semantic_search_results:
toggle: "Showing %{count} results found using AI"
toggle_hidden: "Hiding %{count} results found using AI"
none: "Sorry, our AI search found no matching topics."
new: "Press 'Search' to begin looking for new results with AI"
ai_generated_result: "Search result found using AI"
ai_bot:
pm_warning: "AI chatbot messages are monitored regularly by moderators."
cancel_streaming: "Stop reply"
default_pm_prefix: "[Untitled AI bot PM]"
shortcut_title: "Start a PM with an AI bot"
share: "Share AI conversation"
conversation_shared: "Conversation copied"
ai_label: "AI"
ai_title: "Conversation with AI"
share_modal:
title: "Share AI conversation"
copy: "Copy"
context: "Interactions to share:"
bot_names:
gpt-4: "GPT-4"
gpt-4-turbo: "GPT-4 Turbo"
gpt-3:
5-turbo: "GPT-3.5"
claude-2: "Claude 2"
gemini-pro: "Gemini"
mixtral-8x7B-Instruct-V0:
"1": "Mixtral-8x7B V0.1"
sentiments:
dashboard:
title: "Sentiment"
review:
types:
reviewable_ai_post:
title: "AI-Flagged post"
reviewable_ai_chat_message:
title: "AI-Flagged chat message"