Spaces:
Build error
Build error
File size: 6,723 Bytes
a8b3f00 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 |
const translation = {
welcome: {
firstStepTip: 'To get started,',
enterKeyTip: 'enter your OpenAI API Key below',
getKeyTip: 'Get your API Key from OpenAI dashboard',
placeholder: 'Your OpenAI API Key (eg.sk-xxxx)',
},
apiKeyInfo: {
cloud: {
trial: {
title: 'You are using the {{providerName}} trial quota.',
description: 'The trial quota is provided for your testing purposes. Before the trial quota is exhausted, please set up your own model provider or purchase additional quota.',
},
exhausted: {
title: 'Your trial quota have been used up, please set up your APIKey.',
description: 'You have exhausted your trial quota. Please set up your own model provider or purchase additional quota.',
},
},
selfHost: {
title: {
row1: 'To get started,',
row2: 'setup your model provider first.',
},
},
callTimes: 'Call times',
usedToken: 'Used token',
setAPIBtn: 'Go to setup model provider',
tryCloud: 'Or try the cloud version of Dify with free quote',
},
overview: {
title: 'Overview',
appInfo: {
explanation: 'Ready-to-use AI WebApp',
accessibleAddress: 'Public URL',
preview: 'Preview',
regenerate: 'Regenerate',
regenerateNotice: 'Do you want to regenerate the public URL?',
preUseReminder: 'Please enable WebApp before continuing.',
settings: {
entry: 'Settings',
title: 'WebApp Settings',
webName: 'WebApp Name',
webDesc: 'WebApp Description',
webDescTip: 'This text will be displayed on the client side, providing basic guidance on how to use the application',
webDescPlaceholder: 'Enter the description of the WebApp',
language: 'Language',
workflow: {
title: 'Workflow',
subTitle: 'Workflow Details',
show: 'Show',
hide: 'Hide',
showDesc: 'Show or hide workflow details in WebApp',
},
chatColorTheme: 'Chat color theme',
chatColorThemeDesc: 'Set the color theme of the chatbot',
chatColorThemeInverted: 'Inverted',
invalidHexMessage: 'Invalid hex value',
sso: {
label: 'SSO Authentication',
title: 'WebApp SSO',
description: 'All users are required to login with SSO before using WebApp',
tooltip: 'Contact the administrator to enable WebApp SSO',
},
more: {
entry: 'Show more settings',
copyright: 'Copyright',
copyRightPlaceholder: 'Enter the name of the author or organization',
privacyPolicy: 'Privacy Policy',
privacyPolicyPlaceholder: 'Enter the privacy policy link',
privacyPolicyTip: 'Helps visitors understand the data the application collects, see Dify\'s <privacyPolicyLink>Privacy Policy</privacyPolicyLink>.',
customDisclaimer: 'Custom Disclaimer',
customDisclaimerPlaceholder: 'Enter the custom disclaimer text',
customDisclaimerTip: 'Custom disclaimer text will be displayed on the client side, providing additional information about the application',
},
},
embedded: {
entry: 'Embedded',
title: 'Embed on website',
explanation: 'Choose the way to embed chat app to your website',
iframe: 'To add the chat app any where on your website, add this iframe to your html code.',
scripts: 'To add a chat app to the bottom right of your website add this code to your html.',
chromePlugin: 'Install Dify Chatbot Chrome Extension',
copied: 'Copied',
copy: 'Copy',
},
qrcode: {
title: 'Link QR Code',
scan: 'Scan To Share',
download: 'Download QR Code',
},
customize: {
way: 'way',
entry: 'Customize',
title: 'Customize AI WebApp',
explanation: 'You can customize the frontend of the Web App to fit your scenario and style needs.',
way1: {
name: 'Fork the client code, modify it and deploy to Vercel (recommended)',
step1: 'Fork the client code and modify it',
step1Tip: 'Click here to fork the source code into your GitHub account and modify the code',
step1Operation: 'Dify-WebClient',
step2: 'Deploy to Vercel',
step2Tip: 'Click here to import the repository into Vercel and deploy',
step2Operation: 'Import repository',
step3: 'Configure environment variables',
step3Tip: 'Add the following environment variables in Vercel',
},
way2: {
name: 'Write client-side code to call the API and deploy it to a server',
operation: 'Documentation',
},
},
},
apiInfo: {
title: 'Backend Service API',
explanation: 'Easily integrated into your application',
accessibleAddress: 'Service API Endpoint',
doc: 'API Reference',
},
status: {
running: 'In Service',
disable: 'Disabled',
},
},
analysis: {
title: 'Analysis',
ms: 'ms',
tokenPS: 'Token/s',
totalMessages: {
title: 'Total Messages',
explanation: 'Daily AI interactions count.',
},
totalConversations: {
title: 'Total Conversations',
explanation: 'Daily AI conversations count; prompt engineering/debugging excluded.',
},
activeUsers: {
title: 'Active Users',
explanation: 'Unique users engaging in Q&A with AI; prompt engineering/debugging excluded.',
},
tokenUsage: {
title: 'Token Usage',
explanation: 'Reflects the daily token usage of the language model for the application, useful for cost control purposes.',
consumed: 'Consumed',
},
avgSessionInteractions: {
title: 'Avg. Session Interactions',
explanation: 'Continuous user-AI communication count; for conversation-based apps.',
},
avgUserInteractions: {
title: 'Avg. User Interactions',
explanation: 'Reflects the daily usage frequency of users. This metric reflects user stickiness.',
},
userSatisfactionRate: {
title: 'User Satisfaction Rate',
explanation: 'The number of likes per 1,000 messages. This indicates the proportion of answers that users are highly satisfied with.',
},
avgResponseTime: {
title: 'Avg. Response Time',
explanation: 'Time (ms) for AI to process/respond; for text-based apps.',
},
tps: {
title: 'Token Output Speed',
explanation: 'Measure the performance of the LLM. Count the Tokens output speed of LLM from the beginning of the request to the completion of the output.',
},
},
}
export default translation
|