The flagship translation large model, comprehensively upgraded based on Qwen3, supports translation between 92 languages (including Chinese, English, Japanese, Korean, French, Spanish, German, Thai, Indonesian, Vietnamese, Arabic, etc.). Performance is significantly optimized, terminology customization is more stable, format restoration is more accurate, domain adaptation is stronger, and translations are precise and natural. If you have high requirements for translation quality, it is recommended to choose the qwen-mt-plus model; if you prefer faster translation speed or lower cost, it is recommended to choose the qwen-mt-turbo model. Pricing:
Model
Input (1M tokens/PTC)
Output (1M tokens/PTC)
qwen-mt-plus
0.257
0.772
qwen-mt-turbo
0.1
0.279
Supported Languages
English Full
Chinese Full Name
Language Code
English
English
en
Chinese
Simplified Chinese
zh
Traditional Chinese
Traditional Chinese
zh_tw
Russian
Russian
ru
Japanese
Japanese
ja
Korean
Korean
ko
Spanish
Spanish
es
French
French
fr
Portuguese
Portuguese
pt
German
German
de
Italian
Italian
it
Thai
Thai
th
Vietnamese
Vietnamese
vi
Indonesian
Indonesian
id
Malay
Malay
ms
Arabic
Arabic
ar
Hindi
Hindi
hi
Hebrew
Hebrew
he
Burmese
Burmese
my
Tamil
Tamil
ta
Urdu
Urdu
ur
Bengali
Bengali
bn
Polish
Polish
pl
Dutch
Dutch
nl
Romanian
Romanian
ro
Turkish
Turkish
tr
Khmer
Khmer
km
Lao
Lao
lo
Cantonese
Cantonese
yue
Czech
Czech
cs
Greek
Greek
el
Swedish
Swedish
sv
Hungarian
Hungarian
hu
Danish
Danish
da
Finnish
Finnish
fi
Ukrainian
Ukrainian
uk
Bulgarian
Bulgarian
bg
Serbian
Serbian
sr
Telugu
Telugu
te
Afrikaans
Afrikaans
af
Armenian
Armenian
hy
Assamese
Assamese
as
Asturian
Asturian
ast
Basque
Basque
eu
Belarusian
Belarusian
be
Bosnian
Bosnian
bs
Catalan
Catalan
ca
Cebuano
Cebuano
ceb
Croatian
Croatian
hr
Egyptian Arabic
Egyptian Arabic
arz
Estonian
Estonian
et
Galician
Galician
gl
Georgian
Georgian
ka
Gujarati
Gujarati
gu
Icelandic
Icelandic
is
Javanese
Javanese
jv
Kannada
Kannada
kn
Kazakh
Kazakh
kk
Latvian
Latvian
lv
Lithuanian
Lithuanian
lt
Luxembourgish
Luxembourgish
lb
Macedonian
Macedonian
mk
Maithili
Maithili
mai
Maltese
Maltese
mt
Marathi
Marathi
mr
Mesopotamian Arabic
Mesopotamian Arabic
acm
Moroccan Arabic
Moroccan Arabic
ary
Najdi Arabic
Najdi Arabic
ars
Nepali
Nepali
ne
North Azerbaijani
North Azerbaijani
az
North Levantine Arabic
North Levantine Arabic
apc
Northern Uzbek
Northern Uzbek
uz
Norwegian Bokmål
Norwegian Bokmål
nb
Norwegian Nynorsk
Norwegian Nynorsk
nn
Occitan
Occitan
oc
Odia
Odia
or
Pangasinan
Pangasinan
pag
Sicilian
Sicilian
scn
Sindhi
Sindhi
sd
Sinhala
Sinhala
si
Slovak
Slovak
sk
Slovenian
Slovenian
sl
South Levantine Arabic
South Levantine Arabic
ajp
Swahili
Swahili
sw
Tagalog
Tagalog
tl
Ta’izzi-Adeni Arabic
Ta’izzi-Adeni Arabic
acq
Tosk Albanian
Tosk Albanian
sq
Tunisian Arabic
Tunisian Arabic
aeb
Venetian
Venetian
vec
Waray
Waray
war
Welsh
Welsh
cy
Western Persian
Western Persian
fa
Request
Header Params
Content-Type
string
required
Example:
application/json
Accept
string
required
Example:
application/json
Authorization
string
optional
Example:
Bearer {{YOUR_API_KEY}}
Body Params application/json
messages
array [object {2}]
required
Generate a chat completion message in chat format.
role
string
optional
content
string
optional
model
string
required
AI Model
frequency_penalty
number
optional
A number between -2.0 and 2.0. Positive values penalize based on the existing frequency of new tokens in the text, reducing the likelihood that the model will repeat the same line word for word. See more information about frequency and presence penalties.
max_tokens
integer
optional
The maximum number of tokens generated when chat is completed. The total length of input tokens and generated tokens is limited by the model's context length.
presence_penalty
number
optional
A number between -2.0 and 2.0. Positive values penalize new tokens based on whether they have appeared in the text so far, increasing the likelihood that the model will discuss new topics. See more information about frequency and presence penalties.
temperature
integer
optional
What sampling temperature to use, between 0 and 2. Higher values (such as 0.8) will make the output more random, while lower values (such as 0.2) will make the output more focused and deterministic. We generally recommend altering this or top_p but not both.
top_p
number
optional
An alternative to temperature sampling is called nucleus sampling, where the model considers outcomes with a top_p probability mass of tokens. So 0.1 means only tokens that make up the top 10% probability mass are considered. We generally recommend changing this or temperature but not both.
top_k
number
required
stream
boolean
optional
If set, partial message increments will be sent, just like in ChatGPT. When tokens become available, they will be sent as plain data server-sent eventsdata: [DONE], and the stream will be terminated by the message. For example code, please refer to the OpenAI Cookbook.
{"choices":[{"message":{"content":"After watching this video, I didn't laugh.","role":"assistant"},"finish_reason":"stop","index":0,"logprobs":null}],"object":"chat.completion","usage":{"prompt_tokens":28,"completion_tokens":10,"total_tokens":38},"created":1753450588,"system_fingerprint":null,"model":"qwen-mt-turbo","id":"chatcmpl-25332d3e-3341-9608-8ee5-fb52d7ae20a2"}