SophNet has officially partnered with 302.AI to provide open-source model integration for all 302 users. All open-source models deployed by SophNet are available on 302.AI, with consistent pricing and no rate limits. Supported model list: https://www.sophnet.com/#/model/listFor pricing, please refer to: https://302.ai/pricing_api/
What sampling temperature to use, between 0 and 2. Higher values (such as 0.8) will make the output more random, while lower values (such as 0.2) will make the output more focused and deterministic. We generally recommend altering this or top_p but not both.
top_p
integer
optional
An alternative to temperature sampling is called nucleus sampling, where the model considers results from tokens with a top_p probability mass. So 0.1 means only tokens that make up the top 10% probability mass are considered. We generally recommend changing this or temperature, but not both.
n
integer
optional
How many chat completion options to generate for each input message.
stream
boolean
optional
If set, partial message increments will be sent, just like in ChatGPT. When tokens become available, tokens will be sent as pure data server-sent eventsdata: [DONE], and the stream will be terminated by the message. For example code, please refer to the OpenAI Cookbook.
stop
string
optional
The API will stop generating up to 4 sequences with the most tokens.
max_tokens
integer
optional
The maximum number of tokens generated when chat is completed. The total length of input tokens and generated tokens is limited by the model's context length.
presence_penalty
number
optional
A number between -2.0 and 2.0. Positive values penalize new tokens based on whether they have appeared in the text so far, increasing the likelihood that the model will talk about new topics. See more information about frequency and presence penalties.
frequency_penalty
number
optional
A number between -2.0 and 2.0. Positive values penalize based on the existing frequency of new tokens in the text, reducing the likelihood that the model will repeat the same line verbatim. See more information about frequency and presence penalties.
logit_bias
null
optional
Modify the likelihood of specified tokens appearing in the completion. Accepts a json object that maps tokens (specified by token IDs from the tokenizer) to associated bias values ranging from -100 to 100. Mathematically, the bias is added to the model-generated logits before sampling. The exact effect varies by model, but values between -1 and 1 should decrease or increase the probability of selection; values like -100 or 100 should result in the relevant token being banned or exclusively selected.
user
string
optional
A unique identifier representing your end user, which can help OpenAI monitor and detect abusive behavior. Learn more.
{"id":"chatcmpl-123","object":"chat.completion","created":1677652288,"choices":[{"index":0,"message":{"role":"assistant","content":"\n\nHello there, how may I assist you today?"},"finish_reason":"stop"}],"usage":{"prompt_tokens":9,"completion_tokens":12,"total_tokens":21}}