nomagick/chatglm3-6b-32k 🔢📝 → 📝

▶️ 329 runs 📅 Oct 2023 ⚙️ Cog 0.8.6 🔗 GitHub 📄 Paper ⚖️ License
code-generation long-context text-generation text-translation

About

A 6B parameter open bilingual chat LLM (optimized for 8k+ context) | 开源双语对话语言模型

Example Output

Prompt:

"<|system|>
You are a helpful assistant
<|user|>
请使用英文重复这段话:"为了使模型生成最优输出,当使用 ChatGLM3-6B 时需要使用特定的输入格式,请按照示例格式组织输入。"
<|assistant|>"

Output

"In order to obtain the optimal output from the model, specific input format should be used when ChatGLM3-6B is used, and input should be organized according to the example format."

Performance Metrics

2.47s Prediction Time
86.93s Total Time
All Input Parameters
{
  "top_p": 0.8,
  "prompt": "<|system|>\nYou are a helpful assistant\n<|user|>\n请使用英文重复这段话:\"为了使模型生成最优输出,当使用 ChatGLM3-6B 时需要使用特定的输入格式,请按照示例格式组织输入。\"\n<|assistant|>\n",
  "max_tokens": 2048,
  "temperature": 0.75
}
Input Parameters
top_p Type: numberDefault: 0.8Range: 0 - 1
Top_p
prompt Type: stringDefault: <|system|> You are a helpful assistant <|user|> 请使用英文重复这段话:"为了使模型生成最优输出,当使用 ChatGLM3-6B 时需要使用特定的输入格式,请按照示例格式组织输入。" <|assistant|>
Prompt for completion
max_tokens Type: integerDefault: 2048Range: 1 - 32768
Max new tokens to generate
temperature Type: numberDefault: 0.75Range: 0 - 5
Temperature
Output Schema

Output

Type: arrayItems Type: string

Version Details
Version ID
0b72c2a81326d09323a3a39689323b09ac9759609c0134ae59c9f2f683ff4cdb
Version Created
October 27, 2023
Run on Replicate →