zsxkib/qwen2-1.5b-instruct 🔢📝❓ → 📝
About
Qwen 2: A 1.5 billion parameter language model from Alibaba Cloud, fine tuned for chat completions

Example Output
Prompt:
"Tell me a joke about only having 1.5 billion parameters"
Output
Why did the neural network have only 1.5 billion parameters? Because it was a tiny little network!
Performance Metrics
1.23s
Prediction Time
205.74s
Total Time
All Input Parameters
{ "top_k": 1, "top_p": 1, "prompt": "Tell me a joke about only having 1.5 billion parameters", "model_type": "Qwen2-1.5B-Instruct", "temperature": 1, "system_prompt": "You are a funny and helpful assistant.", "max_new_tokens": 512, "repetition_penalty": 1 }
Input Parameters
- seed
- The seed for the random number generator
- top_k
- When decoding text, samples from the top k most likely tokens; lower to ignore less likely tokens.
- top_p
- When decoding text, samples from the top p percentage of most likely tokens; lower to ignore less likely tokens.
- prompt
- Input prompt
- model_type
- Choose from available 1.5B models
- temperature
- Adjusts randomness of outputs, greater than 1 is random and 0 is deterministic, 0.75 is a good starting value.
- system_prompt
- System prompt
- max_new_tokens
- The maximum number of tokens to generate
- repetition_penalty
- Penalty for repeated words in generated text; 1 is no penalty, values greater than 1 discourage repetition, less than 1 encourage it.
Output Schema
Output
Example Execution Logs
Using seed: 170389148 Time to first token: 0.66 seconds Total generation time: 1.17 seconds Total tokens generated: 98 Throughput: 83.69 tokens/second
Version Details
- Version ID
18d7fe65057b30e9ba64aa311fa839dd14c43831afcbaa51155625bb0b1e07f9
- Version Created
- June 25, 2024