tomasmcm/carl-llama-2-13b 🔢📝 → 📝
About
Source: ajibawa-2023/carl-llama-2-13b ✦ Quant: TheBloke/Carl-Llama-2-13B-AWQ ✦ Carl: A Therapist AI
Example Output
Prompt:
"
This is a conversation with your Therapist AI, Carl. Carl is designed to help you while in stress. It can answer your questions and help you to calm down
Context
You are Carl, A Therapist AI
USER: hello
CARL:
Output
Hi ! I'm glad to see you here. How have you been lately?
Performance Metrics
1.60s
Prediction Time
1.57s
Total Time
All Input Parameters
{
"top_k": 50,
"top_p": 0.95,
"prompt": "This is a conversation with your Therapist AI, Carl. Carl is designed to help you while in stress. It can answer your questions and help you to calm down\n\nContext\nYou are Carl, A Therapist AI\nUSER: hello\nCARL:",
"temperature": 0.8,
"max_new_tokens": 128,
"presence_penalty": 1
}
Input Parameters
- top_k
- The number of highest probability tokens to consider for generating the output. If > 0, only keep the top k tokens with highest probability (top-k filtering).
- top_p
- A probability threshold for generating the output. If < 1.0, only keep the top tokens with cumulative probability >= top_p (nucleus filtering). Nucleus filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751).
- prompt (required)
- Text prompt to send to the model.
- temperature
- The value used to modulate the next token probabilities.
- max_new_tokens
- The maximum number of tokens the model should generate as output.
- presence_penalty
- Presence penalty
Output Schema
Output
Example Execution Logs
Processed prompts: 0%| | 0/1 [00:00<?, ?it/s] Processed prompts: 100%|██████████| 1/1 [00:01<00:00, 1.58s/it] Processed prompts: 100%|██████████| 1/1 [00:01<00:00, 1.58s/it] Generated 19 tokens in 1.5811388492584229 seconds.
Version Details
- Version ID
7c3d624a8eca7ba9f4a1b7e51e1eae14aadf37407395222bb212e6eb39b0c5e1- Version Created
- October 23, 2023