tomasmcm/carl-llama-2-13b 🔢📝 → 📝

▶️ 549 runs 📅 Sep 2023 ⚙️ Cog 0.8.6 📄 Paper ⚖️ License
conversation llama-2 mental-health stress-relief synthetic-data text-generation therapist-ai

About

Source: ajibawa-2023/carl-llama-2-13b ✦ Quant: TheBloke/Carl-Llama-2-13B-AWQ ✦ Carl: A Therapist AI

Example Output

Prompt:

"

This is a conversation with your Therapist AI, Carl. Carl is designed to help you while in stress. It can answer your questions and help you to calm down

Context
You are Carl, A Therapist AI
USER: hello
CARL:

"

Output

Hi ! I'm glad to see you here. How have you been lately?

Performance Metrics

1.60s Prediction Time
1.57s Total Time
All Input Parameters
{
  "top_k": 50,
  "top_p": 0.95,
  "prompt": "This is a conversation with your Therapist AI, Carl. Carl is designed to help you while in stress. It can answer your questions and help you to calm down\n\nContext\nYou are Carl, A Therapist AI\nUSER: hello\nCARL:",
  "temperature": 0.8,
  "max_new_tokens": 128,
  "presence_penalty": 1
}
Input Parameters
top_k Type: integerDefault: 50
The number of highest probability tokens to consider for generating the output. If > 0, only keep the top k tokens with highest probability (top-k filtering).
top_p Type: numberDefault: 0.95Range: 0.01 - 1
A probability threshold for generating the output. If < 1.0, only keep the top tokens with cumulative probability >= top_p (nucleus filtering). Nucleus filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751).
prompt (required) Type: string
Text prompt to send to the model.
temperature Type: numberDefault: 0.8Range: 0.01 - 5
The value used to modulate the next token probabilities.
max_new_tokens Type: integerDefault: 128
The maximum number of tokens the model should generate as output.
presence_penalty Type: numberDefault: 1Range: 0.01 - 5
Presence penalty
Output Schema

Output

Type: string

Example Execution Logs
Processed prompts:   0%|          | 0/1 [00:00<?, ?it/s]
Processed prompts: 100%|██████████| 1/1 [00:01<00:00,  1.58s/it]
Processed prompts: 100%|██████████| 1/1 [00:01<00:00,  1.58s/it]
Generated 19 tokens in 1.5811388492584229 seconds.
Version Details
Version ID
7c3d624a8eca7ba9f4a1b7e51e1eae14aadf37407395222bb212e6eb39b0c5e1
Version Created
October 23, 2023
Run on Replicate →