cbh123/samsum 🔢✓📝 → 📝
About
Example Output
Prompt:
"
Input:
Amanda: I baked cookies. Do you want some?
Jerry: Sure!
Amanda: I'll bring you tomorrow :-) [/INST]
Summary:
"Output
Amanda will bring Jerry some cookies tomorrow..</s>
Performance Metrics
1.06s
Prediction Time
68.91s
Total Time
All Input Parameters
{ "debug": false, "top_k": 50, "top_p": 0.9, "prompt": "Input:\nAmanda: I baked cookies. Do you want some?\nJerry: Sure!\nAmanda: I'll bring you tomorrow :-) [/INST]\n\nSummary: ", "temperature": 0.75, "max_new_tokens": 128, "min_new_tokens": -1 }
Input Parameters
- seed
- Random seed. Leave blank to randomize the seed
- debug
- provide debugging output in logs
- top_k
- When decoding text, samples from the top k most likely tokens; lower to ignore less likely tokens
- top_p
- When decoding text, samples from the top p percentage of most likely tokens; lower to ignore less likely tokens
- prompt (required)
- Prompt to send to the model.
- temperature
- Adjusts randomness of outputs, greater than 1 is random and 0 is deterministic, 0.75 is a good starting value.
- max_new_tokens
- Maximum number of tokens to generate. A word is generally 2-3 tokens
- min_new_tokens
- Minimum number of tokens to generate. To disable, set to -1. A word is generally 2-3 tokens.
- stop_sequences
- A comma-separated list of sequences to stop generation at. For example, '<end>,<stop>' will stop generation at the first instance of 'end' or '<stop>'.
- replicate_weights
- Path to fine-tuned weights produced by a Replicate fine-tune job.
Output Schema
Output
Example Execution Logs
Your formatted prompt is: Input: Amanda: I baked cookies. Do you want some? Jerry: Sure! Amanda: I'll bring you tomorrow :-) [/INST] Summary:
Version Details
- Version ID
233e4149fb21487169fa620c7f87f0126dc77bc0ef2f0bd4f06fe1ec601f4165
- Version Created
- September 1, 2023