organisciak/ocsai-llama2-7b 🔢✓📝 → 📝
About
Example Output
Prompt:
"AUT Prompt:brick
Response:build bookshelves
Score:"
Output
15
Performance Metrics
0.13s
Prediction Time
0.17s
Total Time
All Input Parameters
{
"prompt": "AUT Prompt:brick\nResponse:build bookshelves \nScore:\n\n",
"temperature": 0.01
}
Input Parameters
- seed
- Random seed. Leave blank to randomize the seed
- debug
- provide debugging output in logs
- top_p
- When decoding text, samples from the top p percentage of most likely tokens; lower to ignore less likely tokens
- prompt (required)
- Prompt to send to the model.
- temperature
- Adjusts randomness of outputs, greater than 1 is random and 0 is deterministic, 0.75 is a good starting value.
- return_logits
- if set, only return logits for the first token. only useful for testing, etc.
- max_new_tokens
- Maximum number of tokens to generate. A word is generally 2-3 tokens
- min_new_tokens
- Minimum number of tokens to generate. To disable, set to -1. A word is generally 2-3 tokens.
- stop_sequences
- A comma-separated list of sequences to stop generation at. For example, '<end>,<stop>' will stop generation at the first instance of 'end' or '<stop>'.
- replicate_weights
- Path to fine-tuned weights produced by a Replicate fine-tune job.
- repetition_penalty
- A parameter that controls how repetitive text can be. Lower means more repetitive, while higher means less repetitive. Set to 1.0 to disable.
Output Schema
Output
Example Execution Logs
Your formatted prompt is: AUT Prompt:brick Response:build bookshelves Score: correct lora is already loaded Overall initialize_peft took 0.000 Exllama: False INFO 01-10 18:31:56 async_llm_engine.py:371] Received request 0: prompt: 'AUT Prompt:brick\nResponse:build bookshelves \nScore:\n\n', sampling params: SamplingParams(n=1, best_of=1, presence_penalty=0.0, frequency_penalty=1.0, temperature=0.01, top_p=0.95, top_k=50, use_beam_search=False, length_penalty=1.0, early_stopping=False, stop=['</s>'], ignore_eos=False, max_tokens=128, logprobs=None, skip_special_tokens=True), prompt token ids: None. INFO 01-10 18:31:56 async_llm_engine.py:111] Finished request 0. hostname: model-hs-77dde5d6-e494c37392ea209c-gpu-a40-5fcfffb44d-5wvkg
Version Details
- Version ID
b00751d00cca65ff9213aea7d4fc79b9f91d2af25c5f097bd2d9fd29cc952218- Version Created
- January 10, 2024