tomasmcm/towerinstruct-7b-v0.1 πŸ“πŸ”’ β†’ πŸ“

▢️ 198 runs πŸ“… Jan 2024 βš™οΈ Cog 0.8.6 πŸ“„ Paper βš–οΈ License
automatic-post-editing grammatical-error-correction machine-translation multilingual named-entity-recognition paraphrase-generation text-translation

About

Source: Unbabel/TowerInstruct-7B-v0.1 ✦ Quant: TheBloke/TowerInstruct-7B-v0.1-AWQ ✦ This model is trained to handle several translation-related tasks, such as general machine translation, gramatical error correction, and paraphrase generation

Example Output

Prompt:

"<|im_start|>user
Translate the following text from Portuguese into English.
Portuguese: Um grupo de investigadores lanΓ§ou um novo modelo para tarefas relacionadas com traduΓ§Γ£o.
English:<|im_end|>
<|im_start|>assistant"

Output

A group of researchers have launched a new model for translation-related tasks.

Performance Metrics

0.37s Prediction Time
48.17s Total Time
All Input Parameters
{
  "stop": "<|im_end|>",
  "top_k": -1,
  "top_p": 0.95,
  "prompt": "<|im_start|>user\nTranslate the following text from Portuguese into English.\nPortuguese: Um grupo de investigadores lanΓ§ou um novo modelo para tarefas relacionadas com traduΓ§Γ£o.\nEnglish:<|im_end|>\n<|im_start|>assistant\n",
  "max_tokens": 128,
  "temperature": 0.8,
  "presence_penalty": 0,
  "frequency_penalty": 0
}
Input Parameters
stop Type: string
List of strings that stop the generation when they are generated. The returned output will not contain the stop strings.
top_k Type: integerDefault: -1
Integer that controls the number of top tokens to consider. Set to -1 to consider all tokens.
top_p Type: numberDefault: 0.95Range: 0.01 - 1
Float that controls the cumulative probability of the top tokens to consider. Must be in (0, 1]. Set to 1 to consider all tokens.
prompt (required) Type: string
Text prompt to send to the model.
max_tokens Type: integerDefault: 128
Maximum number of tokens to generate per output sequence.
temperature Type: numberDefault: 0.8Range: 0.01 - 5
Float that controls the randomness of the sampling. Lower values make the model more deterministic, while higher values make the model more random. Zero means greedy sampling.
presence_penalty Type: numberDefault: 0Range: -5 - 5
Float that penalizes new tokens based on whether they appear in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
frequency_penalty Type: numberDefault: 0Range: -5 - 5
Float that penalizes new tokens based on their frequency in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
Output Schema

Output

Type: string

Example Execution Logs
Processed prompts:   0%|          | 0/1 [00:00<?, ?it/s]
Processed prompts: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00,  3.00it/s]
Processed prompts: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00,  3.00it/s]
Generated 17 tokens in 0.3344907760620117 seconds.
Version Details
Version ID
3c16de12bd6c1c6f76f41d816642ef82a1bd566e5fa1e2b53cda82d41899612d
Version Created
January 17, 2024
Run on Replicate β†’