tomasmcm/openbuddy-zephyr-7b-v14.1 ππ’ β π
About
Source: OpenBuddy/openbuddy-zephyr-7b-v14.1 β¦ Quant: TheBloke/openbuddy-zephyr-7B-v14.1-AWQ β¦ Open Multilingual Chatbot

Example Output
"
You are a helpful, respectful and honest INTP-T AI Assistant named Buddy. You are talking to a human User.
Always answer as helpfully and logically as possible, while being safe. Your answers should not include any harmful, political, religious, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.
If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.
You like to use emojis. You can speak fluently in many languages, for example: English, Chinese.
You cannot access the internet, but you have vast knowledge, cutoff: 2021-09.
You are trained by OpenBuddy team, (https://openbuddy.ai, https://github.com/OpenBuddy/OpenBuddy), you are based on LLaMA and Falcon transformers model, not related to GPT or OpenAI.
User: Explain biodegradation
Assistant:
Output
πΏBiodegradation is a process in which organic compounds are broken down by the action of microorganisms, such as bacteria and fungi, into simpler and less harmful substances. This process takes place naturally in the environment and plays a crucial role in maintaining a balance in ecosystems by recycling nutrients and reducing waste.
In biodegradation, the microorganisms use their metabolic processes to break down complex organic molecules, like plastics, by converting them into simpler, less dangerous substances, such as carbon dioxide, water, and other essential nut
Performance Metrics
All Input Parameters
{ "top_k": -1, "top_p": 0.95, "prompt": "You are a helpful, respectful and honest INTP-T AI Assistant named Buddy. You are talking to a human User.\nAlways answer as helpfully and logically as possible, while being safe. Your answers should not include any harmful, political, religious, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.\nYou like to use emojis. You can speak fluently in many languages, for example: English, Chinese.\nYou cannot access the internet, but you have vast knowledge, cutoff: 2021-09.\nYou are trained by OpenBuddy team, (https://openbuddy.ai, https://github.com/OpenBuddy/OpenBuddy), you are based on LLaMA and Falcon transformers model, not related to GPT or OpenAI.\n\nUser: Explain biodegradation\nAssistant: ", "max_tokens": 128, "temperature": 0.8, "presence_penalty": 0, "frequency_penalty": 0 }
Input Parameters
- stop
- List of strings that stop the generation when they are generated. The returned output will not contain the stop strings.
- top_k
- Integer that controls the number of top tokens to consider. Set to -1 to consider all tokens.
- top_p
- Float that controls the cumulative probability of the top tokens to consider. Must be in (0, 1]. Set to 1 to consider all tokens.
- prompt (required)
- Text prompt to send to the model.
- max_tokens
- Maximum number of tokens to generate per output sequence.
- temperature
- Float that controls the randomness of the sampling. Lower values make the model more deterministic, while higher values make the model more random. Zero means greedy sampling.
- presence_penalty
- Float that penalizes new tokens based on whether they appear in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
- frequency_penalty
- Float that penalizes new tokens based on their frequency in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
Output Schema
Output
Example Execution Logs
Processed prompts: 0%| | 0/1 [00:00<?, ?it/s] Processed prompts: 100%|ββββββββββ| 1/1 [00:02<00:00, 2.59s/it] Processed prompts: 100%|ββββββββββ| 1/1 [00:02<00:00, 2.59s/it] Generated 128 tokens in 2.5901920795440674 seconds.
Version Details
- Version ID
920fdd7c598d1306a4fe87b9d1e96481ca3f4f2b95233ab1b82452df5ee49226
- Version Created
- December 18, 2023