adirik/mamba-130m 🔢📝 → 📝
About
Base version of Mamba 130M, a 130 million parameter state space language model
Example Output
Prompt:
"How are you doing today?"
Output
I'm just getting back from a trip to the beach. I've been working on my new book, and it's going really well! It was great seeing you guys again last week at our first meeting with your team in New York City (we're still in NYC). We had such an amazing time there that we decided not only did we want more of your work but also wanted some extra time for us as part-time writers so this is what happened:
We were able get together
We were able get together
Performance Metrics
5.45s
Prediction Time
89.02s
Total Time
All Input Parameters
{
"top_k": 1,
"top_p": 1,
"prompt": "How are you doing today?",
"max_length": 100,
"temperature": 1,
"repetition_penalty": 1.2
}
Input Parameters
- seed
- The seed for the random number generator
- top_k
- When decoding text, samples from the top k most likely tokens; lower to ignore less likely tokens.
- top_p
- When decoding text, samples from the top p percentage of most likely tokens; lower to ignore less likely tokens.
- prompt (required)
- Text prompt to send to the model.
- max_length
- Maximum number of tokens to generate. A word is generally 2-3 tokens.
- temperature
- Adjusts randomness of outputs, greater than 1 is random and 0 is deterministic, 0.75 is a good starting value.
- repetition_penalty
- Penalty for repeated words in generated text; 1 is no penalty, values greater than 1 discourage repetition, less than 1 encourage it.
Output Schema
Output
Version Details
- Version ID
d7fe29cf1ad11cbd2fc28b808a7a7b4fd1ff880c394eaef914e40b6da08934ce- Version Created
- February 5, 2024