๐ง MiniBot-0.9M-Instruct
Instruction-tuned GPT-2 style language model (~900K parameters) optimized for Portuguese conversational tasks.
๐ Overview
MiniBot-0.9M-Instruct is the instruction-tuned version of MiniBot-0.9M-Base, designed to follow prompts more accurately, respond to user inputs, and generate more coherent conversational outputs in Portuguese.
Built on a GPT-2 architecture (~0.9M parameters), this model was fine-tuned on conversational and instruction-style data to improve usability in real-world interactions.
๐ฏ Key Characteristics
| Attribute | Detail |
|---|---|
| ๐ง๐ท Language | Portuguese (primary) |
| ๐ง Architecture | GPT-2 style (Transformer decoder-only) |
| ๐ค Embeddings | GPT-2 compatible |
| ๐ Parameters | ~900K |
| โ๏ธ Base Model | MiniBot-0.9M-Base |
| ๐ฏ Fine-tuning | Instruction tuning (supervised) |
| โ Alignment | Basic prompt-following behavior |
๐ง What Changed from Base?
Instruction tuning introduced significant behavioral improvements with no architectural changes:
| Feature | Base | Instruct |
|---|---|---|
| Prompt understanding | โ | โ |
| Conversational flow | โ ๏ธ Partial | โ |
| Instruction following | โ | โ |
| Overall coherence | Low | Improved |
| Practical usability | Experimental | Functional |
๐ก The model is now significantly more usable in chat scenarios.
๐๏ธ Architecture
The core architecture remains identical to the base model:
- Decoder-only Transformer (GPT-2 style)
- Token embeddings + positional embeddings
- Self-attention + MLP blocks
- Autoregressive generation
No structural changes were made โ only behavioral improvement through fine-tuning.
๐ Fine-Tuning Dataset
The model was fine-tuned on a Portuguese instruction-style conversational dataset composed of:
- ๐ฌ Questions and answers
- ๐ Simple instructions
- ๐ค Assistant-style chat
- ๐ญ Basic roleplay
- ๐ฃ๏ธ Natural conversations
Expected format:
User: Me explique o que รฉ gravidade
Bot: A gravidade รฉ a forรงa que atrai objetos com massa...
Training strategy:
- Supervised Fine-Tuning (SFT)
- Pattern learning for instruction-following
- No RLHF or preference optimization
๐ก Capabilities
โ Strengths
- Following simple instructions
- Answering basic questions
- Conversing more naturally
- Higher coherence in short responses
- More consistent dialogue structure
โ Limitations
- Reasoning is still limited
- May generate incorrect facts
- Does not retain long context
- Sensitive to poorly structured prompts
โ ๏ธ Even with instruction tuning, this remains an extremely small model. Adjust expectations accordingly.
๐ Getting Started
Installation
pip install transformers torch
Usage with Hugging Face Transformers
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "AxionLab-official/MiniBot-0.9M-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
prompt = "User: Me diga uma curiosidade sobre o espaรงo\nBot:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(
**inputs,
max_new_tokens=80,
temperature=0.7,
top_p=0.9,
do_sample=True,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
โ๏ธ Recommended Settings
| Parameter | Recommended Value | Description |
|---|---|---|
temperature |
0.6 โ 0.8 |
Controls randomness |
top_p |
0.85 โ 0.95 |
Nucleus sampling |
do_sample |
True |
Enable sampling |
max_new_tokens |
40 โ 100 |
Response length |
๐ก Instruct models tend to perform better at lower temperatures. Try values around
0.65for more accurate and focused responses.
๐งช Intended Use Cases
| Use Case | Suitability |
|---|---|
| ๐ฌ Lightweight Portuguese chatbots | โ Ideal |
| ๐ฎ NPCs and games | โ Ideal |
| ๐ง Fine-tuning experiments | โ Ideal |
| ๐ NLP education | โ Ideal |
| โก Local / CPU-only applications | โ Ideal |
| ๐ญ Critical production environments | โ Not recommended |
โ ๏ธ Disclaimer
- Extremely small model (~900K parameters)
- No robust alignment (no RLHF)
- May generate incorrect or nonsensical responses
- Not suitable for critical production environments
๐ฎ Future Work
- ๐ง Reasoning-tuned version (
MiniBot-Reason) - ๐ Scaling to 1Mโ10M parameters
- ๐ Larger and more diverse dataset
- ๐ค Improved response alignment
- ๐งฉ Tool-use experiments
๐ License
Distributed under the MIT License. See LICENSE for more details.
๐ค Author
Developed by AxionLab ๐ฌ
- Downloads last month
- -
Model tree for AxionLab-official/MiniBot-0.9M-Instruct
Base model
AxionLab-official/MiniBot-0.9M-Base