๐Ÿง  MiniBot-0.9M-Instruct

Instruction-tuned GPT-2 style language model (~900K parameters) optimized for Portuguese conversational tasks.

Model License Language Parameters


๐Ÿ“Œ Overview

MiniBot-0.9M-Instruct is the instruction-tuned version of MiniBot-0.9M-Base, designed to follow prompts more accurately, respond to user inputs, and generate more coherent conversational outputs in Portuguese.

Built on a GPT-2 architecture (~0.9M parameters), this model was fine-tuned on conversational and instruction-style data to improve usability in real-world interactions.


๐ŸŽฏ Key Characteristics

Attribute Detail
๐Ÿ‡ง๐Ÿ‡ท Language Portuguese (primary)
๐Ÿง  Architecture GPT-2 style (Transformer decoder-only)
๐Ÿ”ค Embeddings GPT-2 compatible
๐Ÿ“‰ Parameters ~900K
โš™๏ธ Base Model MiniBot-0.9M-Base
๐ŸŽฏ Fine-tuning Instruction tuning (supervised)
โœ… Alignment Basic prompt-following behavior

๐Ÿง  What Changed from Base?

Instruction tuning introduced significant behavioral improvements with no architectural changes:

Feature Base Instruct
Prompt understanding โŒ โœ…
Conversational flow โš ๏ธ Partial โœ…
Instruction following โŒ โœ…
Overall coherence Low Improved
Practical usability Experimental Functional

๐Ÿ’ก The model is now significantly more usable in chat scenarios.


๐Ÿ—๏ธ Architecture

The core architecture remains identical to the base model:

  • Decoder-only Transformer (GPT-2 style)
  • Token embeddings + positional embeddings
  • Self-attention + MLP blocks
  • Autoregressive generation

No structural changes were made โ€” only behavioral improvement through fine-tuning.


๐Ÿ“š Fine-Tuning Dataset

The model was fine-tuned on a Portuguese instruction-style conversational dataset composed of:

  • ๐Ÿ’ฌ Questions and answers
  • ๐Ÿ“‹ Simple instructions
  • ๐Ÿค– Assistant-style chat
  • ๐ŸŽญ Basic roleplay
  • ๐Ÿ—ฃ๏ธ Natural conversations

Expected format:

User: Me explique o que รฉ gravidade
Bot: A gravidade รฉ a forรงa que atrai objetos com massa...

Training strategy:

  • Supervised Fine-Tuning (SFT)
  • Pattern learning for instruction-following
  • No RLHF or preference optimization

๐Ÿ’ก Capabilities

โœ… Strengths

  • Following simple instructions
  • Answering basic questions
  • Conversing more naturally
  • Higher coherence in short responses
  • More consistent dialogue structure

โŒ Limitations

  • Reasoning is still limited
  • May generate incorrect facts
  • Does not retain long context
  • Sensitive to poorly structured prompts

โš ๏ธ Even with instruction tuning, this remains an extremely small model. Adjust expectations accordingly.


๐Ÿš€ Getting Started

Installation

pip install transformers torch

Usage with Hugging Face Transformers

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "AxionLab-official/MiniBot-0.9M-Instruct"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

prompt = "User: Me diga uma curiosidade sobre o espaรงo\nBot:"
inputs = tokenizer(prompt, return_tensors="pt")

outputs = model.generate(
    **inputs,
    max_new_tokens=80,
    temperature=0.7,
    top_p=0.9,
    do_sample=True,
)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

โš™๏ธ Recommended Settings

Parameter Recommended Value Description
temperature 0.6 โ€“ 0.8 Controls randomness
top_p 0.85 โ€“ 0.95 Nucleus sampling
do_sample True Enable sampling
max_new_tokens 40 โ€“ 100 Response length

๐Ÿ’ก Instruct models tend to perform better at lower temperatures. Try values around 0.65 for more accurate and focused responses.


๐Ÿงช Intended Use Cases

Use Case Suitability
๐Ÿ’ฌ Lightweight Portuguese chatbots โœ… Ideal
๐ŸŽฎ NPCs and games โœ… Ideal
๐Ÿง  Fine-tuning experiments โœ… Ideal
๐Ÿ“š NLP education โœ… Ideal
โšก Local / CPU-only applications โœ… Ideal
๐Ÿญ Critical production environments โŒ Not recommended

โš ๏ธ Disclaimer

  • Extremely small model (~900K parameters)
  • No robust alignment (no RLHF)
  • May generate incorrect or nonsensical responses
  • Not suitable for critical production environments

๐Ÿ”ฎ Future Work

  • ๐Ÿง  Reasoning-tuned version (MiniBot-Reason)
  • ๐Ÿ“ˆ Scaling to 1Mโ€“10M parameters
  • ๐Ÿ“š Larger and more diverse dataset
  • ๐Ÿค– Improved response alignment
  • ๐Ÿงฉ Tool-use experiments

๐Ÿ“œ License

Distributed under the MIT License. See LICENSE for more details.


๐Ÿ‘ค Author

Developed by AxionLab ๐Ÿ”ฌ


MiniBot-0.9M-Instruct ยท AxionLab ยท MIT License
Downloads last month
-
Safetensors
Model size
986k params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for AxionLab-official/MiniBot-0.9M-Instruct

Finetuned
(1)
this model

Collection including AxionLab-official/MiniBot-0.9M-Instruct