| | --- |
| | license: apache-2.0 |
| | base_model: google/gemma-2-12b-it |
| | tags: |
| | - verilog |
| | - code-generation |
| | - instruction-tuned |
| | - vericoder |
| | --- |
| | |
| | # Gemma-3-12B-IT (VeriCoder Dataset Ablation) |
| |
|
| | This is a fine-tuned version of Gemma-3-12B-IT model trained on VeriCoder dataset. |
| |
|
| | ## Model Details |
| |
|
| | - **Base Model**: Gemma-3-12B-IT |
| | - **Training Dataset**: VeriCoder dataset (126k samples) |
| | - **Model Architecture**: Gemma3ForCausalLM |
| | - **Parameters**: ~11.7B |
| | - **Context Length**: 131,072 tokens |
| | - **Sliding Window**: 1024 |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | |
| | model_name = "LLM4Code/VeriCoder_Gemma12b" |
| | tokenizer = AutoTokenizer.from_pretrained(model_name) |
| | model = AutoModelForCausalLM.from_pretrained(model_name) |
| | |
| | # Example usage |
| | inputs = tokenizer("Your prompt here", return_tensors="pt") |
| | outputs = model.generate(**inputs, max_length=512) |
| | print(tokenizer.decode(outputs[0])) |
| | ``` |
| |
|
| | ## Training Details |
| |
|
| | - **Dataset**: VeriCoder dataset ablation (126k samples) |
| | - **Commit**: ae17392c |
| |
|
| | ## Files |
| |
|
| | The model includes: |
| | - Model weights in SafeTensors format (5 shards) |
| | - Tokenizer files (tokenizer.json, tokenizer.model, tokenizer_config.json) |
| | - Model configuration (config.json) |
| | - Generation configuration (generation_config.json) |
| | - Chat template (chat_template.jinja) |
| | |
| | |