Model_name stringclasses 9 values | Train_size int64 17.5k 80k | Test_size int64 4.38k 19.9k | arg dict | lora null | Parameters int64 109M 576M | Trainable_parameters int64 109M 576M | r null | Memory Allocation stringlengths 6 8 | Training Time stringlengths 5 7 | Performance dict |
|---|---|---|---|---|---|---|---|---|---|---|
FacebookAI/roberta-base | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 124,655,629 | 124,655,629 | null | 2819.64 | 1444.75 | {
"accuracy": 0.8086620883252293,
"f1_macro": 0.8074971081620297,
"f1_weighted": 0.8089926084331718,
"precision": 0.8112376642019833,
"recall": 0.8048878175934993
} |
google-bert/bert-base-uncased | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 109,492,237 | 109,492,237 | null | 2640.15 | 1433.95 | {
"accuracy": 0.8115695022306882,
"f1_macro": 0.8101042959591243,
"f1_weighted": 0.8118172140602734,
"precision": 0.8133373143934648,
"recall": 0.8080391782949717
} |
google/rembert | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 575,935,373 | 575,935,373 | null | 10776.53 | 8466.9 | {
"accuracy": 0.10942904406235901,
"f1_macro": 0.015174686148841219,
"f1_weighted": 0.02158716818978219,
"precision": 0.008417618774027616,
"recall": 0.07692307692307693
} |
FacebookAI/xlm-roberta-large | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 559,903,757 | 559,903,757 | null | 10131.01 | 4703.38 | {
"accuracy": 0.8106170735375207,
"f1_macro": 0.809018812566914,
"f1_weighted": 0.8108204110948648,
"precision": 0.8119560567314967,
"recall": 0.8076222775774148
} |
FacebookAI/roberta-large | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 355,373,069 | 355,373,069 | null | 7396.25 | 1528.33 | {
"accuracy": 0.8172840743896937,
"f1_macro": 0.8163705094398566,
"f1_weighted": 0.8174146001493484,
"precision": 0.819546863393102,
"recall": 0.8143561508863832
} |
google-bert/bert-large-uncased | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 335,155,213 | 335,155,213 | null | 4886.3 | 1500.66 | {
"accuracy": 0.8118702691864254,
"f1_macro": 0.810497140289229,
"f1_weighted": 0.8119116411541117,
"precision": 0.8142730315701795,
"recall": 0.8082541366904293
} |
answerdotai/ModernBERT-large | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 395,844,621 | 395,844,621 | null | 5846.52 | 1720.84 | {
"accuracy": 0.8195899543836784,
"f1_macro": 0.8187302351427349,
"f1_weighted": 0.8198901011478449,
"precision": 0.8238169381133621,
"recall": 0.8152943589028868
} |
andreasmadsen/efficient_mlm_m0.40 | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 355,375,117 | 355,375,117 | null | 5240.13 | 1550.9 | {
"accuracy": 0.8015439370394506,
"f1_macro": 0.8006785018268369,
"f1_weighted": 0.8019864355002386,
"precision": 0.805722054413156,
"recall": 0.7970090688999505
} |
microsoft/deberta-large | 79,963 | 19,949 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 406,225,933 | 406,225,933 | null | 6680.97 | 1982.33 | {
"accuracy": 0.8142262770063662,
"f1_macro": 0.8132404294946559,
"f1_weighted": 0.8146145205476452,
"precision": 0.8174711295837235,
"recall": 0.8104503643591923
} |
answerdotai/ModernBERT-large | 17,524 | 4,381 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 395,844,621 | 395,844,621 | null | 6886.18 | 648.9 | {
"accuracy": 0.8427299703264095,
"f1_macro": 0.717201496341446,
"f1_weighted": 0.8399309483967907,
"precision": 0.753381419061771,
"recall": 0.6984358761194106
} |
answerdotai/ModernBERT-large | 17,524 | 4,381 | {
"auto_find_batch_size": true,
"gradient_accumulation_steps": 4,
"learning_rate": 0.00005,
"logging_steps": 1,
"lr_scheduler_type": "linear",
"num_train_epochs": 1,
"output_dir": "outputs",
"report_to": "none",
"save_strategy": "no",
"save_total_limit": 0,
"seed": 3407,
"warmup_steps": 5,
"weight_decay": 0.01
} | null | 398,476,821 | 398,476,821 | null | 6901.78 | 660.37 | {
"accuracy": 0.5119835653960283,
"f1_macro": 0.08231988245386028,
"f1_weighted": 0.4433997976948826,
"precision": 0.08321908207453645,
"recall": 0.09436029720418769
} |
README.md exists but content is empty.
- Downloads last month
- 8