Dataset Viewer
model_type
stringclasses 3
values | namespace
stringclasses 3
values | model_name
stringclasses 10
values | training_method
stringclasses 4
values | model_size
int64 357M
16B
| trainable_params
int64 20.5k
16B
| url
stringlengths 64
82
| doi
float64 |
---|---|---|---|---|---|---|---|
codegen
|
Salesforce
|
codegen-350M-multi
|
fine-tuning
| 356,712,448 | 356,712,448 |
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
fine-tuning
| 1,015,306,240 | 1,015,306,240 |
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
fine-tuning
| 3,641,174,016 | 3,641,174,016 |
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
fine-tuning
| 6,862,858,240 | 6,862,858,240 |
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
fine-tuning
| 16,032,155,648 | 16,032,155,648 |
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-fine-tuning
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
fine-tuning
| 6,738,546,688 | 6,738,546,688 |
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
fine-tuning
| 7,173,923,840 | 7,173,923,840 |
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
fine-tuning
| 15,957,889,024 | 15,957,889,024 |
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
fine-tuning
| 3,030,371,328 | 3,030,371,328 |
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-fine-tuning
| null |
starcoder2
|
bigcode
|
starcoderbase
|
fine-tuning
| 15,517,456,384 | 15,517,456,384 |
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-fine-tuning
| null |
codegen
|
Salesforce
|
codegen-350M-multi
|
ia3
| 356,712,448 | 143,360 |
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
ia3
| 1,015,306,240 | 229,376 |
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
ia3
| 3,641,174,016 | 458,752 |
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
ia3
| 6,862,858,240 | 917,504 |
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
ia3
| 16,032,155,648 | 1,462,272 |
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-ia3
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
ia3
| 6,738,546,688 | 614,400 |
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
ia3
| 7,173,923,840 | 753,664 |
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
ia3
| 15,957,889,024 | 1,249,280 |
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
ia3
| 3,030,371,328 | 468,480 |
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-ia3
| null |
starcoder2
|
bigcode
|
starcoderbase
|
ia3
| 15,517,456,384 | 1,239,040 |
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-ia3
| null |
codegen
|
Salesforce
|
codegen-350M-multi
|
lora
| 356,712,448 | 1,310,720 |
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
lora
| 1,015,306,240 | 2,097,152 |
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
lora
| 3,641,174,016 | 4,194,304 |
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
lora
| 6,862,858,240 | 8,388,608 |
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
lora
| 16,032,155,648 | 13,369,344 |
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-lora
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
lora
| 6,738,546,688 | 8,388,608 |
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
lora
| 7,173,923,840 | 7,340,032 |
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
lora
| 15,957,889,024 | 12,124,160 |
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
lora
| 3,030,371,328 | 4,546,560 |
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-lora
| null |
starcoder2
|
bigcode
|
starcoderbase
|
lora
| 15,517,456,384 | 8,028,160 |
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-lora
| null |
codegen
|
Salesforce
|
codegen-350M-multi
|
prompt-tuning
| 356,712,448 | 20,480 |
https://huggingface.co/fals3/Salesforce-codegen-350M-multi-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-1B_P
|
prompt-tuning
| 1,015,306,240 | 40,960 |
https://huggingface.co/fals3/Salesforce-codegen2-1B_P-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-3_7B_P
|
prompt-tuning
| 3,641,174,016 | 81,920 |
https://huggingface.co/fals3/Salesforce-codegen2-3_7B_P-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-7B_P
|
prompt-tuning
| 6,862,858,240 | 81,920 |
https://huggingface.co/fals3/Salesforce-codegen2-7B_P-unit-test-prompt-tuning
| null |
codegen
|
Salesforce
|
codegen2-16B_P
|
prompt-tuning
| 16,032,155,648 | 122,880 |
https://huggingface.co/fals3/Salesforce-codegen2-16B_P-unit-test-prompt-tuning
| null |
llama
|
meta-llama
|
CodeLLama-7b-hf
|
prompt-tuning
| 6,738,546,688 | 81,920 |
https://huggingface.co/fals3/meta-llama-CodeLlama-7b-hf-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-7b
|
prompt-tuning
| 7,173,923,840 | 92,160 |
https://huggingface.co/fals3/bigcode-starcoder2-7b-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-15b
|
prompt-tuning
| 15,957,889,024 | 122,880 |
https://huggingface.co/fals3/bigcode-starcoder2-15b-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoder2-3b
|
prompt-tuning
| 3,030,371,328 | 61,440 |
https://huggingface.co/fals3/bigcode-starcoder2-3b-unit-test-prompt-tuning
| null |
starcoder2
|
bigcode
|
starcoderbase
|
prompt-tuning
| 15,517,456,384 | 122,880 |
https://huggingface.co/fals3/bigcode-starcoderbase-unit-test-prompt-tuning
| null |
PEFT Unit Test Generation Experiments
Dataset description
The PEFT Unit Test Generation Experiments dataset contains metadata and details about a set of trained models used for generating unit tests with parameter-efficient fine-tuning (PEFT) methods. This dataset includes models from multiple namespaces and various sizes, trained with different tuning methods to provide a comprehensive resource for unit test generation research.
Dataset Structure
Data Fields
Each example in the dataset corresponds to a specific trained model variant and includes the following features:
Feature Name | Description |
---|---|
model_type |
The type or architecture of the base model (e.g., codegen, starcoder). |
namespace |
The organization or group that created or published the base model (e.g., Salesforce, meta-llama). |
model_name |
The specific name or identifier of the model. |
training_method |
The parameter-efficient fine-tuning method used for training (e.g., full fine-tuning, LoRA, IA³). |
model_size |
The size of the model, typically measured in number of parameters (e.g., 350M, 7B). |
trainable_params |
The number of trainable parameters for the specific tuning method and hyperparameters. |
url |
A direct link to the model repository. |
doi |
The digital object identifier associated with the trained model. |
Dataset Details
Dataset Description
Training Hyperparameters
Model-agnostic Hyperparameters
Hyperparameter | Method | Value |
---|---|---|
Common | ||
Optimizer | - | AdamW |
LR schedule | - | Linear |
LR warmup ratio | - | 0.1 |
Batch size | - | 1 |
Gradient accumulation steps | - | 8 |
# Epochs | - | 3 |
Precision | - | Mixed |
Learning rate | Full fine-tuning | 5E-5 |
LoRA | 3E-4 | |
(IA)3 | 3E-4 | |
Prompt tuning | 3E-3 | |
Method specific | ||
Alpha | LoRA | 32 |
Dropout | LoRA | 0.1 |
Rank | LoRA | 16 |
Virtual tokens | Prompt tuning | 20 |
Model-specific Hyperparameters
Hyperparameter | Method | Model | Value |
---|---|---|---|
Targeted attention modules | LoRA, (IA)3 | codegen-350M-multi | qkv_proj |
Salesforce/codegen2-1B_P | qkv_proj | ||
Salesforce/codegen2-3_7B_P | qkv_proj | ||
Salesforce/codegen2-7B_P | qkv_proj | ||
Salesforce/codegen2-16B_P | qkv_proj | ||
meta-llama/CodeLlama-7b-hf | q_proj, v_proj | ||
bigcode/starcoderbase | c_attn | ||
bigcode/starcoder2-3b | q_proj, v_proj | ||
bigcode/starcoder2-7b | q_proj, v_proj | ||
bigcode/starcoder2-15b | q_proj, v_proj | ||
Targeted feedforward modules | (IA)3 | codegen-350M-multi | fc_out |
Salesforce/codegen2-1B_P | fc_out | ||
Salesforce/codegen2-3_7B_P | fc_out | ||
Salesforce/codegen2-7B_P | fc_out | ||
Salesforce/codegen2-16B_P | fc_out | ||
meta-llama/CodeLlama-7b-hf | down_proj | ||
bigcode/starcoderbase | mlp.c_proj | ||
bigcode/starcoder2-3b | q_proj, c_proj | ||
bigcode/starcoder2-7b | q_proj, c_proj | ||
bigcode/starcoder2-15b | q_proj, c_proj |
Training Runs
- Downloads last month
- 76