--- library_name: transformers license: apache-2.0 --- # llama-161M Trained on 100B tokens. - 1e-3 LR - 0.1 wd - WSD scheduler with 10% decay - 80% code, 10% NL, 10% instruction data ## Model Details | [openai/openai_humaneval](https://huggingface.co/datasets/openai/openai_humaneval) (greedy) | [mbpp](https://huggingface.co/datasets/google-research-datasets/mbpp) (greedy) | | :------------------ | :------------- | | 9.2% | 9.8% |