--- license: apache-2.0 datasets: - smcleod/golang-coder - smcleod/golang-programming-style-best-practices - ExAi/Code-Golang-QA-2k - google/code_x_glue_ct_code_to_text - semeru/code-text-go language: - en tags: - golang - code - go - programming - llama - text-generation-inference --- # Llama 3.1 8b Golang Coder v2 I trained this model (based on Llama 3.1 8b) on a merged dataset I created consisting of 50,627 rows, 13.3M input tokens and 2.2M output tokens. The total training consisted of 1,020,719 input tokens and 445,810 output tokens from 45,565 items in the dataset.