PEFT
Safetensors
dudleymax commited on
Commit
b955aa0
·
verified ·
1 Parent(s): 28b1658

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -11
README.md CHANGED
@@ -34,8 +34,7 @@ license: mit
34
  This model focuses on fine-tuning the Llama-2 7B large language model for Python code generation. The project leverages Ludwig, an open-source toolkit, and a dataset of 500k Python code samples from Hugging Face. The model applies techniques such as prompt templating, zero-shot inference, and few-shot learning, enhancing the model's performance in generating Python code snippets efficiently.
35
 
36
  - **Developed by:** Kevin Geejo, Aniket Yadav, Rishab Pandey
37
- - **Funded by [optional]:** No specific funding agency identified
38
- - **Shared by [optional]:** No additional sharing information provided
39
  - **Model type:** Fine-tuned Llama-2 7B for Python code generation
40
  - **Language(s) (NLP):** Python (for code generation tasks)
41
  - **License:** Not explicitly mentioned, but Llama-2 models are typically governed by Meta AI’s open-source licensing
@@ -45,9 +44,8 @@ This model focuses on fine-tuning the Llama-2 7B large language model for Python
45
 
46
  <!-- Provide the basic links for the model. -->
47
 
48
- - **Repository:** Hugging Face (trained models uploaded, no specific link provided)
49
- - **Paper [optional]:** Not explicitly mentioned
50
- - **Demo [optional]:** No demo link provided
51
 
52
  ## Uses
53
 
@@ -157,13 +155,9 @@ The fine-tuned model showed enhanced proficiency in generating Python code snipp
157
 
158
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute).
159
 
160
- - **Hardware Type:** Not specified
161
- - **Hours used:** Not specified
162
- - **Cloud Provider:** Not specified
163
- - **Compute Region:** Not specified
164
- - **Carbon Emitted:** Not specified
165
 
166
- ## Technical Specifications [optional]
 
167
 
168
  ### Model Architecture and Objective
169
 
 
34
  This model focuses on fine-tuning the Llama-2 7B large language model for Python code generation. The project leverages Ludwig, an open-source toolkit, and a dataset of 500k Python code samples from Hugging Face. The model applies techniques such as prompt templating, zero-shot inference, and few-shot learning, enhancing the model's performance in generating Python code snippets efficiently.
35
 
36
  - **Developed by:** Kevin Geejo, Aniket Yadav, Rishab Pandey
37
+
 
38
  - **Model type:** Fine-tuned Llama-2 7B for Python code generation
39
  - **Language(s) (NLP):** Python (for code generation tasks)
40
  - **License:** Not explicitly mentioned, but Llama-2 models are typically governed by Meta AI’s open-source licensing
 
44
 
45
  <!-- Provide the basic links for the model. -->
46
 
47
+ - **Repository:** Hugging Face
48
+
 
49
 
50
  ## Uses
51
 
 
155
 
156
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute).
157
 
 
 
 
 
 
158
 
159
+
160
+
161
 
162
  ### Model Architecture and Objective
163