lambdabrendan commited on
Commit
677197e
·
verified ·
1 Parent(s): 0a6d456

Update src/app.py

Browse files

Remove token field, switch to single column layout.

Files changed (1) hide show
  1. src/app.py +4 -4
src/app.py CHANGED
@@ -122,10 +122,10 @@ with col.expander("Detailed Information", expanded=False):
122
  st.markdown("""- For LoRa Fine-tuning, This assumes a **16-bit** dtype of trainable parameters. The formula (in terms of GB) is""")
123
  st.latex(r"\text{Memory}_\text{LoRa} \approx \text{Model Size} + \left(\text{ \# trainable Params}_\text{Billions}\times\frac{16}{8} \times 4\right) \times 1.2")
124
 
125
- access_token = st.sidebar.text_input("Access token")
126
 
127
- if access_token:
128
- login(token=access_token)
129
 
130
  #model_name = st.sidebar.text_input("Model name", value="mistralai/Mistral-7B-v0.1")
131
  with st.sidebar.container():
@@ -200,7 +200,7 @@ with col1:
200
  info = _memory_table[_memory_table['dtype'] == dtype].set_index('Variable')
201
  show_gpu_info(info, lora_pct, gpu_vendor)
202
  st.write(memory_table.iloc[[0, 1, 2, 4]])
203
- with col2:
204
  extra = ""
205
  if gpu_vendor == "Apple":
206
  st.warning("This graph is irrelevant for M1/M2 chips as they can't run in parallel.", icon="⚠️")
 
122
  st.markdown("""- For LoRa Fine-tuning, This assumes a **16-bit** dtype of trainable parameters. The formula (in terms of GB) is""")
123
  st.latex(r"\text{Memory}_\text{LoRa} \approx \text{Model Size} + \left(\text{ \# trainable Params}_\text{Billions}\times\frac{16}{8} \times 4\right) \times 1.2")
124
 
125
+ #access_token = st.sidebar.text_input("Access token")
126
 
127
+ #if access_token:
128
+ # login(token=access_token)
129
 
130
  #model_name = st.sidebar.text_input("Model name", value="mistralai/Mistral-7B-v0.1")
131
  with st.sidebar.container():
 
200
  info = _memory_table[_memory_table['dtype'] == dtype].set_index('Variable')
201
  show_gpu_info(info, lora_pct, gpu_vendor)
202
  st.write(memory_table.iloc[[0, 1, 2, 4]])
203
+ with col1:
204
  extra = ""
205
  if gpu_vendor == "Apple":
206
  st.warning("This graph is irrelevant for M1/M2 chips as they can't run in parallel.", icon="⚠️")