Spaces:
Paused
Paused
File size: 1,479 Bytes
ed5e255 21a461d eef98a6 ed5e255 1f4d5df ed5e255 eef98a6 ed5e255 8a983b3 2fbb1cf 91b65b2 639f902 91b65b2 1813b17 ed5e255 91b65b2 a83d4d2 1f4d5df 91b65b2 1f4d5df |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
---
title: Blog
fullWidth: true
emoji: ⚔️
colorFrom: red
colorTo: green
sdk: streamlit
sdk_version: 1.21.0
app_file: app.py
pinned: true
---

# 7/23/23 - Towards A Unified Agent with Foundation Models
https://arxiv.org/abs/2307.09668
Generate synthetic data set for the state that you want, search over the action space until you find a trajectory that reaches a cosine similarity threshold denoted by the state you want, add all those frames and states of the buffer and incorporate into training
You can bootstrap process with priors still search for the desired state
### reward
Reward any trajectory proportionally to a semantically similar state as any state in a run with a victory condition.
Linear or some function reward curve
### Sample curve
Sections of states with more changes in them
# 7/21/23
am going to naively, without evidence, state that you can represent any function in text with a large language model.
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
# measured steps
Probably should have figured out sooner that small measured steps done in consistent application leads to results. Predicting the outcome while getting there can be interesting but is ultimately just an イメージ in your head.
# Stack More Layers Differently:
High-Rank Training Through Low-Rank Updates
https://arxiv.org/pdf/2307.05695.pdf
https://github.com/guitaricet/peft_pretraining |