Skip to content

Commit

Permalink
Create learning_traj_cond_pred.md
Browse files Browse the repository at this point in the history
  • Loading branch information
dmarx authored Mar 5, 2024
1 parent 1b7d445 commit 9edf17c
Showing 1 changed file with 16 additions and 0 deletions.
16 changes: 16 additions & 0 deletions learning_traj_cond_pred.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,16 @@
# Predict learning trajectories conditioned on model embeddings

labels: experimental, contrastive_learning, llm, code_gen, public_good, open_source, open_ai

* ContrastiveLoss(TEXT(model_config), TEXT(model_code)) -> ModelEmbedding space
* wandb dataset: conditioning on ModelEmbedding and some text representation of the early training history, predict the training history out to some horizon into the future

pilot study:

* some trivial model family paired with a problem we can train quickly and easily, like MNIST
* generate configs representing a hyperparameter sweep
* wait... wandb has a sweep feature built in, right? let's just use that.
* generate some dense region of hyperparameter space to train on
* goal: does this strategy at least work for a trivial problem?
* if yes, use some more sophisticated models and see what happens
* use (LoRA? on) a 7B codegen base model to parameterize the contrastively-learned ModelEmbedding space

0 comments on commit 9edf17c

Please sign in to comment.