-
Notifications
You must be signed in to change notification settings - Fork 8
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Create text_gen_as_graph_traversal.md
- Loading branch information
Showing
1 changed file
with
22 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,22 @@ | ||
# sandbox to probe scaling laws for representation learning | ||
|
||
labels: experimental | ||
|
||
|
||
inspired by "generative wiki rabbit hole" | ||
|
||
early graph assumes homogeneity of node and edge types. | ||
|
||
certain nodes actually designate community membership. treating community nodes as conventional results in extremely dense graph, eliminating a lot of the desired structure. | ||
|
||
structure can be recovered by pojecting membership nodes up into an orthogonal dimension, i.e. layering on a hypernetwork channel for each community... | ||
|
||
mmm... i dunno. what i want is for this to be a multi-partite projection | ||
|
||
... | ||
|
||
anyway, underlying mechanisms aside, my thinking is that i could construct toy graphs and see under what conditions a transformer learns the multi-partite topology. | ||
|
||
use a "next token prediction" objective, where sequences are random paths through the graph, and tokens are nodes | ||
|
||
should be able model learning dynamics of NLU under this simplification? |