Skip to main content
Reports
Created by
Created On
Last edited
0
2025-08-23
0
2025-08-26
fixed bench sweep
I've long suspected that there was smth wrong w backprop
0
2025-08-17
bench sweep
I've long suspected that there was smth wrong w backprop
0
2025-08-12
0
2025-08-04
whittle away
some features may be unnecessary.
0
2025-06-30
0
2025-06-02
0
2025-05-07
0
2025-04-30
0
2025-03-31
fast
An attempt to make testing palindromes fast. Just breakneck lr, pretty much.
0
2025-02-26
0
2025-02-12
0
2025-02-04
Sweeps for Paper Figures
just a wee bit o hyperparameter tuning. nothing serious
0
2025-01-18
Paper Figures
Starts with the Key-Recall task. Then, perhaps the associative retrieval task from Ba & Hinton, repeated sequences, or even the baby names or short stories datasets. Might consider hyperparameter sweeps or a search for the ideal effective update magnitude.
0
2025-01-18
0
2025-01-18
0
2024-09-15
0
2024-09-11
0
2024-09-05
0
2024-09-03
0
2024-08-28
0
2024-08-21
regularize to effective lr
a new mechanism keeps the average of the plasticity values to be around 1, so that the effective lr is always the same, overall.
0
2024-08-19
make long_range dataset work
I want to just tinker with what I have so far to just get the model to begin to really show useful results on this dataset. The build of it really feels like it inherently proves what I'm seeking to prove about the model.
0
2024-08-06
limit effective lr
experiments with static plasticity and with trainable
0
2024-08-05
Static Plasticity
Should the plasticity values for each parameter change? Or can they stay the same and still let the model work fine? What should their value be? Randomly initialized, or all zero, or something else?
0
2024-07-25
No Recurrent Connection
Remove the recurrent connection. Try to get the rnn running, anyways, using weight updates for short-term memory. The plasticity value ought to be instrumental, here, some weights being more plastic than others and therefore changing quickly so as to store info.
0
2024-07-17
0
2024-07-24
0
2024-07-10
0
2024-07-03
0
2024-06-28
0
2024-06-25
0
2024-06-21
0
2024-06-17
torn apart
a look at what the rnn is doing step-by-step
0
2024-06-13
0
2024-06-04
0
2024-06-03
0
2024-03-12
0
2024-03-13
0
2024-03-11
wackpropogation
Direct feedback alignment, but the weight updated is modulated by a local rule that preserves signal from the previous layer.
0
2024-03-08
0
2024-03-07
0
2024-02-16
0
2024-02-16
0
2024-01-23