One more non-NLP GPT experiment: Prompted with a training set, can a GPT predict the weights of a MLP?
1/6
Each sequence is built as follows:
- generate eight rectangles at random in [-1,1]^2, their interior is class 1, exterior class 0.
- generate uniformly 250 training points and 250 test points quantize the xs and ys in 101 values.
2/6
- train a one hidden layer MLP with 2 input units, 32 hidden units and 2 output units. Quantize the weights during training more and more aggressively so that at the end they are quantized in 101 values
3/6
- create a sequence with 750 tokens for the training set (x/y/class) followed by a marker, and the quantized weights.
4/6
Here is the graph after 1 and 7 epochs of the ratio between the test error rate of the trained MLP and the GPT-generated MLP (e.g. 700 out of 1000 GPT-generated MLPs have a test error lesser than 2x the backprop one)
No extraordinary, but still cool IMO.
6/6
Oct 18, 2023 · 6:37 AM UTC