Oh! I should have clarified that I’m not the author of that paper, I just read it and wanted to share.
That idea of a supervisor AI agent is essentially the idea, but instead of supervising the weights or internal activations, it’s actually supervising the tokens of output instead. Very cool stuff!
I have some more notes that I forgot to copy over here, so I’ll share those tomorrow 😅
Oh! I should have clarified that I’m not the author of that paper, I just read it and wanted to share.
That idea of a supervisor AI agent is essentially the idea, but instead of supervising the weights or internal activations, it’s actually supervising the tokens of output instead. Very cool stuff!
I have some more notes that I forgot to copy over here, so I’ll share those tomorrow 😅