We assume a pre-trained model & a dataset that covers many possible model edits
Then, we meta-train a model editor that predicts a model update that:
- edits the model
- otherwise keeps the model behavior the same
(2/4)
You can train model editors for massive models (e.g. GPT-J, T5-11B) in <1 day on a single GPU.
Edits with the resulting model editor are extremely fast, with edit success rate of 80-90%.
(3/4)
While I think that MEND makes significant progress on the problem of model editing, still lots of open challenges including:
- deeper edit generalization
- making many model edits
It was also a lot of fun to collaborate with @stanfordnlp!
(4/4)
• • •
Missing some Tweet in this thread? You can try to
force a refresh
2/ Student feedback is a fundamental problem in scaling education.
Providing good feedback is hard: existing approaches provide canned responses, cryptic error messages, or simply provide the answer.
3/ Providing feedback is also hard for ML: not a ton of data, teachers frequently change their assignments, and student solutions are open-ended and long-tailed.
Supervised learning doesn’t work. We weren’t sure if this problem can even be solved using ML.
To get reward functions that generalize, we train domain-agnostic video discriminators (DVD) with:
* a lot of diverse human data, and
* a narrow & small amount of robot demos
The idea is super simple: predict if two videos are performing the same task or not.
(2/5)
This discriminator can be used as a reward by feeding in a human video of the desired task and a video of the robot’s behavior.
We use it by planning with a learned visual dynamics model.
(3/5)
To think about this question, we first look at how equivariances are represented in neural nets.
They can be seen as certain weight-sharing & weight-sparsity patterns. For example, consider convolutions.
(2/8)
We reparametrize a weight matrix into a sharing matrix & underlying filter parameters
It turns out this can provably represent any equivariant structure + filter parameters, for all group-equivariant convolutions with finite groups.
(3/8)