#modelediting

2024-03-13

Rome edits model weights to replace a fact the model knew with another

But, it sometimes ruined everything in the process

Not anymore, this was an implementation problem and

R-Rome solved it

☠️arrr- rome☠️

arxiv.org/abs/2403.07175
#nlproc #modelEditing #llm #llms #ml

2024-01-19

How can we improve LM factuality and editability with nothing but the LM itself?

Introducing Deductive Closure Training (DCT):

1. generate statements and their implications
2. identify a logically consistent subset
3. distill this subset back to LM

lingo-mit.github.io/deductive-

#NLP #modelEditing #LLM #LLMs #data #bias #factuality

2023-07-25

You edit a model telling it Baiden is the U.S. president
Now you ask:
Who lives in the white house?
What do you think it answers? (hint: not Baiden)

@mega arxiv.org/abs/2307.12976
#nlproc #ModelEditing #machinelearning

2023-07-16

This entity wasn't in the pretraining😢
Don't cry, little ML researcher

Take new term definitions
Continue them with follow-up sentences
Distill your model (D-KL) to continue the same, without the definition
You know the new terms now
Go prompt them tiger🐯

arxiv.org/abs/2306.09306

#NLProc #modelRecycling #ModelEditing #machinelearning

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst