EMNLP2025 Xu: tree of prompting
Evaluate the quote attribution score as a way to prioritize more factual quotes.
EMNLP2025 Fan: medium is not the message
Unwanted feature such as language a medium who found in embedding, use linear concept of eraser to learn a projection that minimize information on unwanted features
EMNLP2025 Hong: variance sensitivity induces attention entropy collapse
Softmax is highly sensitive to variance which is why pre-training loss spikes without QK norm
EMNLP2025 Vashurin: lengthy variant estimation of uncertainty
We find that uncertainty of language models decrease as a function of length but quality decreases also as a function of length (because the model struggle), thus, we correct uncertainty measurement based on length such as if we run for too long the model should show up as more uncertain
EMNLP2025 Hutson: measuring informative of open and questions
Using top K outputs in LM, initialize belief state, chooses next action based on those that preturb the belief statement the most
EMNLP2025 Du: similarity individual values by backstory generation
If you want language models to specialize to individual values, generate a backstory for the persona
EMNLP2025 Karamanolakis: interactive machine teaching
Synthetic labeling via instance labels is difficult, because you have to make people label everything. Constitution methods are hard because making constitutions is hard. Instead, use instance labels to extract constitution and feed both into synthetic labeling
