Similar Items: Pretraining Exposure Explains Popularity Judgments in Large Language Models
- A Causal Language Modeling Detour Improves Encoder Continued Pretraining
- EMO: Pretraining Mixture of Experts for Emergent Modularity
- Scalable Token-Level Hallucination Detection in Large Language Models
- Logical Consistency as a Bridge: Improving LLM Hallucination Detection via Label Constraint Modeling between Responses and Self-Judgments
- DPN-LE: Dual Personality Neuron Localization and Editing for Large Language Models
- Detecting Hallucinations in Large Language Models via Internal Attention Divergence Signals