Recent work explores multiple facets of large language models. One direction focuses on improving reasoning through reinforcement learning, noting that strategic compute investment via RL methods can be more effective than simply scaling model size and data. Another direction investigates interpretability, including attention superposition, cross-layer attention representations, and the varying reasons models refuse jailbreaks. Finally, research introduces fine-tuning defenses, StruQ and SecAlign, designed to mitigate prompt injection vulnerabilities in LLM-integrated applications, significantly reducing the success of such attacks while preserving utility.