ICLR 2026
ICLR is the International Conference on Learning Representations. In this edition, held in Rio de Janeiro, I will present two papers: Cooperative Sheaf Neural Networks and Orthogonal Gradient Projection for Continual LLM Unlearning.
ICLR is the International Conference on Learning Representations. In this edition, held in Rio de Janeiro, I will present two papers: Cooperative Sheaf Neural Networks and Orthogonal Gradient Projection for Continual LLM Unlearning.
Juan Belieni, Ana Carolina Erthal, Eliezer de Souza da Silva, Diego Mesquita
Machine unlearning enables the removal of specific knowledge from trained models without full retraining. While effective methods exist for single deletion requests, handling sequential requests in large language models (LLMs) remains underexplored. In this setting, we observe that gradient interference between successive unlearning steps degrades prior objectives. We propose ONPO (Orthogonal Negative Preference Optimization), which projects each step’s update onto the orthogonal complement of a low-dimensional subspace spanned by cached gradients from previous unlearning requests. This preserves prior unlearning objectives with minimal per-step overhead. On the TOFU benchmark, ONPO achieves a better trade-off between forgetting quality and model utility than existing methods.
André Ribeiro, Ana Luiza Tenorio, Juan Belieni, Amauri Souza, Diego Mesquita
Sheaf neural networks (SNNs) leverage cellular sheaves to induce flexible diffusion processes on graphs, generalizing the diffusion mechanism of classical graph neural networks. While SNNs have been shown to cope well with heterophilic tasks and alleviate oversmoothing, we argue that there is further room for improving sheaf diffusion. More specifically, we show that SNNs do not allow nodes to independently choose how they cooperate with their neighbors, i.e., whether they convey and/or gather information to/from their neighbors. To address this issue, we first introduce the notion of cellular sheaves over directed graphs and characterize their in- and out-degree Laplacians. We then leverage our construction to propose Cooperative Sheaf Neural Network (CSNN). Additionally, we formally characterize its receptive field and prove that it allows nodes to selectively attend (listen) to arbitrarily far nodes while ignoring all others in their path, which is key to alleviating oversquashing. Our results on synthetic data empirically substantiate our claims, showing that CSNN can handle long-range interactions while avoiding oversquashing. We also show that CSNN performs strongly in heterophilic node classification and long-range graph classification benchmarks.
ML4Good is a bootcamp focused on AI Safety upskilling, including workshops on interpretability, alignment and governance of artificial intelligence. In this edition, I participated as a teaching assistant.
ML4Good is a bootcamp focused on AI Safety upskilling, including workshops on interpretability, alignment and governance of artificial intelligence. In this edition, I developed a Mechanistic Interpretability Course.
Condor Camp is an amazing camp on AI safety. There, I learned and discussed topics related to AI governance and technical AI safety. I was also introduced to the effective altruism philosophy. It was probably the best experience regarding career planning as well.
Tropical ProbAi is a event about probabilistic artificial intelligence that occurred at FGV/EMAp in Rio de Janeiro at the end of January 2024. It had lecturers from MIT, Meta, USP, etc.