Hi.
I am Marco.

Check out the nnj repo
Me deciding to create nnj

Publications

Outcome of my research from 2021 to present.

  • Uncertainty for huge models

    Sketched Lanczos uncertainty score: a low-memory summary of the Fisher information. ArXiv. Neurips.

    Current uncertainty quantification is memory and compute expensive, which hinders practical uptake. To counter, we develop Sketched Lanczos Uncertainty (SLU): an architecture-agnostic uncertainty score that can be applied to pre-trained neural networks with minimal overhead. We combine Lanczos' algorithm with dimensionality reduction techniques to compute a sketch of the leading eigenvectors of the Fisher information. Empirically, SLU yields well-calibrated uncertainties even for large Vision Attention models with 200M parameters.

  • Bayes meets reparameterizations

    Reparameterization invariance in approximate Bayesian inference. (SPOTLIGHT) ArXiv. Neurips.

    Current approximate posteriors in Bayesian neural networks (BNNs) exhibit a crucial limitation: they fail to maintain invariance under reparameterization, i.e. BNNs assign different posterior densities to different parametrizations of identical functions. This creates a fundamental flaw in the application of Bayesian principles as it breaks the correspondence between uncertainty over the parameters with uncertainty over the parametrized function. In this paper, we investigate this issue by developing a new geometric view of reparametrizations.

  • Bayesian Metric Learning

    Bayesian Metric Learning for Uncertainty Quantification in Image Retreival. ArXiv. Neurips 2023.

    We propose the first Bayesian encoder for metric learning. Rather than relying on neural amortization as done in prior works, we learn a distribution over the network weights with the Laplace Approximation. We actualize this by first proving that the contrastive loss is a valid log-posterior. We then propose three methods that ensure a positive definite Hessian. Lastly, we present a novel decomposition of the Generalized Gauss-Newton approximation. Empirically, we show that our Laplacian Metric Learner (LAM) estimates well-calibrated uncertainties, reliably detects out-of-distribution examples, and yields state-of-the-art predictive performance.

  • Laplacian AutoEncoder

    Laplacian Autoencoders for Learning Stochastic Representations. ArXiv. Neurips 2022.

    Established methods for unsupervised representation learning such as variational autoencoders produce none or poorly calibrated uncertainty estimates making it difficult to evaluate if learned representations are stable and reliable. In this work, we present a Bayesian autoencoder for unsupervised representation learning, which is trained using a novel variational lower-bound of the autoencoder evidence. This is maximized using Monte Carlo EM with a variational distribution that takes the shape of a Laplace approximation. We develop a new Hessian approximation that scales linearly with data size allowing us to model high-dimensional data. Empirically, we show that our Laplacian autoencoder estimates well-calibrated uncertainties in both latent and output space.

  • Bayesian Skip Connections

    Laplacian Segmentation Networks: Improved Epistemic Uncertainty from Spatial Aleatoric Uncertainty. ArXiv. Miccai 2024.

    Out of distribution (OOD) medical images are frequently encountered, e.g. because of site- or scanner differences, or image corruption. OOD images come with a risk of incorrect image segmentation, potentially negatively affecting downstream diagnoses or treatment. To ensure robustness to such incorrect segmentations, we propose Laplacian Segmentation Networks, (LSN) that jointly model epistemic (model) and aleatoric (data) uncertainty in image segmentation. We capture data uncertainty with a spatially correlated logit distribution. For model uncertainty, we propose the first Laplace approximation of the weight posterior that scales to large neural networks with skip connections that have high-dimensional outputs. Empirically, we demonstrate that modelling spatial pixel correlation allows the Laplacian Segmentation Network to successfully assign high epistemic uncertainty to out-of-distribution objects appearing within images.

  • Curious Explorer

    Curious Explorer: a provable exploration strategy in Policy Learning. ArXiv. IEEE 2024.

    Having access to an exploring restart distribution is critical with policy gradient methods in Reinforcement Learning. However, this assumption can be unfeasible in certain environments. In these cases, classical policy gradient algorithms can have very poor convergence properties and sample efficiency. In this paper, we develop Curious Explorer, a novel and simple iterative state space exploration strategy. Curious Explorer uses intrinsic rewards assigned to the set of poorly visited states produces a sequence of policies, each one more exploratory than the previous one in an informed way. We provide theoretical upper bounds on how often an optimal policy visits poorly visited states.

Projects

Collection of random cool stuff.

About

Attempt to resume myself in 10 lines + a bunch of links.

I am Marco Miani, mathematician and passionate about everything that is simple but not trivial. I graduated at Scuola Normale Superiore in Pisa. I'm now a PhD researcher in the Cognitive Systems section at Technical University of Denmark in Copenaghen.

I’m currently working on deep learning theory.

Outside of science, I enjoy trampolines, fpv drones and onewheels.