Michael J. Catanzaro

Senior Scientist, Geometric Data Analytics, Inc.

mike.jpg

Hello!

I am a mathematician turned data scientist interested in solving problems in big data and machine learning. Lately I’ve been working on simplifying machine learning problems using the tools of algebraic topology. Before this, I studied topological data analysis and its applications on a variety of applied problems, ranging from understanding deep perception models to analyzing fMRI data. Once upon a time, I studied stochastic topology and empirical currents as they arise in physics and statistical mechanics.

I am currently a Senior Scientist at Geometric Data Analytics, Inc. Prior to this, I was an Assistant Professor in the mathematics department at Iowa State University, a postdoc at the Unviersity of Florida, and a graduate student at Wayne State University. I earned my PhD in 2016 under the supervision of John Klein and Vladimir Chernyak and was mentored by Peter Bubenik as a postdoc.

I like to do lots of other things besides math, machine learning, and physics.

selected publications

  1. swot_preview.png
    Deep Learning Methods for Inference of Sea Surface Kinematics from SWOT Altimetry
    James B. Polly, Kenneth Ball, Michael Catanzaro, and Jay Hineman
    OCEANS 2024 - Halifax, 2024

    Observations of sea surface height (SSH) provided by the Surface Water and Ocean Topography (SWOT) satellite capture features at length scales smaller than the scales on which geostrophic balance is a valid approximation. At these smaller length scales (less than order 100 km) the SSH signature of submesoscale balanced motions are comparable with the signatures of internal gravity waves (IGW). In addition to these non-dynamical features in the SWOT SSH observations, measurements from the SWOT satellite contain noise and errors from a variety of sources. These issues complicate the usual approximation of surface velocity using geostrophic balance as instrument noise and non-dynamical features need to be filtered out. To address these challenges, a recent study [1] has demon-strated the effectiveness of a machine learning approach to infer vorticity from SSH using model data on structured grids. There remain a number of actions needed to extend this method to SWOT measurements. The dimensions and structure of the SWOT observation data differs significantly from the regularly structured model data used in [1]. We have created tools that transform SWOT observations into data structures that are compatible with the existing UNet developed in [1], which was not designed to handle SWOT observations. We revisit the model selection and model architecture of this UNet. We have created methods for handling gaps in SWOT observations including the nadir gap, missing/bad data, and the shape of the swath. These tools also provide some interpolation techniques so that a region of interest larger than the swath may be populated with observations from multiple orbital passes. We present results summarizing the effectiveness of these tools, and discuss the next steps in scaling out these workflows to support global SWOT observations, including validation and fine-tuning.

  2. frontiers_preview.png
    Implications of Data Topology for Deep Generative Models
    Jin Yinzhu, Rory McDaniel, Joseph N. Tatro, Michael J. Catanzaro, Abraham D. Smith, Paul Bendich, Matthew B. Dwyer, and P. Thomas Fletcher
    Frontiers in Computer Science, 2024

    Many deep generative models, such as variational autoencoders (VAEs) and generative adversarial networks (GANs), learn an immersion mapping from a standard normal distribution in a low-dimensional latent space into a higher-dimensional data space. As such, these mappings are only capable of producing simple data topologies, i.e., those equivalent to an immersion of Euclidean space. In this work, we demonstrate the limitations of such latent space generative models when trained on data distributions with non-trivial topologies. We do this by training these models on synthetic image datasets with known topologies (spheres, torii, etc.). We then show how this results in failures of both data generation as well as data interpolation. Next, we compare this behavior to two classes of deep generative models that in principle allow for more complex data topologies. First, we look at chart autoencoders (CAEs), which construct a smooth data manifold from multiple latent space chart mappings. Second, we explore score-based models, e.g., denoising diffusion probabilistic models, which estimate gradients of the data distribution without resorting to an explicit mapping to a latent space. Our results show that these models do demonstrate improved ability over latent space models in modeling data distributions with complex topologies, however, challenges still remain.

  3. parallax_preview.png
    Topological Parallax: A Geometric Specification for Deep Perception Models
    Abraham Smith, Michael Catanzaro, Gabrielle Angeloro, Nirav Patel, and Paul Bendich
    In Advances in Neural Information Processing Systems , 2023

    For safety and robustness of AI systems, we introduce topological parallax as a theoretical and computational tool that compares a trained model to a reference dataset to determine whether they have similar multiscale geometric structure. Our proofs and examples show that this geometric similarity between dataset and model is essential to trustworthy interpolation and perturbation, and we conjecture that this new concept will add value to the current debate regarding the unclear relationship between "overfitting" and "generalization" in applications of deep- learning. In typical DNN applications, an explicit geometric description of the model is impossible, but parallax can estimate topological features (components, cycles, voids, etc.) in the model by examining the effect on the Rips complex of geodesic distortions using the reference dataset. Thus, parallax indicates whether the model shares similar multiscale geometric features with the dataset. Parallax presents theoretically via topological data analysis [TDA] as a bi-filtered persistence module, and the key properties of this module are stable under pertur- bation of the reference dataset.

Feel free to send me an email!