Simplex-Zero: Direct Generative Modeling of Sparse Compositional Data

Aihua Li1, Li Ma2,
1Duke University, 2University of Chicago

Abstract

We present Simplex-Zero, a deep generative framework that directly generates sparse compositional data on the probability simplex. It overcomes the geometric distortions and the inability to model exact zeros of traditional transformation-based methods. Simplex-Zero utilizes a convolutional neural network, informed by a tree-structural sorting of the data, to achieve effective feature extraction and high-fidelity generative performance.


Contribution 1: Tree-informed CNN

tree

Our approach first addresses the data's inherent tree structure to exploit hierarchical feature correlations. We permute the input vector using a Depth-First Search of the tree, which places related features in proximity. This allows a subsequent convolutional neural network to effectively process the data as spatially structured, enabling it to learn salient, low-dimensional representations.



Contribution 2: Simplex-Zero Activation Function

We introduce Simplex-Zero Activation (SZA) Function, a novel operator designed to produce outputs that are simultaneously non-negative, compositional (sum-to-one), and sparse. By also maintaining Lipschitz continuity, our function ensures stable and efficient training. This approach directly contrasts with traditional pre-processing transformations, which introduce geometric distortions and irreversibly destroy the data's natural sparsity.

SZA Function

sza

Contribution 3: Generative Models for Sparse Compositional Data

Our method seamlessly integrates with existing generative modeling frameworks to enable the direct generation of sparse compositional data. This approach yields a 4x speedup over complex manifold-specific models while improving generation quality by 81% against traditional transformation-based baselines.

Flow Matching: Direct flow in the vector field towards target probability simplex

True flow GIF
Learned flow GIF

Denoising Diffusion Model: Diffusion with efficient projection on the probability simplex

True diffusion GIF
Learned diffusion GIF

Variational AutoEncoder: Enhanced Classification in Representation Space

PCA over Latent Representations (SZA)

latent1

PCA over Latent Representations (CLR)

latent2

High-Fidelity Generation

This approach yields a 4x speedup over complex manifold-specific models while improving generation quality by 81% against traditional transformation-based baselines.

Stacked-Bar Plot of Compositional Data

tree

BibTeX

@article{li2025SimplexZero,
      author = {A. Li and L. Ma},
      title  = {Simplex-Zero: Direct Generative Modeling of Sparse Compositional Data},
      year   = {2025},
      }