Julian G. Zilly

Julian G. Zilly

Applied Scientist, Flexion Robotics  ·  Zurich, Switzerland

julian@julianzilly.com · GitHub · Google Scholar · arXiv · ORCID · LinkedIn

Research interests

What if ideas got taken seriously enough to see them through?

by day

Generative AI, robotics, and machine learning. The work is architectural: how such systems are designed, shipped, and kept running at scale to reach users. In production, resource trade-offs like correctness versus latency dominate.

by night

The information-theoretic foundations of quantum mechanics. Existence is distinguishability. A physical state is its pattern of distinctions against all other states. From finite precision, finite-dimensional quantum mechanics follows. The core is formalized in Lean 4.

Guiding principle: Theory → Production.

Currently: submitting Existence as Distinguishability to Physical Review X; follow-up work on its physical consequences and on generalizing the kernel framework to pure mathematics; and a methodology paper in which a Lean formalization, a semantic graph, and a LaTeX manuscript jointly form the primary research artifact. The published article becomes a rendering of the work, not its source.

5 distinguishable states · 10 pairwise distinctions · Hilbert dim 5
One statement, three renderings. Drag N — all three stay in sync, because they are views of the same object.
5
Lean 4

      
LaTeX
semantic

Bio

I am a researcher and engineer based in Zurich. In May 2026 I am joining Flexion Robotics as an Applied Scientist. From 2023 to March 2026 I was an Applied Scientist II at AWS in New York, working on generative AI: hallucination detection for Q Business, agentic natural-language data analysis for Quick Suite, and enterprise retrieval systems.

I earned my PhD in 2022 from ETH Zurich, advised by Prof. Emilio Frazzoli and Prof. Stefano Soatto, on plasticity, invariance, and priors in deep neural networks. Before the PhD I spent six months as a research scientist at IDSIA under Prof. Jürgen Schmidhuber (co-inventor of LSTM). MSc Robotics, ETH 2016, Excellence Scholarship; BSc Mechanical Engineering, ETH 2013, Outstanding Bachelor Award.

Beyond day-to-day engineering I maintain an active research line on the information-theoretic foundations of quantum mechanics, deriving the Schrödinger equation and the Born rule from a single ontological principle, and on the corresponding Lean 4 formalization. My published work has accumulated over 1000 citations across ICML, NeurIPS, and ICRA venues.

Selected work

Existence as Distinguishability: Quantum Mechanics from Finite Graded Equality

Julian G. Zilly  ·  arXiv:2603.11900  ·  Lean 4 formalization  ·  2026

A derivation of finite-dimensional quantum mechanics from a single ontological principle: existence is constituted by distinguishability. Comparison has finite resolution, distinctions are graded by a kernel $K(x,y) \in [0,1]$ replacing binary equality, and two axioms (finite capacity with saturation; universal relationality) determine the state space $\mathbb{C}P^{N-1}$, complex coefficients, the Born rule, unitary dynamics, and tensor products uniquely. Standard QM is the $N\!\to\!\infty$ limit; finite $N$ is a natural UV cutoff. The core mathematical derivation is machine-checked in Lean 4 with zero sorry steps.

On Plasticity, Invariance, and Mutually Frozen Weights in Sequential Task Learning

Julian G. Zilly, Alessandro Achille, Andrea Censi, Emilio Frazzoli  ·  NeurIPS 2021  ·  Scholar

Pre-training does not always help downstream tasks. We introduce the notion of mutually frozen weights: weight configurations where gradients on one task are annihilated by the geometry imposed by another, producing a hidden form of plasticity loss. The work characterizes when and why sequential task structure induces representation rigidity, giving a concrete answer to the plasticity question that underlies continual and transfer learning.

Recurrent Highway Networks

Julian G. Zilly, Rupesh Kumar Srivastava, Jan Koutník, Jürgen Schmidhuber  ·  arXiv:1607.03474  ·  ICML 2017  ·  NVIDIA V100 award (Srivastava)  ·  590+ citations

A recurrent architecture that extends LSTM with deep transition functions between successive timesteps, allowing step-wise computational depth without step-count-dependent vanishing gradients. Establishes new state-of-the-art language-modeling results on Penn Treebank and Hutter Prize Wikipedia, and introduces a GCU-style analysis of gradient flow across compositional transitions.

The AI Driving Olympics at NeurIPS 2018

Julian G. Zilly, Jacopo Tani, Breandan Considine, Bhairav Mehta, Andrea F. Daniele, Manfred Diaz, Gianmarco Bernasconi, Claudio Ruch, Jan Hakenberg, Florian Golemo, A. Kirsten Bowser, Matthew R. Walter, Ruslan Hristov, Sunil Mallya, Emilio Frazzoli, Andrea Censi, Liam Paull  ·  arXiv:1903.02503  ·  NeurIPS'18 Competition Track, 2019

The first AI Driving Olympics (AIDO), a reproducible autonomous-driving competition built on the Duckietown platform. Tasks of increasing complexity, from lane following through control with multiple vehicles and navigation with fleet management, tie together a perception-planning-control pipeline in a benchmarkable form. 100+ international teams participated across the NeurIPS and ICRA editions.

Earlier work on glaucoma detection via entropy sampling (best paper award, Computerized Medical Imaging and Graphics, 2017), and the PhD dissertation Plasticity, Invariance, and Priors in Deep Neural Networks (ETH Zurich, 2022). Full list on Google Scholar.

News