Icon for Bryon Tjanaka
Bryon Tjanaka profile picture

Bryon Tjanaka

I am a software engineer at Waymo developing world models to enhance simulation of long-tail scenarios. Previously, I completed my Ph.D. with Stefanos Nikolaidis in the ICAROS Lab at USC, where I researched applications of quality diversity optimization to reinforcement learning and robotics. Overall, I am passionate about delivering polished solutions to challenging machine learning problems. My work includes pyribs, which has become one of the most popular open source libraries for quality diversity, and Discount Model Search, which was published as an oral at ICLR 2026. In my spare time, I enjoy competing in ballroom dance, visiting museums, and hiking.

News

  1. Discount Model Search has been accepted to ICLR 2026 as an Oral Presentation!

    February 5, 2026

  2. I have defended my Ph.D. and graduated from USC! Next, I will be joining Waymo.

    December 2025

Show more
  1. Today I will present pyribs: A Bare-Bones Python Library for Quality Diversity Optimization at GECCO 2023!

    July 17, 2023

  2. I have received an NVIDIA Academic Hardware Grant.

    March 10, 2022

  3. I have received an NSF Graduate Research Fellowship.

    March 23, 2021

  4. Download pyribs, a quality diversity library I released with ICAROS.

    February 5, 2021

  5. I have started my Ph.D. at USC.

    August 24, 2020

Selected Publications

  • Discount Model Search for Quality Diversity Optimization in High-Dimensional Measure Spaces

    B. Tjanaka, H. Chen, M. C. Fontaine, S. Nikolaidis

    International Conference on Learning Representations (ICLR), April 2026,
    Oral Presentation, Acceptance rate: 1.2%

    Abstract

    Quality diversity (QD) optimization searches for a collection of solutions that optimize an objective while attaining diverse outputs of a user-specified, vector-valued measure function. Contemporary QD algorithms are typically limited to low-dimensional measures because high-dimensional measures are prone to distortion, where many solutions found by the QD algorithm map to similar measures. For example, the state-of-the-art CMA-MAE algorithm guides measure space exploration with a histogram in measure space that records so-called discount values. However, CMA-MAE stagnates in domains with high-dimensional measure spaces because solutions with similar measures fall into the same histogram cell and hence receive the same discount value. To address these limitations, we propose Discount Model Search (DMS), which guides exploration with a model that provides a smooth, continuous representation of discount values. In high-dimensional measure spaces, this model enables DMS to distinguish between solutions with similar measures and thus continue exploration. We show that DMS facilitates new capabilities for QD algorithms by introducing two new domains where the measure space is the high-dimensional space of images, which enables users to specify their desired measures by providing a dataset of images rather than hand-designing the measure function. Results in these domains and on high-dimensional benchmarks show that DMS outperforms CMA-MAE and other existing black-box QD algorithms.

    Diagram of hikers found by Discount Model Search.
  • pyribs: A Bare-Bones Python Library for Quality Diversity Optimization

    B. Tjanaka, M. C. Fontaine, D. H. Lee, Y. Zhang, N. R. Balam, N. Dennler, S. S. Garlanka, N. D. Klapsis, S. Nikolaidis

    Genetic and Evolutionary Computation Conference (GECCO), July 2023, Acceptance rate: 34.7%

    Abstract

    Recent years have seen a rise in the popularity of quality diversity (QD) optimization, a branch of optimization that seeks to find a collection of diverse, high-performing solutions to a given problem. To grow further, we believe the QD community faces two challenges: developing a framework to represent the field's growing array of algorithms, and implementing that framework in software that supports a range of researchers and practitioners. To address these challenges, we have developed pyribs, a library built on a highly modular conceptual QD framework. By replacing components in the conceptual framework, and hence in pyribs, users can compose algorithms from across the QD literature; equally important, they can identify unexplored algorithm variations. Furthermore, pyribs makes this framework simple, flexible, and accessible, with a user-friendly API supported by extensive documentation and tutorials. This paper overviews the creation of pyribs, focusing on the conceptual framework that it implements and the design principles that have guided the library's development.

    Diagram of the RIBS framework described in the paper.
  • Approximating Gradients for Differentiable Quality Diversity in Reinforcement Learning

    B. Tjanaka, M. C. Fontaine, J. Togelius, S. Nikolaidis

    Genetic and Evolutionary Computation Conference (GECCO), July 2022, Acceptance rate: 37%

    Abstract

    Consider the problem of training robustly capable agents. One approach is to generate a diverse collection of agent polices. Training can then be viewed as a quality diversity (QD) optimization problem, where we search for a collection of performant policies that are diverse with respect to quantified behavior. Recent work shows that differentiable quality diversity (DQD) algorithms greatly accelerate QD optimization when exact gradients are available. However, agent policies typically assume that the environment is not differentiable. To apply DQD algorithms to training agent policies, we must approximate gradients for performance and behavior. We propose two variants of the current state-of-the-art DQD algorithm that compute gradients via approximation methods common in reinforcement learning (RL). We evaluate our approach on four simulated locomotion tasks. One variant achieves results comparable to the current state-of-the-art in combining QD and RL, while the other performs comparably in two locomotion tasks. These results provide insight into the limitations of current DQD algorithms in domains where gradients must be approximated. Source code is available at https://github.com/icaros-usc/dqd-rl.

    Diagram of how the CMA-MEGA is applied in reinforcement learning settings. This is also the first figure in our paper.