penzai0.2.4
Published
Penzai: A JAX research toolkit for building, editing, and visualizing neural networks.
pip install penzai
Package Downloads
Authors
Project URLs
Requires Python
>=3.10
Dependencies
- absl-py
>=1.4.0
- jax
>=0.4.23
- numpy
>=1.25.2
- ordered_set
>=4.1.0
- treescope
>=0.1.3
- typing_extensions
>=4.2
- pylint
>=2.6.0; extra == "dev"
- pyink
>=24.3.0; extra == "dev"
- ipython
; extra == "dev"
- jupyter
; extra == "dev"
- pytest
>=8.2.2; extra == "dev"
- pytype
; extra == "dev"
- ipython
; extra == "docs"
- flax
>=0.8.2; extra == "docs"
- optax
; extra == "docs"
- setuptools
; extra == "docs"
- sphinx
<7.3.0,>=6.0.0; extra == "docs"
- sphinx-book-theme
>=1.0.1; extra == "docs"
- sphinxcontrib-katex
; extra == "docs"
- ipython
>=8.8.0; extra == "docs"
- myst-nb
>=1.0.0; extra == "docs"
- myst-parser
>=3.0.1; extra == "docs"
- matplotlib
>=3.5.0; extra == "docs"
- packaging
==24.1; extra == "docs"
- sphinx-collections
>=0.0.1; extra == "docs"
- sphinx_contributors
; extra == "docs"
- sphinx-hoverxref
; extra == "docs"
- jax
[cpu]>=0.4.23; extra == "docs"
- ipython
; extra == "extras"
- flax
>=0.8.2; extra == "extras"
- optax
; extra == "extras"
- torch
; extra == "extras"
- transformers
>=4.41.2; extra == "extras"
- ipython
; extra == "notebook"
- flax
>=0.8.2; extra == "notebook"
- optax
; extra == "notebook"
- orbax-checkpoint
; extra == "notebook"
- palettable
; extra == "notebook"
Penzai
盆 ("pen", tray) 栽 ("zai", planting) - an ancient Chinese art of forming trees and landscapes in miniature, also called penjing and an ancestor of the Japanese art of bonsai.
Penzai is a JAX library for writing models as legible, functional pytree data structures, along with tools for visualizing, modifying, and analyzing them. Penzai focuses on making it easy to do stuff with models after they have been trained, making it a great choice for research involving reverse-engineering or ablating model components, inspecting and probing internal activations, performing model surgery, debugging architectures, and more. (But if you just want to build and train a model, you can do that too!)
With Penzai, your neural networks could look like this:
Penzai is structured as a collection of modular tools, designed together but each useable independently:
-
A superpowered interactive Python pretty-printer:
- Treescope (
pz.ts
): A drop-in replacement for the ordinary IPython/Colab renderer, originally a part of Penzai but now available as a standalone package. It's designed to help understand Penzai models and other deeply-nested JAX pytrees, with built-in support for visualizing arbitrary-dimensional NDArrays.
- Treescope (
-
A set of JAX tree and array manipulation utilities:
-
penzai.core.selectors
(pz.select
): A pytree swiss-army-knife, generalizing JAX's.at[...].set(...)
syntax to arbitrary type-driven pytree traversals, and making it easy to do complex rewrites or on-the-fly patching of Penzai models and other data structures. -
penzai.core.named_axes
(pz.nx
): A lightweight named axis system which lifts ordinary JAX functions to vectorize over named axes, and allows you to seamlessly switch between named and positional programming styles without having to learn a new array API.
-
-
A declarative combinator-based neural network library, where models are represented as easy-to-modify data structures:
penzai.nn
(pz.nn
): An alternative to other neural network libraries like Flax, Haiku, Keras, or Equinox, which exposes the full structure of your model's forward pass using declarative combinators. Like Equinox, models are represented as JAX PyTrees, which means you can see everything your model does by pretty printing it, and inject new runtime logic withjax.tree_util
. However,penzai.nn
models may also contain mutable variables at the leaves of the tree, allowing them to keep track of mutable state and parameter sharing.
-
A modular implementation of common Transformer architectures, to support research into interpretability, model surgery, and training dynamics:
penzai.models.transformer
: A reference Transformer implementation that can load the pre-trained weights for the Gemma, Llama, Mistral, and GPT-NeoX / Pythia architectures. Built using modular components and named axes, to simplify complex model-manipulation workflows.
Documentation on Penzai can be found at https://penzai.readthedocs.io.
[!IMPORTANT] Penzai 0.2 includes a number of breaking changes to the neural network API. These changes are intended to simplify common workflows by introducing first-class support for mutable state and parameter sharing and removing unnecessary boilerplate. You can read about the differences between the old "V1" API and the current "V2" API in the "Changes in the V2 API" overview.
If you are currently using the V1 API and have not yet converted to the V2 system, you can instead keep the old behavior by importing from the
penzai.deprecated.v1
submodule, e.g. ::from penzai.deprecated.v1 import pz from penzai.deprecated.v1.example_models import simple_mlp
Getting Started
If you haven't already installed JAX, you should do that first, since the installation process depends on your platform. You can find instructions in the JAX documentation. Afterward, you can install Penzai using
pip install penzai
and import it using
import penzai
from penzai import pz
(penzai.pz
is an alias namespace, which makes it easier to reference
common Penzai objects.)
When working in an Colab or IPython notebook, we recommend also configuring Treescope (Penzai's companion pretty-printer) as the default pretty printer, and enabling some utilities for interactive use:
import treescope
treescope.basic_interactive_setup(autovisualize_arrays=True)
Here's how you could initialize and visualize a simple neural network:
from penzai.models import simple_mlp
mlp = simple_mlp.MLP.from_config(
name="mlp",
init_base_rng=jax.random.key(0),
feature_sizes=[8, 32, 32, 8]
)
# Models and arrays are visualized automatically when you output them from a
# Colab/IPython notebook cell:
mlp
Here's how you could capture and extract the activations after the elementwise nonlinearities:
@pz.pytree_dataclass
class AppendIntermediate(pz.nn.Layer):
saved: pz.StateVariable[list[Any]]
def __call__(self, x: Any, **unused_side_inputs) -> Any:
self.saved.value = self.saved.value + [x]
return x
var = pz.StateVariable(value=[], label="my_intermediates")
# Make a copy of the model that saves its activations:
saving_model = (
pz.select(mlp)
.at_instances_of(pz.nn.Elementwise)
.insert_after(AppendIntermediate(var))
)
output = saving_model(pz.nx.ones({"features": 8}))
intermediates = var.value
To learn more about how to build and manipulate neural networks with Penzai, we recommend starting with the "How to Think in Penzai" tutorial or one of the other tutorials in the Penzai documentation.
Citation
If you have found Penzai to be useful for your research, please consider citing the following writeup (also avaliable on arXiv):
@article{johnson2024penzai,
author={Daniel D. Johnson},
title={{Penzai} + {Treescope}: A Toolkit for Interpreting, Visualizing, and Editing Models As Data},
year={2024},
journal={ICML 2024 Workshop on Mechanistic Interpretability}
}
This is not an officially supported Google product.