Often it is useful to have structures in a program containing a mixture of JAX arrays and non-JAX types (e.g. strings, ...). But, this makes it difficult to pass these objects through JAX transformations. Drinx solves this by allowing dataclass fields to be declared as static. Moreover, drinx introduces numerous quality-of-life features when working with dataclasses in JAX.
You can install drinx simply via
pip install drinxIf you want to use the GPU-acceleration from JAX, you can install afterwards:
pip install jax[cuda]Below you can find some examples to get you quickly started with drinx. But, beware, there are so much more features available, which are documented in detail in our Documentation
Use @drinx.dataclass as a drop-in replacement for @dataclasses.dataclass.
The class is automatically frozen and registered as a JAX pytree:
import jax
import jax.numpy as jnp
import drinx
@drinx.dataclass
class Params:
weights: jax.Array
bias: jax.Array
params = Params(weights=jnp.ones((3,)), bias=jnp.zeros((3,)))
# Works transparently with JAX transforms
doubled = jax.tree_util.tree_map(lambda x: x * 2, params)Fields that should not be traced by JAX (e.g. shapes, dtypes, hyperparameters)
are marked with static_field or field(static=True). Changing a static
field triggers recompilation under jit:
@drinx.dataclass
class Model:
weights: jax.Array
hidden_size: int = drinx.static_field(default=128)
@jax.jit
def forward(model, x):
# hidden_size is a compile-time constant; weights are traced
return model.weights[:model.hidden_size] @ x
model = Model(weights=jnp.ones((128, 32)))Subclass DataClass instead of using the decorator. The transform is applied
automatically โ no @dataclass needed:
class Model(drinx.DataClass):
weights: jax.Array
learning_rate: float = drinx.static_field(default=1e-3)
model = Model(weights=jnp.ones((10,)))Dataclass options are forwarded via the class definition, or alternatively by using a combination of inheritance and decorator.
class Config(drinx.DataClass, kw_only=True, order=True):
hidden_size: int = drinx.static_field(default=128)
num_layers: int = drinx.static_field(default=4)
# This is the recommended way: Typechecker will recognize the kw_only argument correctly
@drinx.dataclass(kw_only=True, order=True)
class Config(drinx.DataClass):
hidden_size: int = drinx.static_field(default=128)
num_layers: int = drinx.static_field(default=4)Because drinx dataclasses are frozen, fields cannot be mutated in place.
aset performs a functional update and returns a new instance. It supports
nested paths using -> as a separator, integer indices [n], and string
dictionary keys ['k'].
Note that this function is only available when inheriting the drinx.Dataclass, but not from the decorator.
class Inner(drinx.DataClass):
w: jax.Array
class Outer(drinx.DataClass):
inner: Inner
bias: jax.Array
outer = Outer(inner=Inner(w=jnp.ones((3,))), bias=jnp.zeros((1,)))
# Update a top-level field
outer2 = outer.aset("bias", jnp.ones((1,)))
# Update a nested field
outer3 = outer.aset("inner->w", jnp.zeros((3,)))Drinx dataclasses work with all JAX transforms out of the box:
class State(drinx.DataClass):
x: jax.Array
step_size: float = drinx.static_field(default=0.1)
# jit
@jax.jit
def update(state):
# updated_copy is convenience wrapper for altering top-level attributes
return state.updated_copy(x=state.x - state.step_size)
def loss(state):
return jnp.sum(state.x ** 2)
grads = jax.grad(loss)(State(x=jnp.array([1.0, 2.0, 3.0])))
@jax.vmap
def scale(state):
return state.x * 2
batched = State(x=jnp.array([[1.0, 2.0], [3.0, 4.0]]))
result = scale(batched) # shape (2, 2)tree_diagram and tree_summary let you inspect any JAX pytree at a glance:
class Encoder(drinx.DataClass):
w: jax.Array
b: jax.Array
class Model(drinx.DataClass):
encoder: Encoder
head: jax.Array
model = Model(encoder=Encoder(w=jnp.ones((16, 32)), b=jnp.zeros((16,))), head=jnp.ones((4, 16)))
print(drinx.tree_diagram(model))
# Model
# โโโ .encoder:Encoder
# โ โโโ .w=f32[16,32] โ [1.0, 1.0], ฮผ=1.0, ฯ=0.0
# โ โโโ .b=f32[16] โ [0.0, 0.0], ฮผ=0.0, ฯ=0.0
# โโโ .head=f32[4,16] โ [1.0, 1.0], ฮผ=1.0, ฯ=0.0
print(drinx.tree_summary(model))
# โโโโโโโโโโโโโโโโฌโโโโโโโโโโโฌโโโโโโโโฌโโโโโโโโโ
# โName โType โCount โSize โ
# โโโโโโโโโโโโโโโโผโโโโโโโโโโโผโโโโโโโโผโโโโโโโโโค
# โ.encoder.w โf32[16,32]โ512 โ2.00KB โ
# โโโโโโโโโโโโโโโโผโโโโโโโโโโโผโโโโโโโโผโโโโโโโโโค
# โ.encoder.b โf32[16] โ16 โ64.00B โ
# โโโโโโโโโโโโโโโโผโโโโโโโโโโโผโโโโโโโโผโโโโโโโโโค
# โ.head โf32[4,16] โ64 โ256.00B โ
# โโโโโโโโโโโโโโโโผโโโโโโโโโโโผโโโโโโโโผโโโโโโโโโค
# โฮฃ โTree โ592 โ2.31KB โ
# โโโโโโโโโโโโโโโโดโโโโโโโโโโโดโโโโโโโโดโโโโโโโโโFor more examples and a detailed documentation, check out the API here.
There exist a number of other libraries, which also integrate dataclass functionality with JAX.
The most similar libraries to drinx are jax_dataclass and pytreeclass, though pytreeclass is unfortunately no longer actively maintained.
Other libraries that are not specifically focusing on dataclasses, but have some functionality for them included, are the chex.dataclass, flax.struct or tjax.dataclass.
The main differences between the libraries are:
- Static Fields::
drinx.field(static=True)ordrinx.static_field()can be used to mark dataclass attributes as static. Thepytreeclassandflaxlibrary support a similar system.jax_dataclasssupports static fields through marking an attribute asAnnotated[..., Static].chexandtjaxdo not support static attributes. - Attribute Updates:
drinximplements the.at["attribute"].set()syntax for functional updates of top-level attributes and.aset()for updates of nested structures. The.at[].set()syntax is also supported bypytreeclass, which heavily inspired our implementation.tjaximplements a context manager which allows for updates of frozen classes, but this is non-functional and makes the usage in jit transforms difficult.flaximplements a.replace()function for changing top-level attributes, but not nested updates.chexonly support updates through creating a new object. - Static Type Checking: Both
drinxandjax_dataclassare thin wrappers around the python dataclass and consequently have full type checking support. All other libraries have some limitations with regards for type checking, for examplekw_onlydoes not work intjaxorflax. - Visualization:
drinximplements some nice visualizations throughdrinx.tree_diagramanddrinx.tree_summarywhich is heavily inspired by thepytreeclasslibrary. Other libraries do not implement visualization tools out of the box.
TODO: add citation once published
Also check out my other repositories:
- ๐ก FDTDX - Electromagnetic FDTD Simulations in JAX.
- ๐ฎ BONNI - Bayesian Optimization via Neural Network surrogates and Interior Point Optimization
