Skip to content

AdamWFP8 support#699

Draft
saluzafa wants to merge 4 commits intoostris:mainfrom
saluzafa:main
Draft

AdamWFP8 support#699
saluzafa wants to merge 4 commits intoostris:mainfrom
saluzafa:main

Conversation

@saluzafa
Copy link

@saluzafa saluzafa commented Feb 5, 2026

Description

This PR adds support for AdamWFP8, the goal was to benchmark this optimizer for Z-Image Base LoRA trainings.
The optimizer Python script come from https://github.com/None9527/None_Z-image-Turbo_trainer/blob/omni/src/zimage_trainer/optimizers/adamw_fp8.py 🙂

Implement AdamW optimizer with FP8 state storage using PyTorch's native float8 format for improved precision and memory efficiency.
@troycjj
Copy link

troycjj commented Feb 9, 2026

The loss is still around 0.4, and nothing has improved.

@saluzafa saluzafa marked this pull request as draft February 10, 2026 07:52
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants