-
Notifications
You must be signed in to change notification settings - Fork 19.6k
Keras <> NNX integration #21252
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Keras <> NNX integration #21252
Conversation
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## master #21252 +/- ##
==========================================
- Coverage 82.65% 77.57% -5.09%
==========================================
Files 565 565
Lines 54802 55062 +260
Branches 8508 8552 +44
==========================================
- Hits 45297 42714 -2583
- Misses 7413 10337 +2924
+ Partials 2092 2011 -81
Flags with carried forward coverage won't be shown. Click here to find out more. ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
keras/src/random/random_test.py
Outdated
import jax.numpy as jnp | ||
|
||
x = ops.ones(3) | ||
|
||
@jax.jit | ||
@nnx.jit |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Would the integration prevent the use of jax.jit
with Keras layers?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yes! it would only work with nnx.jit for now ( They might be working on adding support for jax.jit)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Added nnx as a opt in with this flag - os.environ["KERAS_NNX_ENABLED"]
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This needs a testing plan
- How do we make sure things work without flax installed on CI?
- How do we make sure things work when NNX is enabled?
- When NNX is not enabled?
We need to make sure we have an automated testing path for these different options we are writing logic for or they will silently break at some point.
Added a github workflow action for nnx backend. Note this will FAIL - because this needs a new release of flax to work. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the PR!
- name: Install Flax for NNX backend | ||
if: matrix.backend == 'nnx' | ||
run: | | ||
pip install flax --progress-bar off --upgrade |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Not needed, it's already part of the requirements file.
@@ -16,7 +16,7 @@ jobs: | |||
fail-fast: false | |||
matrix: | |||
python-version: ['3.10'] | |||
backend: [tensorflow, jax, torch, numpy, openvino] | |||
backend: [tensorflow, jax, torch, numpy, openvino, nnx] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think you need to do something else in addition (or instead of this?), right now, this does not turn on NNX.
elif "mutable" not in nnx_metadata: | ||
nnx_metadata["mutable"] = actual_nnx_mutable | ||
|
||
# Initialize nnx.Variable first. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we need a placeholder value? Does it allocate HBM? The issue is that we may be using twice the memory we need by initializing the variable twice.
else: | ||
_placeholder_value = jnp.array(0.0, dtype=jnp.float32) | ||
|
||
# Call nnx.Variable.__init__ directly. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What about the sharding on the placeholder_value, do we need it?
|
||
return self._maybe_autocast(current_value) | ||
|
||
def __hash__(self): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Per the discussion we had, we can't do that. It's not compatible with the semantics of __eq__
that we have.
I think my preferred fix would be to change NNX to not use variables as keys, but instead use id(variable)
in dictionaries.
The PR integrates NNX into JAX backend!
The following snippet shows how you would enable the nnx backend
Demo colab here : https://colab.sandbox.google.com/drive/1mK-4qbce2HGRIkcb4v5n4niWGDezL_6n#scrollTo=m-ZH9Mpnphfz
Added a github workflow action for nnx backend. Note this will fail - because this needs a new release of flax to work.