You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Mar 11, 2026. It is now read-only.
Describe the feature and the current behavior/state.
Hi, I think it's good if someone can support Gradient Accumulate optimizer for this repo, this feature is really helpful for those who train the large model with a low resource such as Bert, etc. The usage should be similar with tfa.optimizer.SWA:
There is an implementation of gradient accumulator but for custom training loop rather than Keras model fit here link.
Relevant information
Are you willing to contribute it (yes/no): no
Are you willing to maintain it going forward? (yes/no): no
Is there a relevant academic paper? (if so, where):
Is there already an implementation in another framework? (if so, where): here but for custom training loop.
Was it part of tf.contrib? (if so, where): no
Which API type would this fall under (layer, metric, optimizer, etc.)
optimizer Who will benefit with this feature?
all tensorflow users. Any other info.
Describe the feature and the current behavior/state.
Hi, I think it's good if someone can support Gradient Accumulate optimizer for this repo, this feature is really helpful for those who train the large model with a low resource such as Bert, etc. The usage should be similar with
tfa.optimizer.SWA:There is an implementation of gradient accumulator but for
custom training looprather than Keras model fit here link.Relevant information
custom training loop.Which API type would this fall under (layer, metric, optimizer, etc.)
optimizer
Who will benefit with this feature?
all tensorflow users.
Any other info.