bacabaebaf
* Added abstract precision plugin to expose clip_gradients function, use within accelerator to clip gradients * Exclude model from override, keep optimizer (needed for sharded clip gradients), add override for O2 support apex * Fix doc * Applied codereview changes * Refactored clip function to encapsulate tpu changes with tpu accelerator. Default to standard clip function for vanilla torch * Pass correct grad clip val * Moved var to property * Apply code review suggestions |
||
---|---|---|
.. | ||
__init__.py | ||
apex.py | ||
ddp_plugin.py | ||
native_amp.py | ||
plugin_connector.py | ||
precision_plugin.py |