Upgrade from 1.8 to the 2.0¶
Regular User¶
If |
Then |
Ref |
---|---|---|
used |
set |
#12804 |
used |
call |
#12184 |
imported |
import |
#12740 |
If |
Then |
Ref |
---|---|---|
used Python 3.7 |
upgrade to Python 3.8 or higher |
#16579 |
used PyTorch 1.10 |
upgrade to PyTorch 1.11 or higher |
#16492 |
used Trainer’s flag |
use |
#16171 |
used Trainer’s flag |
use |
#16171 |
used Trainer’s flag |
use |
#16171 |
used Trainer’s flag |
use |
#16171 |
used Trainer’s flag |
pass the path to the |
#10061 |
used Trainer’s flag |
use |
#16184 |
called the |
use Trainer’s flag``devices=”auto”`` |
#16184 |
called the |
use Trainer’s flag``devices=”auto”`` |
#16184 |
used Trainer’s flag |
use the |
#16729 |
imported profiles from |
import from |
#16359 |
used |
move to a standalone |
https://lightning.ai/docs/pytorch/latest/advanced/training_tricks.html#batch-size-finder |
used Trainer’s flag |
use |
|
used Trainer’s flag |
use callbacks |
Advanced User¶
If |
Then |
Ref |
---|---|---|
imported |
import |
#13031 |
imported |
import |
#13043 |
imported |
import |
#13767 |
imported profiler classes from |
import |
#12308 |
used |
use |
#13636 |
used |
use |
|
used the |
switch to |
|
used the Lightning Hydra multi-run integration |
removed support for it as it caused issues with processes hanging |
#15689 |
used |
use |
#9921 |
If |
Then |
Ref |
---|---|---|
used the |
switch to |
#15953 |
used Trainer’s flag |
use DDP with |
#16748 |
implemented |
port your logic to |
#16520 |
implemented |
port your logic to |
#16520 |
implemented |
port your logic to |
#16520 |
used Trainer’s flag |
switch to |
#16800 |
used Trainer’s flag |
implement particular offload logic in your custom metric or turn it on in |
#16358 |
used Trainer’s flag |
overwrite |
#16745 #16745 |
used Trainer’s flag |
use |
|
relied on the |
switch to manual optimization |
#16537 Manual Optimization |
relied on the |
switch to manual optimization |
#16538 Manual Optimization |
were using |
switch to PyTorch native mixed precision |
#16039 Precision |
used Trainer’s flag |
use PyTorch native mixed precision |
#16039 Precision |
used Trainer’s flag |
use PyTorch native mixed precision |
#16039 Precision |
used Trainer’s flag |
use PyTorch native mixed precision |
#16039 Precision |
used Trainer’s attribute |
use PyTorch native mixed precision |
#16039 Precision |
used Trainer’s attribute |
use PyTorch native mixed precision |
#16039 Precision |
used Trainer’s attribute |
use PyTorch native mixed precision |
#16039 Precision |
use the |
consider using PyTorch’s native FSDP implementation or outsourced implementation into own project |
|
used |
use native FSDP instead |
#16400 FSDP |
used |
use native FSDP instead |
#16400 FSDP |
used |
use native FSDP instead |
#16400 FSDP |
used |
use native FSDP instead |
#16400 FSDP |
used |
use native FSDP instead |
#16400 FSDP |
used |
use native FSDP instead |
#16400 FSDP |
used |
pass this option and via dictionary of |
#14998 |
used |
pass this option and via dictionary of |
#14998 |
have customized loops |
implement your training loop with Fabric. |
#14998 Fabric |
have customized loops |
implement your training loop with Fabric. |
#14998 Fabric |
have customized loops |
implement your training loop with Fabric. |
#14998 Fabric |
used the Trainer’s |
implement your training loop with Fabric |
#14998 Fabric |
used the Trainer’s |
implement your training loop with Fabric |
#14998 Fabric |
used the Trainer’s |
implement your training loop with Fabric |
#14998 Fabric |
used the Trainer’s |
implement your training loop with Fabric |
#14998 Fabric |
used the |
being marked as protected |
|
used |
use manual optimization |
#16539 |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used declaring optimizer frequencies in the dictionary returned from |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used |
use manual optimization |
#16539 Manual Optimization |
used Trainer’s |
use manual optimization |
|
used |
#16355 |
|
used training integration with Horovod |
install standalone package/project |
|
used training integration with ColossalAI |
install standalone package/project |
https://lightning.ai/docs/pytorch/latest/advanced/third_party/colossalai.html |
used |
use Torch’s Quantization directly |
#16750 |
had any logic except reducing the DP outputs in |
port it to |
#16791 |
had any logic except reducing the DP outputs in |
port it to |
#16791 |
had any logic except reducing the DP outputs in |
port it to |
#16791 |
used |
switch to general |
#16809 |
used the automatic addition of a moving average of the |
use |
#16192 |
rely on the |
access them via |
#16655 |
need to pass a dictionary to |
pass them independently. |
#16389 |
Developer¶
If |
Then |
Ref |
---|---|---|
derived from |
derive from |
#12014 |
derived from |
derive from |
#12150 |
derived from |
derive from |
#12106 |
If |
Then |
Ref |
---|---|---|
passed the |
passed the (required) |
#16386 |
used |
use DDP or DeepSpeed instead |
#16748 DDP |
used |
use DDP or DeepSpeed instead |
#16386 DDP |
called |
use DDP or DeepSpeed instead |
#16386 DDP |
used or derived from |
use DDP instead |
#16386 DDP |
used the pl.plugins.ApexMixedPrecisionPlugin`` plugin |
use PyTorch native mixed precision |
#16039 |
used the |
switch to the |
#16039 |
used the |
implement your training loop with Fabric |
#16803 |
used the |
implement your training loop with Fabric |
#16803 |
used the |
check the same using |
#16703 |
used any function from |
switch to |
#14514 #14550 |
imported functions from |
import them from |
#14492 #14753 |
imported functions from |
import them from |
#14515 |
imported functions from |
import them from |
#14516 #14537 |
used any code from |
use the base classes |
#16424 |
used any code from |
rely on Pytorch’s native functions |
#16390 |
used any code from |
it was removed |
#16440 |
used any code from |
it was removed |
#16439 |
used any code from |
it was removed |
#16422 |
were using truncated backpropagation through time (TBPTT) with |
use manual optimization |
#16172 Manual Optimization |
were using truncated backpropagation through time (TBPTT) with |
use manual optimization |
#16172 Manual Optimization |
were using truncated backpropagation through time (TBPTT) and passing |
use manual optimization |
#16172 Manual Optimization |
used |
it was removed |
|
used |
it was removed |
|
used |
it was removed |
|
used |
it was removed |
|
used |
it was removed |
|
used |
it was removed |
#16437 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
used |
switch to using |
#16708 |
derived from |
switch to PyTorch native equivalent |
#15364 |
used |
customize your logger |
#16204 |
if you derived from mixin’s method |
rely on |
#16999 |
used |
switch to |
#16436 |
used |
implement own logic with Fabric |
#16516 #16533 |
used or derived from public |
it is set as protected |
#16826 |
used the |
use manual optimization |
#16726 Manual Optimization |
used the |
use manual optimization |
#16726 Manual Optimization |
used the |
use manual optimization |
#16726 Manual Optimization |
used |
use |
#16726 |
used |
rely on Trainer precision attribute |
#16203 |
used |
you shall pass the |
#16462 |
relied on |
pass dataloders directly |
#16714 |
relied on |
pass dataloders directly |
#16714 |
accessed |
rely on Trainer internal loops’ properties |
#16760 |
accessed |
rely on Trainer internal loops’ properties |
#16760 |
accessed |
rely on Trainer internal loops’ properties |
#16760 |
accessed |
rely on Trainer internal loops’ properties |
#16760 |
used |
rely on precision plugin |
#16759 |
used |
it was removed |
#16618 |
used |
it was removed |
#16618 |