Source code for pytorch_lightning.callbacks.device_stats_monitor
# Copyright The PyTorch Lightning team.## Licensed under the Apache License, Version 2.0 (the "License");# you may not use this file except in compliance with the License.# You may obtain a copy of the License at## http://www.apache.org/licenses/LICENSE-2.0## Unless required by applicable law or agreed to in writing, software# distributed under the License is distributed on an "AS IS" BASIS,# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.# See the License for the specific language governing permissions and# limitations under the License."""Device Stats Monitor====================Monitors and logs device stats during training."""fromtypingimportAny,Dict,Optionalimportpytorch_lightningasplfrompytorch_lightning.callbacks.callbackimportCallbackfrompytorch_lightning.utilities.exceptionsimportMisconfigurationExceptionfrompytorch_lightning.utilities.importsimport_PSUTIL_AVAILABLEfrompytorch_lightning.utilities.rank_zeroimportrank_zero_deprecation,rank_zero_warnfrompytorch_lightning.utilities.typesimportSTEP_OUTPUT
[docs]classDeviceStatsMonitor(Callback):r""" Automatically monitors and logs device stats during training stage. ``DeviceStatsMonitor`` is a special callback as it requires a ``logger`` to passed as argument to the ``Trainer``. Args: cpu_stats: if ``None``, it will log CPU stats only if the accelerator is CPU. It will raise a warning if ``psutil`` is not installed till v1.9.0. If ``True``, it will log CPU stats regardless of the accelerator, and it will raise an exception if ``psutil`` is not installed. If ``False``, it will not log CPU stats regardless of the accelerator. Raises: MisconfigurationException: If ``Trainer`` has no logger. Example: >>> from pytorch_lightning import Trainer >>> from pytorch_lightning.callbacks import DeviceStatsMonitor >>> device_stats = DeviceStatsMonitor() # doctest: +SKIP >>> trainer = Trainer(callbacks=[device_stats]) # doctest: +SKIP """def__init__(self,cpu_stats:Optional[bool]=None)->None:self._cpu_stats=cpu_stats
[docs]defsetup(self,trainer:"pl.Trainer",pl_module:"pl.LightningModule",stage:Optional[str]=None,)->None:ifstage!="fit":returnifnottrainer.loggers:raiseMisconfigurationException("Cannot use `DeviceStatsMonitor` callback with `Trainer(logger=False)`.")# warn in setup to warn oncedevice=trainer.strategy.root_deviceifself._cpu_statsisNoneanddevice.type=="cpu"andnot_PSUTIL_AVAILABLE:# TODO: raise an exception from v1.9rank_zero_warn("`DeviceStatsMonitor` will not log CPU stats as `psutil` is not installed."" To install `psutil`, run `pip install psutil`."" It will raise an exception if `psutil` is not installed post v1.9.0.")self._cpu_stats=False
def_get_and_log_device_stats(self,trainer:"pl.Trainer",key:str)->None:ifnottrainer._logger_connector.should_update_logs:returndevice=trainer.strategy.root_deviceifself._cpu_statsisFalseanddevice.type=="cpu":# cpu stats are disabledreturndevice_stats=trainer.accelerator.get_device_stats(device)ifself._cpu_statsanddevice.type!="cpu":# Don't query CPU stats twice if CPU is acceleratorfrompytorch_lightning.accelerators.cpuimportget_cpu_statsdevice_stats.update(get_cpu_stats())forloggerintrainer.loggers:separator=logger.group_separatorprefixed_device_stats=_prefix_metric_keys(device_stats,f"{self.__class__.__qualname__}.{key}",separator)logger.log_metrics(prefixed_device_stats,step=trainer.fit_loop.epoch_loop._batches_that_stepped)
def_prefix_metric_keys(metrics_dict:Dict[str,float],prefix:str,separator:str)->Dict[str,float]:return{prefix+separator+k:vfork,vinmetrics_dict.items()}defprefix_metric_keys(metrics_dict:Dict[str,float],prefix:str)->Dict[str,float]:rank_zero_deprecation("`pytorch_lightning.callbacks.device_stats_monitor.prefix_metrics`"" is deprecated in v1.6 and will be removed in v1.8.")sep=""return_prefix_metric_keys(metrics_dict,prefix,sep)
To analyze traffic and optimize your experience, we serve cookies on this site. By clicking or navigating, you agree to allow our usage of cookies. Read PyTorch Lightning's Privacy Policy.