2020from pytorch_lightning .callbacks .callback import Callback
2121from pytorch_lightning .callbacks .finetuning import BackboneFinetuning
2222from pytorch_lightning .demos .boring_classes import BoringModel
23+ from pytorch_lightning .loggers import CSVLogger
2324from pytorch_lightning .utilities .exceptions import MisconfigurationException
2425from tests_pytorch .helpers .datamodules import ClassifDataModule
2526from tests_pytorch .helpers .runif import RunIf
@@ -32,7 +33,12 @@ def test_lr_monitor_single_lr(tmpdir):
3233
3334 lr_monitor = LearningRateMonitor ()
3435 trainer = Trainer (
35- default_root_dir = tmpdir , max_epochs = 2 , limit_val_batches = 0.1 , limit_train_batches = 0.5 , callbacks = [lr_monitor ]
36+ default_root_dir = tmpdir ,
37+ max_epochs = 2 ,
38+ limit_val_batches = 0.1 ,
39+ limit_train_batches = 0.5 ,
40+ callbacks = [lr_monitor ],
41+ logger = CSVLogger (tmpdir ),
3642 )
3743 trainer .fit (model )
3844
@@ -70,6 +76,7 @@ def configure_optimizers(self):
7076 limit_train_batches = 5 ,
7177 log_every_n_steps = 1 ,
7278 callbacks = [lr_monitor ],
79+ logger = CSVLogger (tmpdir ),
7380 )
7481 trainer .fit (model )
7582
@@ -96,6 +103,7 @@ def configure_optimizers(self):
96103 limit_train_batches = 5 ,
97104 log_every_n_steps = 1 ,
98105 callbacks = [lr_monitor ],
106+ logger = CSVLogger (tmpdir ),
99107 )
100108 with pytest .warns (RuntimeWarning , match = "optimizers do not have momentum." ):
101109 trainer .fit (model )
@@ -117,7 +125,12 @@ def configure_optimizers(self):
117125
118126 lr_monitor = LearningRateMonitor ()
119127 trainer = Trainer (
120- default_root_dir = tmpdir , max_epochs = 2 , limit_val_batches = 0.1 , limit_train_batches = 0.5 , callbacks = [lr_monitor ]
128+ default_root_dir = tmpdir ,
129+ max_epochs = 2 ,
130+ limit_val_batches = 0.1 ,
131+ limit_train_batches = 0.5 ,
132+ callbacks = [lr_monitor ],
133+ logger = CSVLogger (tmpdir ),
121134 )
122135
123136 trainer .fit (model )
@@ -154,6 +167,7 @@ def configure_optimizers(self):
154167 limit_train_batches = 5 ,
155168 log_every_n_steps = 1 ,
156169 callbacks = [lr_monitor ],
170+ logger = CSVLogger (tmpdir ),
157171 )
158172 trainer .fit (model )
159173
@@ -179,6 +193,7 @@ def configure_optimizers(self):
179193 limit_train_batches = 5 ,
180194 log_every_n_steps = 1 ,
181195 callbacks = [lr_monitor ],
196+ logger = CSVLogger (tmpdir ),
182197 )
183198 with pytest .warns (RuntimeWarning , match = "optimizers do not have momentum." ):
184199 trainer .fit (model )
@@ -226,6 +241,7 @@ def configure_optimizers(self):
226241 limit_train_batches = 7 ,
227242 limit_val_batches = 0.1 ,
228243 callbacks = [lr_monitor ],
244+ logger = CSVLogger (tmpdir ),
229245 )
230246 trainer .fit (model )
231247
@@ -269,6 +285,7 @@ def configure_optimizers(self):
269285 limit_train_batches = 7 ,
270286 limit_val_batches = 0.1 ,
271287 callbacks = [lr_monitor ],
288+ logger = CSVLogger (tmpdir ),
272289 )
273290 trainer .fit (model )
274291
@@ -305,7 +322,12 @@ def configure_optimizers(self):
305322
306323 lr_monitor = LearningRateMonitor ()
307324 trainer = Trainer (
308- default_root_dir = tmpdir , max_epochs = 2 , limit_val_batches = 0.1 , limit_train_batches = 0.5 , callbacks = [lr_monitor ]
325+ default_root_dir = tmpdir ,
326+ max_epochs = 2 ,
327+ limit_val_batches = 0.1 ,
328+ limit_train_batches = 0.5 ,
329+ callbacks = [lr_monitor ],
330+ logger = CSVLogger (tmpdir ),
309331 )
310332 trainer .fit (model , datamodule = dm )
311333
@@ -330,6 +352,7 @@ def configure_optimizers(self):
330352 callbacks = [lr_monitor ],
331353 enable_progress_bar = False ,
332354 enable_model_summary = False ,
355+ logger = CSVLogger (tmpdir ),
333356 )
334357 trainer .fit (TestModel ())
335358 assert list (lr_monitor .lrs ) == ["my_logging_name" ]
@@ -349,6 +372,7 @@ def configure_optimizers(self):
349372 limit_val_batches = 2 ,
350373 limit_train_batches = 2 ,
351374 callbacks = [lr_monitor ],
375+ logger = CSVLogger (tmpdir ),
352376 enable_progress_bar = False ,
353377 enable_model_summary = False ,
354378 )
@@ -384,6 +408,7 @@ def configure_optimizers(self):
384408 limit_val_batches = 2 ,
385409 limit_train_batches = 2 ,
386410 callbacks = [lr_monitor ],
411+ logger = CSVLogger (tmpdir ),
387412 enable_progress_bar = False ,
388413 enable_model_summary = False ,
389414 )
@@ -475,6 +500,7 @@ def finetune_function(self, pl_module, epoch: int, optimizer, opt_idx: int):
475500 limit_val_batches = 0 ,
476501 limit_train_batches = 2 ,
477502 callbacks = [TestFinetuning (), lr_monitor , Check ()],
503+ logger = CSVLogger (tmpdir ),
478504 enable_progress_bar = False ,
479505 enable_model_summary = False ,
480506 enable_checkpointing = False ,
@@ -533,6 +559,7 @@ def configure_optimizers(self):
533559 limit_val_batches = 2 ,
534560 limit_train_batches = 2 ,
535561 callbacks = [lr_monitor ],
562+ logger = CSVLogger (tmpdir ),
536563 enable_progress_bar = False ,
537564 enable_model_summary = False ,
538565 )
0 commit comments