From 6c0fcb723596b046bc85b0fbb667ee4e6782fd82 Mon Sep 17 00:00:00 2001 From: lucas_miranda <lucasmiranda42@gmail.com> Date: Mon, 12 Oct 2020 19:43:53 +0200 Subject: [PATCH] Added support for tensorboard HParams while tuning hyperparameters --- deepof/train_utils.py | 6 ------ 1 file changed, 6 deletions(-) diff --git a/deepof/train_utils.py b/deepof/train_utils.py index b17037f3..d4ec7457 100644 --- a/deepof/train_utils.py +++ b/deepof/train_utils.py @@ -10,7 +10,6 @@ Simple utility functions used in deepof example scripts. These are not part of t from datetime import datetime from kerastuner import BayesianOptimization -from kerastuner import HyperParameters from kerastuner_tensorboard_logger import TensorBoardLogger from typing import Tuple, Union, Any, List import deepof.hypermodels @@ -20,8 +19,6 @@ import os import pickle import tensorflow as tf -hp = HyperParameters() - def load_hparams(hparams): """Loads hyperparameters from a custom dictionary pickled on disc. @@ -164,13 +161,10 @@ def tune_search( predictor=predictor, ) - try: if "ELBO" in loss and kl_wu > 0: callbacks.append(hypermodel.kl_warmup_callback) if "MMD" in loss and mmd_wu > 0: callbacks.append(hypermodel.mmd_warmup_callback) - except AttributeError: - pass else: return False -- GitLab