diff --git a/numpy_ml/neural_nets/models/w2v.py b/numpy_ml/neural_nets/models/w2v.py index 7cfbef1..55829d0 100644 --- a/numpy_ml/neural_nets/models/w2v.py +++ b/numpy_ml/neural_nets/models/w2v.py @@ -295,6 +295,7 @@ def _build_noise_distribution(self): def _train_epoch(self, corpus_fps, encoding): total_loss = 0 + smooth_loss = 0 batch_generator = self.minibatcher(corpus_fps, encoding) for ix, (X, target) in enumerate(batch_generator): loss = self._train_batch(X, target) diff --git a/numpy_ml/rl_models/trainer.py b/numpy_ml/rl_models/trainer.py index f2bdede..2d47aba 100644 --- a/numpy_ml/rl_models/trainer.py +++ b/numpy_ml/rl_models/trainer.py @@ -69,6 +69,7 @@ def train( t0 = time() render_every = np.inf if render_every is None else render_every sf = smooth_factor + smooth_tot = 0 for ep in range(n_episodes): tot_rwd, duration, n_steps = self._train_episode(max_steps)