diff --git a/README.md b/README.md index 429839f2..b258e414 100644 --- a/README.md +++ b/README.md @@ -10,6 +10,8 @@ [![Donate](https://img.shields.io/badge/donate-Liberapay-yellow)](https://liberapay.com/TensorforceTeam/donate) +**This project is not maintained any longer!** + #### Introduction diff --git a/docs/requirements.txt b/docs/requirements.txt index 31627ba2..45ed003a 100644 --- a/docs/requirements.txt +++ b/docs/requirements.txt @@ -1,5 +1,5 @@ m2r -mistune == 0.8.4 # temporary +mistune == 2.0.3 # temporary msgpack msgpack-numpy recommonmark diff --git a/requirements-all.txt b/requirements-all.txt index 5b74667e..bf7b0e3b 100644 --- a/requirements-all.txt +++ b/requirements-all.txt @@ -1,11 +1,11 @@ -gym >= 0.21.0 +gym >= 0.21.0, <0.23 h5py >= 3.6.0 matplotlib >= 3.5.1 msgpack >= 1.0.3 msgpack-numpy >= 0.4.7.1 numpy ~= 1.21.5 Pillow >= 9.0.0 -tensorflow == 2.8.0 +tensorflow == 2.12.1 tqdm >= 4.62.3 tensorflow-addons >= 0.15.0 hpbandster >= 0.7.4 diff --git a/requirements.txt b/requirements.txt index 5e53df18..1d9ed7f1 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,9 +1,9 @@ -gym >= 0.21.0 +gym >= 0.21.0, <0.23 h5py >= 3.6.0 matplotlib >= 3.5.1 msgpack >= 1.0.3 msgpack-numpy >= 0.4.7.1 numpy ~= 1.21.5 Pillow >= 9.0.0 -tensorflow == 2.8.0 +tensorflow == 2.12.1 tqdm >= 4.62.3 diff --git a/tensorforce/core/optimizers/tf_optimizer.py b/tensorforce/core/optimizers/tf_optimizer.py index ecd48d52..a0568149 100755 --- a/tensorforce/core/optimizers/tf_optimizer.py +++ b/tensorforce/core/optimizers/tf_optimizer.py @@ -22,14 +22,14 @@ tensorflow_optimizers = dict( - adadelta=tf.keras.optimizers.Adadelta, - adagrad=tf.keras.optimizers.Adagrad, - adam=tf.keras.optimizers.Adam, - adamax=tf.keras.optimizers.Adamax, - ftrl=tf.keras.optimizers.Ftrl, - nadam=tf.keras.optimizers.Nadam, - rmsprop=tf.keras.optimizers.RMSprop, - sgd=tf.keras.optimizers.SGD + adadelta=tf.keras.optimizers.legacy.Adadelta, + adagrad=tf.keras.optimizers.legacy.Adagrad, + adam=tf.keras.optimizers.legacy.Adam, + adamax=tf.keras.optimizers.legacy.Adamax, + ftrl=tf.keras.optimizers.legacy.Ftrl, + nadam=tf.keras.optimizers.legacy.Nadam, + rmsprop=tf.keras.optimizers.legacy.RMSprop, + sgd=tf.keras.optimizers.legacy.SGD ) diff --git a/tensorforce/core/utils/dicts.py b/tensorforce/core/utils/dicts.py index dcc95638..7dab1f84 100644 --- a/tensorforce/core/utils/dicts.py +++ b/tensorforce/core/utils/dicts.py @@ -32,10 +32,7 @@ def __init__(self, arg=None, *, value_type=None, overwrite=None, **kwargs): ) def __setattr__(self, name, value): - if name.startswith('_self_'): - super(NestedDict, self).__setattr__(name, value) - else: - super().__setattr__(name, value) + super(NestedDict, self).__setattr__(name, value) def __setitem__(self, key, value): if key is None: diff --git a/tensorforce/environments/openai_gym.py b/tensorforce/environments/openai_gym.py index 575e64ac..898e7d3c 100644 --- a/tensorforce/environments/openai_gym.py +++ b/tensorforce/environments/openai_gym.py @@ -113,7 +113,7 @@ def create_level(cls, level, max_episode_steps, reward_threshold, **kwargs): # Modified specification if requires_register: entry_point = gym.envs.registry.env_specs[level].entry_point - _kwargs = dict(gym.envs.registry.env_specs[level]._kwargs) + _kwargs = dict(gym.envs.registry.env_specs[level].kwargs) nondeterministic = gym.envs.registry.env_specs[level].nondeterministic if '-v' in level and level[level.rindex('-v') + 2:].isdigit(): diff --git a/tensorforce/execution/runner.py b/tensorforce/execution/runner.py index 261aa5d6..15f75eb9 100644 --- a/tensorforce/execution/runner.py +++ b/tensorforce/execution/runner.py @@ -275,6 +275,7 @@ def run( additional information (averaged over number of episodes given via mean_horizon):