python_code
stringlengths
0
869k
import os HPARAMS_REGISTRY = {} DEFAULT_OUT_DIR = os.path.expandvars('$HOME/dist-aug') class Hyperparams(dict): def __getattr__(self, attr): try: return self[attr] except KeyError: return None def __setattr__(self, attr, value): self[attr] = value good_base...
from mpi4py import MPI import numpy as np import tensorflow as tf import blocksparse as bs from blocksparse import nccl def mpi_init(initializer): 'Variable initializer for MPI. Used such that allreduce ' 'syncs variables at the beginning of training. ' 'This is better than multiplying the values by 0, wh...
import argparse import subprocess import shlex import os import glob import json import time from matplotlib import pyplot as plt import numpy as np LOGDIR = os.path.expanduser('~/bigtrans_logs') GRAPHDIR = os.path.expanduser('~/bigtrans_graphs') BUCKET = '<input bucket>' os.makedirs(LOGDIR, exist_ok=True) os.maked...
import pickle import os import numpy as np import imageio try: from sklearn.cross_validation import train_test_split except ModuleNotFoundError: from sklearn.model_selection import train_test_split from mpi_utils import mpi_size, mpi_rank from janky_stuff import JankySubsampler mpisize = mpi_size() mpirank = ...
import os import itertools import json import tempfile import numpy as np import tensorflow as tf import blocksparse as bs import time import subprocess from mpi_utils import mpi_rank def logger(log_prefix): 'Prints the arguments out to stdout, .txt, and .jsonl files' jsonl_path = f'{log_prefix}.jsonl' t...
# This page is copied from the TensorFlow source code. # https://github.com/tensorflow/tpu/blob/master/models/official/efficientnet/autoaugment.py # Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in comp...
# the base model is the optimized version of the Sparse Transformer # presented at https://arxiv.org/abs/1904.10509 # if hacking on the model, be sure to use the mpi init functions # (random_or_zeros_init, constant_or_zeros_init, etc) or # else the models wont be synced across ranks from collections import namedtuple i...
''' Optimizers should take the arguments grads, variables, learning_rate, grad_scale, max_grad_norm, and **kwargs. ''' from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf import blocksparse as bs from mpi_utils import mpi_rank def ...
import numpy as np class JankySampler: def __init__(self, arr, seed=None): self.arr = arr self.nprng = np.random.RandomState(seed) self.reset() def reset(self): self.drawn = 0 self.idx = self.nprng.permutation(len(self.arr)) def draw(self, n): ''' ...
import re from setuptools import setup, find_packages import sys if sys.version_info.major != 3: print('This Python is only compatible with Python 3, but you are running ' 'Python {}. The installation will likely fail.'.format(sys.version_info.major)) extras = { 'test': [ 'filelock', ...
import numpy as np import matplotlib matplotlib.use('TkAgg') # Can change to 'Agg' for non-interactive mode import matplotlib.pyplot as plt plt.rcParams['svg.fonttype'] = 'none' from baselines.common import plot_util X_TIMESTEPS = 'timesteps' X_EPISODES = 'episodes' X_WALLTIME = 'walltime_hrs' Y_REWARD = 'reward' Y_...
import sys import re import multiprocessing import os.path as osp import gym from collections import defaultdict import tensorflow as tf import numpy as np from baselines.common.vec_env import VecFrameStack, VecNormalize, VecEnv from baselines.common.vec_env.vec_video_recorder import VecVideoRecorder from baselines.co...
import os import sys import shutil import os.path as osp import json import time import datetime import tempfile from collections import defaultdict from contextlib import contextmanager DEBUG = 10 INFO = 20 WARN = 30 ERROR = 40 DISABLED = 50 class KVWriter(object): def writekvs(self, kvs): raise NotImpl...
''' This code is used to evalaute the imitators trained with different number of trajectories and plot the results in the same figure for easy comparison. ''' import argparse import os import glob import gym import matplotlib.pyplot as plt import numpy as np import tensorflow as tf from baselines.gail import run_muj...
''' Reference: https://github.com/openai/imitation I follow the architecture from the official repository ''' import tensorflow as tf import numpy as np from baselines.common.mpi_running_mean_std import RunningMeanStd from baselines.common import tf_util as U def logsigmoid(a): '''Equivalent to tf.log(tf.sigmoid(...
''' The code is used to train BC imitator, or pretrained GAIL imitator ''' import argparse import tempfile import os.path as osp import gym import logging from tqdm import tqdm import tensorflow as tf from baselines.gail import mlp_policy from baselines import bench from baselines import logger from baselines.common...
''' Disclaimer: this code is highly based on trpo_mpi at @openai/baselines and @openai/imitation ''' import argparse import os.path as osp import logging from mpi4py import MPI from tqdm import tqdm import numpy as np import gym from baselines.gail import mlp_policy from baselines.common import set_global_seeds, tf_...
''' Disclaimer: The trpo part highly rely on trpo_mpi at @openai/baselines ''' import time import os from contextlib import contextmanager from mpi4py import MPI from collections import deque import tensorflow as tf import numpy as np import baselines.common.tf_util as U from baselines.common import explained_varian...
''' This code is highly based on https://github.com/carpedm20/deep-rl-tensorflow/blob/master/agents/statistic.py ''' import tensorflow as tf import numpy as np import baselines.common.tf_util as U class stats(): def __init__(self, scalar_keys=[], histogram_keys=[]): self.scalar_keys = scalar_keys ...
''' from baselines/ppo1/mlp_policy.py and add simple modification (1) add reuse argument (2) cache the `stochastic` placeholder ''' import tensorflow as tf import gym import baselines.common.tf_util as U from baselines.common.mpi_running_mean_std import RunningMeanStd from baselines.common.distributions import make_pd...
''' Data structure of the input .npz: the data is save in python dictionary format with keys: 'acs', 'ep_rets', 'rews', 'obs' the values of each item is a list storing the expert trajectory sequentially a transition can be: (data['obs'][t], data['acs'][t], data['obs'][t+1]) and get reward data['rews'][t] ''' from base...
from .monitor import Monitor import gym import json def test_monitor(): import pandas import os import uuid env = gym.make("CartPole-v1") env.seed(0) mon_file = "/tmp/baselines-test-%s.monitor.csv" % uuid.uuid4() menv = Monitor(env, mon_file) menv.reset() for _ in range(1000): ...
__all__ = ['Monitor', 'get_monitor_files', 'load_results'] from gym.core import Wrapper import time from glob import glob import csv import os.path as osp import json class Monitor(Wrapper): EXT = "monitor.csv" f = None def __init__(self, env, filename, allow_early_resets=False, reset_keywords=(), info_k...
# flake8: noqa F403 from baselines.bench.benchmarks import * from baselines.bench.monitor import *
import re import os SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__)) _atari7 = ['BeamRider', 'Breakout', 'Enduro', 'Pong', 'Qbert', 'Seaquest', 'SpaceInvaders'] _atariexpl7 = ['Freeway', 'Gravitar', 'MontezumaRevenge', 'Pitfall', 'PrivateEye', 'Solaris', 'Venture'] _BENCHMARKS = [] remove_version_re = re.comp...
#!/usr/bin/env python3 import os from baselines.common.cmd_util import make_mujoco_env, mujoco_arg_parser from baselines.common import tf_util as U from baselines import logger import gym def train(num_timesteps, seed, model_path=None): env_id = 'Humanoid-v2' from baselines.ppo1 import mlp_policy, pposgd_simp...
import baselines.common.tf_util as U import tensorflow as tf import gym from baselines.common.distributions import make_pdtype class CnnPolicy(object): recurrent = False def __init__(self, name, ob_space, ac_space, kind='large'): with tf.variable_scope(name): self._init(ob_space, ac_space, ...
#!/usr/bin/env python3 from mpi4py import MPI from baselines.common import set_global_seeds from baselines import logger from baselines.common.cmd_util import make_robotics_env, robotics_arg_parser import mujoco_py def train(env_id, num_timesteps, seed): from baselines.ppo1 import mlp_policy, pposgd_simple i...
from baselines.common import Dataset, explained_variance, fmt_row, zipsame from baselines import logger import baselines.common.tf_util as U import tensorflow as tf, numpy as np import time from baselines.common.mpi_adam import MpiAdam from baselines.common.mpi_moments import mpi_moments from mpi4py import MPI from col...
#!/usr/bin/env python3 from baselines.common.cmd_util import make_mujoco_env, mujoco_arg_parser from baselines.common import tf_util as U from baselines import logger def train(env_id, num_timesteps, seed): from baselines.ppo1 import mlp_policy, pposgd_simple U.make_session(num_cpu=1).__enter__() def poli...
from baselines.common.mpi_running_mean_std import RunningMeanStd import baselines.common.tf_util as U import tensorflow as tf import gym from baselines.common.distributions import make_pdtype class MlpPolicy(object): recurrent = False def __init__(self, name, *args, **kwargs): with tf.variable_scope(na...
#!/usr/bin/env python3 from mpi4py import MPI from baselines.common import set_global_seeds from baselines import bench import os.path as osp from baselines import logger from baselines.common.atari_wrappers import make_atari, wrap_deepmind from baselines.common.cmd_util import atari_arg_parser def train(env_id, num_...
import time import functools import numpy as np import tensorflow as tf from baselines import logger from baselines.common import set_global_seeds from baselines.common.policies import build_policy from baselines.common.tf_util import get_session, save_variables, load_variables from baselines.common.vec_env.vec_frame_...
import numpy as np from baselines.common.runners import AbstractEnvRunner from baselines.common.vec_env.vec_frame_stack import VecFrameStack from gym import spaces class Runner(AbstractEnvRunner): def __init__(self, env, model, nsteps): super().__init__(env=env, model=model, nsteps=nsteps) assert...
import numpy as np class Buffer(object): # gets obs, actions, rewards, mu's, (states, masks), dones def __init__(self, env, nsteps, size=50000): self.nenv = env.num_envs self.nsteps = nsteps # self.nh, self.nw, self.nc = env.observation_space.shape self.obs_shape = env.observati...
def atari(): return dict( lrschedule='constant' )
import numpy as np import tensorflow as tf from baselines.common.policies import nature_cnn from baselines.a2c.utils import fc, batch_to_seq, seq_to_batch, lstm, sample class AcerCnnPolicy(object): def __init__(self, sess, ob_space, ac_space, nenv, nsteps, nstack, reuse=False): nbatch = nenv * nsteps ...
import tensorflow as tf from baselines.common.models import get_network_builder class Model(object): def __init__(self, name, network='mlp', **network_kwargs): self.name = name self.network_builder = get_network_builder(network)(**network_kwargs) @property def vars(self): return t...
import numpy as np class RingBuffer(object): def __init__(self, maxlen, shape, dtype='float32'): self.maxlen = maxlen self.start = 0 self.length = 0 self.data = np.zeros((maxlen,) + shape).astype(dtype) def __len__(self): return self.length def __getitem__(self, i...
from copy import copy from functools import reduce import numpy as np import tensorflow as tf import tensorflow.contrib as tc from baselines import logger from baselines.common.mpi_adam import MpiAdam import baselines.common.tf_util as U from baselines.common.mpi_running_mean_std import RunningMeanStd try: from m...
from baselines.common.tests.util import smoketest def _run(argstr): smoketest('--alg=ddpg --env=Pendulum-v0 --num_timesteps=0 ' + argstr) def test_popart(): _run('--normalize_returns=True --popart=True') def test_noise_normal(): _run('--noise_type=normal_0.1') def test_noise_ou(): _run('--noise_type=...
import numpy as np class AdaptiveParamNoiseSpec(object): def __init__(self, initial_stddev=0.1, desired_action_stddev=0.1, adoption_coefficient=1.01): self.initial_stddev = initial_stddev self.desired_action_stddev = desired_action_stddev self.adoption_coefficient = adoption_coefficient ...
import os import time from collections import deque import pickle from baselines.ddpg.ddpg_learner import DDPG from baselines.ddpg.models import Actor, Critic from baselines.ddpg.memory import Memory from baselines.ddpg.noise import AdaptiveParamNoiseSpec, NormalActionNoise, OrnsteinUhlenbeckActionNoise from baselines...
import numpy as np def tile_images(img_nhwc): """ Tile N images into one big PxQ image (P,Q) are chosen to be as close as possible, and if N is square, then P=Q. input: img_nhwc, list or array of images, ndim=4 once turned into array n = batch index, h = height, w = width, c = channel ...
from mpi4py import MPI import numpy as np from baselines.common import zipsame def mpi_mean(x, axis=0, comm=None, keepdims=False): x = np.asarray(x) assert x.ndim > 0 if comm is None: comm = MPI.COMM_WORLD xsum = x.sum(axis=axis, keepdims=keepdims) n = xsum.size localsum = np.zeros(n+1, x.dtyp...
import numpy as np def cg(f_Ax, b, cg_iters=10, callback=None, verbose=False, residual_tol=1e-10): """ Demmel p 312 """ p = b.copy() r = b.copy() x = np.zeros_like(b) rdotr = r.dot(r) fmtstr = "%10i %10.3g %10.3g" titlestr = "%10s %10s %10s" if verbose: print(titlestr % ("iter...
import numpy as np import tensorflow as tf from baselines.a2c import utils from baselines.a2c.utils import conv, fc, conv_to_fc, batch_to_seq, seq_to_batch from baselines.common.mpi_running_mean_std import RunningMeanStd mapping = {} def register(name): def _thunk(func): mapping[name] = func retur...
import gym import numpy as np import os import pickle import random import tempfile import zipfile def zipsame(*seqs): L = len(seqs[0]) assert all(len(seq) == L for seq in seqs[1:]) return zip(*seqs) class EzPickle(object): """Objects that are pickled and unpickled via their constructor argument...
""" Helpers for scripts like run_atari.py. """ import os try: from mpi4py import MPI except ImportError: MPI = None import gym from gym.wrappers import FlattenObservation, FilterObservation from baselines import logger from baselines.bench import Monitor from baselines.common import set_global_seeds from base...
import numpy as np import os os.environ.setdefault('PATH', '') from collections import deque import gym from gym import spaces import cv2 cv2.ocl.setUseOpenCL(False) from .wrappers import TimeLimit class NoopResetEnv(gym.Wrapper): def __init__(self, env, noop_max=30): """Sample initial states by taking ra...
import baselines.common.tf_util as U import tensorflow as tf import numpy as np try: from mpi4py import MPI except ImportError: MPI = None class MpiAdam(object): def __init__(self, var_list, *, beta1=0.9, beta2=0.999, epsilon=1e-08, scale_grad_by_procs=True, comm=None): self.var_list = var_list ...
# flake8: noqa F403 from baselines.common.console_util import * from baselines.common.dataset import Dataset from baselines.common.math_util import * from baselines.common.misc_util import *
from __future__ import print_function from contextlib import contextmanager import numpy as np import time import shlex import subprocess # ================================================================ # Misc # ================================================================ def fmt_row(width, row, header=False): ...
import numpy as np import tensorflow as tf # pylint: ignore-module import copy import os import functools import collections import multiprocessing def switch(condition, then_expression, else_expression): """Switches between two operations depending on a scalar value (int or bool). Note that both `then_expres...
import numpy as np class Dataset(object): def __init__(self, data_map, deterministic=False, shuffle=True): self.data_map = data_map self.deterministic = deterministic self.enable_shuffle = shuffle self.n = next(iter(data_map.values())).shape[0] self._next_id = 0 self...
import tensorflow as tf import numpy as np import baselines.common.tf_util as U from baselines.a2c.utils import fc from tensorflow.python.ops import math_ops class Pd(object): """ A particular probability distribution """ def flatparam(self): raise NotImplementedError def mode(self): ...
import gym class TimeLimit(gym.Wrapper): def __init__(self, env, max_episode_steps=None): super(TimeLimit, self).__init__(env) self._max_episode_steps = max_episode_steps self._elapsed_steps = 0 def step(self, ac): observation, reward, done, info = self.env.step(ac) sel...
import operator class SegmentTree(object): def __init__(self, capacity, operation, neutral_element): """Build a Segment Tree data structure. https://en.wikipedia.org/wiki/Segment_tree Can be used as regular array, but with two important differences: a) setting item's...
import tensorflow as tf import numpy as np from baselines.common.tf_util import get_session class RunningMeanStd(object): # https://en.wikipedia.org/wiki/Algorithms_for_calculating_variance#Parallel_algorithm def __init__(self, epsilon=1e-4, shape=()): self.mean = np.zeros(shape, 'float64') sel...
import numpy as np import tensorflow as tf from gym.spaces import Discrete, Box, MultiDiscrete def observation_placeholder(ob_space, batch_size=None, name='Ob'): ''' Create placeholder to feed observations into of the size appropriate to the observation space Parameters: ---------- ob_space: gym....
import os, subprocess, sys def mpi_fork(n, bind_to_core=False): """Re-launches the current script with workers Returns "parent" for original parent, "child" for MPI children """ if n<=1: return "child" if os.getenv("IN_MPI") is None: env = os.environ.copy() env.update( ...
import numpy as np from abc import ABC, abstractmethod class AbstractEnvRunner(ABC): def __init__(self, *, env, model, nsteps): self.env = env self.model = model self.nenv = nenv = env.num_envs if hasattr(env, 'num_envs') else 1 self.batch_ob_shape = (nenv*nsteps,) + env.observation...
from collections import deque import cv2 cv2.ocl.setUseOpenCL(False) from .atari_wrappers import WarpFrame, ClipRewardEnv, FrameStack, ScaledFloatFrame from .wrappers import TimeLimit import numpy as np import gym class StochasticFrameSkip(gym.Wrapper): def __init__(self, env, n, stickprob): gym.Wrapper._...
import numpy as np import tensorflow as tf from baselines.common import tf_util as U from baselines.common.tests.test_with_mpi import with_mpi from baselines import logger try: from mpi4py import MPI except ImportError: MPI = None class MpiAdamOptimizer(tf.train.AdamOptimizer): """Adam optimizer that avera...
from collections import defaultdict import os, numpy as np import platform import shutil import subprocess import warnings import sys try: from mpi4py import MPI except ImportError: MPI = None def sync_from_root(sess, variables, comm=None): """ Send the root node's parameters to every worker. Arg...
try: from mpi4py import MPI except ImportError: MPI = None import tensorflow as tf, baselines.common.tf_util as U, numpy as np class RunningMeanStd(object): # https://en.wikipedia.org/wiki/Algorithms_for_calculating_variance#Parallel_algorithm def __init__(self, epsilon=1e-2, shape=()): self....
import numpy as np import scipy.signal def discount(x, gamma): """ computes discounted sums along 0th dimension of x. inputs ------ x: ndarray gamma: float outputs ------- y: ndarray with same shape as x, satisfying y[t] = x[t] + gamma*x[t+1] + gamma^2*x[t+2] + ... + gam...
import matplotlib.pyplot as plt import os.path as osp import json import os import numpy as np import pandas from collections import defaultdict, namedtuple from baselines.bench import monitor from baselines.logger import read_json, read_csv def smooth(y, radius, mode='two_sided', valid_only=False): ''' Smooth...
import tensorflow as tf from baselines.common import tf_util from baselines.a2c.utils import fc from baselines.common.distributions import make_pdtype from baselines.common.input import observation_placeholder, encode_observation from baselines.common.tf_util import adjust_shape from baselines.common.mpi_running_mean_s...
from baselines.common import mpi_util from baselines import logger from baselines.common.tests.test_with_mpi import with_mpi try: from mpi4py import MPI except ImportError: MPI = None @with_mpi() def test_mpi_weighted_mean(): comm = MPI.COMM_WORLD with logger.scoped_configure(comm=comm): if com...
"""This file is used for specifying various schedules that evolve over time throughout the execution of the algorithm, such as: - learning rate for the optimizer - exploration epsilon for the epsilon greedy exploration strategy - beta parameter for beta parameter in prioritized replay Each schedule has a function `...
import pytest try: import mujoco_py _mujoco_present = True except BaseException: mujoco_py = None _mujoco_present = False @pytest.mark.skipif( not _mujoco_present, reason='error loading mujoco - either mujoco / mujoco key not present, or LD_LIBRARY_PATH is not pointing to mujoco library' ) def...
import numpy as np from baselines.common.schedules import ConstantSchedule, PiecewiseSchedule def test_piecewise_schedule(): ps = PiecewiseSchedule([(-5, 100), (5, 200), (10, 50), (100, 50), (200, -50)], outside_value=500) assert np.isclose(ps.value(-10), 500) assert np.isclose(ps.value(0), 150) ass...
import tensorflow as tf import numpy as np from baselines.common.vec_env.dummy_vec_env import DummyVecEnv N_TRIALS = 10000 N_EPISODES = 100 _sess_config = tf.ConfigProto( allow_soft_placement=True, intra_op_parallelism_threads=1, inter_op_parallelism_threads=1 ) def simple_test(env_fn, learn_fn, min_rewa...
import pytest import gym import tensorflow as tf from baselines.common.vec_env.subproc_vec_env import SubprocVecEnv from baselines.run import get_learn_function from baselines.common.tf_util import make_session algos = ['a2c', 'acer', 'acktr', 'deepq', 'ppo2', 'trpo_mpi'] @pytest.mark.parametrize('algo', algos) def ...
# tests for tf_util import tensorflow as tf from baselines.common.tf_util import ( function, initialize, single_threaded_session ) def test_function(): with tf.Graph().as_default(): x = tf.placeholder(tf.int32, (), name="x") y = tf.placeholder(tf.int32, (), name="y") z = 3 * x ...
import pytest from baselines.common.tests.envs.fixed_sequence_env import FixedSequenceEnv from baselines.common.tests.util import simple_test from baselines.run import get_learn_function from baselines.common.tests import mark_slow common_kwargs = dict( seed=0, total_timesteps=50000, ) learn_kwargs = { ...
import os, pytest mark_slow = pytest.mark.skipif(not os.getenv('RUNSLOW'), reason='slow')
# smoke tests of plot_util from baselines.common import plot_util as pu from baselines.common.tests.util import smoketest def test_plot_util(): nruns = 4 logdirs = [smoketest('--alg=ppo2 --env=CartPole-v0 --num_timesteps=10000') for _ in range(nruns)] data = pu.load_results(logdirs) assert len(data) =...
import pytest # from baselines.acer import acer_simple as acer from baselines.common.tests.envs.mnist_env import MnistEnv from baselines.common.tests.util import simple_test from baselines.run import get_learn_function from baselines.common.tests import mark_slow # TODO investigate a2c and ppo2 failures - is it due t...
import numpy as np from baselines.common.segment_tree import SumSegmentTree, MinSegmentTree def test_tree_set(): tree = SumSegmentTree(4) tree[2] = 1.0 tree[3] = 3.0 assert np.isclose(tree.sum(), 4.0) assert np.isclose(tree.sum(0, 2), 0.0) assert np.isclose(tree.sum(0, 3), 1.0) assert n...
import pytest import gym from baselines.run import get_learn_function from baselines.common.tests.util import reward_per_episode_test from baselines.common.tests import mark_slow pytest.importorskip('mujoco_py') common_kwargs = dict( network='mlp', seed=0, ) learn_kwargs = { 'her': dict(total_timesteps=...
import pytest from baselines.common.tests.envs.identity_env import DiscreteIdentityEnv, BoxIdentityEnv, MultiDiscreteIdentityEnv from baselines.run import get_learn_function from baselines.common.tests.util import simple_test from baselines.common.tests import mark_slow common_kwargs = dict( total_timesteps=30000,...
import pytest import gym from baselines.run import get_learn_function from baselines.common.tests.util import reward_per_episode_test from baselines.common.tests import mark_slow common_kwargs = dict( total_timesteps=30000, network='mlp', gamma=1.0, seed=0, ) learn_kwargs = { 'a2c' : dict(nsteps=...
import os import sys import subprocess import cloudpickle import base64 import pytest from functools import wraps try: from mpi4py import MPI except ImportError: MPI = None def with_mpi(nproc=2, timeout=30, skip_if_no_mpi=True): def outer_thunk(fn): @wraps(fn) def thunk(*args, **kwargs): ...
import os import gym import tempfile import pytest import tensorflow as tf import numpy as np from baselines.common.tests.envs.mnist_env import MnistEnv from baselines.common.vec_env.dummy_vec_env import DummyVecEnv from baselines.run import get_learn_function from baselines.common.tf_util import make_session, get_ses...
from baselines.common.tests.envs.identity_env import DiscreteIdentityEnv def test_discrete_nodelay(): nsteps = 100 eplen = 50 env = DiscreteIdentityEnv(10, episode_len=eplen) ob = env.reset() for t in range(nsteps): action = env.action_space.sample() next_ob, rew, done, info = env....
import numpy as np from gym import Env from gym.spaces import Discrete class FixedSequenceEnv(Env): def __init__( self, n_actions=10, episode_len=100 ): self.action_space = Discrete(n_actions) self.observation_space = Discrete(1) self.np_random = np....
import numpy as np from abc import abstractmethod from gym import Env from gym.spaces import MultiDiscrete, Discrete, Box from collections import deque class IdentityEnv(Env): def __init__( self, episode_len=None, delay=0, zero_first_rewards=True ): self...
import os.path as osp import numpy as np import tempfile from gym import Env from gym.spaces import Discrete, Box class MnistEnv(Env): def __init__( self, episode_len=None, no_images=None ): import filelock from tensorflow.examples.tutorials.mnist import in...
""" Tests for asynchronous vectorized environments. """ import gym import numpy as np import pytest from .dummy_vec_env import DummyVecEnv from .shmem_vec_env import ShmemVecEnv from .subproc_vec_env import SubprocVecEnv from baselines.common.tests.test_with_mpi import with_mpi def assert_venvs_equal(venv1, venv2, n...
""" Tests for asynchronous vectorized environments. """ import gym import pytest import os import glob import tempfile from .dummy_vec_env import DummyVecEnv from .shmem_vec_env import ShmemVecEnv from .subproc_vec_env import SubprocVecEnv from .vec_video_recorder import VecVideoRecorder @pytest.mark.parametrize('kl...
""" An interface for asynchronous vectorized environments. """ import multiprocessing as mp import numpy as np from .vec_env import VecEnv, CloudpickleWrapper, clear_mpi_env_vars import ctypes from baselines import logger from .util import dict_to_obs, obs_space_info, obs_to_dict _NP_TO_CT = {np.float32: ctypes.c_fl...