Skip to main content
1 vote
1 answer
56 views

I'm creating a custom PyEnvironment in TensorFlow Agents to simulate the track and field decathlon. I've managed to create a functioning environment in the sense that I can use _step and _reset, but ...
Perry's user avatar
  • 31
1 vote
0 answers
158 views

i trying training tf agent in my environment but have problome when running PyDriver.run code is here env spec code self._action_spec = array_spec.BoundedArraySpec( shape=(), dtype=np.float64, ...
변상진's user avatar
1 vote
0 answers
160 views

I'm trying to modify the MountainCarContinuous-v0 environment from suite_gym() because training is getting stuck in a local minima. The default reward function penalizes large actions which are ...
brian_ds's user avatar
  • 377
1 vote
0 answers
163 views

When running the Tensorflow agents notebook for the Soft Actor-Critic Minitaur, https://www.tensorflow.org/agents/tutorials/7_SAC_minitaur_tutorial, the following directories are created under /tmp: +...
Daniel von Eschwege's user avatar
1 vote
1 answer
369 views

I am trying to save a model with tensorflow-agents. First I define the following: collect_policy = tf_agent.collect_policy saver = PolicySaver(collect_policy, batch_size=None) and then save the model ...
Enrique's user avatar
  • 10.2k
0 votes
1 answer
59 views

In the environment tutorial of tensorflow agents (https://www.tensorflow.org/agents/tutorials/2_environments_tutorial), the state is stored as an integer. When the state is required, it is converted ...
user3053216's user avatar
2 votes
1 answer
287 views

I'm trying to experiment with using tf_agents' PPOAgent in the CartPole-v1 environment, but I am recieving the following error upon declaring the agent itself: ValueError: actor_network output spec ...
Old_Frog's user avatar
3 votes
1 answer
1k views

I successfully followed this official tensorflow tutorial for training an agent to solve the 'CartPole-v0' gym environment. I only diverged from the tutorial in that I did not use reverb, because it's ...
Gaberocksall's user avatar
1 vote
1 answer
232 views

I'm working on Multi-Armed-Bandit problem, using LinearUCBAgent and LinearThompsonSamplingAgent but they both return a single action for an observation. What I need is the probability for all the ...
Kushal Jain's user avatar
0 votes
1 answer
701 views

I am building a PPO agent side by side with the TF-Agents DQN tutorial. The idea was checking the basics structures needed for a simple tf-agent to work, and adapting it to a PPO agent. I am also ...
HWerneck's user avatar
0 votes
1 answer
633 views

I am using Google Colab to run ViZDoom combined with TensorFlow (specifically, the TF-Agents library). Most of the times when I start the Colab notebook with my code I get the following error: ...
HWerneck's user avatar
2 votes
1 answer
532 views

I want to model my environment such that each action is made of 3 possible sub-actions. I've defined the _action_spec of my tf_agents.environments.py_environment.PyEnvironment as: self._action_spec = ...
Lostefra's user avatar
  • 370
1 vote
1 answer
715 views

In the tensorflow documentation for TF-Agents Environments there is an example of an environment for a simple (blackjack inspired) card game. The init looks like the following: class CardGameEnv(...
sapo_cosmico's user avatar
  • 6,592
1 vote
1 answer
469 views

I noticed something weird happening when converting a Python environment into a TF environment using tf_agents.environments.TFPyEnvironment and I'd like to ask you what general changes occur. To ...
Daviiid's user avatar
  • 119
0 votes
1 answer
286 views

I am trying to use TF-agents in a simple multi-agent non-cooperative parallel game. To simplify, I have two agents, defined with TF-agents. I defined a custom gym environment that takes as input the ...
Benares's user avatar
  • 1,288

15 30 50 per page