@@ -204,7 +204,7 @@ def __init__(
204
204
self ._states = np .zeros ((self ._capacity , self ._state_dim ), dtype = np .float32 )
205
205
self ._actions = np .zeros ((self ._capacity , self ._action_dim ), dtype = np .float32 )
206
206
self ._rewards = np .zeros (self ._capacity , dtype = np .float32 )
207
- self ._terminals = np .zeros (self ._capacity , dtype = np . bool )
207
+ self ._terminals = np .zeros (self ._capacity , dtype = bool )
208
208
self ._next_states = np .zeros ((self ._capacity , self ._state_dim ), dtype = np .float32 )
209
209
self ._returns = np .zeros (self ._capacity , dtype = np .float32 )
210
210
self ._advantages = np .zeros (self ._capacity , dtype = np .float32 )
@@ -373,7 +373,7 @@ def __init__(
373
373
self ._actions = [np .zeros ((self ._capacity , action_dim ), dtype = np .float32 ) for action_dim in self ._action_dims ]
374
374
self ._rewards = [np .zeros (self ._capacity , dtype = np .float32 ) for _ in range (self .agent_num )]
375
375
self ._next_states = np .zeros ((self ._capacity , self ._state_dim ), dtype = np .float32 )
376
- self ._terminals = np .zeros (self ._capacity , dtype = np . bool )
376
+ self ._terminals = np .zeros (self ._capacity , dtype = bool )
377
377
378
378
assert len (agent_states_dims ) == self .agent_num
379
379
self ._agent_states_dims = agent_states_dims
0 commit comments