-
Notifications
You must be signed in to change notification settings - Fork 99
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feat Sebulba recurrent IQL #1148
base: develop
Are you sure you want to change the base?
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I've looked through everything except the system file and it looks good, Sebulba utils especially! Just some relatively minor style changes
|
||
# todo: remove the ppo dependencies when we make sebulba for other systems |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is a good point though, maybe there's something we can do about it 🤔
Maybe a protocol like that has action, obs, reward
, not sure if there's any other common attributes?
terminated = np.repeat( | ||
terminated[..., np.newaxis], repeats=self.num_agents, axis=-1 | ||
) # (B,) --> (B, N) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Does this already happen for smax and lbf?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Also: if terminated is repeated for num_agents in the wrapper, should we still be repeating dones for num_agents in the systems? Because I see you didn't remove this from sebulba PPO. Or does this not actually effect the dones?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great work here! Really minor changes required. Happy to merge this pending some benchmarks
target: Array, | ||
) -> Tuple[Array, Metrics]: | ||
# axes switched here to scan over time | ||
hidden_state, obs_term_or_trunc = prep_inputs_to_scannedrnn(obs, term_or_trunc) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
A general comment, I think this would be a lot easier to read if we used done
to mean term_or_trunc
which I think is a reasonable thing. Would have to make the change in anakin also though :/
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great work Louay! Just two questions from my side as you can see in the comments.
rewards = np.zeros((num_envs, num_agents), dtype=float) | ||
teminated = np.zeros(num_envs, dtype=float) | ||
rewards = np.zeros((num_envs, self.num_agents), dtype=float) | ||
terminated = np.zeros((num_envs, self.num_agents), dtype=float) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I assume with this change we would also need to change sebulba PPO? Since currently it does this same operation. We should decide if generally its better doing this in the system or the wrapper.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good point, I think we do this in the wrappers for the anakin systems
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great to me, just a few minor things 🙏
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🔥
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good! Just one request and one question.
What?
A recurrent IQL implementation using the Sebulba architecture.
Why?
Offline Sebulba base and non-jax envs in Mava.
How?
Mixed the Sebulba structure from PPO with the learner code from Anakin IQL.