Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

model over-fitted? #2300

Unanswered
majid5776 asked this question in Q&A
Jul 22, 2024 · 2 comments
Discussion options

I've design my custom navigation env that have obstacles on it and the agents don't hit each other.
when I run my mappo_ippo.py I got strange outputs. can my model overfitted?
my custom_env is:
https://drive.google.com/file/d/1yw1rOpJcmoU99zcz-2wEGqV_ZnfOT1qF/view?usp=sharing
my config of mappo_ippo is:
max_steps:200
n_iters:625
n_agents and n_targets:3
backend:csv
entropy_eps:0.0001
remain confs is the same.

when I look at my csv and my videos they are surprising:
in my video the 20 first epochs they reach the goals very easy but after that episode they stop.
in my csv the train_mean_reward increasing non-stoply but my critic loss also increasing.
is this meaning my model overfitted?

@matteobettini

You must be logged in to vote

Replies: 2 comments

Comment options

I'm tagging Matteo which is our PoC for MARL things and the owner of the mappo_ippo script!
He's at ICML atm so he's probably not going to be very active on GH these days...

You must be logged in to vote
0 replies
Comment options

It could be that the reward for not colliding is taking over and preventing the navigation success.

The reward increasing is a good sign in general in case the reward function makes sense.

But in general I am not able to make diagnostic comments about your custome environment, sorry.

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
🙏
Q&A
Labels
None yet
3 participants
Morty Proxy This is a proxified and sanitized view of the page, visit original site.