-
Notifications
You must be signed in to change notification settings - Fork 11
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to reproduce claimed results for VoxFormer on KITTI360-SSCBench #8
Comments
#9 could be the possible cause, you may check the IoU scores for classes beyond 10. |
Thank you for the suggestion! Sadly I already accounted for that! |
@ahayler I'm wondering where is the provided script indicating the baseline number? I've searched the SSCBench and VoxFormer repos for that specific number but couldn't locate it. I want to make sure I'm using the right script. |
In the mobilestereonet prediction.py script you can find on top:
With 0.54 * 721 = 388.1823 being the value for semantic KITTI it seemed like 0.6 * 552.554261 = 331.53255659999996 is the baseline value for KITTI360. This was only an informed guess, so correct me if I am wrong. Let me know, whether or not you manage to reproduce the results from the paper :) |
@ahayler Thank you for your response. I have an additional question regarding:
How could I match the frames between SSCBench and KITTI360 to get stereo image pairs? I found the frame ID ranges do not align between the 2 datasets (e.g. sequence 0 in SSCBench contains frames 0 - 10,482, while sequence 0 in KITTI360 contains frames 0 - 11,517). |
Dear all,
I am trying to replicate the results of VoxFormer on your KITTI360 benchmark by using your checkpoint. So far I have achieved sensible results, but with an occupancy IoU of about 34.5 instead of your 38.6. Here is how I proceeded to generate the needed query proposals for stage 2:
--dataset kitti360 --baseline 331.53255659999996
.(the baseline number was indicated in the script. Outputs look sensible)Note: As stereo images were not provided in SSCBench and the pose files in the dataset also do not necessarily match the frame Ids provided, I created a mapping of the SSCBench frame Ids to the KITTI360 frame ids.
Is my process correct? Why might my results be worse? Could you either provide detailed instructions on how to replicate your results or the predictions?
Kind regards,
Adrian
The text was updated successfully, but these errors were encountered: