Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

monocualr results #12

Open
WendaLigit opened this issue Jul 4, 2023 · 4 comments
Open

monocualr results #12

WendaLigit opened this issue Jul 4, 2023 · 4 comments

Comments

@WendaLigit
Copy link

Hi, thanks for your impressive work!
I found that in Table 2, there are no results on monocular settings for training (Only stereo and monocular plus stereo).
Did you do the experiments on the monocular settings?

@Dwawayu
Copy link
Member

Dwawayu commented Jul 5, 2023

Hi! Thanks for your interest and kind words, and sorry for the delayed response.
We have included some results on monocular settings in the supplementary material. However, the training may not succeed since the predefined planes may not happen to be in a suitable position due to scale ambiguity, so we use a pre-trained PoseNet provided by Monodepth2 as a basic solution.

If you are interested in it, we can offer some minor suggestions.

  1. AQUANet addresses this problem by giving a pseudo depth map in the first five epochs.
  2. Another solution in my experiments is to apply a loss that forces the probability in the ground region of the middle ground plane to be 1 in the first epoch. Like:
loss_init = (grid_y > 0.) * |p[56] - 1.|

I hope these help, please let me know if you have any other concerns.

@WendaLigit
Copy link
Author

Thanks for your detailed information.

@Zhangwenyao1
Copy link

Hi! Thanks for your interest and kind words, and sorry for the delayed response. We have included some results on monocular settings in the supplementary material. However, the training may not succeed since the predefined planes may not happen to be in a suitable position due to scale ambiguity, so we use a pre-trained PoseNet provided by Monodepth2 as a basic solution.

If you are interested in it, we can offer some minor suggestions.

  1. AQUANet addresses this problem by giving a pseudo depth map in the first five epochs.
  2. Another solution in my experiments is to apply a loss that forces the probability in the ground region of the middle ground plane to be 1 in the first epoch. Like:
loss_init = (grid_y > 0.) * |p[56] - 1.|

I hope these help, please let me know if you have any other concerns.

Thanks for your great work! can you provide the monocular experimental code?

@Dwawayu
Copy link
Member

Dwawayu commented Mar 12, 2024

Thanks for your great work! can you provide the monocular experimental code?

Hi! Thank you for your interest and kind words. The command for the monocular experiment is here. If you want to use pretrained PoseNet as I did, please download mono+stereo_640x192 from monodepth2 and point towards the folder with the flag '--load_weights_folder'.
Hope this helps, and feel free to reach out if you have any more questions!

CUDA_VISIBLE_DEVICES=0,1,2,3 OMP_NUM_THREADS=1 torchrun --nproc_per_node=4 train.py \
--warp_type homography_warp \
--data_path ./kitti \
--split eigen_zhou \
--log_dir ./log \
--png \
--batch_size 8 --num_workers 8 \
--learning_rate 1e-4 \
--model_name Mono_pretrainpose \
--use_denseaspp \
--num_ep 8 \
--net_type ResNet \
--pc_net vgg19 --alpha_pc 0.1 \
--alpha_smooth 0.04 \
--gamma_smooth 2 \
--use_mixture_loss \
--plane_residual \
--xz_levels 14 \
--novel_frame_ids 1 -1 \
--no_stereo \
--automask \
--load_weights_folder ./log/mono+stereo_640x192 \
--models_to_load pose pose_encoder

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants