Categorical Depth Distribution Network for Monocular 3D Object Detection

Related tags

Deep LearningCaDDN
Overview

CaDDN

CaDDN is a monocular-based 3D object detection method. This repository is based off of [OpenPCDet].

Categorical Depth Distribution Network for Monocular 3D Object Detection
Cody Reading, Ali Harakeh, Julia Chae, and Steven L. Waslander
[Paper]

Overview

Changelog

[2021-03-16] CaDDN v0.3.0 is released.

Introduction

What does CaDDN do?

CaDDN is a general PyTorch-based method for 3D object detection from monocular images. At the time of submission, CaDDN achieved first 1st place among published monocular methods on the Kitti 3D object detection benchmark. We welcome contributions to this project.

CaDDN design pattern

We inherit the design pattern from [OpenPCDet].

  • Data-Model separation with unified point cloud coordinate for easily extending to custom datasets:

  • Unified 3D box definition: (x, y, z, dx, dy, dz, heading).

Model Zoo

KITTI 3D Object Detection Baselines

Selected supported methods are shown in the below table. The results are the 3D detection performance of Car class on the val set of KITTI dataset.

  • All models are trained with 2 Tesla T4 GPUs and are available for download.
  • The training time is measured with 2 Tesla T4 GPUs and PyTorch 1.4.
training time [email protected] [email protected] [email protected] download
CaDDN ~76 hours 23.77 16.07 13.61 model-774M

Installation

Please refer to INSTALL.md for the installation of CaDDN.

Getting Started

Please refer to GETTING_STARTED.md to learn more usage about this project.

License

CaDDN is released under the Apache 2.0 license.

Acknowledgement

CaDDN is an open source project for monocular-based 3D scene perception. We would like to thank the authors of OpenPCDet for their open-source release of their 3D object detection codebase.

Citation

If you find this project useful in your research, please consider citing:

@article{CaDDN,
    title={Categorical Depth DistributionNetwork for Monocular 3D Object Detection},
    author={Cody Reading and
            Ali Harakeh and
            Julia Chae and
            Steven L. Waslander},
    journal = {CVPR},
    year={2021}
}

Contribution

Welcome to be a member of the CaDDN development team by contributing to this repo, and feel free to contact us for any potential contributions.

Comments
  • Out of memory, can not train on 11G machine

    Out of memory, can not train on 11G machine

    Hi,I set batch_size = 2, and I can not train resnet50 on 2080ti due to out of memory.Do you have any way to reduce the memory usage of the code? I tried to reduce the number of blocks in resnet50,but loss=nan, and error:

    WARNING:root:NaN or Inf found in input tensor.
    

    Thank you.

    opened by DuZzzs 14
  • Can I train the Nuscene model on OpenPCDet?

    Can I train the Nuscene model on OpenPCDet?

    Hello, I'm trying to use the CaDDN to train the Nuscene model. But it seems that it can not be directly applied. Do you implement the Nuscene model using CaDDN? If not, could you please give me some advice on what to do if I'd like to train the Nuscene model with CaDDN? As I know, the Nuscene dataset has six cameras, but the CaDDN only ran on the front camera. Thank you.

    opened by rockywind 7
  • Cuda Out of Memory with Cam150 image (above 2M image size)

    Cuda Out of Memory with Cam150 image (above 2M image size)

    I train CaDDN on A100 with Kitti Dataset it works fluently. But with our private dataset (captured with lidar and camera 150 degree) (after conversion to Kitti), it produces errors: File "/home/ubuntu/CaDDN/CaDDN/pcdet/models/backbones_3d/ffe/depth_ffe.py", line 93, in create_frustum_features frustum_features = depth_probs * image_features RuntimeError: CUDA out of memory. Tried to allocate 5.53 GiB (GPU 0; 39.59 GiB total capacity; 32.48 GiB already allocated; 4.92 GiB free; 32.99 GiB reserved in total by PyTorch). depth probs: torch.Size([2, 1, 80, 302, 480]) image feature shape: torch.Size([2, 64, 1, 302, 480])

    Pytorch 1.9.0 cudatoolkit 11.1 torchaudio 0.9.0
    torchvision 0.5.0
    It is the same error with cudatoolkit 11.0, pytorch 1.7.1 . Thank you very much !!!

    opened by namnv78 7
  • Can I train the Waymo model on OpenPCDet?

    Can I train the Waymo model on OpenPCDet?

    HELLO, I'm trying to use the OpenPCDet to train the Waymo model. But it seems that it can not be directly applied, do you implement the Waymo model using OpenPCDet? If not, could you please give me some advice on what to do if I'd like to train the Waymo model with OpenPCDet. Thank you.

    opened by Cc-Hy 6
  • How many gpus were used ?

    How many gpus were used ?

    hi, Thank you for sharing your work.

    when training on Kitti benchmark, do you mean by batch size of 4 per GPU? which equals to batch size of 8 when using two GPUs? Am I getting this right..?

    opened by ksh11023 6
  • Run test on pretrained model

    Run test on pretrained model

    i am getting the following error by running this command:

    python ./tools/test.py --cfg_file ./tools/cfgs/kitti_models/CaDDN.yaml --launcher pytorch --eval_all

    error:

    
    Traceback (most recent call last):
      File "./tools/test.py", line 199, in <module>
        main()
      File "./tools/test.py", line 190, in main
        model = build_network(model_cfg=cfg.MODEL, num_class=len(cfg.CLASS_NAMES), dataset=test_set)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/__init__.py", line 14, in build_network
        model = build_detector(
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/__init__.py", line 23, in build_detector
        model = __all__[model_cfg.NAME](
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/caddn.py", line 7, in __init__
        self.module_list = self.build_networks()
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/detector3d_template.py", line 50, in build_networks
        module, model_info_dict = getattr(self, 'build_%s' % module_name)(
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/detector3d_template.py", line 74, in build_ffe
        ffe_module = ffe.__all__[self.model_cfg.FFE.NAME](
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/depth_ffe.py", line 25, in __init__
        self.ddn = ddn.__all__[ddn_cfg.NAME](
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_deeplabv3.py", line 21, in __init__
        super().__init__(constructor=constructor, **kwargs)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_template.py", line 34, in __init__
        self.model = self.get_model(constructor=constructor)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_template.py", line 63, in get_model
        pretrained_dict = self.filter_pretrained_dict(model_dict=model_dict, pretrained_dict=pretrained_dict)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_template.py", line 88, in filter_pretrained_dict
        pretrained_num_classes = pretrained_dict["classifier.4.weight"].shape[0]
    KeyError: 'classifier.4.weight'
    

    to be more precise: i downloaded the cadnn.pth from the link in README.md and set the path in the config file. also pretrained_dict only contains: pretrained_dict : dict_keys(['epoch', 'it', 'model_state', 'optimizer_state', 'version'])

    installation is based on what you mentioned in INSTALL.md

    opened by EmbeddedAndMore 6
  • How can I transform depth map to real depth ?

    How can I transform depth map to real depth ?

    I notice that you turn the depth map to real depth by dividing 256. But when I change the depth to point cloud, I found this value not correct. Could you tell me why you divide 256 to get the depth?

    opened by kaixinbear 5
  • Rendering 3D bounding boxes onto test images

    Rendering 3D bounding boxes onto test images

    Hello,

    Thanks for your help on the previous issue. I'm running the tools/test.py script with the intent to generate 3D bounding boxes on the test images as a result image set. I'm using the KITTI dataset. Is there a flag I can set for test.py, or use another software package, or follow some other additional steps to be able to render those 3D bounding boxes?

    Also, my next step is to try using a custom generated imageset. I'll be using the pre-trained model even though it's from the perspective of a street light. They would be frames from a video capture. I was going to try to follow the structure of the KITTI dataset and see if I can repurpose the script that generated the datainfos. Do you have any suggestions for the best way to go about using a custom set of images like this?

    Ultimately, it's my goal to see 3D bounding box rendering working on those images.

    Thanks!

    opened by dpwolfe 5
  • Why does it show the loading model's EPOCH is 3?

    Why does it show the loading model's EPOCH is 3?

    Thank you for your help in advance! Why does it show the loading model's EPOCH is 3? 2021-10-22 20:57:44,189 INFO ==> Checkpoint trained from version: pcdet+0.3.0+0000000 2021-10-22 20:57:47,894 INFO ==> Done (loaded 903/903) 2021-10-22 20:57:48,489 INFO *************** EPOCH 3 EVALUATION *****************

    opened by rockywind 4
  • Question about the Voxel Features

    Question about the Voxel Features

    Congratulations on your great work!

    I have read your paper and have several questions that bother me:

    In your work,

    1. Firstly, the voxel grid is first generated.
    2. Secondly, use the gird_to_lidar, lidar_to_cam, cam_to_img transformation to find the correspondence between the grid coordinates and the image coordinates.
    3. Thirdly, use grid_sample to sample features from Frustum to Voxel.
    4. Finally, Voxelcollapse to BEV features

    Since, in my opinion, the BEV features represent the world coordinates. My question is, why not just use BEV features to generate 'BEV grid', which represents the real world (lidar) coordinates? So, the gird_to_lidar step can be omitted. Am I right?

    I am still confused about the 'Voxel Features'. I don'y know what is it used for?

    Thank you very much, looking forward to your replay!

    opened by taylover-pei 4
  • Can I change the lidar coordinate to camera coordinate?

    Can I change the lidar coordinate to camera coordinate?

    Hi, I have eight camera data. The 3D lable is similar to KITTI style, it is based on camera coordinate. I think the network would be confused if converted to lidar coordinate. But, the generated voxel and anchors setting are based on lidar coordinate. What can I do to change the code based on camera coordinate?

    opened by rockywind 4
  • `run.sh` caddn -> codyreading/caddn

    `run.sh` caddn -> codyreading/caddn

    When following our tutorial on using Docker, the pulled Docker Image name did not match the command used to run run.sh, resulting in an error with the docker run command. https://github.com/TRAILab/CaDDN/blob/master/docker/DOCKER.md#get-a-docker-image

    docker pull codyreading/caddn
    
    CMD="docker run -it \
        --runtime=nvidia \
        --net=host \
        --privileged=true \
        --ipc=host \
        --volume="/tmp/.X11-unix:/tmp/.X11-unix:rw" \
        --volume="$XAUTHORITY:/root/.Xauthority:rw" \
        --env="DISPLAY" \
        --env="QT_X11_NO_MITSHM=1" \
        --hostname="inside-DOCKER" \
        --name="CaDDN" \
        --volume $PROJ_DIR/checkpoints:/CaDDN/checkpoints \
        --volume $PROJ_DIR/data:/CaDDN/data \
        --volume $PROJ_DIR/output:/CaDDN/output \
        --volume $PROJ_DIR/tools:/CaDDN/tools \
        --volume $PROJ_DIR/.git:/CaDDN/.git \
        $DATA_VOLUMES \
        $PCDET_VOLUMES \
        --rm \
        caddn bash <--- codyreading/caddn
    
    opened by PINTO0309 0
  • Error while training the model

    Error while training the model "ValueError: num_samples should be a positive integer value, but got num_samples=0"

    Code we used for training: !python /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/train.py --cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml

    Log we got: 2022-11-19 21:23:27,419 INFO Start logging 2022-11-19 21:23:27,420 INFO CUDA_VISIBLE_DEVICES=ALL 2022-11-19 21:23:27,420 INFO cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml 2022-11-19 21:23:27,420 INFO batch_size 4 2022-11-19 21:23:27,420 INFO epochs 80 2022-11-19 21:23:27,420 INFO workers 8 2022-11-19 21:23:27,420 INFO extra_tag default 2022-11-19 21:23:27,421 INFO ckpt None 2022-11-19 21:23:27,421 INFO pretrained_model None 2022-11-19 21:23:27,421 INFO launcher none 2022-11-19 21:23:27,421 INFO tcp_port 18888 2022-11-19 21:23:27,421 INFO sync_bn False 2022-11-19 21:23:27,421 INFO fix_random_seed False 2022-11-19 21:23:27,421 INFO ckpt_save_interval 1 2022-11-19 21:23:27,421 INFO local_rank 0 2022-11-19 21:23:27,421 INFO max_ckpt_save_num 30 2022-11-19 21:23:27,421 INFO merge_all_iters_to_one_epoch False 2022-11-19 21:23:27,421 INFO set_cfgs None 2022-11-19 21:23:27,421 INFO max_waiting_mins 0 2022-11-19 21:23:27,422 INFO start_epoch 0 2022-11-19 21:23:27,422 INFO save_to_file False 2022-11-19 21:23:27,422 INFO cfg.ROOT_DIR: /content/gdrive/MyDrive/IITH_VCA/CaDDN 2022-11-19 21:23:27,422 INFO cfg.LOCAL_RANK: 0 2022-11-19 21:23:27,422 INFO cfg.CLASS_NAMES: ['Car', 'Pedestrian', 'Cyclist'] 2022-11-19 21:23:27,422 INFO
    cfg.DATA_CONFIG = edict() 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.DATASET: KittiDataset 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.DATA_PATH: /gdrive/MyDrive/3D_Obj/CaDDN/data/kitti 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.POINT_CLOUD_RANGE: [2, -30.08, -3.0, 46.8, 30.08, 1.0] 2022-11-19 21:23:27,422 INFO
    cfg.DATA_CONFIG.DATA_SPLIT = edict() 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.DATA_SPLIT.train: train 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.DATA_SPLIT.test: test 2022-11-19 21:23:27,423 INFO
    cfg.DATA_CONFIG.INFO_PATH = edict() 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.INFO_PATH.train: ['kitti_infos_trainval.pkl'] 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.INFO_PATH.test: ['kitti_infos_test.pkl'] 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.FOV_POINTS_ONLY: True 2022-11-19 21:23:27,423 INFO
    cfg.DATA_CONFIG.DATA_AUGMENTOR = edict() 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.DISABLE_AUG_LIST: ['placeholder'] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.AUG_CONFIG_LIST: [{'NAME': 'random_image_flip', 'ALONG_AXIS_LIST': ['horizontal']}] 2022-11-19 21:23:27,424 INFO
    cfg.DATA_CONFIG.POINT_FEATURE_ENCODING = edict() 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.encoding_type: absolute_coordinates_encoding 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.used_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.src_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.DATA_PROCESSOR: [{'NAME': 'mask_points_and_boxes_outside_range', 'REMOVE_OUTSIDE_BOXES': True}, {'NAME': 'calculate_grid_size', 'VOXEL_SIZE': [0.16, 0.16, 0.16]}] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.BASE_CONFIG: cfgs/dataset_configs/kitti_dataset.yaml 2022-11-19 21:23:27,425 INFO
    cfg.DATA_CONFIG.IMAGE = edict() 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.IMAGE.ENABLED: True 2022-11-19 21:23:27,425 INFO
    cfg.DATA_CONFIG.IMAGE.COLLATE = edict() 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.MODE: Pad 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.CONSTANT_VALUES: [0.485, 0.456, 0.406] 2022-11-19 21:23:27,425 INFO
    cfg.DATA_CONFIG.DEPTH_MAP = edict() 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.DEPTH_MAP.ENABLED: True 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.DEPTH_MAP.DOWNSAMPLE_FACTOR: 4 2022-11-19 21:23:27,426 INFO
    cfg.DATA_CONFIG.CALIB = edict() 2022-11-19 21:23:27,426 INFO cfg.DATA_CONFIG.CALIB.ENABLED: True 2022-11-19 21:23:27,426 INFO
    cfg.MODEL = edict() 2022-11-19 21:23:27,426 INFO cfg.MODEL.NAME: CaDDN 2022-11-19 21:23:27,426 INFO
    cfg.MODEL.FFE = edict() 2022-11-19 21:23:27,426 INFO cfg.MODEL.FFE.NAME: DepthFFE 2022-11-19 21:23:27,426 INFO
    cfg.MODEL.FFE.DDN = edict() 2022-11-19 21:23:27,426 INFO cfg.MODEL.FFE.DDN.NAME: DDNDeepLabV3 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.DDN.BACKBONE_NAME: ResNet101 2022-11-19 21:23:27,427 INFO
    cfg.MODEL.FFE.DDN.ARGS = edict() 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.DDN.ARGS.feat_extract_layer: layer1 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.DDN.ARGS.pretrained_path: ../checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:23:27,427 INFO
    cfg.MODEL.FFE.CHANNEL_REDUCE = edict() 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.in_channels: 256 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.out_channels: 64 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.kernel_size: 1 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.stride: 1 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.bias: False 2022-11-19 21:23:27,428 INFO
    cfg.MODEL.FFE.DISCRETIZE = edict() 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.mode: LID 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.num_bins: 80 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.depth_min: 2.0 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.depth_max: 46.8 2022-11-19 21:23:27,428 INFO
    cfg.MODEL.FFE.DDN_LOSS = edict() 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DDN_LOSS.weight: 3.0 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DDN_LOSS.alpha: 0.25 2022-11-19 21:23:27,429 INFO cfg.MODEL.FFE.DDN_LOSS.gamma: 2.0 2022-11-19 21:23:27,429 INFO cfg.MODEL.FFE.DDN_LOSS.fg_weight: 13 2022-11-19 21:23:27,429 INFO cfg.MODEL.FFE.DDN_LOSS.bg_weight: 1 2022-11-19 21:23:27,429 INFO
    cfg.MODEL.F2V = edict() 2022-11-19 21:23:27,429 INFO cfg.MODEL.F2V.NAME: FrustumToVoxel 2022-11-19 21:23:27,429 INFO
    cfg.MODEL.F2V.SAMPLER = edict() 2022-11-19 21:23:27,429 INFO cfg.MODEL.F2V.SAMPLER.mode: bilinear 2022-11-19 21:23:27,429 INFO cfg.MODEL.F2V.SAMPLER.padding_mode: zeros 2022-11-19 21:23:27,429 INFO
    cfg.MODEL.MAP_TO_BEV = edict() 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.NAME: Conv2DCollapse 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.NUM_BEV_FEATURES: 64 2022-11-19 21:23:27,430 INFO
    cfg.MODEL.MAP_TO_BEV.ARGS = edict() 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.ARGS.kernel_size: 1 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.ARGS.stride: 1 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.ARGS.bias: False 2022-11-19 21:23:27,430 INFO
    cfg.MODEL.BACKBONE_2D = edict() 2022-11-19 21:23:27,430 INFO cfg.MODEL.BACKBONE_2D.NAME: BaseBEVBackbone 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.LAYER_NUMS: [10, 10, 10] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.LAYER_STRIDES: [2, 2, 2] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.NUM_FILTERS: [64, 128, 256] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.UPSAMPLE_STRIDES: [1, 2, 4] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.NUM_UPSAMPLE_FILTERS: [128, 128, 128] 2022-11-19 21:23:27,431 INFO
    cfg.MODEL.DENSE_HEAD = edict() 2022-11-19 21:23:27,431 INFO cfg.MODEL.DENSE_HEAD.NAME: AnchorHeadSingle 2022-11-19 21:23:27,431 INFO cfg.MODEL.DENSE_HEAD.CLASS_AGNOSTIC: False 2022-11-19 21:23:27,431 INFO cfg.MODEL.DENSE_HEAD.USE_DIRECTION_CLASSIFIER: True 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.DIR_OFFSET: 0.78539 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.DIR_LIMIT_OFFSET: 0.0 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.NUM_DIR_BINS: 2 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.ANCHOR_GENERATOR_CONFIG: [{'class_name': 'Car', 'anchor_sizes': [[3.9, 1.6, 1.56]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-1.78], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.6, 'unmatched_threshold': 0.45}, {'class_name': 'Pedestrian', 'anchor_sizes': [[0.8, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}, {'class_name': 'Cyclist', 'anchor_sizes': [[1.76, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}] 2022-11-19 21:23:27,432 INFO
    cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG = edict() 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NAME: AxisAlignedTargetAssigner 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.POS_FRACTION: -1.0 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.SAMPLE_SIZE: 512 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NORM_BY_NUM_EXAMPLES: False 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.MATCH_HEIGHT: False 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.BOX_CODER: ResidualCoder 2022-11-19 21:23:27,433 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG = edict() 2022-11-19 21:23:27,433 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS = edict() 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.cls_weight: 1.0 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.loc_weight: 2.0 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.dir_weight: 0.2 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.code_weights: [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0] 2022-11-19 21:23:27,434 INFO
    cfg.MODEL.POST_PROCESSING = edict() 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.RECALL_THRESH_LIST: [0.3, 0.5, 0.7] 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.SCORE_THRESH: 0.1 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.OUTPUT_RAW_SCORE: False 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.EVAL_METRIC: kitti 2022-11-19 21:23:27,434 INFO
    cfg.MODEL.POST_PROCESSING.NMS_CONFIG = edict() 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.MULTI_CLASSES_NMS: False 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_TYPE: nms_gpu 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_THRESH: 0.01 2022-11-19 21:23:27,435 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_PRE_MAXSIZE: 4096 2022-11-19 21:23:27,435 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_POST_MAXSIZE: 500 2022-11-19 21:23:27,435 INFO
    cfg.OPTIMIZATION = edict() 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.BATCH_SIZE_PER_GPU: 4 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.NUM_EPOCHS: 80 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.OPTIMIZER: adam_onecycle 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.LR: 0.001 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.WEIGHT_DECAY: 0.01 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.MOMENTUM: 0.9 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.MOMS: [0.95, 0.85] 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.PCT_START: 0.4 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.DIV_FACTOR: 10 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.DECAY_STEP_LIST: [35, 45] 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.LR_DECAY: 0.1 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.LR_CLIP: 1e-07 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.LR_WARMUP: False 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.WARMUP_EPOCH: 1 2022-11-19 21:23:27,437 INFO cfg.OPTIMIZATION.GRAD_NORM_CLIP: 10 2022-11-19 21:23:27,437 INFO cfg.TAG: CaDDN 2022-11-19 21:23:27,437 INFO cfg.EXP_GROUP_PATH: content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models 2022-11-19 21:23:28,525 INFO Loading KITTI dataset 2022-11-19 21:23:28,526 INFO Total samples for KITTI dataset: 0 Traceback (most recent call last): File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/train.py", line 198, in main() File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/train.py", line 112, in main total_epochs=args.epochs File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/pcdet/datasets/init.py", line 69, in build_dataloader drop_last=False, sampler=sampler, timeout=0 File "/usr/local/lib/python3.7/dist-packages/torch/utils/data/dataloader.py", line 213, in init sampler = RandomSampler(dataset) File "/usr/local/lib/python3.7/dist-packages/torch/utils/data/sampler.py", line 94, in init "value, but got num_samples={}".format(self.num_samples)) ValueError: num_samples should be a positive integer value, but got num_samples=0

    Screen shots of the same: image

    Error screen shot: image

    opened by prabhata07 1
  • KeyError: 'model_state'

    KeyError: 'model_state'

    Facing KeyError: 'model_state' while trying to test with "Pretrained model" Command used:

    !python /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py --cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml --batch_size 1 --ckpt /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth

    Environemnt: Collab

    Model: as given in the github repo deeplabv3_resnet101_coco-586e9e4e.pth

    Error log:

    2022-11-19 21:16:39,034 INFO Start logging 2022-11-19 21:16:39,035 INFO CUDA_VISIBLE_DEVICES=ALL 2022-11-19 21:16:39,035 INFO cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml 2022-11-19 21:16:39,035 INFO batch_size 1 2022-11-19 21:16:39,035 INFO workers 4 2022-11-19 21:16:39,035 INFO extra_tag default 2022-11-19 21:16:39,035 INFO ckpt /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:16:39,035 INFO launcher none 2022-11-19 21:16:39,036 INFO tcp_port 18888 2022-11-19 21:16:39,036 INFO local_rank 0 2022-11-19 21:16:39,036 INFO set_cfgs None 2022-11-19 21:16:39,036 INFO max_waiting_mins 30 2022-11-19 21:16:39,036 INFO start_epoch 0 2022-11-19 21:16:39,036 INFO eval_tag default 2022-11-19 21:16:39,036 INFO eval_all False 2022-11-19 21:16:39,036 INFO ckpt_dir None 2022-11-19 21:16:39,036 INFO save_to_file False 2022-11-19 21:16:39,037 INFO cfg.ROOT_DIR: /content/gdrive/MyDrive/IITH_VCA/CaDDN 2022-11-19 21:16:39,037 INFO cfg.LOCAL_RANK: 0 2022-11-19 21:16:39,037 INFO cfg.CLASS_NAMES: ['Car', 'Pedestrian', 'Cyclist'] 2022-11-19 21:16:39,037 INFO
    cfg.DATA_CONFIG = edict() 2022-11-19 21:16:39,037 INFO cfg.DATA_CONFIG.DATASET: KittiDataset 2022-11-19 21:16:39,037 INFO cfg.DATA_CONFIG.DATA_PATH: /gdrive/MyDrive/3D_Obj/CaDDN/data/kitti 2022-11-19 21:16:39,037 INFO cfg.DATA_CONFIG.POINT_CLOUD_RANGE: [2, -30.08, -3.0, 46.8, 30.08, 1.0] 2022-11-19 21:16:39,037 INFO
    cfg.DATA_CONFIG.DATA_SPLIT = edict() 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_SPLIT.train: train 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_SPLIT.test: test 2022-11-19 21:16:39,038 INFO
    cfg.DATA_CONFIG.INFO_PATH = edict() 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.INFO_PATH.train: ['kitti_infos_trainval.pkl'] 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.INFO_PATH.test: ['kitti_infos_test.pkl'] 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.FOV_POINTS_ONLY: True 2022-11-19 21:16:39,038 INFO
    cfg.DATA_CONFIG.DATA_AUGMENTOR = edict() 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.DISABLE_AUG_LIST: ['placeholder'] 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.AUG_CONFIG_LIST: [{'NAME': 'random_image_flip', 'ALONG_AXIS_LIST': ['horizontal']}] 2022-11-19 21:16:39,038 INFO
    cfg.DATA_CONFIG.POINT_FEATURE_ENCODING = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.encoding_type: absolute_coordinates_encoding 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.used_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.src_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.DATA_PROCESSOR: [{'NAME': 'mask_points_and_boxes_outside_range', 'REMOVE_OUTSIDE_BOXES': True}, {'NAME': 'calculate_grid_size', 'VOXEL_SIZE': [0.16, 0.16, 0.16]}] 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.BASE_CONFIG: cfgs/dataset_configs/kitti_dataset.yaml 2022-11-19 21:16:39,039 INFO
    cfg.DATA_CONFIG.IMAGE = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.IMAGE.ENABLED: True 2022-11-19 21:16:39,039 INFO
    cfg.DATA_CONFIG.IMAGE.COLLATE = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.MODE: Pad 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.CONSTANT_VALUES: [0.485, 0.456, 0.406] 2022-11-19 21:16:39,039 INFO
    cfg.DATA_CONFIG.DEPTH_MAP = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.DEPTH_MAP.ENABLED: True 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.DEPTH_MAP.DOWNSAMPLE_FACTOR: 4 2022-11-19 21:16:39,040 INFO
    cfg.DATA_CONFIG.CALIB = edict() 2022-11-19 21:16:39,040 INFO cfg.DATA_CONFIG.CALIB.ENABLED: True 2022-11-19 21:16:39,040 INFO
    cfg.MODEL = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.NAME: CaDDN 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.NAME: DepthFFE 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE.DDN = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.NAME: DDNDeepLabV3 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.BACKBONE_NAME: ResNet101 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE.DDN.ARGS = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.ARGS.feat_extract_layer: layer1 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.ARGS.pretrained_path: ../checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE.CHANNEL_REDUCE = edict() 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.in_channels: 256 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.out_channels: 64 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.kernel_size: 1 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.stride: 1 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.bias: False 2022-11-19 21:16:39,041 INFO
    cfg.MODEL.FFE.DISCRETIZE = edict() 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.mode: LID 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.num_bins: 80 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.depth_min: 2.0 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.depth_max: 46.8 2022-11-19 21:16:39,041 INFO
    cfg.MODEL.FFE.DDN_LOSS = edict() 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DDN_LOSS.weight: 3.0 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DDN_LOSS.alpha: 0.25 2022-11-19 21:16:39,042 INFO cfg.MODEL.FFE.DDN_LOSS.gamma: 2.0 2022-11-19 21:16:39,042 INFO cfg.MODEL.FFE.DDN_LOSS.fg_weight: 13 2022-11-19 21:16:39,042 INFO cfg.MODEL.FFE.DDN_LOSS.bg_weight: 1 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.F2V = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.F2V.NAME: FrustumToVoxel 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.F2V.SAMPLER = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.F2V.SAMPLER.mode: bilinear 2022-11-19 21:16:39,042 INFO cfg.MODEL.F2V.SAMPLER.padding_mode: zeros 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.MAP_TO_BEV = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.MAP_TO_BEV.NAME: Conv2DCollapse 2022-11-19 21:16:39,042 INFO cfg.MODEL.MAP_TO_BEV.NUM_BEV_FEATURES: 64 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.MAP_TO_BEV.ARGS = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.MAP_TO_BEV.ARGS.kernel_size: 1 2022-11-19 21:16:39,043 INFO cfg.MODEL.MAP_TO_BEV.ARGS.stride: 1 2022-11-19 21:16:39,043 INFO cfg.MODEL.MAP_TO_BEV.ARGS.bias: False 2022-11-19 21:16:39,043 INFO
    cfg.MODEL.BACKBONE_2D = edict() 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.NAME: BaseBEVBackbone 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.LAYER_NUMS: [10, 10, 10] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.LAYER_STRIDES: [2, 2, 2] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.NUM_FILTERS: [64, 128, 256] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.UPSAMPLE_STRIDES: [1, 2, 4] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.NUM_UPSAMPLE_FILTERS: [128, 128, 128] 2022-11-19 21:16:39,043 INFO
    cfg.MODEL.DENSE_HEAD = edict() 2022-11-19 21:16:39,043 INFO cfg.MODEL.DENSE_HEAD.NAME: AnchorHeadSingle 2022-11-19 21:16:39,043 INFO cfg.MODEL.DENSE_HEAD.CLASS_AGNOSTIC: False 2022-11-19 21:16:39,043 INFO cfg.MODEL.DENSE_HEAD.USE_DIRECTION_CLASSIFIER: True 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.DIR_OFFSET: 0.78539 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.DIR_LIMIT_OFFSET: 0.0 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.NUM_DIR_BINS: 2 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.ANCHOR_GENERATOR_CONFIG: [{'class_name': 'Car', 'anchor_sizes': [[3.9, 1.6, 1.56]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-1.78], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.6, 'unmatched_threshold': 0.45}, {'class_name': 'Pedestrian', 'anchor_sizes': [[0.8, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}, {'class_name': 'Cyclist', 'anchor_sizes': [[1.76, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}] 2022-11-19 21:16:39,044 INFO
    cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG = edict() 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NAME: AxisAlignedTargetAssigner 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.POS_FRACTION: -1.0 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.SAMPLE_SIZE: 512 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NORM_BY_NUM_EXAMPLES: False 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.MATCH_HEIGHT: False 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.BOX_CODER: ResidualCoder 2022-11-19 21:16:39,044 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG = edict() 2022-11-19 21:16:39,045 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS = edict() 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.cls_weight: 1.0 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.loc_weight: 2.0 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.dir_weight: 0.2 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.code_weights: [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0] 2022-11-19 21:16:39,045 INFO
    cfg.MODEL.POST_PROCESSING = edict() 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.RECALL_THRESH_LIST: [0.3, 0.5, 0.7] 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.SCORE_THRESH: 0.1 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.OUTPUT_RAW_SCORE: False 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.EVAL_METRIC: kitti 2022-11-19 21:16:39,045 INFO
    cfg.MODEL.POST_PROCESSING.NMS_CONFIG = edict() 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.MULTI_CLASSES_NMS: False 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_TYPE: nms_gpu 2022-11-19 21:16:39,046 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_THRESH: 0.01 2022-11-19 21:16:39,046 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_PRE_MAXSIZE: 4096 2022-11-19 21:16:39,046 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_POST_MAXSIZE: 500 2022-11-19 21:16:39,046 INFO
    cfg.OPTIMIZATION = edict() 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.BATCH_SIZE_PER_GPU: 4 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.NUM_EPOCHS: 80 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.OPTIMIZER: adam_onecycle 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.LR: 0.001 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.WEIGHT_DECAY: 0.01 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.MOMENTUM: 0.9 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.MOMS: [0.95, 0.85] 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.PCT_START: 0.4 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.DIV_FACTOR: 10 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.DECAY_STEP_LIST: [35, 45] 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.LR_DECAY: 0.1 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.LR_CLIP: 1e-07 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.LR_WARMUP: False 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.WARMUP_EPOCH: 1 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.GRAD_NORM_CLIP: 10 2022-11-19 21:16:39,048 INFO cfg.TAG: CaDDN 2022-11-19 21:16:39,048 INFO cfg.EXP_GROUP_PATH: content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models 2022-11-19 21:16:39,048 INFO Loading KITTI dataset 2022-11-19 21:16:39,048 INFO Total samples for KITTI dataset: 0 /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:16:42,793 INFO ==> Loading parameters from checkpoint /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth to GPU Traceback (most recent call last): File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py", line 200, in main() File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py", line 196, in main eval_single_ckpt(model, test_loader, args, eval_output_dir, logger, epoch_id, dist_test=dist_test) File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py", line 58, in eval_single_ckpt model.load_params_from_file(filename=args.ckpt, logger=logger, to_cpu=dist_test) File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/pcdet/models/detectors/detector3d_template.py", line 359, in load_params_from_file model_state_disk = checkpoint['model_state'] KeyError: 'model_state'

    opened by prabhata07 1
  • the result is not reproduced

    the result is not reproduced

    I trained the model with 2 Tesla V100 GPUs in the default configuration and got [email protected], [email protected], [email protected] 21.203, 15.1487,12.7407 respectively, which is lower than the 23.77,16.07, 13.61 you declared. But when I use the pre-trained weight, I can get 23.77,16.07, 13.61. So, are there any training tricks that have not been published?

    opened by leo038 0
Releases(v0.3.1)
Owner
Toronto Robotics and AI Laboratory
Toronto Robotics and AI Laboratory
Py-FEAT: Python Facial Expression Analysis Toolbox

Py-FEAT is a suite for facial expressions (FEX) research written in Python. This package includes tools to detect faces, extract emotional facial expressions (e.g., happiness, sadness, anger), facial

Computational Social Affective Neuroscience Laboratory 147 Jan 06, 2023
ConvMAE: Masked Convolution Meets Masked Autoencoders

ConvMAE ConvMAE: Masked Convolution Meets Masked Autoencoders Peng Gao1, Teli Ma1, Hongsheng Li2, Jifeng Dai3, Yu Qiao1, 1 Shanghai AI Laboratory, 2 M

Alpha VL Team of Shanghai AI Lab 345 Jan 08, 2023
Implementation for our ICCV 2021 paper: Dual-Camera Super-Resolution with Aligned Attention Modules

DCSR: Dual Camera Super-Resolution Implementation for our ICCV 2021 oral paper: Dual-Camera Super-Resolution with Aligned Attention Modules paper | pr

Tengfei Wang 110 Dec 20, 2022
Semi-supervised Domain Adaptation via Minimax Entropy

Semi-supervised Domain Adaptation via Minimax Entropy (ICCV 2019) Install pip install -r requirements.txt The code is written for Pytorch 0.4.0, but s

Vision and Learning Group 243 Jan 09, 2023
A Python package for causal inference using Synthetic Controls

Synthetic Control Methods A Python package for causal inference using synthetic controls This Python package implements a class of approaches to estim

Oscar Engelbrektson 107 Dec 28, 2022
Semantic Segmentation Architectures Implemented in PyTorch

pytorch-semseg Semantic Segmentation Algorithms Implemented in PyTorch This repository aims at mirroring popular semantic segmentation architectures i

Meet Shah 3.3k Dec 29, 2022
Crowd-Kit is a powerful Python library that implements commonly-used aggregation methods for crowdsourced annotation and offers the relevant metrics and datasets

Crowd-Kit: Computational Quality Control for Crowdsourcing Documentation Crowd-Kit is a powerful Python library that implements commonly-used aggregat

Toloka 125 Dec 30, 2022
Self-Supervised Learning of Event-based Optical Flow with Spiking Neural Networks

Self-Supervised Learning of Event-based Optical Flow with Spiking Neural Networks Work accepted at NeurIPS'21 [paper, video]. If you use this code in

TU Delft 43 Dec 07, 2022
This repo contains research materials released by members of the Google Brain team in Tokyo.

Brain Tokyo Workshop 🧠 🗼 This repo contains research materials released by members of the Google Brain team in Tokyo. Past Projects Weight Agnostic

Google 1.2k Jan 02, 2023
BlueFog Tutorials

BlueFog Tutorials Welcome to the BlueFog tutorials! In this repository, we've put together a collection of awesome Jupyter notebooks. These notebooks

4 Oct 27, 2021
[CVPR 2021] Forecasting the panoptic segmentation of future video frames

Panoptic Segmentation Forecasting Colin Graber, Grace Tsai, Michael Firman, Gabriel Brostow, Alexander Schwing - CVPR 2021 [Link to paper] We propose

Niantic Labs 44 Nov 29, 2022
Pytorch Implementation of Residual Vision Transformers(ResViT)

ResViT Official Pytorch Implementation of Residual Vision Transformers(ResViT) which is described in the following paper: Onat Dalmaz and Mahmut Yurt

ICON Lab 41 Dec 08, 2022
This tool uses Deep Learning to help you draw and write with your hand and webcam.

This tool uses Deep Learning to help you draw and write with your hand and webcam. A Deep Learning model is used to try to predict whether you want to have 'pencil up' or 'pencil down'.

lmagne 169 Dec 10, 2022
Code for 'Blockwise Sequential Model Learning for Partially Observable Reinforcement Learning' (AAAI 2022)

Blockwise Sequential Model Learning Code for 'Blockwise Sequential Model Learning for Partially Observable Reinforcement Learning' (AAAI 2022) For ins

2 Jun 17, 2022
The implementation our EMNLP 2021 paper "Enhanced Language Representation with Label Knowledge for Span Extraction".

LEAR The implementation our EMNLP 2021 paper "Enhanced Language Representation with Label Knowledge for Span Extraction". **The code is in the "master

杨攀 93 Jan 07, 2023
This repository lets you interact with Lean through a REPL.

lean-gym This repository lets you interact with Lean through a REPL. See Formal Mathematics Statement Curriculum Learning for a presentation of lean-g

OpenAI 87 Dec 28, 2022
Object classification with basic computer vision techniques

naive-image-classification Object classification with basic computer vision techniques. Final assignment for the computer vision course I took at univ

2 Jul 01, 2022
A simple implementation of Kalman filter in single object tracking

kalman-filter-in-single-object-tracking A simple implementation of Kalman filter in single object tracking https://www.bilibili.com/video/BV1Qf4y1J7D4

130 Dec 26, 2022
Docker containers of baseline agents for the Crafter environment

Crafter Baselines This repository contains Docker containers for running various baselines on the Crafter environment. Reward Agents DreamerV2 based o

Danijar Hafner 17 Sep 25, 2022