Categorical Depth Distribution Network for Monocular 3D Object Detection

Related tags

Deep LearningCaDDN
Overview

CaDDN

CaDDN is a monocular-based 3D object detection method. This repository is based off of [OpenPCDet].

Categorical Depth Distribution Network for Monocular 3D Object Detection
Cody Reading, Ali Harakeh, Julia Chae, and Steven L. Waslander
[Paper]

Overview

Changelog

[2021-03-16] CaDDN v0.3.0 is released.

Introduction

What does CaDDN do?

CaDDN is a general PyTorch-based method for 3D object detection from monocular images. At the time of submission, CaDDN achieved first 1st place among published monocular methods on the Kitti 3D object detection benchmark. We welcome contributions to this project.

CaDDN design pattern

We inherit the design pattern from [OpenPCDet].

  • Data-Model separation with unified point cloud coordinate for easily extending to custom datasets:

  • Unified 3D box definition: (x, y, z, dx, dy, dz, heading).

Model Zoo

KITTI 3D Object Detection Baselines

Selected supported methods are shown in the below table. The results are the 3D detection performance of Car class on the val set of KITTI dataset.

  • All models are trained with 2 Tesla T4 GPUs and are available for download.
  • The training time is measured with 2 Tesla T4 GPUs and PyTorch 1.4.
training time [email protected] [email protected] [email protected] download
CaDDN ~76 hours 23.77 16.07 13.61 model-774M

Installation

Please refer to INSTALL.md for the installation of CaDDN.

Getting Started

Please refer to GETTING_STARTED.md to learn more usage about this project.

License

CaDDN is released under the Apache 2.0 license.

Acknowledgement

CaDDN is an open source project for monocular-based 3D scene perception. We would like to thank the authors of OpenPCDet for their open-source release of their 3D object detection codebase.

Citation

If you find this project useful in your research, please consider citing:

@article{CaDDN,
    title={Categorical Depth DistributionNetwork for Monocular 3D Object Detection},
    author={Cody Reading and
            Ali Harakeh and
            Julia Chae and
            Steven L. Waslander},
    journal = {CVPR},
    year={2021}
}

Contribution

Welcome to be a member of the CaDDN development team by contributing to this repo, and feel free to contact us for any potential contributions.

Comments
  • Out of memory, can not train on 11G machine

    Out of memory, can not train on 11G machine

    Hi,I set batch_size = 2, and I can not train resnet50 on 2080ti due to out of memory.Do you have any way to reduce the memory usage of the code? I tried to reduce the number of blocks in resnet50,but loss=nan, and error:

    WARNING:root:NaN or Inf found in input tensor.
    

    Thank you.

    opened by DuZzzs 14
  • Can I train the Nuscene model on OpenPCDet?

    Can I train the Nuscene model on OpenPCDet?

    Hello, I'm trying to use the CaDDN to train the Nuscene model. But it seems that it can not be directly applied. Do you implement the Nuscene model using CaDDN? If not, could you please give me some advice on what to do if I'd like to train the Nuscene model with CaDDN? As I know, the Nuscene dataset has six cameras, but the CaDDN only ran on the front camera. Thank you.

    opened by rockywind 7
  • Cuda Out of Memory with Cam150 image (above 2M image size)

    Cuda Out of Memory with Cam150 image (above 2M image size)

    I train CaDDN on A100 with Kitti Dataset it works fluently. But with our private dataset (captured with lidar and camera 150 degree) (after conversion to Kitti), it produces errors: File "/home/ubuntu/CaDDN/CaDDN/pcdet/models/backbones_3d/ffe/depth_ffe.py", line 93, in create_frustum_features frustum_features = depth_probs * image_features RuntimeError: CUDA out of memory. Tried to allocate 5.53 GiB (GPU 0; 39.59 GiB total capacity; 32.48 GiB already allocated; 4.92 GiB free; 32.99 GiB reserved in total by PyTorch). depth probs: torch.Size([2, 1, 80, 302, 480]) image feature shape: torch.Size([2, 64, 1, 302, 480])

    Pytorch 1.9.0 cudatoolkit 11.1 torchaudio 0.9.0
    torchvision 0.5.0
    It is the same error with cudatoolkit 11.0, pytorch 1.7.1 . Thank you very much !!!

    opened by namnv78 7
  • Can I train the Waymo model on OpenPCDet?

    Can I train the Waymo model on OpenPCDet?

    HELLO, I'm trying to use the OpenPCDet to train the Waymo model. But it seems that it can not be directly applied, do you implement the Waymo model using OpenPCDet? If not, could you please give me some advice on what to do if I'd like to train the Waymo model with OpenPCDet. Thank you.

    opened by Cc-Hy 6
  • How many gpus were used ?

    How many gpus were used ?

    hi, Thank you for sharing your work.

    when training on Kitti benchmark, do you mean by batch size of 4 per GPU? which equals to batch size of 8 when using two GPUs? Am I getting this right..?

    opened by ksh11023 6
  • Run test on pretrained model

    Run test on pretrained model

    i am getting the following error by running this command:

    python ./tools/test.py --cfg_file ./tools/cfgs/kitti_models/CaDDN.yaml --launcher pytorch --eval_all

    error:

    
    Traceback (most recent call last):
      File "./tools/test.py", line 199, in <module>
        main()
      File "./tools/test.py", line 190, in main
        model = build_network(model_cfg=cfg.MODEL, num_class=len(cfg.CLASS_NAMES), dataset=test_set)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/__init__.py", line 14, in build_network
        model = build_detector(
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/__init__.py", line 23, in build_detector
        model = __all__[model_cfg.NAME](
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/caddn.py", line 7, in __init__
        self.module_list = self.build_networks()
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/detector3d_template.py", line 50, in build_networks
        module, model_info_dict = getattr(self, 'build_%s' % module_name)(
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/detectors/detector3d_template.py", line 74, in build_ffe
        ffe_module = ffe.__all__[self.model_cfg.FFE.NAME](
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/depth_ffe.py", line 25, in __init__
        self.ddn = ddn.__all__[ddn_cfg.NAME](
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_deeplabv3.py", line 21, in __init__
        super().__init__(constructor=constructor, **kwargs)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_template.py", line 34, in __init__
        self.model = self.get_model(constructor=constructor)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_template.py", line 63, in get_model
        pretrained_dict = self.filter_pretrained_dict(model_dict=model_dict, pretrained_dict=pretrained_dict)
      File "/home/mm/MasterThesis/CaDDN/pcdet/models/backbones_3d/ffe/ddn/ddn_template.py", line 88, in filter_pretrained_dict
        pretrained_num_classes = pretrained_dict["classifier.4.weight"].shape[0]
    KeyError: 'classifier.4.weight'
    

    to be more precise: i downloaded the cadnn.pth from the link in README.md and set the path in the config file. also pretrained_dict only contains: pretrained_dict : dict_keys(['epoch', 'it', 'model_state', 'optimizer_state', 'version'])

    installation is based on what you mentioned in INSTALL.md

    opened by EmbeddedAndMore 6
  • How can I transform depth map to real depth ?

    How can I transform depth map to real depth ?

    I notice that you turn the depth map to real depth by dividing 256. But when I change the depth to point cloud, I found this value not correct. Could you tell me why you divide 256 to get the depth?

    opened by kaixinbear 5
  • Rendering 3D bounding boxes onto test images

    Rendering 3D bounding boxes onto test images

    Hello,

    Thanks for your help on the previous issue. I'm running the tools/test.py script with the intent to generate 3D bounding boxes on the test images as a result image set. I'm using the KITTI dataset. Is there a flag I can set for test.py, or use another software package, or follow some other additional steps to be able to render those 3D bounding boxes?

    Also, my next step is to try using a custom generated imageset. I'll be using the pre-trained model even though it's from the perspective of a street light. They would be frames from a video capture. I was going to try to follow the structure of the KITTI dataset and see if I can repurpose the script that generated the datainfos. Do you have any suggestions for the best way to go about using a custom set of images like this?

    Ultimately, it's my goal to see 3D bounding box rendering working on those images.

    Thanks!

    opened by dpwolfe 5
  • Why does it show the loading model's EPOCH is 3?

    Why does it show the loading model's EPOCH is 3?

    Thank you for your help in advance! Why does it show the loading model's EPOCH is 3? 2021-10-22 20:57:44,189 INFO ==> Checkpoint trained from version: pcdet+0.3.0+0000000 2021-10-22 20:57:47,894 INFO ==> Done (loaded 903/903) 2021-10-22 20:57:48,489 INFO *************** EPOCH 3 EVALUATION *****************

    opened by rockywind 4
  • Question about the Voxel Features

    Question about the Voxel Features

    Congratulations on your great work!

    I have read your paper and have several questions that bother me:

    In your work,

    1. Firstly, the voxel grid is first generated.
    2. Secondly, use the gird_to_lidar, lidar_to_cam, cam_to_img transformation to find the correspondence between the grid coordinates and the image coordinates.
    3. Thirdly, use grid_sample to sample features from Frustum to Voxel.
    4. Finally, Voxelcollapse to BEV features

    Since, in my opinion, the BEV features represent the world coordinates. My question is, why not just use BEV features to generate 'BEV grid', which represents the real world (lidar) coordinates? So, the gird_to_lidar step can be omitted. Am I right?

    I am still confused about the 'Voxel Features'. I don'y know what is it used for?

    Thank you very much, looking forward to your replay!

    opened by taylover-pei 4
  • Can I change the lidar coordinate to camera coordinate?

    Can I change the lidar coordinate to camera coordinate?

    Hi, I have eight camera data. The 3D lable is similar to KITTI style, it is based on camera coordinate. I think the network would be confused if converted to lidar coordinate. But, the generated voxel and anchors setting are based on lidar coordinate. What can I do to change the code based on camera coordinate?

    opened by rockywind 4
  • `run.sh` caddn -> codyreading/caddn

    `run.sh` caddn -> codyreading/caddn

    When following our tutorial on using Docker, the pulled Docker Image name did not match the command used to run run.sh, resulting in an error with the docker run command. https://github.com/TRAILab/CaDDN/blob/master/docker/DOCKER.md#get-a-docker-image

    docker pull codyreading/caddn
    
    CMD="docker run -it \
        --runtime=nvidia \
        --net=host \
        --privileged=true \
        --ipc=host \
        --volume="/tmp/.X11-unix:/tmp/.X11-unix:rw" \
        --volume="$XAUTHORITY:/root/.Xauthority:rw" \
        --env="DISPLAY" \
        --env="QT_X11_NO_MITSHM=1" \
        --hostname="inside-DOCKER" \
        --name="CaDDN" \
        --volume $PROJ_DIR/checkpoints:/CaDDN/checkpoints \
        --volume $PROJ_DIR/data:/CaDDN/data \
        --volume $PROJ_DIR/output:/CaDDN/output \
        --volume $PROJ_DIR/tools:/CaDDN/tools \
        --volume $PROJ_DIR/.git:/CaDDN/.git \
        $DATA_VOLUMES \
        $PCDET_VOLUMES \
        --rm \
        caddn bash <--- codyreading/caddn
    
    opened by PINTO0309 0
  • Error while training the model

    Error while training the model "ValueError: num_samples should be a positive integer value, but got num_samples=0"

    Code we used for training: !python /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/train.py --cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml

    Log we got: 2022-11-19 21:23:27,419 INFO Start logging 2022-11-19 21:23:27,420 INFO CUDA_VISIBLE_DEVICES=ALL 2022-11-19 21:23:27,420 INFO cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml 2022-11-19 21:23:27,420 INFO batch_size 4 2022-11-19 21:23:27,420 INFO epochs 80 2022-11-19 21:23:27,420 INFO workers 8 2022-11-19 21:23:27,420 INFO extra_tag default 2022-11-19 21:23:27,421 INFO ckpt None 2022-11-19 21:23:27,421 INFO pretrained_model None 2022-11-19 21:23:27,421 INFO launcher none 2022-11-19 21:23:27,421 INFO tcp_port 18888 2022-11-19 21:23:27,421 INFO sync_bn False 2022-11-19 21:23:27,421 INFO fix_random_seed False 2022-11-19 21:23:27,421 INFO ckpt_save_interval 1 2022-11-19 21:23:27,421 INFO local_rank 0 2022-11-19 21:23:27,421 INFO max_ckpt_save_num 30 2022-11-19 21:23:27,421 INFO merge_all_iters_to_one_epoch False 2022-11-19 21:23:27,421 INFO set_cfgs None 2022-11-19 21:23:27,421 INFO max_waiting_mins 0 2022-11-19 21:23:27,422 INFO start_epoch 0 2022-11-19 21:23:27,422 INFO save_to_file False 2022-11-19 21:23:27,422 INFO cfg.ROOT_DIR: /content/gdrive/MyDrive/IITH_VCA/CaDDN 2022-11-19 21:23:27,422 INFO cfg.LOCAL_RANK: 0 2022-11-19 21:23:27,422 INFO cfg.CLASS_NAMES: ['Car', 'Pedestrian', 'Cyclist'] 2022-11-19 21:23:27,422 INFO
    cfg.DATA_CONFIG = edict() 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.DATASET: KittiDataset 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.DATA_PATH: /gdrive/MyDrive/3D_Obj/CaDDN/data/kitti 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.POINT_CLOUD_RANGE: [2, -30.08, -3.0, 46.8, 30.08, 1.0] 2022-11-19 21:23:27,422 INFO
    cfg.DATA_CONFIG.DATA_SPLIT = edict() 2022-11-19 21:23:27,422 INFO cfg.DATA_CONFIG.DATA_SPLIT.train: train 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.DATA_SPLIT.test: test 2022-11-19 21:23:27,423 INFO
    cfg.DATA_CONFIG.INFO_PATH = edict() 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.INFO_PATH.train: ['kitti_infos_trainval.pkl'] 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.INFO_PATH.test: ['kitti_infos_test.pkl'] 2022-11-19 21:23:27,423 INFO cfg.DATA_CONFIG.FOV_POINTS_ONLY: True 2022-11-19 21:23:27,423 INFO
    cfg.DATA_CONFIG.DATA_AUGMENTOR = edict() 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.DISABLE_AUG_LIST: ['placeholder'] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.AUG_CONFIG_LIST: [{'NAME': 'random_image_flip', 'ALONG_AXIS_LIST': ['horizontal']}] 2022-11-19 21:23:27,424 INFO
    cfg.DATA_CONFIG.POINT_FEATURE_ENCODING = edict() 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.encoding_type: absolute_coordinates_encoding 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.used_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.src_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.DATA_PROCESSOR: [{'NAME': 'mask_points_and_boxes_outside_range', 'REMOVE_OUTSIDE_BOXES': True}, {'NAME': 'calculate_grid_size', 'VOXEL_SIZE': [0.16, 0.16, 0.16]}] 2022-11-19 21:23:27,424 INFO cfg.DATA_CONFIG.BASE_CONFIG: cfgs/dataset_configs/kitti_dataset.yaml 2022-11-19 21:23:27,425 INFO
    cfg.DATA_CONFIG.IMAGE = edict() 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.IMAGE.ENABLED: True 2022-11-19 21:23:27,425 INFO
    cfg.DATA_CONFIG.IMAGE.COLLATE = edict() 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.MODE: Pad 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.CONSTANT_VALUES: [0.485, 0.456, 0.406] 2022-11-19 21:23:27,425 INFO
    cfg.DATA_CONFIG.DEPTH_MAP = edict() 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.DEPTH_MAP.ENABLED: True 2022-11-19 21:23:27,425 INFO cfg.DATA_CONFIG.DEPTH_MAP.DOWNSAMPLE_FACTOR: 4 2022-11-19 21:23:27,426 INFO
    cfg.DATA_CONFIG.CALIB = edict() 2022-11-19 21:23:27,426 INFO cfg.DATA_CONFIG.CALIB.ENABLED: True 2022-11-19 21:23:27,426 INFO
    cfg.MODEL = edict() 2022-11-19 21:23:27,426 INFO cfg.MODEL.NAME: CaDDN 2022-11-19 21:23:27,426 INFO
    cfg.MODEL.FFE = edict() 2022-11-19 21:23:27,426 INFO cfg.MODEL.FFE.NAME: DepthFFE 2022-11-19 21:23:27,426 INFO
    cfg.MODEL.FFE.DDN = edict() 2022-11-19 21:23:27,426 INFO cfg.MODEL.FFE.DDN.NAME: DDNDeepLabV3 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.DDN.BACKBONE_NAME: ResNet101 2022-11-19 21:23:27,427 INFO
    cfg.MODEL.FFE.DDN.ARGS = edict() 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.DDN.ARGS.feat_extract_layer: layer1 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.DDN.ARGS.pretrained_path: ../checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:23:27,427 INFO
    cfg.MODEL.FFE.CHANNEL_REDUCE = edict() 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.in_channels: 256 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.out_channels: 64 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.kernel_size: 1 2022-11-19 21:23:27,427 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.stride: 1 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.bias: False 2022-11-19 21:23:27,428 INFO
    cfg.MODEL.FFE.DISCRETIZE = edict() 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.mode: LID 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.num_bins: 80 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.depth_min: 2.0 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DISCRETIZE.depth_max: 46.8 2022-11-19 21:23:27,428 INFO
    cfg.MODEL.FFE.DDN_LOSS = edict() 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DDN_LOSS.weight: 3.0 2022-11-19 21:23:27,428 INFO cfg.MODEL.FFE.DDN_LOSS.alpha: 0.25 2022-11-19 21:23:27,429 INFO cfg.MODEL.FFE.DDN_LOSS.gamma: 2.0 2022-11-19 21:23:27,429 INFO cfg.MODEL.FFE.DDN_LOSS.fg_weight: 13 2022-11-19 21:23:27,429 INFO cfg.MODEL.FFE.DDN_LOSS.bg_weight: 1 2022-11-19 21:23:27,429 INFO
    cfg.MODEL.F2V = edict() 2022-11-19 21:23:27,429 INFO cfg.MODEL.F2V.NAME: FrustumToVoxel 2022-11-19 21:23:27,429 INFO
    cfg.MODEL.F2V.SAMPLER = edict() 2022-11-19 21:23:27,429 INFO cfg.MODEL.F2V.SAMPLER.mode: bilinear 2022-11-19 21:23:27,429 INFO cfg.MODEL.F2V.SAMPLER.padding_mode: zeros 2022-11-19 21:23:27,429 INFO
    cfg.MODEL.MAP_TO_BEV = edict() 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.NAME: Conv2DCollapse 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.NUM_BEV_FEATURES: 64 2022-11-19 21:23:27,430 INFO
    cfg.MODEL.MAP_TO_BEV.ARGS = edict() 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.ARGS.kernel_size: 1 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.ARGS.stride: 1 2022-11-19 21:23:27,430 INFO cfg.MODEL.MAP_TO_BEV.ARGS.bias: False 2022-11-19 21:23:27,430 INFO
    cfg.MODEL.BACKBONE_2D = edict() 2022-11-19 21:23:27,430 INFO cfg.MODEL.BACKBONE_2D.NAME: BaseBEVBackbone 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.LAYER_NUMS: [10, 10, 10] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.LAYER_STRIDES: [2, 2, 2] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.NUM_FILTERS: [64, 128, 256] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.UPSAMPLE_STRIDES: [1, 2, 4] 2022-11-19 21:23:27,431 INFO cfg.MODEL.BACKBONE_2D.NUM_UPSAMPLE_FILTERS: [128, 128, 128] 2022-11-19 21:23:27,431 INFO
    cfg.MODEL.DENSE_HEAD = edict() 2022-11-19 21:23:27,431 INFO cfg.MODEL.DENSE_HEAD.NAME: AnchorHeadSingle 2022-11-19 21:23:27,431 INFO cfg.MODEL.DENSE_HEAD.CLASS_AGNOSTIC: False 2022-11-19 21:23:27,431 INFO cfg.MODEL.DENSE_HEAD.USE_DIRECTION_CLASSIFIER: True 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.DIR_OFFSET: 0.78539 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.DIR_LIMIT_OFFSET: 0.0 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.NUM_DIR_BINS: 2 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.ANCHOR_GENERATOR_CONFIG: [{'class_name': 'Car', 'anchor_sizes': [[3.9, 1.6, 1.56]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-1.78], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.6, 'unmatched_threshold': 0.45}, {'class_name': 'Pedestrian', 'anchor_sizes': [[0.8, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}, {'class_name': 'Cyclist', 'anchor_sizes': [[1.76, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}] 2022-11-19 21:23:27,432 INFO
    cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG = edict() 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NAME: AxisAlignedTargetAssigner 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.POS_FRACTION: -1.0 2022-11-19 21:23:27,432 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.SAMPLE_SIZE: 512 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NORM_BY_NUM_EXAMPLES: False 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.MATCH_HEIGHT: False 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.BOX_CODER: ResidualCoder 2022-11-19 21:23:27,433 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG = edict() 2022-11-19 21:23:27,433 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS = edict() 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.cls_weight: 1.0 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.loc_weight: 2.0 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.dir_weight: 0.2 2022-11-19 21:23:27,433 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.code_weights: [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0] 2022-11-19 21:23:27,434 INFO
    cfg.MODEL.POST_PROCESSING = edict() 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.RECALL_THRESH_LIST: [0.3, 0.5, 0.7] 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.SCORE_THRESH: 0.1 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.OUTPUT_RAW_SCORE: False 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.EVAL_METRIC: kitti 2022-11-19 21:23:27,434 INFO
    cfg.MODEL.POST_PROCESSING.NMS_CONFIG = edict() 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.MULTI_CLASSES_NMS: False 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_TYPE: nms_gpu 2022-11-19 21:23:27,434 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_THRESH: 0.01 2022-11-19 21:23:27,435 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_PRE_MAXSIZE: 4096 2022-11-19 21:23:27,435 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_POST_MAXSIZE: 500 2022-11-19 21:23:27,435 INFO
    cfg.OPTIMIZATION = edict() 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.BATCH_SIZE_PER_GPU: 4 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.NUM_EPOCHS: 80 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.OPTIMIZER: adam_onecycle 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.LR: 0.001 2022-11-19 21:23:27,435 INFO cfg.OPTIMIZATION.WEIGHT_DECAY: 0.01 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.MOMENTUM: 0.9 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.MOMS: [0.95, 0.85] 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.PCT_START: 0.4 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.DIV_FACTOR: 10 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.DECAY_STEP_LIST: [35, 45] 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.LR_DECAY: 0.1 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.LR_CLIP: 1e-07 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.LR_WARMUP: False 2022-11-19 21:23:27,436 INFO cfg.OPTIMIZATION.WARMUP_EPOCH: 1 2022-11-19 21:23:27,437 INFO cfg.OPTIMIZATION.GRAD_NORM_CLIP: 10 2022-11-19 21:23:27,437 INFO cfg.TAG: CaDDN 2022-11-19 21:23:27,437 INFO cfg.EXP_GROUP_PATH: content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models 2022-11-19 21:23:28,525 INFO Loading KITTI dataset 2022-11-19 21:23:28,526 INFO Total samples for KITTI dataset: 0 Traceback (most recent call last): File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/train.py", line 198, in main() File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/train.py", line 112, in main total_epochs=args.epochs File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/pcdet/datasets/init.py", line 69, in build_dataloader drop_last=False, sampler=sampler, timeout=0 File "/usr/local/lib/python3.7/dist-packages/torch/utils/data/dataloader.py", line 213, in init sampler = RandomSampler(dataset) File "/usr/local/lib/python3.7/dist-packages/torch/utils/data/sampler.py", line 94, in init "value, but got num_samples={}".format(self.num_samples)) ValueError: num_samples should be a positive integer value, but got num_samples=0

    Screen shots of the same: image

    Error screen shot: image

    opened by prabhata07 1
  • KeyError: 'model_state'

    KeyError: 'model_state'

    Facing KeyError: 'model_state' while trying to test with "Pretrained model" Command used:

    !python /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py --cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml --batch_size 1 --ckpt /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth

    Environemnt: Collab

    Model: as given in the github repo deeplabv3_resnet101_coco-586e9e4e.pth

    Error log:

    2022-11-19 21:16:39,034 INFO Start logging 2022-11-19 21:16:39,035 INFO CUDA_VISIBLE_DEVICES=ALL 2022-11-19 21:16:39,035 INFO cfg_file /content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models/CaDDN.yaml 2022-11-19 21:16:39,035 INFO batch_size 1 2022-11-19 21:16:39,035 INFO workers 4 2022-11-19 21:16:39,035 INFO extra_tag default 2022-11-19 21:16:39,035 INFO ckpt /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:16:39,035 INFO launcher none 2022-11-19 21:16:39,036 INFO tcp_port 18888 2022-11-19 21:16:39,036 INFO local_rank 0 2022-11-19 21:16:39,036 INFO set_cfgs None 2022-11-19 21:16:39,036 INFO max_waiting_mins 30 2022-11-19 21:16:39,036 INFO start_epoch 0 2022-11-19 21:16:39,036 INFO eval_tag default 2022-11-19 21:16:39,036 INFO eval_all False 2022-11-19 21:16:39,036 INFO ckpt_dir None 2022-11-19 21:16:39,036 INFO save_to_file False 2022-11-19 21:16:39,037 INFO cfg.ROOT_DIR: /content/gdrive/MyDrive/IITH_VCA/CaDDN 2022-11-19 21:16:39,037 INFO cfg.LOCAL_RANK: 0 2022-11-19 21:16:39,037 INFO cfg.CLASS_NAMES: ['Car', 'Pedestrian', 'Cyclist'] 2022-11-19 21:16:39,037 INFO
    cfg.DATA_CONFIG = edict() 2022-11-19 21:16:39,037 INFO cfg.DATA_CONFIG.DATASET: KittiDataset 2022-11-19 21:16:39,037 INFO cfg.DATA_CONFIG.DATA_PATH: /gdrive/MyDrive/3D_Obj/CaDDN/data/kitti 2022-11-19 21:16:39,037 INFO cfg.DATA_CONFIG.POINT_CLOUD_RANGE: [2, -30.08, -3.0, 46.8, 30.08, 1.0] 2022-11-19 21:16:39,037 INFO
    cfg.DATA_CONFIG.DATA_SPLIT = edict() 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_SPLIT.train: train 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_SPLIT.test: test 2022-11-19 21:16:39,038 INFO
    cfg.DATA_CONFIG.INFO_PATH = edict() 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.INFO_PATH.train: ['kitti_infos_trainval.pkl'] 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.INFO_PATH.test: ['kitti_infos_test.pkl'] 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.FOV_POINTS_ONLY: True 2022-11-19 21:16:39,038 INFO
    cfg.DATA_CONFIG.DATA_AUGMENTOR = edict() 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.DISABLE_AUG_LIST: ['placeholder'] 2022-11-19 21:16:39,038 INFO cfg.DATA_CONFIG.DATA_AUGMENTOR.AUG_CONFIG_LIST: [{'NAME': 'random_image_flip', 'ALONG_AXIS_LIST': ['horizontal']}] 2022-11-19 21:16:39,038 INFO
    cfg.DATA_CONFIG.POINT_FEATURE_ENCODING = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.encoding_type: absolute_coordinates_encoding 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.used_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.POINT_FEATURE_ENCODING.src_feature_list: ['x', 'y', 'z', 'intensity'] 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.DATA_PROCESSOR: [{'NAME': 'mask_points_and_boxes_outside_range', 'REMOVE_OUTSIDE_BOXES': True}, {'NAME': 'calculate_grid_size', 'VOXEL_SIZE': [0.16, 0.16, 0.16]}] 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.BASE_CONFIG: cfgs/dataset_configs/kitti_dataset.yaml 2022-11-19 21:16:39,039 INFO
    cfg.DATA_CONFIG.IMAGE = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.IMAGE.ENABLED: True 2022-11-19 21:16:39,039 INFO
    cfg.DATA_CONFIG.IMAGE.COLLATE = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.MODE: Pad 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.IMAGE.COLLATE.CONSTANT_VALUES: [0.485, 0.456, 0.406] 2022-11-19 21:16:39,039 INFO
    cfg.DATA_CONFIG.DEPTH_MAP = edict() 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.DEPTH_MAP.ENABLED: True 2022-11-19 21:16:39,039 INFO cfg.DATA_CONFIG.DEPTH_MAP.DOWNSAMPLE_FACTOR: 4 2022-11-19 21:16:39,040 INFO
    cfg.DATA_CONFIG.CALIB = edict() 2022-11-19 21:16:39,040 INFO cfg.DATA_CONFIG.CALIB.ENABLED: True 2022-11-19 21:16:39,040 INFO
    cfg.MODEL = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.NAME: CaDDN 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.NAME: DepthFFE 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE.DDN = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.NAME: DDNDeepLabV3 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.BACKBONE_NAME: ResNet101 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE.DDN.ARGS = edict() 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.ARGS.feat_extract_layer: layer1 2022-11-19 21:16:39,040 INFO cfg.MODEL.FFE.DDN.ARGS.pretrained_path: ../checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:16:39,040 INFO
    cfg.MODEL.FFE.CHANNEL_REDUCE = edict() 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.in_channels: 256 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.out_channels: 64 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.kernel_size: 1 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.stride: 1 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.CHANNEL_REDUCE.bias: False 2022-11-19 21:16:39,041 INFO
    cfg.MODEL.FFE.DISCRETIZE = edict() 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.mode: LID 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.num_bins: 80 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.depth_min: 2.0 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DISCRETIZE.depth_max: 46.8 2022-11-19 21:16:39,041 INFO
    cfg.MODEL.FFE.DDN_LOSS = edict() 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DDN_LOSS.weight: 3.0 2022-11-19 21:16:39,041 INFO cfg.MODEL.FFE.DDN_LOSS.alpha: 0.25 2022-11-19 21:16:39,042 INFO cfg.MODEL.FFE.DDN_LOSS.gamma: 2.0 2022-11-19 21:16:39,042 INFO cfg.MODEL.FFE.DDN_LOSS.fg_weight: 13 2022-11-19 21:16:39,042 INFO cfg.MODEL.FFE.DDN_LOSS.bg_weight: 1 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.F2V = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.F2V.NAME: FrustumToVoxel 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.F2V.SAMPLER = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.F2V.SAMPLER.mode: bilinear 2022-11-19 21:16:39,042 INFO cfg.MODEL.F2V.SAMPLER.padding_mode: zeros 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.MAP_TO_BEV = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.MAP_TO_BEV.NAME: Conv2DCollapse 2022-11-19 21:16:39,042 INFO cfg.MODEL.MAP_TO_BEV.NUM_BEV_FEATURES: 64 2022-11-19 21:16:39,042 INFO
    cfg.MODEL.MAP_TO_BEV.ARGS = edict() 2022-11-19 21:16:39,042 INFO cfg.MODEL.MAP_TO_BEV.ARGS.kernel_size: 1 2022-11-19 21:16:39,043 INFO cfg.MODEL.MAP_TO_BEV.ARGS.stride: 1 2022-11-19 21:16:39,043 INFO cfg.MODEL.MAP_TO_BEV.ARGS.bias: False 2022-11-19 21:16:39,043 INFO
    cfg.MODEL.BACKBONE_2D = edict() 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.NAME: BaseBEVBackbone 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.LAYER_NUMS: [10, 10, 10] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.LAYER_STRIDES: [2, 2, 2] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.NUM_FILTERS: [64, 128, 256] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.UPSAMPLE_STRIDES: [1, 2, 4] 2022-11-19 21:16:39,043 INFO cfg.MODEL.BACKBONE_2D.NUM_UPSAMPLE_FILTERS: [128, 128, 128] 2022-11-19 21:16:39,043 INFO
    cfg.MODEL.DENSE_HEAD = edict() 2022-11-19 21:16:39,043 INFO cfg.MODEL.DENSE_HEAD.NAME: AnchorHeadSingle 2022-11-19 21:16:39,043 INFO cfg.MODEL.DENSE_HEAD.CLASS_AGNOSTIC: False 2022-11-19 21:16:39,043 INFO cfg.MODEL.DENSE_HEAD.USE_DIRECTION_CLASSIFIER: True 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.DIR_OFFSET: 0.78539 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.DIR_LIMIT_OFFSET: 0.0 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.NUM_DIR_BINS: 2 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.ANCHOR_GENERATOR_CONFIG: [{'class_name': 'Car', 'anchor_sizes': [[3.9, 1.6, 1.56]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-1.78], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.6, 'unmatched_threshold': 0.45}, {'class_name': 'Pedestrian', 'anchor_sizes': [[0.8, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}, {'class_name': 'Cyclist', 'anchor_sizes': [[1.76, 0.6, 1.73]], 'anchor_rotations': [0, 1.57], 'anchor_bottom_heights': [-0.6], 'align_center': False, 'feature_map_stride': 2, 'matched_threshold': 0.5, 'unmatched_threshold': 0.35}] 2022-11-19 21:16:39,044 INFO
    cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG = edict() 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NAME: AxisAlignedTargetAssigner 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.POS_FRACTION: -1.0 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.SAMPLE_SIZE: 512 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.NORM_BY_NUM_EXAMPLES: False 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.MATCH_HEIGHT: False 2022-11-19 21:16:39,044 INFO cfg.MODEL.DENSE_HEAD.TARGET_ASSIGNER_CONFIG.BOX_CODER: ResidualCoder 2022-11-19 21:16:39,044 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG = edict() 2022-11-19 21:16:39,045 INFO
    cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS = edict() 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.cls_weight: 1.0 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.loc_weight: 2.0 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.dir_weight: 0.2 2022-11-19 21:16:39,045 INFO cfg.MODEL.DENSE_HEAD.LOSS_CONFIG.LOSS_WEIGHTS.code_weights: [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0] 2022-11-19 21:16:39,045 INFO
    cfg.MODEL.POST_PROCESSING = edict() 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.RECALL_THRESH_LIST: [0.3, 0.5, 0.7] 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.SCORE_THRESH: 0.1 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.OUTPUT_RAW_SCORE: False 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.EVAL_METRIC: kitti 2022-11-19 21:16:39,045 INFO
    cfg.MODEL.POST_PROCESSING.NMS_CONFIG = edict() 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.MULTI_CLASSES_NMS: False 2022-11-19 21:16:39,045 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_TYPE: nms_gpu 2022-11-19 21:16:39,046 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_THRESH: 0.01 2022-11-19 21:16:39,046 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_PRE_MAXSIZE: 4096 2022-11-19 21:16:39,046 INFO cfg.MODEL.POST_PROCESSING.NMS_CONFIG.NMS_POST_MAXSIZE: 500 2022-11-19 21:16:39,046 INFO
    cfg.OPTIMIZATION = edict() 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.BATCH_SIZE_PER_GPU: 4 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.NUM_EPOCHS: 80 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.OPTIMIZER: adam_onecycle 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.LR: 0.001 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.WEIGHT_DECAY: 0.01 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.MOMENTUM: 0.9 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.MOMS: [0.95, 0.85] 2022-11-19 21:16:39,046 INFO cfg.OPTIMIZATION.PCT_START: 0.4 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.DIV_FACTOR: 10 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.DECAY_STEP_LIST: [35, 45] 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.LR_DECAY: 0.1 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.LR_CLIP: 1e-07 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.LR_WARMUP: False 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.WARMUP_EPOCH: 1 2022-11-19 21:16:39,047 INFO cfg.OPTIMIZATION.GRAD_NORM_CLIP: 10 2022-11-19 21:16:39,048 INFO cfg.TAG: CaDDN 2022-11-19 21:16:39,048 INFO cfg.EXP_GROUP_PATH: content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/cfgs/kitti_models 2022-11-19 21:16:39,048 INFO Loading KITTI dataset 2022-11-19 21:16:39,048 INFO Total samples for KITTI dataset: 0 /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth 2022-11-19 21:16:42,793 INFO ==> Loading parameters from checkpoint /content/gdrive/MyDrive/IITH_VCA/CaDDN/checkpoints/deeplabv3_resnet101_coco-586e9e4e.pth to GPU Traceback (most recent call last): File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py", line 200, in main() File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py", line 196, in main eval_single_ckpt(model, test_loader, args, eval_output_dir, logger, epoch_id, dist_test=dist_test) File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/tools/test.py", line 58, in eval_single_ckpt model.load_params_from_file(filename=args.ckpt, logger=logger, to_cpu=dist_test) File "/content/gdrive/MyDrive/IITH_VCA/CaDDN/pcdet/models/detectors/detector3d_template.py", line 359, in load_params_from_file model_state_disk = checkpoint['model_state'] KeyError: 'model_state'

    opened by prabhata07 1
  • the result is not reproduced

    the result is not reproduced

    I trained the model with 2 Tesla V100 GPUs in the default configuration and got [email protected], [email protected], [email protected] 21.203, 15.1487,12.7407 respectively, which is lower than the 23.77,16.07, 13.61 you declared. But when I use the pre-trained weight, I can get 23.77,16.07, 13.61. So, are there any training tricks that have not been published?

    opened by leo038 0
Releases(v0.3.1)
Owner
Toronto Robotics and AI Laboratory
Toronto Robotics and AI Laboratory
Deep and online learning with spiking neural networks in Python

Introduction The brain is the perfect place to look for inspiration to develop more efficient neural networks. One of the main differences with modern

Jason Eshraghian 447 Jan 03, 2023
code for Grapadora research paper experimentation

Road feature embedding selection method Code for research paper experimentation Abstract Traffic forecasting models rely on data that needs to be sens

Eric López Manibardo 0 May 26, 2022
Gym environment for FLIPIT: The Game of "Stealthy Takeover"

gym-flipit Gym environment for FLIPIT: The Game of "Stealthy Takeover" invented by Marten van Dijk, Ari Juels, Alina Oprea, and Ronald L. Rivest. Desi

Lisa Oakley 2 Dec 15, 2021
Code needed to reproduce the examples found in "The Temporal Robustness of Stochastic Signals"

The Temporal Robustness of Stochastic Signals Code needed to reproduce the examples found in "The Temporal Robustness of Stochastic Signals" Case stud

0 Oct 28, 2021
Unofficial Implementation of Oboe (SIGCOMM'18').

Oboe-Reproduce This is the unofficial implementation of the paper "Oboe: Auto-tuning video ABR algorithms to network conditions, Zahaib Akhtar, Yun Se

Tianchi Huang 13 Nov 04, 2022
CondenseNet: Light weighted CNN for mobile devices

CondenseNets This repository contains the code (in PyTorch) for "CondenseNet: An Efficient DenseNet using Learned Group Convolutions" paper by Gao Hua

Shichen Liu 690 Nov 30, 2022
The Few-Shot Bot: Prompt-Based Learning for Dialogue Systems

Few-Shot Bot: Prompt-Based Learning for Dialogue Systems This repository includes the dataset, experiments results, and code for the paper: Few-Shot B

Andrea Madotto 103 Dec 28, 2022
3D position tracking for soccer players with multi-camera videos

This repo contains a full pipeline to support 3D position tracking of soccer players, with multi-view calibrated moving/fixed video sequences as inputs.

Yuchang Jiang 72 Dec 27, 2022
Detectron2 is FAIR's next-generation platform for object detection and segmentation.

Detectron2 is Facebook AI Research's next generation software system that implements state-of-the-art object detection algorithms. It is a ground-up r

Facebook Research 23.3k Jan 08, 2023
Liquid Warping GAN with Attention: A Unified Framework for Human Image Synthesis

Liquid Warping GAN with Attention: A Unified Framework for Human Image Synthesis, including human motion imitation, appearance transfer, and novel view synthesis. Currently the paper is under review

2.3k Jan 05, 2023
PoseCamera is python based SDK for human pose estimation through RGB webcam.

PoseCamera PoseCamera is python based SDK for human pose estimation through RGB webcam. Install install posecamera package through pip pip install pos

WonderTree 7 Jul 20, 2021
Pose estimation with MoveNet Lightning

Pose Estimation With MoveNet Lightning MoveNet is the TensorFlow pre-trained model that identifies 17 different key points of the human body. It is th

Yash Vora 2 Jan 04, 2022
Hierarchical Motion Encoder-Decoder Network for Trajectory Forecasting (HMNet)

Hierarchical Motion Encoder-Decoder Network for Trajectory Forecasting (HMNet) Our paper: https://arxiv.org/abs/2111.13324 We will release the complet

15 Oct 17, 2022
Multi-modal Text Recognition Networks: Interactive Enhancements between Visual and Semantic Features

Multi-modal Text Recognition Networks: Interactive Enhancements between Visual and Semantic Features | paper | Official PyTorch implementation for Mul

48 Dec 28, 2022
automated systems to assist guarding corona Virus precautions for Closed Rooms (e.g. Halls, offices, etc..)

Automatic-precautionary-guard automated systems to assist guarding corona Virus precautions for Closed Rooms (e.g. Halls, offices, etc..) what is this

badra 0 Jan 06, 2022
Simulation of the solar system using various nummerical methods

solar-system Simulation of the solar system using various nummerical methods Download the repo Make shure matplotlib, scipy etc. are installed execute

Caspar 7 Jul 15, 2022
Official PyTorch implementation of "Synthesis of Screentone Patterns of Manga Characters"

Manga Character Screentone Synthesis Official PyTorch implementation of "Synthesis of Screentone Patterns of Manga Characters" presented in IEEE ISM 2

Tsubota 2 Nov 20, 2021
The implementation of "Bootstrapping Semantic Segmentation with Regional Contrast".

ReCo - Regional Contrast This repository contains the source code of ReCo and baselines from the paper, Bootstrapping Semantic Segmentation with Regio

Shikun Liu 128 Dec 30, 2022
This program uses trial auth token of Azure Cognitive Services to do speech synthesis for you.

🗣️ aspeak A simple text-to-speech client using azure TTS API(trial). 😆 TL;DR: This program uses trial auth token of Azure Cognitive Services to do s

Levi Zim 359 Jan 05, 2023
ByteTrack(Multi-Object Tracking by Associating Every Detection Box)のPythonでのONNX推論サンプル

ByteTrack-ONNX-Sample ByteTrack(Multi-Object Tracking by Associating Every Detection Box)のPythonでのONNX推論サンプルです。 ONNXに変換したモデルも同梱しています。 変換自体を試したい方はByteT

KazuhitoTakahashi 16 Oct 26, 2022