Giter VIP home page Giter VIP logo

simtrack's People

Contributors

chenxuluo avatar xiaodongyang avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

simtrack's Issues

motion updating?

can you tell me where you do the motion updating in your code? I only find the ego-motion update in your val_nusc_tracking.py

No recurrence nuscenes experiments

**Dear authors:
I really appreciate your great work, and i'm tring to use it in some of my projects. I train your model on 4 3090 and this model is trained for 187 epochs on nuscenes, but the tracking result is bad, just like this picture, What is the cause?
image

Thanks for your answer!**

It seems that point_pillars_tracking.py is needed

when i train the model using command line in the tutorial, I got this :
Traceback (most recent call last): File "./tools/train.py", line 13, in <module> from det3d.models import build_detector File "/home/lz/task3/simtrack/simtrack/det3d/models/__init__.py", line 13, in <module> from .detectors import * # noqa: F401,F403 File "/home/lz/task3/simtrack/simtrack/det3d/models/detectors/__init__.py", line 3, in <module> from .point_pillars_tracking import PointPillarsTracking ModuleNotFoundError: No module named 'det3d.models.detectors.point_pillars_tracking'
The error occurs because there is no point_pillars_tracking.py in /home/lz/task3/simtrack/simtrack/det3d/models/detectors, but a voxelnet.py which has been commented out.
Can anybodyhelp me ?

Train pointpillars model with dynamic voxelizer

Nice work. I noticed that your implemented DynamicPillarFeatureNet requests the raw points as an input and not the output of the voxelizer. I wonder how to train using this pipeline?

Thanks!

parameters init?

I notice that in "simtrack/det3d/models/detectors/single_stage.py" line 32, init_weights is Annotated, why? how you init the parameters?

The difference with centerpoint.

I have one question, does the model structure of the simTrack is completely the same as CenterPoint?
And the only difference is the post-processing. Am I right? Do I miss something important?

metric jitter on nuscene valset

The metirc result is not deterministic.

cmd:
python tools/val_nusc_tracking.py examples/point_pillars/configs/nusc_all_pp_centernet_tracking.py --checkpoint model_zoo/simtrack_pillar.pth --work_dir word_dirs/baseline

For fast, I just use nuscenes v1.0-mini set to test.
It returns the different metric results when I run the above cmd.

image
image
image

Visualization of the result

Dear authors:
I really appreciate your great work, and i'm tring to use it in some of my projects. I am wondering how to visualize your result just like the first gif of your README.
Thanks for your answer!

KeyError: 'PointPillars is not in the detector registry'

hello, when I tried to train a model, I just got this error, the traceback is following:

2022-04-24 18:16:11,886 - INFO - Distributed training: False
2022-04-24 18:16:11,886 - INFO - torch.backends.cudnn.benchmark: False
2022-04-24 18:16:11,900 - INFO - Backup source files to SAVE_DIR/det3d
Traceback (most recent call last):
File "./tools/train.py", line 146, in
main()
File "./tools/train.py", line 119, in main
model = build_detector(cfg.model, train_cfg=cfg.train_cfg, test_cfg=cfg.test_cfg)
File "/simtrack/det3d/models/builder.py", line 53, in build_detector
return build(cfg, DETECTORS, dict(train_cfg=train_cfg, test_cfg=test_cfg))
File "/simtrack/det3d/models/builder.py", line 21, in build
return build_from_cfg(cfg, registry, default_args)
File "/simtrack/det3d/utils/registry.py", line 66, in build_from_cfg
"{} is not in the {} registry".format(obj_type, registry.name)
KeyError: 'PointPillars is not in the detector registry'
Traceback (most recent call last):
File "/anaconda3/envs/simtrack/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"main", mod_spec)
File "/anaconda3/envs/simtrack/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/anaconda3/envs/simtrack/lib/python3.6/site-packages/torch/distributed/launch.py", line 260, in
main()
File "/anaconda3/envs/simtrack/lib/python3.6/site-packages/torch/distributed/launch.py", line 256, in main
cmd=cmd)
subprocess.CalledProcessError: Command '[/anaconda3/envs/simtrack/bin/python', '-u', './tools/train.py', '--local_rank=0', 'examples/point_pillars/configs/nusc_all_pp_centernet_tracking.py', '--work_dir', 'SAVE_DIR']' returned non-zero exit status 1.

the order is: python -m torch.distributed.launch --nproc_per_node=1 ./tools/train.py examples/point_pillars/configs/nusc_all_pp_centernet_tracking.py --work_dir SAVE_DIR

About motion compensation

Thanks your fantastic work first, and I have a problem while view the code, did you do anything about motion compensation or undistort the pointcloud? I saw you use the matrix between two sweep lidar points to concat the points, but I did not see motion process when concat two sweeps lidar points, maybe I missed. If do not do this, is there any distort problem for the objects with big velocity?

Look forward to your kind reply,Thank you

motion update branch

I am now trying to reproduce this work. I am curious about the motion update branch, especially how to implement update if multiple frames are lost. The update here seems to only predict the detection target in the previous frame, but How to predict when multiple frames are lost. The results of the paper show the prediction results when multiple frames of the vehicle are lost. Can you tell me how this operates or what part of the code is there?

val_nusc_tracking.py AssertionError

I used the default model to eval, but triger a warming and a AssertionError, sincerely look forward to your answer. Also, when will you release the full version of the code?

python ./tools/val_nusc_tracking.py examples/point_pillars/configs/nusc_all_pp_centernet_tracking.py --checkpoint model_zoo/simtrack_pillar.pth --work_dir /data/simtrack_output/

Use HM Bias: -2.19
====== Loading NuScenes tables for version v1.0-trainval...
23 category,
8 attribute, 4 visibility, 64386 instance,
12 sensor,
10200 calibrated_sensor,
2631083 ego_pose,
68 log,
850 scene, 34149 sample,
2631083 sample_data,
1166187 sample_annotation,
4 map,
Done loading in 135.2 seconds.

Reverse indexing ...
Done reverse indexing in 10.4 seconds.

/data/simtrack/det3d/core/bbox/geometry.py:160: NumbaWarning:
Compilation is falling back to object mode WITH looplifting enabled because Function "points_in_convex_polygon_jit" failed type inference due to: No implementation of function Function() found for signature:

getitem(array(float64, 3d, C), Tuple(slice<a:b>, list(int64)<iv=None>, slice<a:b>))

There are 22 candidate implementations:

  • Of which 20 did not match due to:
    Overload of function 'getitem': File: : Line N/A.
    With argument(s): '(array(float64, 3d, C), Tuple(slice<a:b>, list(int64)<iv=None>, slice<a:b>))':
    No match.
  • Of which 2 did not match due to:
    Overload in function 'GetItemBuffer.generic': File: numba/core/typing/arraydecl.py: Line 166.
    With argument(s): '(array(float64, 3d, C), Tuple(slice<a:b>, list(int64)<iv=None>, slice<a:b>))':
    Rejected as the implementation raised a specific error:
    NumbaTypeError: unsupported array index type list(int64)<iv=None> in Tuple(slice<a:b>, list(int64)<iv=None>, slice<a:b>)
    raised from /home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/numba/core/typing/arraydecl.py:73

During: typing of intrinsic-call at /data/simtrack/det3d/core/bbox/geometry.py (179)

File "det3d/core/bbox/geometry.py", line 179:
def points_in_convex_polygon_jit(points, polygon, clockwise=True):

:,
[num_points_of_polygon - 1] + list(range(num_points_of_polygon - 1)),
^

@numba.jit
/data/simtrack/det3d/core/bbox/geometry.py:160: NumbaWarning:
Compilation is falling back to object mode WITHOUT looplifting enabled because Function "points_in_convex_polygon_jit" failed type inference due to: Cannot determine Numba type of <class 'numba.core.dispatcher.LiftedLoop'>
File "det3d/core/bbox/geometry.py", line 196: def points_in_convex_polygon_jit(points, polygon, clockwise=True):

cross = 0.0 for i in range(num_points):
^
@numba.jit
/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/numba/core/object_mode_passes.py:152: NumbaWarning: Function "points_in_convex_polygon_jit" was compiled in object mode without forceobj=True, but has lifted loops. File "det3d/core/bbox/geometry.py", line 171:
def points_in_convex_polygon_jit(points, polygon, clockwise=True):

# first convert polygon to directed lines
num_points_of_polygon = polygon.shape[1]
^
state.func_ir.loc))
/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/numba/core/object_mode_passes.py:162: NumbaDeprecationWarning:
Fall-back from the nopython compilation path to the object mode compilation path has been detected, this is deprecated behaviour.

For more information visit https://numba.readthedocs.io/en/stable/reference/deprecation.html#deprecation-of-object-mode-fall-back-behaviour-when-using-jit

File "det3d/core/bbox/geometry.py", line 171:
def points_in_convex_polygon_jit(points, polygon, clockwise=True):

# first convert polygon to directed lines
num_points_of_polygon = polygon.shape[1]
^

state.func_ir.loc))

Loading NuScenes tables for version v1.0-trainval...
23 category,
8 attribute,
4 visibility,
64386 instance,
12 sensor,
10200 calibrated_sensor,
2631083 ego_pose,
68 log,
850 scene,
34149 sample,
2631083 sample_data,
1166187 sample_annotation,
4 map,
Done loading in 35.9 seconds.

Reverse indexing ...
Done reverse indexing in 9.5 seconds.

Finish generate predictions for testset, save to /data/simtrack_output/tracking_results.json

Loading NuScenes tables for version v1.0-trainval...
23 category, 8 attribute, 4 visibility,
64386 instance,
12 sensor, 10200 calibrated_sensor,
2631083 ego_pose,68 log,
850 scene,
34149 sample, 2631083 sample_data, 1166187 sample_annotation,
4 map,
Done loading in 36.0 seconds.

Reverse indexing ...
Done reverse indexing in 8.5 seconds. ======
Initializing nuScenes tracking evaluation
Loaded results from /data/simtrack_output/tracking_results.json. Found detections for 6019 samples.
Loading annotations for val split from nuScenes version: v1.0-trainval
100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 6019/6019 [00:06<00:00, 871.16it/s]
Loaded ground truth annotations for 6019 samples.
Filtering tracks
=> Original number of boxes: 227984
=> After distance based filtering: 190099
=> After LIDAR points based filtering: 190099
=> After bike rack filtering: 189972
Filtering ground truth tracks
=> Original number of boxes: 142261
=> After distance based filtering: 103564
=> After LIDAR points based filtering: 93885
=> After bike rack filtering: 93875
Accumulating metric data...
Computing metrics for class bicycle...

Computed thresholds
[15/167]
MOTAR MOTP Recall Frames GT GT-Mtch GT-Miss GT-IDS Pred Pred-TP Pred-FP Pred-IDS
thr_0.1681 0.000 0.278 0.507 1923 1993 971 982 40 2431 971 1420 40

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS   

thr_0.1975 0.000 0.271 0.488 1769 1993 939 1021 33 1999 939 1027 33

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS                  

thr_0.2212 0.127 0.266 0.462 1686 1993 893 1073 27 1700 893 780 27

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.2547 0.450 0.262 0.441 1546 1993 857 1114 22 1350 857 471 22

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS                    

thr_0.2824 0.535 0.262 0.414 1514 1993 804 1167 22 1200 804 374 22

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.2922 0.551 0.260 0.395 1501 1993 766 1205 22 1132 766 344 22

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.3012 0.538 0.277 0.368 1490 1993 712 1260 21 1062 712 329 21

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.3335 0.603 0.266 0.346 1470 1993 673 1303 17 957 673 267 17

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.3816 0.741 0.267 0.316 1422 1993 617 1364 12 789 617 160 12

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.4070 0.769 0.248 0.293 1413 1993 575 1410 8 716 575 133 8

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.4231 0.781 0.241 0.276 1407 1993 544 1443 6 669 544 119 6

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.4741 0.841 0.231 0.243 1385 1993 479 1508 6 561 479 76 6

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.4873 0.837 0.223 0.221 1385 1993 435 1553 5 511 435 71 5

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.5002 0.860 0.206 0.199 1378 1993 394 1596 3 452 394 55 3

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.5331 0.926 0.202 0.183 1351 1993 363 1628 2 392 363 27 2

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.5464 0.944 0.199 0.153 1347 1993 303 1688 2 322 303 17 2

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.5668 0.940 0.206 0.134 1347 1993 266 1726 1 283 266 16 1

            MOTAR   MOTP    Recall  Frames  GT      GT-Mtch GT-Miss GT-IDS  Pred    Pred-TP Pred-FP Pred-IDS

thr_0.5879 0.956 0.207 0.104 1343 1993 206 1786 1 216 206 9 1

Traceback (most recent call last):
File "./tools/val_nusc_tracking.py", line 202, in
tracking()
File "./tools/val_nusc_tracking.py", line 148, in tracking
dataset.evaluation_tracking(copy.deepcopy(predictions), output_dir=args.work_dir, testset=False)
File "/data/simtrack/det3d/datasets/nuscenes/nuscenes.py", line 382, in evaluation_tracking
metrics_summary = nusc_eval.main()
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/nuscenes/eval/tracking/evaluate.py", line 205, in main
metrics, metric_data_list = self.evaluate()
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/nuscenes/eval/tracking/evaluate.py", line 135, in evaluate
accumulate_class(class_name)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/nuscenes/eval/tracking/evaluate.py", line 131, in accumulate_class
curr_md = curr_ev.accumulate()
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/nuscenes/eval/tracking/algo.py", line 156, in accumulate
assert unachieved_thresholds + duplicate_thresholds + len(thresh_metrics) == self.num_thresholds
AssertionError

8 2080ti GPU using default cfg to train,but triger CUDA out of memory

2022-03-17 14:21:36,906 - INFO - Start running, host: yangjinrong@tracking-q5x64-32246-worker-0, work_dir: /data/simtrack_output
2022-03-17 14:21:36,907 - INFO - workflow: [('train', 1), ('val', 1)], max: 20 epochs
Traceback (most recent call last):
File "./tools/train.py", line 141, in
main()
File "./tools/train.py", line 136, in main
logger=logger,
File "/data/simtrack/det3d/torchie/apis/train.py", line 206, in train_detector
trainer.run(data_loaders, cfg.workflow, cfg.total_epochs, local_rank=cfg.local_rank)
File "/data/simtrack/det3d/torchie/trainer/trainer.py", line 527, in run
epoch_runner(data_loaders[i], self.epoch, **kwargs)
File "/data/simtrack/det3d/torchie/trainer/trainer.py", line 393, in train
self.model, data_batch, train_mode=True, **kwargs
File "/data/simtrack/det3d/torchie/trainer/trainer.py", line 356, in batch_processor
losses = model(example, return_loss=True)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/modules/module.py", line 722, in _call_impl
result = self.forward(*input, **kwargs)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/parallel/distributed.py", line 511, in forward
output = self.module(*inputs[0], **kwargs[0])
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/modules/module.py", line 722, in _call_impl
result = self.forward(*input, **kwargs)
File "/data/simtrack/det3d/models/detectors/point_pillars.py", line 48, in forward
x = self.extract_feat(data)
File "/data/simtrack/det3d/models/detectors/point_pillars.py", line 29, in extract_feat
x = self.neck(x)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/modules/module.py", line 722, in _call_impl
result = self.forward(*input, **kwargs)
File "/data/simtrack/det3d/models/necks/rpn.py", line 142, in forward
ups.append(self.deblocksi - self._upsample_start_idx)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/modules/module.py", line 722, in _call_impl
result = self.forward(*input, **kwargs)
File "/data/simtrack/det3d/models/utils/misc.py", line 82, in forward
input = module(input)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/modules/module.py", line 722, in _call_impl
result = self.forward(*input, **kwargs)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/modules/activation.py", line 102, in forward
return F.relu(input, inplace=self.inplace)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/nn/functional.py", line 1119, in relu
result = torch.relu(input)
RuntimeError: CUDA out of memory. Tried to allocate 64.00 MiB (GPU 2; 10.76 GiB total capacity; 9.76 GiB already allocated; 47.44 MiB free; 9.88 GiB reserved in total by PyTorch)
^CProcess Process-10:
^CProcess Process-9:
Process Process-9:
Process Process-9:
Process Process-9:
Process Process-3:
Process Process-2:
Process Process-5:
Traceback (most recent call last):
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/subprocess.py", line 1019, in wait
Process Process-2:
Process Process-10:
Process Process-5:
Process Process-10:
Process Process-2:
Process Process-10:
Process Process-5:
Process Process-1:
Process Process-7:
Process Process-5:
Process Process-9:
Process Process-5:
Process Process-4:
Process Process-8:
Process Process-8:
Process Process-8:
Process Process-4:
Process Process-4:
Process Process-1:
Process Process-3:
Process Process-1:
Process Process-6:
Process Process-1:
Process Process-7:
Process Process-7:
return self._wait(timeout=timeout)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/subprocess.py", line 1653, in _wait
(pid, sts) = self._try_wait(0)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/subprocess.py", line 1611, in _try_wait
(pid, sts) = os.waitpid(self.pid, wait_flags)
KeyboardInterrupt

During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/runpy.py", line 193, in _run_module_as_main
"main", mod_spec)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/distributed/launch.py", line 261, in
main()
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/site-packages/torch/distributed/launch.py", line 254, in main
process.wait()
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/subprocess.py", line 1032, in wait
self._wait(timeout=sigint_timeout)
File "/home/yangjinrong/miniconda3/envs/det3d/lib/python3.7/subprocess.py", line 1647, in _wait
time.sleep(delay)
KeyboardInterruptq

Unable to download pretrained model

Hi, I'm trying to download the pretrained model using the link in INSTALL.md. That link seem to point to a Azure storage service. I tried many accounts and all got error

AADSTS50177: User account '***' from identity provider 'live.com' does not exist in tenant 'Massachusetts Institute of Technology' and cannot access the application '00000003-0000-0ff1-ce00-000000000000'(Office 365 SharePoint Online) in that tenant. The account needs to be added as an external user in the tenant first. Sign out and sign in again with a different Azure Active Directory user account.

Have you made this file public?

A100 for training simtrack but not reproduce the paper results

Dear author,I still have some questions.

  1. We use A100 for training, 4 gpu cards, 4 epochs , the rest of the parameters have not changed, the training 200 epochs still does not achieve the effect. What training skills do you have? such as the learning rate modify and so on
  2. We train the model that comes with simtrack. After training for 20 epochs, we continue to train based on the model, and the loss will continue to decline. why did you stop training?
    I sincerely look forward to your reply.

No module named 'det3d', how to install det3d?

I followed install.md, does anyone know how to install det3d? Thanks.

(simtrack) z@z:~/dev/simtrack$ python ./tools/val_nusc_tracking.py examples/point_pillars/configs/nusc_all_pp_centernet_tracking.py --checkpoint model_zoo/simtrack_pillar.pth  --work_dir work_dirs/
Traceback (most recent call last):
  File "./tools/val_nusc_tracking.py", line 8, in <module>
    from det3d.datasets import build_dataloader, build_dataset
ModuleNotFoundError: No module named 'det3d'

How to use v1.0-mini of nuscenes dataset?

How to use v1.0-mini of nuscenes dataset? I can train v1.0-mini, but it is not supported during the test. I want to know if there is a solution. How long did you spend training nuscenes dataset? I only have one GPU. Should i set -- nproc_ per_ Set to 1? About ./model_ zoo/simtrack_ Pillar.pth, I get the following results: Super (open_zipfile_reader, self)_ init__ (torch.C.PyTorchFileReader(name_or_buffer))
RuntimeError: version
<= kMaxSupportedFileFormatVersion INTERNAL ASSERT FAILED at /opt/conda/conda-bld/pytorch_ 1579061855666/work/caffe2/serialize/inline_ container.cc:132, please report a bug to PyTorch. Attempted to read a PyTorch file with version 3, but the maximum supported version for reading is 2. Your PyTorch installation may be too old. (init at /opt/conda/conda-bld/pytorch_1579061855666/work/caffe2/serialize/inline_container.cc:132)
frame #0: c10::Error::Error(c10::SourceLocation, std::string const&) + 0x47 (0x7f24ac0a7627 in /home/ubuntu/anaconda3/lib/python3.8/site-packages/torch/lib/libc10.so)
Can you provide a trained (pillar based) model with unlimited pytorch version and the specific environment version? Thank you very much for your help.

tracking_batch_hm = (batch_hm + prev_hm[task_id]) / 2.0

Hi,author!
tracking_batch_hm = (batch_hm + prev_hm[task_id]) / 2.0

I don't understand the actual physical meaning of “tracking_batch_hm”.I also don’t understand why we need to execute this way instead of directly using batch_hm or prev_hm?

I have another question that if the displacement of an object is relatively large, then the position of this object in the previous centerness map will be farther away from that in the current centerness map.(In other words, there is no intersection between the representation of this object in the previous centerness map and the current centerness map.)

So after NMS operation, will this object be considered as a new object?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.