Skip to main content

Transferring Dense Object Detection Models To Event-Based Data

  • Conference paper
  • First Online:
Advanced Intelligent Virtual Reality Technologies

Part of the book series: Smart Innovation, Systems and Technologies ((SIST,volume 330))

Abstract

Event-based image representations are fundamentally different to traditional dense images. This poses a challenge to apply current state-of-the-art models for object detection as they are designed for dense images. In this work we evaluate the YOLO object detection model on event data. To this end we replace dense-convolution layers by either sparse convolutions or asynchronous sparse convolutions which enables direct processing of event-based images and compare the performance and runtime to feeding event-histograms into dense-convolutions. Here, hyper-parameters are shared across all variants to isolate the effect sparse-representation has on detection performance. At this, we show that current sparse-convolution implementations cannot translate their theoretical lower computation requirements into an improved runtime.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 219.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 279.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 279.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    https://github.com/zzzheng/pytorch-yolo-v1.

  2. 2.

    https://github.com/paroj/rpg_asynet.

  3. 3.

    https://github.com/thtrieu/darkflow/issues/957.

  4. 4.

    https://docs.python.org/3/library/profile.html.

References

  1. Cannici, M., Ciccone, M., Romanoni, A., Matteucci, M.: Asynchronous convolutional networks for object detection in neuromorphic cameras. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)

    Google Scholar 

  2. Gallego, G., Delbruck, T., Orchard, G.M., Bartolozzi, C., Taba, B., Censi, A., Leutenegger, S., Davison, A., Conradt, J., Daniilidis, K., Scaramuzza, D.: Event-based vision: A survey. IEEE Trans. Patt. Anal. Mach. Intell. (2020)

    Google Scholar 

  3. Gehrig, D., Gehrig, M., Hidalgo-Carrió, J., Scaramuzza, D.: Video to events: recycling video datasets for event cameras. IEEE Conf. Comput. Vis. Patt. Recog. (CVPR) (2020)

    Google Scholar 

  4. Geiger, A.: The kitti vision benchmark suite (2017). http://www.cvlibs.net/datasets/kitti/eval_object.php

  5. Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? the kitti vision benchmark suite. Conf. Comp. Vis. Patt. Recogn. (CVPR) (2012)

    Google Scholar 

  6. Graham, B., Engelcke, M., van der Maaten, L.: 3d Semantic segmentation with submanifold sparse convolutional networks. CVPR (2018)

    Google Scholar 

  7. Maass, W.: Networks of spiking neurons: the third generation of neural network models. Neur. Netw. 10(9), 1659–1671 (1997)

    Article  Google Scholar 

  8. Messikommer, N., Gehrig, D., Loquercio, A., Scaramuzza, D.: Event-based asynchronous sparse convolutional networks (2020). http://rpg.ifi.uzh.ch/docs/ECCV20_Messikommer.pdf

  9. Rebecq, H., Ranftl, R., Koltun, V., Scaramuzza, D.: Events-to-video: bringing modern computer vision to event cameras. IEEE Conf. Comput. Vis. Pattern Recog. (CVPR) (2019)

    Google Scholar 

  10. Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pavel Rojtberg .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Mechler, V., Rojtberg, P. (2023). Transferring Dense Object Detection Models To Event-Based Data. In: Nakamatsu, K., Patnaik, S., Kountchev, R., Li, R., Aharari, A. (eds) Advanced Intelligent Virtual Reality Technologies. Smart Innovation, Systems and Technologies, vol 330. Springer, Singapore. https://doi.org/10.1007/978-981-19-7742-8_3

Download citation

Publish with us

Policies and ethics