Fr. 157.00

Multi-Level Bayesian Models for Environment Perception

English · Paperback / Softback

Shipping usually within 1 to 2 weeks (title will be printed to order)

Description

Read more

This book deals with selected problems of machine perception, using various 2D and 3D imaging sensors. It proposes several new original methods, and also provides a detailed state-of-the-art overview of existing techniques for automated, multi-level interpretation of the observed static or dynamic environment. To ensure a sound theoretical basis of the new models, the surveys and algorithmic developments are performed in well-established Bayesian frameworks.  Low level scene understanding functions are formulated as various image segmentation problems, where the advantages of probabilistic inference techniques such as Markov Random Fields (MRF) or Mixed Markov Models are considered. For the object level scene analysis, the book mainly relies on the literature of Marked Point Process (MPP) approaches, which consider strong geometric and prior interaction constraints in object population modeling. In particular, key developments are introduced in the spatial hierarchical decomposition of the observed scenarios, and in the temporal extension of complex MRF and MPP models.  Apart from utilizing conventional optical sensors, case studies are provided on passive radar (ISAR) and Lidar-based Bayesian environment perception tasks. It is shown, via several experiments, that the proposed contributions embedded into a strict mathematical toolkit can significantly improve the results in real world 2D/3D test images and videos, for applications in video surveillance, smart city monitoring, autonomous driving, remote sensing, and optical industrial inspection.

List of contents

Introduction.- Fundamentals. - Bayesian models for Dynamic Scene Analysis.- Multi-layer label fusion models.- Multitemporal data analysis with Marked Point Processes. - Multi-level object population analysis with an EMPP model.- Concluding Remarks.- References.- Index.

About the author










Dr. Csaba Benedek is a scientific advisor with the Machine Perception Research Laboratory at the Institute for Computer Science and Control (SZTAKI), Eötvös Loránd Research Network (ELKH) in Budapest, Hungary, and a professor with the Faculty of Information Technology and Bionics of the Péter Pázmány Catholic University (PPCU). He obtained his PhD from PPCU in 2008, and his DSc from the Hungarian of Academy of Sciences (HAS) in 2020. Dr. Benedek has been the president of the Hungarian Image Processing and Pattern Recognition Society (Képaf), and the Hungarian Governing Board Member of the International Association for Pattern Recognition (IAPR). He has been a Senior Member of the IEEE, an Associate Editor of the journal Digital Signal Processing (Elsevier) and a Guest Editor of Remote Sensing (MDPI). His awards include the Bolyai plaquette from HAS (2019), a Researcher Acknowledgement from the HAS Secretary-General (2018), the Imreh Csanád plaquette (2019), and the Michelberger Master Award from the Hungarian Academy of Engineering (2020). In recent years, he has managed various national and international research projects. His research interests include Bayesian image and point cloud segmentation, object extraction, change detection, machine learning applications and GIS data analysis. 


Product details

Authors Csaba Benedek
Publisher Springer, Berlin
 
Languages English
Product format Paperback / Softback
Released 03.05.2023
 
EAN 9783030836566
ISBN 978-3-0-3083656-6
No. of pages 202
Dimensions 155 mm x 11 mm x 235 mm
Illustrations XIII, 202 p. 101 illus., 70 illus. in color.
Subject Natural sciences, medicine, IT, technology > Mathematics > Probability theory, stochastic theory, mathematical statistics

Customer reviews

No reviews have been written for this item yet. Write the first review and be helpful to other users when they decide on a purchase.

Write a review

Thumbs up or thumbs down? Write your own review.

For messages to CeDe.ch please use the contact form.

The input fields marked * are obligatory

By submitting this form you agree to our data privacy statement.