This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| en:safeav:curriculum:maps-b [2025/11/03 12:00] – raivo.sell | en:safeav:curriculum:maps-b [2025/11/05 11:05] (current) – airi | ||
|---|---|---|---|
| Line 1: | Line 1: | ||
| + | ====== Module: Perception, Mapping, and Localization (Part 1) ====== | ||
| + | |||
| + | ^ **Study level** | Bachelor | | ||
| + | ^ **ECTS credits** | 1 ECTS | | ||
| + | ^ **Study forms** | Hybrid or fully online | | ||
| + | ^ **Module aims** | The aim of the module is to introduce perception, mapping and localisation methods for autonomous systems. The course develops students’ ability to combine data from multiple sensors to detect and interpret the environment, | ||
| + | ^ **Pre-requirements** | Basic knowledge of linear algebra, probability and signal processing, as well as programming skills. Familiarity with control systems, kinematics, Linux/ROS environments or computer vision libraries is recommended but not mandatory. | | ||
| + | ^ **Learning outcomes** | **Knowledge**\\ • Describe perception, mapping, and localization processes in autonomous systems.\\ • Explain principles of sensor fusion, simultaneous localization and mapping.\\ • Understand AI-based perception, including object detection, classification, | ||
| + | ^ **Topics** | 1. Cameras, LiDARs, radars, and IMUs in perception and mapping.\\ 2. Sensor calibration, | ||
| + | ^ **Type of assessment** | The prerequisite of a positive grade is a positive evaluation of module topics and presentation of practical work results with required documentation | | ||
| + | ^ **Learning methods** | **Lecture** — Theoretical background on perception, mapping, and AI-based scene understanding.\\ **Lab works** — Implementation of sensor fusion and mapping algorithms using ROS2, Python, and simulated data.\\ **Individual assignments** — Analysis of perception pipeline performance and report preparation.\\ **Self-learning** — Study of academic papers, datasets, and open-source AI perception frameworks. | | ||
| + | ^ **AI involvement** | AI tools can assist in code debugging, model training, and visualization of perception results. Students must cite AI-generated assistance transparently and verify the correctness of outcomes. | | ||
| + | ^ **Recommended tools and environments** | SLAM, CNN, OpenCV, PyTorch, TensorFlow, KITTI, NuScenes | | ||
| + | ^ **Verification and Validation focus** | | | ||
| + | ^ **Relevant standards and regulatory frameworks** | ISO 26262, ISO 21448 (SOTIF) | | ||
| + | |||
| + | |||