Environment Classification System for Mobile Robots: How to distinguish Between Indoor from Outdoor Environment

The research of environment classification for mobile robots focuses on using 1D non-visual sensors such as thermometer, humidity sensor, air pressure sensor, or ceiling height sensor to detect the transition between defined environments. The robot’s behavior is then adapted based on this information.

The crucial part of this research is an image classifier trained to recognize the environment’s type. In particular, in our paper [1], we focused on the classification of an indoor and outdoor environment. It is the most extensive comparison of description methods and classifiers (including neural networks) performed on a single dataset to our best knowledge.

In our papers [2] and [3], the system for environment change detection and environment classification is proposed. The system uses non-visual sensors to generate triggers when the transition between environments is detected. Triggers activate the classification of the environment. Thus, its main purpose is to reduce the time requirements of the environment classification compared to the system, which depends only on images.

References:

[1] Neduchal, P., Gruber, I., & Železný, M. (2020, October). Indoor vs. Outdoor Scene Classification for Mobile Robots. In International Conference on Interactive Collaborative Robotics (pp. 243-252). Springer, Cham.

[2] Neduchal, P., & Železný, M. (2020, September). Environment Classification Approach for Mobile Robots. In Proceedings of 15th International Conference on Electromechanics and Robotics” Zavalishin’s Readings” (pp. 421-432). Springer, Singapore.

[3] Neduchal, P., Bureš, L., & Železný, M. (2019). Environment detection system for localization and mapping purposes. IFAC-PapersOnLine, 52(27), 323-328.


Neural Sign Language Synthesis: Words Are Our Glosses

Jan Zelinka, Jakub Kanis

Our work deals with a text-to-video sign language synthesis. Instead of direct video production, we focused on skeletal model production. Our main goal was to design a fully end-to-end automatic sign language synthesis system trained only on available free data (daily TV broadcasting). Thus, we excluded any manual video annotation. Furthermore, our designed approach even does not rely on any video segmentation. A proposed feed-forward transformer and recurrent transformer were investigated. To improve the performance of our sequence-to-sequence transformer, soft non-monotonic attention was employed in our training process. A benefit of character-level features was compared with word-level features. We focused our experiments on a weather forecasting dataset in the Czech Sign Language.


J. Zelinka and J. Kanis, “Neural Sign Language Synthesis: Words Are Our Glosses,” 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 2020, pp. 3384-3392, doi: 10.1109/WACV45572.2020.9093516.