Smart powered wheelchairs offer the possibility of enhanced mobility to a large and growing population - most notably older adults - and a key feature of such a chair is collision avoidance. Sensors are required to detect nearby obstacles; however, complete sensor coverage of the immediate neighbourhood is challenging for reasons including financial, computational, aesthetic, user identity and sensor reliability. It is also desirable to predict the future motion of the wheelchair based on potential input signals; however, direct modeling and control of commercial wheelchairs is not possible because of proprietary internals and interfaces. In this paper we design a dynamic egocentric occupancy map which maintains information about local obstacles even when they are outside the field of view of the sensor system, and we construct a neural network model of the mapping between joystick inputs and wheelchair motion. Using this map and model infrastructure, we can evaluate a variety of risk assessment metrics for collaborative control of a smart wheelchair. One such metric is demonstrated on a wheelchair with a single RGB-D camera in two scenarios: a doorway traversal where the near edge of the doorframe is no longer visible to the camera as the chair makes its turn, and a longer navigation through a typical cluttered office environment.
|Original language||English (US)|
|Title of host publication||IROS 2014 Conference Digest - IEEE/RSJ International Conference on Intelligent Robots and Systems|
|Publisher||Institute of Electrical and Electronics Engineers Inc.|
|Number of pages||6|
|State||Published - Oct 31 2014|
|Event||2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2014 - Chicago, United States|
Duration: Sep 14 2014 → Sep 18 2014
|Name||IEEE International Conference on Intelligent Robots and Systems|
|Other||2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2014|
|Period||9/14/14 → 9/18/14|
Bibliographical notePublisher Copyright:
© 2014 IEEE.