The application of machine learning in multi sensor data fusion for activity recognition in mobile device space
Asmaa Al-Marhoubi
Sara Saravi
Eran Edirisinghe
2134/21854
https://repository.lboro.ac.uk/articles/conference_contribution/The_application_of_machine_learning_in_multi_sensor_data_fusion_for_activity_recognition_in_mobile_device_space/9403055
The present generation of mobile handheld devices comes equipped with a large number of sensors. The key sensors include the Ambient Light Sensor, Proximity Sensor, Gyroscope, Compass and the Accelerometer. Many mobile applications are driven based on the readings obtained from either one or two of these sensors. However the presence of
multiple-sensors will enable the determination of more detailed activities that are carried out by the user of a mobile
device, thus enabling smarter mobile applications to be developed that responds more appropriately to user behavior and
device usage. In the proposed research we use recent advances in machine learning to fuse together the data obtained from all key sensors of a mobile device. We investigate the possible use of single and ensemble classifier based approaches to identify a mobile device’s behavior in the space it is present. Feature selection algorithms are used to remove non-discriminant features that often lead to poor classifier performance. As the sensor readings are noisy and include a significant proportion of missing values and outliers, we use machine learning based approaches to clean the raw data obtained from the sensors, before use. Based on selected practical case studies, we demonstrate the ability to
accurately recognize device behavior based on multi-sensor data fusion.
2016-06-28 13:46:59
Activity Recognition
Mobile phone
Mobile Sensors
Gyroscope
Accelerometer
Multi Sensor
Mechanical Engineering not elsewhere classified