Gait Activity Classification Using CNN-LSTM Networks

Multi-modal Sensor Fusion for Enhanced Activity Recognition

Objective

The primary objective of this research is to develop and evaluate a hybrid deep learning architecture that effectively classifies human gait activities using multi-modal sensor data. Specifically, this study aims to:

  • Leverage convolutional neural networks (CNN) for automatic feature extraction from sequential sensor data
  • Utilize Long Short-Term Memory (LSTM) networks to capture temporal dependencies in gait patterns
  • Investigate the performance improvement gained by fusing inertial measurement unit (IMU) data with GPS information
  • Compare classification accuracy across subject-specific and multi-subject training scenarios
  • Classify gait activities based on terrain inclination and surface characteristics
Research Question: Can the integration of GPS data with IMU sensors significantly enhance the accuracy of gait activity classification in a CNN-LSTM framework, and how does this improvement vary between subject-specific and generalized models?

Approach

Model Architecture

This study employs a hybrid CNN-LSTM architecture that combines the strengths of both neural network types. The CNN component processes subsequences of input data to automatically extract relevant features, while the LSTM backend interprets these features as temporal sequences to capture long-term dependencies in gait patterns.

Data Processing Pipeline

  • Sliding Window Segmentation: Time-series data is partitioned into overlapping windows for activity classification
  • Data Augmentation: Gaussian noise is added to the training set to improve model robustness and generalization
  • Train-Test Split: A balanced 50-50 split ensures comprehensive coverage of all gait activities in both training and testing sets
  • Feature Normalization: Sensor readings are standardized to facilitate efficient model training

Experimental Design

Multiple experimental trials were conducted to evaluate different configurations:

  • Subject-Specific Models: Trained and tested on individual subjects to maximize personalization
  • Multi-Subject Models: Trained across multiple subjects to assess generalization capabilities
  • Sensor Modalities: Compared IMU-only versus IMU+GPS data fusion
  • Classification Targets: Two complexity levels - 4-class (inclination only) and 7-class (inclination + terrain)
Key Innovation: The integration of GPS spatial information with IMU temporal dynamics provides complementary data streams that enhance the model's ability to distinguish between similar gait patterns occurring in different environmental contexts.

Dataset

Data Collection

The dataset comprises synchronized IMU and GPS sensor recordings captured during various gait activities across different terrains and inclinations. Data was collected from multiple subjects performing standardized locomotion tasks in controlled and natural environments.

Sensor Configuration

  • IMU Sensors: Tri-axial accelerometer and gyroscope measurements capturing linear and angular motion
  • GPS Data: Location coordinates, altitude, and velocity information providing spatial context
  • Sampling Rate: High-frequency data collection to capture subtle gait variations

Activity Classes

Classification Level Number of Classes Description
Inclination Only 4 classes Level ground, uphill, downhill, variable incline
Inclination + Terrain 7 classes Combinations of incline with surface types (paved, gravel, grass, etc.)
Original Labels 22 classes Consolidated into the above categories for classification

Data Characteristics

Dataset Size: The dataset includes extensive recordings from both single-subject and multi-subject trials, with balanced representation across all activity classes to prevent bias. The 50-50 train-test split was specifically designed to ensure that all gait activities appear in the test set for comprehensive evaluation.

Results & Discussion

Performance Improvements with GPS Integration

The experimental results demonstrate substantial performance gains when GPS data is incorporated alongside IMU measurements in the CNN-LSTM architecture. The improvements are particularly pronounced in subject-specific models, though multi-subject models also show meaningful enhancements.

Subject-Specific (4-class)

+23.8%

F1 Score improvement for inclination classification

Subject-Specific (7-class)

+30.4%

F1 Score improvement for inclination + terrain classification

Multi-Subject (4-class)

+13.4%

F1 Score improvement for inclination classification

Multi-Subject (7-class)

+9.1%

F1 Score improvement for inclination + terrain classification

Key Findings

  • Multi-modal Sensor Fusion Effectiveness: The integration of GPS data consistently improved classification performance across all experimental conditions, validating the hypothesis that spatial context enhances gait activity recognition
  • Subject-Specific vs. Multi-Subject Trade-offs: While subject-specific models achieved higher accuracy gains (+23.8% to +30.4%), they may be more susceptible to overfitting. Multi-subject models showed more modest improvements (+9.1% to +13.4%) but demonstrate better generalization potential
  • Classification Complexity Impact: The 7-class problem (inclination + terrain) showed higher improvement rates for subject-specific models, suggesting that GPS data is particularly valuable for disambiguating complex activity patterns
  • CNN-LSTM Architecture Suitability: The hybrid architecture effectively leverages both spatial feature extraction (CNN) and temporal pattern recognition (LSTM) capabilities, making it well-suited for time-series gait data

Discussion

The substantial performance improvements observed with GPS integration can be attributed to several factors. GPS data provides complementary spatial information that helps distinguish between activities that may produce similar IMU patterns but occur in different environmental contexts. For instance, walking on level ground versus a slight incline might generate comparable acceleration patterns, but GPS altitude data can disambiguate these scenarios.

The higher improvement rates in subject-specific models suggest that individual gait patterns interact uniquely with environmental factors. This finding has important implications for personalized activity monitoring systems, where user-specific calibration could significantly enhance accuracy. However, the risk of overfitting in subject-specific models must be carefully managed through appropriate regularization and validation strategies.

The positive results for multi-subject models, while more modest, are encouraging for developing generalized activity recognition systems. These models could be deployed without extensive per-user calibration while still benefiting from multi-modal sensor fusion. The 9.1% to 13.4% improvements represent meaningful gains that could translate to more reliable activity tracking in practical applications.

Practical Implications: These results demonstrate that incorporating GPS sensors into wearable activity monitoring devices can significantly enhance classification accuracy without requiring complex algorithmic changes. The CNN-LSTM architecture provides a scalable framework for processing multi-modal sensor streams in real-time applications.

Future Directions

Future research could explore several promising avenues:

  • Investigating attention mechanisms to automatically weight the contribution of different sensor modalities
  • Exploring transfer learning approaches to leverage subject-specific gains while maintaining generalization
  • Evaluating performance with additional sensor types (e.g., barometric pressure, heart rate)
  • Testing real-time implementation feasibility on resource-constrained wearable devices
  • Extending classification to more granular activity types and transition states

Interested in This Research?

I'm passionate about advancing the field of activity recognition and wearable sensing technologies. If you'd like to discuss this research, explore potential collaborations, or have thoughtful comments and questions, I'd love to hear from you!

View My GitHub Contact Me