Deep-Learning-for-Human-Action-Recognition

Abstract

The study leverages the Human Action Recognition (HAR) dataset, which can be accessed here. Alongside this documentation, the trained models and associated Jupyter Notebook (.ipynb) files have been attached for reproducibility and further exploration.

Deep Learning for Human Action Recognition

Objective

The goal of this project is to develop a Convolutional Neural Network (CNN) model to classify human activities from images. The model is trained to recognize and label activities from 15 predefined categories based on visual content.

Dataset

The dataset comprises over 12,000 labeled images categorized into 15 human activity classes:

Each class contains 840 training images, and the dataset is balanced with no missing or duplicate values.

Here are some sample images from the dataset: Sample Image

Methodology

1. Data Exploration

2. Data Preprocessing

3. CNN Model Design

4. VGG16-Based Model Design

5. Model Training and Evaluation

Results

The performance of the models is summarized in the table below:

Metric CNN Model VGG Model
Training Accuracy 79.44% 81.17%
Validation Accuracy 54.05% 46.91%
Final Training Loss 0.634 0.548
Validation Loss 1.835 2.560
Test Accuracy (from Contest) 56.53% 48.29%
Practice Rank 8th N/A

Usage

  1. Clone this repository.
  2. Load the ipynb files in Jupyter Notebook or Google Colab.
  3. Ensure you have downloaded the HAR dataset from the provided link.
  4. Follow the code to train or evaluate the models.

Project Structure

├── images
│   ├── sample_image.png
├── best_cnn_model.keras
├── best_vgg_model.keras
├── cnn_model_test.csv
├── vgg_model_test.csv
├── README.md
└── har-detection-testing-accuracy-56.ipynb

Acknowledgments

This project demonstrates the application of deep learning techniques in human activity recognition using image data, highlighting both challenges and potential improvements for practical deployment.