GCHAR: An efficient Group-based Context—aware human activity recognition on smartphone

Liang Cao, Yufeng Wang*, Bo Zhang, Qun Jin, Athanasios V. Vasilakos


研究成果: Article査読

109 被引用数 (Scopus)


With smartphones increasingly becoming ubiquitous and being equipped with various sensors, nowadays, there is a trend towards implementing HAR (Human Activity Recognition) algorithms and applications on smartphones, including health monitoring, self-managing system and fitness tracking. However, one of the main issues of the existing HAR schemes is that the classification accuracy is relatively low, and in order to improve the accuracy, high computation overhead is needed. In this paper, an efficient Group-based Context-aware classification method for human activity recognition on smartphones, GCHAR is proposed, which exploits hierarchical group-based scheme to improve the classification efficiency, and reduces the classification error through context awareness rather than the intensive computation. Specifically, GCHAR designs the two-level hierarchical classification structure, i.e., inter-group and inner-group, and utilizes the previous state and transition logic (so-called context awareness) to detect the transitions among activity groups. In comparison with other popular classifiers such as RandomTree, Bagging, J48, BayesNet, KNN and Decision Table, thorough experiments on the realistic dataset (UCI HAR repository) demonstrate that GCHAR achieves the best classification accuracy, reaching 94.1636%, and time consumption in training stage of GCHAR is four times shorter than the simple Decision Table and is decreased by 72.21% in classification stage in comparison with BayesNet.

ジャーナルJournal of Parallel and Distributed Computing
出版ステータスPublished - 2018 8月

ASJC Scopus subject areas

  • ソフトウェア
  • 理論的コンピュータサイエンス
  • ハードウェアとアーキテクチャ
  • コンピュータ ネットワークおよび通信
  • 人工知能


「GCHAR: An efficient Group-based Context—aware human activity recognition on smartphone」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。