This paper proposes a low-cost multisensory glove (≈ 140 USD) equipped with commercial piezoresistive force sensors (FSRs) and inertial measurement units (IMUs) for object recognition. A set of 28 daily life objects is used to evaluate the glove by applying grasping actions. The raw signals acquired through these actions are employed directly to train a shallow one-dimensional convolutional neural network. The network is deployed on a low-cost edge system to perform online object recognition, thereby presenting an end-to-end system. The proposed system achieves a classification accuracy of 99.29% while consuming only 59.85 mW and 0.4875 mJ of power and energy per inference respectively. Overall, the proposed system seems a suitable candidate for various biomedical applications such as post-stroke rehabilitation.

Multisensory Wearable Glove for Object Recognition Based Embedded Machine Learning

Yaacoub M.;Al Haj Ali H.;Gianoglio C.;
2024-01-01

Abstract

This paper proposes a low-cost multisensory glove (≈ 140 USD) equipped with commercial piezoresistive force sensors (FSRs) and inertial measurement units (IMUs) for object recognition. A set of 28 daily life objects is used to evaluate the glove by applying grasping actions. The raw signals acquired through these actions are employed directly to train a shallow one-dimensional convolutional neural network. The network is deployed on a low-cost edge system to perform online object recognition, thereby presenting an end-to-end system. The proposed system achieves a classification accuracy of 99.29% while consuming only 59.85 mW and 0.4875 mJ of power and energy per inference respectively. Overall, the proposed system seems a suitable candidate for various biomedical applications such as post-stroke rehabilitation.
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11567/1239157
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
social impact