Abstract
With the widespread use of unmanned aerial vehicles (UAVs), there is an increasing demand for the development of their control technology. The key interaction technology between humans and UAVs needs to focus on the human body language, which comprises rich interactive information, as it is the most natural, intuitive, and easy to master approach of interpersonal communication for humans. Therefore, the research on human motion estimation for UAV control is of considerable practical significance. Recently, deep learning has made breakthroughs in speech, image recognition and, other fields, and has crushed the performance of traditional methods in many fields. However, in the field of human motion estimation, deep learning has been progressing slowly. To overcome the limitations of the traditional methods and explore the application of deep learning methods in the field of motion estimation, this study proposes a method to estimate human arm motion using deep learning networks. We proposed a bidirectional two-layer LSTM fusion network to estimate the forearms' motion according to the hand position measured by HTC Vive. The performance was verified using a real data set. The average Euclidean distance similarity can reach up to 56%. In comparison with the traditional methods, the proposed method demonstrated wider applicability and better performance.
| Original language | English |
|---|---|
| Title of host publication | Proceedings - 2019 IEEE International Congress on Cybermatics |
| Subtitle of host publication | 12th IEEE International Conference on Internet of Things, 15th IEEE International Conference on Green Computing and Communications, 12th IEEE International Conference on Cyber, Physical and Social Computing and 5th IEEE International Conference on Smart Data, iThings/GreenCom/CPSCom/SmartData 2019 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 381-384 |
| Number of pages | 4 |
| ISBN (Electronic) | 9781728129808 |
| DOIs | |
| State | Published - Jul 2019 |
| Event | 12th IEEE International Conference on Internet of Things, 15th IEEE International Conference on Green Computing and Communications, 12th IEEE International Conference on Cyber, Physical and Social Computing and 5th IEEE International Conference on Smart Data, iThings/GreenCom/CPSCom/SmartData 2019 - Atlanta, United States Duration: 14 Jul 2019 → 17 Jul 2019 |
Publication series
| Name | Proceedings - 2019 IEEE International Congress on Cybermatics: 12th IEEE International Conference on Internet of Things, 15th IEEE International Conference on Green Computing and Communications, 12th IEEE International Conference on Cyber, Physical and Social Computing and 5th IEEE International Conference on Smart Data, iThings/GreenCom/CPSCom/SmartData 2019 |
|---|
Conference
| Conference | 12th IEEE International Conference on Internet of Things, 15th IEEE International Conference on Green Computing and Communications, 12th IEEE International Conference on Cyber, Physical and Social Computing and 5th IEEE International Conference on Smart Data, iThings/GreenCom/CPSCom/SmartData 2019 |
|---|---|
| Country/Territory | United States |
| City | Atlanta |
| Period | 14/07/19 → 17/07/19 |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 7 Affordable and Clean Energy
Keywords
- Deep learning
- HTC Vive
- Motion estimation
- UAV control
Fingerprint
Dive into the research topics of 'Motion estimation approach for UAV controls using bidirectional two-layer LSTMs'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver