Abstract
A sensor system with ultra-high sensitivity, high resolution, rapid response time, and a high signal-to-noise ratio can produce raw data that is exceedingly rich in information, including signals that have the appearances of "noise". The "noise"feature directly correlates to measurands in orthogonal dimensions, and are simply manifestations of the off-diagonal elements of 2nd-order tensors that describe the spatial anisotropy of matter in physical structures and spaces. The use of machine learning techniques to extract useful meanings from the rich information afforded by ultra-sensitive one-dimensional sensors may offer the potential for probing mundane events for novel embedded phenomena. Inspired by our very recent invention of ultra-sensitive optical-based inclinometers, this work aims to answer a transformative question for the first time: can a single-dimension point sensor with ultra-high sensitivity, fidelity, and signal-to-noise ratio identify an arbitrary mechanical impact event in three-dimensional space? This work is expected to inspire researchers in the fields of sensing and measurement to promote the development of a new generation of powerful sensors or sensor networks with expanded functionalities and enhanced intelligence, which may provide rich n-dimensional information, and subsequently, data-driven insights into significant problems.
Recommended Citation
C. Zhu et al., "One-Dimensional Sensor Learns to Sense Three-Dimensional Space," Optics Express, vol. 28, no. 13, pp. 19374 - 19389, Optical Society of America, Jun 2020.
The definitive version is available at https://doi.org/10.1364/OE.395282
Department(s)
Electrical and Computer Engineering
Research Center/Lab(s)
Intelligent Systems Center
International Standard Serial Number (ISSN)
1094-4087
Document Type
Article - Journal
Document Version
Final Version
File Type
text
Language(s)
English
Rights
© 2020 Optical Society of America, All rights reserved.
Publication Date
22 Jun 2020
PubMed ID
32672216
Comments
Army Research Laboratory (W911NF-14-2-0034); Leonard Wood Institute, Grant LWI-2018-006