A 260µW infrared gesture recognition system-on-chip for smart devices

نویسندگان

  • Sechang Oh
  • Ngoc Le Ba
  • Suyoung Bang
  • Junwon Jeong
  • David Blaauw
  • Tony T. Kim
  • Dennis Sylvester
چکیده

Abstract This paper presents a low-power infrared motion detection system suitable for smart devices such as wearables. The SoC incorporates instrumentation chopper amplifiers (ICA), LPFs, ADCs, and a DSP. The low-noise ICAs amplify very low frequency μV-level thermopile outputs with 2.0 NEF and provide programmable gain modes. To reduce standby power the ICA uses lower current when the system is in idle mode. Wakeup can be triggered by detection of a simple gesture. For the LPF, source degeneration by pseudo-resistors and gm division techniques are used for both improved linearity and 30Hz bandwidth. The DSP employs a motion history image technique to achieve low-power detection. The system consumes 260μW in active mode and 46μW in idle mode while processing 16×4 infrared data at 30fps. A complete system demonstration is shown. Introduction Recent demand for natural human-computer interfaces such as gesture recognition has increased, particularly for compact wearable devices. Cameras are currently the most common platform for gesture sensing, but they are highly sensitive to environmental light conditions. Extended range capacitive sensing [1] and ultrasonic techniques [2] have been explored but they consume significant energy due to their excitation source. In contrast, an infrared sensing system, in which a thermopile array directly converts incoming infrared radiation energy into electrical energy, is an appealing lowpower choice since the sensor array itself is passive. It can be fabricated in CMOS technology and generates voltages linearly proportional to the temperature difference between an object and the background environment [3, 4]. However, array sensitivity is just a few μV/oC and its time constant is several ms. To achieve ultra-low power gesture recognition, we propose an SoC including a low-noise instrumentation chopper amplifier for low frequency signals, a lowpower LPF for filtering out-band noise including the chopper frequency and its harmonics, an ADC, and a motion history image based [5] low-power DSP. Proposed Gesture Recognition SoC This paper targets a gesture sensing system using a thermopile array (Figs. 1 and 2). A hand emits infrared radiation with wavelength representing its temperature; this forms an image incident upon a 16×4 thermopile array. Each thermopile signal connects to an AFE path that consists of an ICA and LPF. The four row ADCs digitize the amplified/filtered signals using time-division multiplexing and the DSP then analyzes the waveform to detect gestures. Fig. 3 shows the proposed ICA. Since the gesture signals are significantly impacted by 1/f noise, they are chopped to remove this 1/f noise and then sent through two amplifiers. Overall gain needs to be up to 80dB for a power-efficient high dynamic range system. C1/C2 (15pF/150fF) and C3/C4 (C4=20fF) set the gains for the Low-noise Amplifier (LNA) and Programmable-gain Amplifier (PGA), respectively. C3 is programmable (200fF−3pF) for system flexibility. OTA1 and OTA2 are implemented with inverter-based cascode amplifiers to maximize gm and gain at a given current. The commonmode feedback (CMFB) amplifiers consume a fraction of the power using ratioed transistor sizes. As in typical noise-limited designs, the first amplifier stage consumes the majority of the total power (up to 2.5μA current) to achieve sub-μV noise while the PGA consumes just 90nA, constrained by the chopper bandwidth. Transistor sizes are chosen for optimal noise efficiency factor (NEF) and chopper frequency is 1kHz. The ICA high-pass corner is set by (R3C5) in the DC servo loop. R1 and R2 paths set the input common mode voltages and cancels the offsets. Fast-settling switches (FS1-3) are selectively turned on to reduce settling time when ICA settings are changed, decreasing the corresponding resistance by 100× in simulation. The ICA outputs show ripple at the 1kHz chopping frequency and its harmonics. These are removed with the proposed Gm-C LPF in Fig. 4. The two biquads are connected in series to form a 4 order filter. Since the gesture information resides in a low frequency range, the LPF bandwidth is set to 30Hz to achieve high SNR. CLPF is a capacitor array and is set to 8.9pF to approximately match AFE and thermopile pixel size. Considering fLPF3dB=gm/(2πCLPF), gm in the nS range is required. To achieve this bias current must be extremely low, leading to potentially poor linearity. Thus, source degeneration and gm division techniques [6] are used in the LPF. The Gm-stage input current is divided by the series-parallel current mirror to effectively obtain gm/32. To enhance linearity, input pair sources are degenerated by pseudo-resistors whose gates are controlled by inputs. Simulation results show the resulting gm is linear within ±100mV input range (defined by full width at half-maximum). The CMFB amplifier replicates voltages in the main Gm stage and sets the common mode output voltage. LPF outputs in each row are time-multiplexed via a 16:1 analog multiplexer and connected to a differential 8b SAR ADC (Fig. 5). The ADC sampling rate is 1kS/s. Fig. 6 describes the overall structure of the proposed motion recognition DSP. There are three separate memories to store frame data. The first memory contains the motion history image (MHI), which is the difference between the current and previous frames (Fig. 7). The second and third memories are used to store two continuous frames once motion is detected. Detection modules use data from the three memories to analyze the gesture. Fig. 8 shows the top level design for the proposed gesture detection algorithm. Motion is detected by counting the number of pixels having significant change in value (i.e., ADC output code) between the current and previous frame. If there is no motion for a period of time the processor goes into an idle mode with only a simpler motion detecting circuit enabled to save power. When motion is detected, a sweeping algorithm uses two motion history image frames to analyze the motion. In this process each row and column of the MHI frames are first summed. The type of movement (diagonal, up-down, or left-right) is then discerned based on the number of peaks found in the rowand column-wise sums. In a diagonal sweep both row and column sums will exhibit clear peaks (i.e., four total peaks detected) whereas in up-down or left-right sweeps only two peaks are observed due to constant behavior in either the horizontal or vertical direction. This is shown in Fig. 7, which illustrates the principle of detection for sweeping gestures. Up-down or left-right direction can be determined based on the relative positions of negative to positive peaks, as seen in Fig. 7. This approach allows the DSP to accurately identify specific gestures. Measured Results The proposed gesture recognition SoC is implemented in 65nm CMOS. The ICA input noise density is 31nV/√Hz in active mode and 130nV/ √Hz in idle mode (Fig. 9), and chopping successfully suppresses 1/f noise. LPF bandwidth is adjustable between 10−150Hz by CLPF changes (Fig. 10). Fig. 11 shows the LPF noise spectrum and HD3 is 48.9dB at InLPF=0.1Vpp. Fig. 12 shows ADC performance. The system is demonstrated with an external 16×4 thermopile and lens, and Fig. 13 shows detection of a hand sweeping across the field of view. Fig. 15 summarizes measured results and compares with recent works. This work represents the first SoC for gesture sensing applications using a thermopile array. Its size (8.1mm, Fig. 14) and power (260μW and 46μW) are suitable for emerging smart devices. References [1] Yingzhe Hu, ISSCC 2014. [2] R. Przybyla, ISSCC 2011. [3] M. Hirota, SPIE 2003. [4] H. Kawanishi, ISSCC 2008. [5] C. Hsieh, ICSPS 2010. [6] A. Arnaud, JSSC 2006. [7] Qinwen Fan, JSSC 2011. [8] Yen-Po Chen, VLSIC 2014. [9] P. Bruschi, JSSC 2007. [10] S.-Y. Lee, TBCAS 2009.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Human System Interaction through Distributed Devices in Intelligent Space

Intelligent Space (iSpace) is a space with ubiquitous sensors and actuators (Lee & Hashimoto, 2002). iSpace observes the space using the distributed sensors, extracts useful information from the obtained data and provides various services to the users through the actuators. iSpace can be considered as an “invisible” robot that is united with the environment since it can carry out the fundamenta...

متن کامل

Gesture Profile for Web Services: An Event-Driven Architecture to Support Gestural Interfaces for Smart Environments

Gestural interfaces have lately become extremely popular due to the introduction on the market of low-cost acquisition devices such as iPhone, Wii, and Kinect. Such devices allow practitioners to design, experiment, and evaluate novel interfaces and interactions for new smart environments. However, gesture recognition algorithms are currently the appanage of machine learning experts which somet...

متن کامل

Human Computer Interaction Using Vision-Based Hand Gesture Recognition

With the rapid emergence of 3D applications and virtual environments in computer systems; the need for a new type of interaction device arises. This is because the traditional devices such as mouse, keyboard, and joystick become inefficient and cumbersome within these virtual environments. In other words, evolution of user interfaces shapes the change in the Human-Computer Interaction (HCI). In...

متن کامل

Wearable machine learning for recognizing and controlling smart devices

Augmenting people with wearable technology can enhance their natural sensing, actuation, and communication capabilities. Interaction with smart devices can become easier and less explicit when combining multiple wearables instead of using device-specific apps on a single smartphone. We demonstrate a prototype for smart device control by combining quasi real-time visual device recognition on a s...

متن کامل

Human Computer Interaction Using Vision-Based Hand Gesture Recognition

With the rapid emergence of 3D applications and virtual environments in computer systems; the need for a new type of interaction device arises. This is because the traditional devices such as mouse, keyboard, and joystick become inefficient and cumbersome within these virtual environments. In other words, evolution of user interfaces shapes the change in the Human-Computer Interaction (HCI). In...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016