Differences

This shows you the differences between two versions of the page.

Link to this comparison view

elaina_chai [2020/03/31 13:16] – created 0.0.0.0elaina_chai [2020/04/02 17:08] (current) reric
Line 1: Line 1:
-====== Elaina_Chai ======+====== Elaina Chai ======
  
 {{wiki:Echai portrait.JPG?200x200|Echai portrait.JPG}}  {{wiki:Echai portrait.JPG?200x200|Echai portrait.JPG}} 
Line 9: Line 9:
 Admitted to Ph.D. Candidacy: 2014-2015  Admitted to Ph.D. Candidacy: 2014-2015 
  
-<br> **Email: **echai AT stanford DOT edu +**Email: **echai AT stanford DOT edu 
  
-**Research Interests**: Deep Neural Network Hardware Acceleration on FPGA's, LSTM Networks, Low energy optimizations, Information Theory/Source Coding<br> +**Research Interests**: Deep Neural Network Hardware Acceleration on FPGA's, LSTM Networks, Low energy optimizations, Information Theory/Source Coding 
  
-<br> +\\
  
-Over the last few years, large increases in dataset sizes and comute resources has fueled the development of deep learning algorithms leading to unprecedented gains in performance in tasks such as object detection and speech recognition. However, these algorithms were primarily designed for server environments, and were trained on highly curated data from a limited range of sensor modalities. Attempts have been made, such as in &nbsp;[[1,2]] to extend these algorithms such as LSTM Networks to to multi-modal sensor fusion applications such as activity recognition. However, while these efforts have demonstrated success in terms of classficiation error rates, deploying these algorithms in energy constrained embedded platforms still remains a challenge in terms of latency and energy, compared to the more classical SVM-based approaches. <br> +Over the last few years, large increases in dataset sizes and comute resources has fueled the development of deep learning algorithms leading to unprecedented gains in performance in tasks such as object detection and speech recognition. However, these algorithms were primarily designed for server environments, and were trained on highly curated data from a limited range of sensor modalities. Attempts have been made, such as in [1,2] to extend these algorithms such as LSTM Networks to to multi-modal sensor fusion applications such as activity recognition. However, while these efforts have demonstrated success in terms of classficiation error rates, deploying these algorithms in energy constrained embedded platforms still remains a challenge in terms of latency and energy, compared to the more classical SVM-based approaches. \\ 
  
-<br>[[3]] has shown that by rethinking how data is extracted from camara system front-ends, one can better leverage the information in the raw sensor data (typcially lost in the data pre-processing of a conventional system), and as a result, latency and energy consumption is significantly reduced in an machine learning inference system. We are exploring how to extend this approach to the development of the deep learning algorithm in the processing back-end. +[3] has shown that by rethinking how data is extracted from camara system front-ends, one can better leverage the information in the raw sensor data (typcially lost in the data pre-processing of a conventional system), and as a result, latency and energy consumption is significantly reduced in an machine learning inference system. We are exploring how to extend this approach to the development of the deep learning algorithm in the processing back-end. 
  
-<br> +\\ 
  
-Breaking down the barriers across the machine learning system stack, between the raw sensor development and the deep learning back-end, we can better exploit the stochastic properties of deep learning kernels across the entire embedded hardware inference platform, thereby potentially realizing much higher gains in latency and energy metrics, than if any signle part was developed in isolation. The goal of this research is to develop techniques to more optimally combine raw sensor data from multiple modalities&nbsp;with the development of LSTM neural network architectures for use in embedded reconfigurable systems, and to propose a reconfigurable hardware platform for rapid integration, development and validation of these techniques with state-of-the-art LSTM Neural Network back-ends across a wide range of heterogenous sensor front ends. +Breaking down the barriers across the machine learning system stack, between the raw sensor development and the deep learning back-end, we can better exploit the stochastic properties of deep learning kernels across the entire embedded hardware inference platform, thereby potentially realizing much higher gains in latency and energy metrics, than if any signle part was developed in isolation. The goal of this research is to develop techniques to more optimally combine raw sensor data from multiple modalities with the development of LSTM neural network architectures for use in embedded reconfigurable systems, and to propose a reconfigurable hardware platform for rapid integration, development and validation of these techniques with state-of-the-art LSTM Neural Network back-ends across a wide range of heterogenous sensor front ends. 
  
-<br> +\\
  
-<br> +\\
  
-{{ wiki:Echai 2017.png ?931x313|Echai 2017.png}}<br> +{{ wiki:Echai 2017.png ?931x313|Echai 2017.png}} \\
  
-<br> +\\ 
  
-[[1]] Shao. H., Jiang, H. Wang, F. Zhao, H. An enhancement deep feature fusion method for rotating machinery fault diagnosis. Knowledge-Based Systems 110 200-220 (2017)<br> +\\
  
-[[2]] Yao, S., Hu, S., ZhaoY., Zhang, A., Abdelzaher, T"Deepsense: A unified deep learning framework for time-series mobile sensing data processing." Proceedings of the 26th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, (2017)<br> +[1ShaoH., JiangHWangFZhaoHAn enhancement deep feature fusion method for rotating machinery fault diagnosis. Knowledge-Based Systems 110 200-220 (2017) \\
  
-[[3]] Omid-Zohoor, A., Young, C., Ta, C, Murmann, B., “Towards Always-On Mobile Object Detection: Energy vs Performance Tradeoffs for Embedded HOG Feature Extraction” IEEE Trans Circuits Syst. Video Technol. Vol. PP, (2017)<br>+[2] Yao, S., Hu, S., Zhao, Y., Zhang, A., Abdelzaher, T. "Deepsense: A unified deep learning framework for time-series mobile sensing data processing." Proceedings of the 26th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, (2017) \\ 
 + 
 +[3] Omid-Zohoor, A., Young, C., Ta, C, Murmann, B., “Towards Always-On Mobile Object Detection: Energy vs Performance Tradeoffs for Embedded HOG Feature Extraction” IEEE Trans Circuits Syst. Video Technol. Vol. PP, (2017) \\