Abstract
In this paper, we developed a deep learning-based compression model to reduce the data rate of multichannel action potentials in neural recording experiments. The proposed compression model is built upon a deep compressive autoencoder (CAE) with discrete latent embeddings. The encoder network of CAE is equipped with residual transformations to extract representative features from spikes, which are mapped into the latent embedding space and updated via vector quantization (VQ). The indexes of VQ codebook are further entropy coded as the compressed signals. The decoder network reconstructs spikes with high quality from the latent embeddings. Experimental results on both synthetic and in-vivo datasets show that the proposed model consistently outperforms conventional methods that utilize hand-crafted features and/or signal-agnostic transformations by achieving much higher compression ratios (20-500 ×) and better or comparable signal reconstruction accuracies. Furthermore, we have estimated the hardware cost of the CAE model and shown the feasibility of its on-chip integration with neural recording circuits. The proposed model can reduce the required data transmission bandwidth in large-scale recording experiments and maintain good signal qualities, which will be helpful to design power-efficient and liahtweiaht wireless neural interfaces.
Original language | English (US) |
---|---|
Title of host publication | 2018 IEEE Biomedical Circuits and Systems Conference, BioCAS 2018 - Proceedings |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Electronic) | 9781538636039 |
DOIs | |
State | Published - Dec 20 2018 |
Event | 2018 IEEE Biomedical Circuits and Systems Conference, BioCAS 2018 - Cleveland, United States Duration: Oct 17 2018 → Oct 19 2018 |
Publication series
Name | 2018 IEEE Biomedical Circuits and Systems Conference, BioCAS 2018 - Proceedings |
---|
Other
Other | 2018 IEEE Biomedical Circuits and Systems Conference, BioCAS 2018 |
---|---|
Country/Territory | United States |
City | Cleveland |
Period | 10/17/18 → 10/19/18 |
Bibliographical note
Funding Information:This work was supported by a DARPA grant HR0011-17-2-0060 and the University of Minnesota internal funding through MnDRIVE. We would like to thank Dr. Wing-kin Tam for useful discussion on designing DNN models, and Dr. Nick
Publisher Copyright:
© 2018 IEEE.
Keywords
- Data compression
- Deep learning
- Neural signal processing
- Vector quantization