14
4094 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020 Low-Complexity On-Demand Reconstruction for Compressively Sensed Problematic Signals Ching-Yao Chou , Student Member, IEEE, Kai-Chieh Hsu , Student Member, IEEE, Bo-Hong Cho, Student Member, IEEE, Kuan-Chun Chen , and An-Yeu (Andy) Wu , Fellow, IEEE Abstract—Compressed Sensing (CS) is a revolutionary technol- ogy for realizing low-power sensor nodes through sub-Nyquist sam- pling, and the CS reconstruction engines have been widely studied to fulfill the energy efficiency for real-time processing. However, in most cases, we only want to analyze the problematic signals which account for a very low percentage. Therefore, large efforts will be wasted if we recover uninterested signals. On the other hand, in order to identify the high-risk signals, additional hardware and computation overhead are required for classification other than CS reconstruction. In this paper, to achieve low-complexity on-demand CS reconstruction, we propose a two-stage classification-aided re- construction (TS-CAR) framework. The compressed signals can be classified with a sparse coding based classifier, which provides the hardware sharing potential with reconstruction. Furthermore, to accelerate the reconstruction speed, a cross-domain sparse trans- form is applied from classification to reconstruction. TS-CAR is implemented in electrocardiography based atrial fibrillation (AF) detection. The average computational cost of TS-CAR is 2.25× fewer compared to traditional frameworks when AF percentage is among 10% to 50%. Finally, we implement TS-CAR in TSMC 40 nm technology. The post-layout results show that the proposed intelligent CS reconstruction engine can provide a competitive area- and energy-efficiency compared to state-of-the-art CS and machine learning engines. Index Terms—Compressed sensing, on-demand reconstruction, compressed learning, sparse transform, hardware sharing. I. INTRODUCTION W IRELESS health-care monitoring offers great potential to substantially improve patient health, quality of life and outcomes [1]. Since the electrocardiography (ECG) signal recorded from the electrical activity of the heart over a period of time has been utilized for diagnosis for many diseases, the ECG Manuscript received May 2, 2019; revised September 16, 2019, March 29, 2020, and June 16, 2020; accepted June 25, 2020. Date of publication July 2, 2020; date of current version July 24, 2020. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Tokunbo Ogunfunmi. This work was supported in part by the Ministry of Science and Technology of Taiwan (MOST 106-2221-E-002 -204 -MY3, MOST 108-2633- E-002-001), in part by National Taiwan University (NTU-108L104039), and in part by Intel Corporation. (Corresponding author: An-Yeu (Andy) Wu.) Ching-Yao Chou, Kuan-Chun Chen, and An-Yeu (Andy) Wu are with the Graduate Institute of Electronics Engineering, National Taiwan Univer- sity, Taipei 10617, Taiwan (e-mail: [email protected]; kcchen@ access.ee.ntu.edu.tw; [email protected]). Kai-Chieh Hsu and Bo-Hong Cho are with the Electrical Engi- neering, National Taiwan University, Taipei 10617, Taiwan (e-mail: [email protected]; [email protected]). Digital Object Identifier 10.1109/TSP.2020.3006766 telemonitoring [2] is recognized as a promising technique to re- alize telemedicine. Unfortunately, ECG sensors encounter band- width limit [1], and the traditional measured-and-compressed compression approaches cause high power consumption [3]. To extend the lifetime of resource-limited sensor nodes, com- pressed sensing (CS) [4] is proposed as a revolutionary technique to efficiently acquire signals through sub-Nyquist sampling [3], [5]. In addition, the reconstruction of compressively sensed physiological signals with hardware solution has been widely studied to fulfill the energy efficiency for real-time processing at the resource-constrained edge devices [6]–[8]. The prior CS solvers reconstruct every compressively sensed signal, so doctors can make diagnoses based on reconstructed signals. However, in most cases, doctors only want to analyze the problematic physiological signals. Hence, large efforts are wasted in the prior CS solvers due to the fact that the re- construction is extremely energy-intensive [9] and uninterested (or normal) signals are usually far more than interested (or problematic) signals [10]. To on-demand reconstruct problematic CS signals, the com- pressed learning (CL) frameworks [11]–[15] have been pro- posed. CL performs the detection in the compressed domain and can then be combined with the CS solver, as shown in Fig. 1(a), which we called CL-REC. Nevertheless, to identify which signals are problematic, CL-REC required additional machine learning (ML) chip [16], [17] for classification other than CS chip for reconstruction. On the other hand, the residual learning (ResL) frame- works [18]–[21], as shown in Fig. 1(b), can share the hardware between classification and reconstruction by comparing the re- construction error. However, the built-in classification requires almost full reconstruction for every compresseively sensed signal. In this paper, to achieve low-complexity on-demand CS re- construction with hardware sharing between two stages, we first propose the CL based on task-driven dictionary learning plus reconstruction (CL-TDDL + REC) framework. Second, we enhance the CL-TDDL + REC and propose the two-stage classification-aided reconstruction (TS-CAR) framework, as shown in Fig. 1(c). In TS-CAR framework, the on-demand reconstruction not only shares the hardware with classification but also reuse the information from classification stage and gets accelerated. The main contributions of this paper are as follows. 1) To achieve low-complexity on-demand CS reconstruc- tion, a two-stage classification-aided reconstruction 1053-587X © 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See https://www.ieee.org/publications/rights/index.html for more information. Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

4094 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020

Low-Complexity On-Demand Reconstruction forCompressively Sensed Problematic Signals

Ching-Yao Chou , Student Member, IEEE, Kai-Chieh Hsu , Student Member, IEEE,Bo-Hong Cho, Student Member, IEEE, Kuan-Chun Chen , and An-Yeu (Andy) Wu , Fellow, IEEE

Abstract—Compressed Sensing (CS) is a revolutionary technol-ogy for realizing low-power sensor nodes through sub-Nyquist sam-pling, and the CS reconstruction engines have been widely studiedto fulfill the energy efficiency for real-time processing. However, inmost cases, we only want to analyze the problematic signals whichaccount for a very low percentage. Therefore, large efforts will bewasted if we recover uninterested signals. On the other hand, inorder to identify the high-risk signals, additional hardware andcomputation overhead are required for classification other than CSreconstruction. In this paper, to achieve low-complexity on-demandCS reconstruction, we propose a two-stage classification-aided re-construction (TS-CAR) framework. The compressed signals can beclassified with a sparse coding based classifier, which provides thehardware sharing potential with reconstruction. Furthermore, toaccelerate the reconstruction speed, a cross-domain sparse trans-form is applied from classification to reconstruction. TS-CAR isimplemented in electrocardiography based atrial fibrillation (AF)detection. The average computational cost of TS-CAR is 2.25×fewer compared to traditional frameworks when AF percentageis among 10% to 50%. Finally, we implement TS-CAR in TSMC40 nm technology. The post-layout results show that the proposedintelligent CS reconstruction engine can provide a competitivearea- and energy-efficiency compared to state-of-the-art CS andmachine learning engines.

Index Terms—Compressed sensing, on-demand reconstruction,compressed learning, sparse transform, hardware sharing.

I. INTRODUCTION

W IRELESS health-care monitoring offers great potentialto substantially improve patient health, quality of life

and outcomes [1]. Since the electrocardiography (ECG) signalrecorded from the electrical activity of the heart over a period oftime has been utilized for diagnosis for many diseases, the ECG

Manuscript received May 2, 2019; revised September 16, 2019, March 29,2020, and June 16, 2020; accepted June 25, 2020. Date of publication July 2,2020; date of current version July 24, 2020. The associate editor coordinatingthe review of this manuscript and approving it for publication was Prof. TokunboOgunfunmi. This work was supported in part by the Ministry of Science andTechnology of Taiwan (MOST 106-2221-E-002 -204 -MY3, MOST 108-2633-E-002-001), in part by National Taiwan University (NTU-108L104039), and inpart by Intel Corporation. (Corresponding author: An-Yeu (Andy) Wu.)

Ching-Yao Chou, Kuan-Chun Chen, and An-Yeu (Andy) Wu are withthe Graduate Institute of Electronics Engineering, National Taiwan Univer-sity, Taipei 10617, Taiwan (e-mail: [email protected]; [email protected]; [email protected]).

Kai-Chieh Hsu and Bo-Hong Cho are with the Electrical Engi-neering, National Taiwan University, Taipei 10617, Taiwan (e-mail:[email protected]; [email protected]).

Digital Object Identifier 10.1109/TSP.2020.3006766

telemonitoring [2] is recognized as a promising technique to re-alize telemedicine. Unfortunately, ECG sensors encounter band-width limit [1], and the traditional measured-and-compressedcompression approaches cause high power consumption [3].

To extend the lifetime of resource-limited sensor nodes, com-pressed sensing (CS) [4] is proposed as a revolutionary techniqueto efficiently acquire signals through sub-Nyquist sampling [3],[5]. In addition, the reconstruction of compressively sensedphysiological signals with hardware solution has been widelystudied to fulfill the energy efficiency for real-time processingat the resource-constrained edge devices [6]–[8].

The prior CS solvers reconstruct every compressively sensedsignal, so doctors can make diagnoses based on reconstructedsignals. However, in most cases, doctors only want to analyzethe problematic physiological signals. Hence, large efforts arewasted in the prior CS solvers due to the fact that the re-construction is extremely energy-intensive [9] and uninterested(or normal) signals are usually far more than interested (orproblematic) signals [10].

To on-demand reconstruct problematic CS signals, the com-pressed learning (CL) frameworks [11]–[15] have been pro-posed. CL performs the detection in the compressed domainand can then be combined with the CS solver, as shown inFig. 1(a), which we called CL-REC. Nevertheless, to identifywhich signals are problematic, CL-REC required additionalmachine learning (ML) chip [16], [17] for classification otherthan CS chip for reconstruction.

On the other hand, the residual learning (ResL) frame-works [18]–[21], as shown in Fig. 1(b), can share the hardwarebetween classification and reconstruction by comparing the re-construction error. However, the built-in classification requiresalmost full reconstruction for every compresseively sensedsignal.

In this paper, to achieve low-complexity on-demand CS re-construction with hardware sharing between two stages, wefirst propose the CL based on task-driven dictionary learningplus reconstruction (CL-TDDL + REC) framework. Second,we enhance the CL-TDDL + REC and propose the two-stageclassification-aided reconstruction (TS-CAR) framework, asshown in Fig. 1(c). In TS-CAR framework, the on-demandreconstruction not only shares the hardware with classificationbut also reuse the information from classification stage and getsaccelerated. The main contributions of this paper are as follows.

1) To achieve low-complexity on-demand CS reconstruc-tion, a two-stage classification-aided reconstruction

1053-587X © 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 2: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

CHOU et al.: LOW-COMPLEXITY ON-DEMAND RECONSTRUCTION FOR COMPRESSIVELY SENSED PROBLEMATIC SIGNALS 4095

Fig. 1. Different frameworks of on-demand CS reconstruction: (a) CL-SVM/CNN + Reconstruction, (b) Residual Learning, and (c) ProposedTwo-Stage Classification-Aided Reconstruction. (The clocks represent thecomputational cost.)

(TS-CAR) framework is proposed. TS-CAR is able toreconstruct target signals because it can conduct classifi-cation on highly compressed signals. Therefore, TS-CARachieves the on-demand reconstruction. Furthermore, thereconstruction process can be greatly accelerated becauseof the cross-domain sparse transform applied from classi-fication to reconstruction.

2) This framework is implemented in ECG-based AFdetection. In addition to better classification performanceand reconstruction quality, the average computational costof on-demand reconstruction required by the proposedTS-CAR is 2.37× fewer compared to ResL and 2.25×fewer compared to CL-REC when the percentage of AFsignals is among 10% to 50%.

3) Based on the proposed TS-CAR framework, the first2-in-1 intelligent CS solver is proposed. The proposedCS solver shares the hardware between two stages becauseof the same sparse coding computation. Furthermore, thesolver is based on fast iterative shrinkage-thresholdingalgorithm (FISTA). To our best knowledge, it is the firsttime FISTA is implemented in integrated circuit design.TS-CAR is implemented in TSMC 40 nm technology,achieving area-energy efficiency 1.64× compared to ResLand 6.83× compared to CL-REC.

The rest of this paper is organized as follows. In Section II,we present the background on compressed sensing, CS re-construction algorithms, and dictionary learning. We also in-troduce some closely related CS frameworks. In Section III,we present the proposed TS-CAR framework. The numericalexperiments and analysis of computational cost are shown inSection IV. Section V presents the hardware architecture andVLSI implementation results. Finally, we conclude this paper inSection VI.

A. Notations

Vectors and matrices are denoted by bold lowercase anduppercase letters, respectively. α[i] denotes the i-th entry ofa vector α and A[i, j] denotes the (i, j)-th entry of a matrixA. When Λ is a finite set of indices, αΛ denotes the vectorthat carries the entries of α indexed by Λ and AΛ is the matrixwhose columns are those of A indexed by Λ. The subscript,t, of values, vectors or matrices suggests they are in the t-thiteration. We denote the transpose of vectors and matrices by(·)T . We write the absolute value by | · |, �p-norm of a vector by‖ · ‖p and the Frobenius norm of a matrix by ‖ · ‖F . The functionmax(·) outputs the maximum value of the arguments and sgn(·)extracts the sign of a real number. A � B denotes that B −Ais positive semi-definite.

II. BACKGROUND

A. Preliminaries

1) Compressed Sensing (CS): CS [4], [22] is a revolutionarytechnique, which allows acquiring and compressing signalssimultaneously. If the signal is of sparsity in a specific domain,CS can guarantee almost perfect reconstruction even with fewermeasurements than Nyquist rate requirement. We can modelthe compressed sensing signal and the original signal by a CSmeasurement matrix as shown in below

x = Φx, (1)

where x ∈ RM , Φ ∈ RM×N , x ∈ RN , N is the original di-mension and M is the compressed dimension. The entries ofCS measurement matrix can be chosen randomly as long as thecolumns in Φ is incoherent with the basis in which the signalhas a sparse representation [4]. In fact, the criteria can be met bychoosing these entries from identically independent distribution(i.i.d.) of Bernoulli(0.5) or Gaussian(0, 1) [22].

2) CS Reconstruction Algorithms: The reconstruction of xfrom x is an underdetermined problem. However, if x hassparsity in a specific domain as (2), there is a high probabilitythat it can be perfectly reconstructed.

x = Dα, (2)

where D ∈ RN×d is the sparse representation basis of x andα ∈ Rd is the sparse coefficients (SC). One method for recon-struction is using the following �1-norm optimization

α = argminα∈Rd

‖α‖1 s.t. ‖x−Φ ·Dα‖2 ≤ ε (3)

via algorithms, such as orthogonal matching pursuit (OMP) [23],subspace pursuit (SP) [24] and fast iterative shrinkage-thresholding algorithm (FISTA) [25].

Throughout this paper, we apply FISTA to obtain the sparsecoefficients because its better performance compared to OMPand SP, which is shown in Section IV-C. In addition, FISTA is ahardware friendly algorithm for our proposed framework, whichwe will elaborate in Section V-A. FISTA can be utilized to solve

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 3: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

4096 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020

�1-norm optimization via this formulation

α(x,D, λ) = argminα∈Rd

1

2‖x−Dα‖22 + λ‖α‖1, (4)

where λ is the sparsity penalty coefficient. Since FISTA is agradient descent based algorithm, it can be used to solve anumber of signals at the same time if we modify (4) as

A(X,D, λ) = argminA∈Rd×n

L(X,D, λ) (5)

L(X,D, λ) � 1

2‖X −D ·A‖2F + λ‖A‖1, (6)

where X = [x1 · · ·xn] ∈ RN×n.3) Dictionary Learning (DL): In real world, signals usually

have sparsity on a specific domain. However, how to find thebasis where signals share the sparse representation is a bigchallenge and dictionary learning (DL) is proposed to dealwith such problem. The traditional DL focused on finding thedictionary, where the target signal can have a sparse coefficientsreconstructing the target signal perfectly [26]–[28]. On the otherhand, dictionary learning can be viewed as a transform layer toattain SC having predictive information, so the classifier basedon the SC can be of low complexity and high accuracy [29], [30].The classifier can be a support vector machine and separatelyoptimized [31] or a linear model and co-optimized with thedictionary [29].

B. Related Work

1) Compressed Learning (CL): The compressed learning(CL) frameworks directly perform the inference, such as clas-sification or regression, in the compressed domain, which arebeneficial both in the compressed sensing and machine learn-ing points of view. From the CS viewpoint, it eliminates theabundant cost of recovering irrelevant data; in other words, CLis like a sieve and makes it possible to only recover the desiredsignals. From the ML viewpoint, CS can be regarded as efficientuniversal sparse dimensionality reduction from the data domainto the measurement domain.

The Johnson-Lindenstrauss (JL) embedding of a data setensures the Euclidean distances between all pairs of points arepreserved within a small distortion after random projection [32].Hence, random projections, which are used in CS, preserveinner products of signals. Moreover, since inner products are atthe core of several machine learning algorithms, the theoreticalguarantees and the application of CL have thus recently receivedattention.

Theoretical guarantees of statistical learning in compresseddomain are derived by employing results from JL embeddingand from CS as building blocks [11], [12]. Specially, Calderbanket al. [11] derive bounds to show that performance of a linearSVM in the compressed domain is close to the performanceof the best linear classifier in the uncompressed domain. Inaddition, sharp bounds are found by discovering that not alldistance are important, and not all details of the data matter inlearning tasks [33], [34].

Correlation-based features are extracted from CS EEG forseizure detection and from compressed videos for action recog-nition in [13] and [14], respectively. Different from correlationfeatures which are linear, non-linear features are extracted di-rectly from the compressively-sensed data using CNN in [15].Instead of learning in the measurement domain, the discrim-inative information is sifted from the compressed signals be-fore making inference in the sub-eigenspace [35]–[37]. As aresult, the preprocessing of the CS measurements can reducethe memory overhead and computational cost, while enhancingthe classification performance due to noise mitigation.

CL can then be combined with the reconstruction engine toachieve the on-demand CS reconstruction, which we called CL-REC. Nevertheless, CL-REC requires additional hardware suchas ML chip other than CS chip. Furthermore, the classificationprocess is irrelevant to the reconstruction process, and cannothelp to accelerate the reconstruction.

2) Residual Learning (ResL): The residual learning (ResL)frameworks aim to classify the signals by comparing the re-construction error from a learned discriminative dictionarywhich consists of class-specific sub-dictionaries for each class.DLSI [19] learns the sub-dictionaries with a novel penalty termto make the sub-dictionaries incoherent. COPAR [20] furtherpoints out the non-overlapping subspaces assumption is unre-alistic in practice. Therefore, a shared dictionary also needsto be learned to represent the common features. Moreover,LRSDL [21] contends that the subspace spanned by columnsof the shared dictionary must have low rank. Otherwise, class-specific features may also get represented by the shared dic-tionary, diminishing the classification ability. In addition tolearning discriminative dictionary in the Nyquist domain, [18]further transforms the learned dictionary according to the CSmeasurement matrix. In this way, the compressed signals canbe reconstructed with built-in classification. Although ResLframeworks can share hardware between reconstruction andclassification with reusing sparse coefficients, they need fullreconstruction for every signal. The comparison of the relatedCS frameworks is summarized in Table I.

III. PROPOSED TWO-STAGE CLASSIFICATION-AIDED

RECONSTRUCTION FRAMEWORK

Fig. 2 shows the proposed two-stage classification-aided re-construction (TS-CAR) framework. The green region in Fig. 2depicts the block diagram of classification stage, which will beelaborated in Section III-A. On the other hand, the red region inFig. 2 describes the reconstruction stage, and it will be furtherillustrated in Section III-B. The arrows between two stagesshow how the information from classification is reused in thereconstruction stage.

In addition to viewing TS-CAR from the task perspectives,Fig. 2 also demonstrates the off-line and on-line of proposedTS-CAR framework. At the same time, Algorithm 3 summa-rizes how we can obtain the dictionaries, classifier and sparsetransform from given data and parameters, while Algorithm 4sums up how we classify the compressed signals and conductclassification-aided reconstruction if necessary.

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 4: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

CHOU et al.: LOW-COMPLEXITY ON-DEMAND RECONSTRUCTION FOR COMPRESSIVELY SENSED PROBLEMATIC SIGNALS 4097

TABLE ICOMPARISON OF CS RECONSTRUCTION FRAMEWORKS

Fig. 2. Proposed TS-CAR framework, including CL-TDDL for the classifica-tion stage and CAR for the on-demand reconstruction stage.

A. Classification Stage: Compressed Learning-TDDL

In this subsection, we consider the binary classification ofnormal ECG signals and AF ECG signals. However, this clas-sification framework can be extended to other physiologicalsignals as well as multi-class classification. We first elaboratehow classification dictionary and classifier model are learned inproposed TS-CAR. After we have these models, we present thecompressed inference. The off-line and on-line for classificationare shown in the green region of Fig. 2.

1) Classification Dictionary and Model Learning: Consider-ing we have n sample data X ∈ RN×n and their labels y ∈ Rn,we want to learn a low-complexity classification model whichcan be utilized in edge computing. If the signals have sparsity,predictive DL can serve as an energy-efficient classifier buthave better classification accuracy than the conventional MLalgorithms, such as SVM and CNN [30]. In addition, predic-tive DL shares the same sparse coding computation with CSreconstruction, so it provides the hardware sharing potentialfor our proposed two-stage classification-aided reconstructionframework. Thus, we utilize predictive DL as our classificationengine.

Predictive DL consists of two layers. The first layer serves asa feature extraction layer to transform target signals to sparsecoefficients, while the second layer is a classifier based on thesparse coefficients from previous layer. Predictive DL optimizesthe following empirical supervised loss

(Dclass,W ) = argminD∈Dc,W∈W

fc(D,W , λclass), (7)

where Dc � {D = [d1 · · ·ddclass] ∈ RN×dclass s.t. ‖dj‖2 ≤

1, ∀j = 1, 2, . . . , dclass},W � Rdclass×2, λclass is the sparsitypenalty used in classification, dclass is the number of atoms inthe classification dictionary and

fc(D,W , λclass) �1

n

n∑

i=1

�c (yi,W ,αi) +ν

2‖W ‖2F , (8)

with αi = α(xi,D, λclass) defined in (4), ν being the regular-ization coefficient, and �c defined as

�c (yi,W ,αi) � ‖W Tαi − yi‖2Fyi is the one-hot encoding of yi. (9)

In fact, �c can be chosen as any supervised loss and here weadopt the linear loss with one-hot encoding.

A naive way to optimize (7) is to optimize Dclass and W se-quentially. Although this method enables easy training process,the accuracy cannot be optimal. In [29], the authors proposedtask-driven dictionary learning (TDDL) to co-optimize Dclass

and W simultaneously with the help of back propagation.Therefore, we adopt TDDL as our optimizer and modify it tobe the mini-batch version (MB-TDDL), which is summarizedin Algorithm 1.

2) Compressed Inference: The reconstruction of compressedsignals usually takes much energy, which is an critical issue foredge devices. However, in most cases, we only want to furtheranalyze the interested signals, such as problematic physiologicalsignals. Therefore, if we can classify the target signals intonormal and problematic when they are highly compressed, wecan conduct the reconstruction for problematic signals only. Inother words, we want to execute the compressed inference priorto the reconstruction, so the on-demand reconstruction can berealized.

Given an unknown compressed sensing signal x as obtainedby (1), we want to classify it into normal or AF via Dclass

and W derived in (7). However, since Dclass is learned in theNyquist domain, we need to first transform Dclass into anotherbasis where x has sparsity. The new dictionary for x can beobtained by

Θclass = Φ ·Dclass. (10)

x can have sparse coefficients from Θclass as long as Φ andDclass are incoherent enough [4]. Therefore, we can obtain the

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 5: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

4098 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020

Algorithm 1: Mini-batch Task-Driven Dictionary Learning.Require

Data: X ∈ RN×n, Label: y ∈ Rn, Initial Dictionary:D0 ∈ D, Initial Weight: W 0 ∈ W , �1-penalty: λ, Thenumber of iterations: T , Regularization: ν, Step size: η

1: for t = 1 to T do2: Pick nb data and concatenate as

Xt = [x1 · · ·xnb].

3: Compute At = [α1 · · ·αnb] using �1-norm

optimization (e.g. FISTA):

At ← argminA∈Rd×nb

L(Xt,Dt−1, λ).

4: Compute the active set and Bt = [β1 · · ·βnb]:

5: for i = 1 to nb do

Λ← {j ∈ {1, 2, . . . , d} : At[j, i] �= 0},βΛc = 0,

βΛ = (DTt−1ΛDt−1Λ)

−1∇αΛ�c(yi,W t−1,αi),

βi ← β.

6: end for7: Update the parameters by a projected gradient step:

Dt ← ΠD[Dt−1 − η

(−Dt−1BtATt

+ (Xt −Dt−1At)BTt

)],

W t ← ΠW [W t−1 − η (∇W �c (yt,W t−1,At)

+ νW t−1)] ,

where ΠW and ΠD are orthogonal projections on thesetsW and D, respectively.

8: end for9: return DT , W T

predictive sparse coefficients of x by

αclass = argminα∈Rdclass

F (α; x,Θclass, λclass),

F (α; x,Θ, λ) = f(α; x,Θ) + g(α;λ)

� 1

2‖x−Θα‖22 + λ‖α‖1. (11)

For simplicity, we will drop dependencies on x,Θ, λ, whichshould be clear by the context.

Here we adopt FISTA as our optimizer and it works asfollowing. The gradient of the reconstruction loss is

∇αf(α) = ΘTclassΘclassα−ΘT

classx. (12)

Since ΘTclassx is unchanged among iterations, we can calculate

ΘTclassx before starting iteratively optimization. Given the step

size, η, we can start to execute FISTA for �1-optimization. Wefirst initialize the sparse coefficient vector (α) and the shrunkvector (β) by zero vector (0) or by given vector. In each it-eration, t, we update ∇αf(αt−1) in (12) by only calculationof ΘT

classΘclassαt−1. Then, the updated sparse coefficients

Algorithm 2: Sparse Coding Computation Using FISTA.Require

Compressed Signal: x, Dictionary: Θ, Initial SparseCoefficients: α0, �1-penalty: λ, The number ofiterations: T , Step size: η

1: κ0 ← 1, β0 ← α0.2: for t = 1 to T do3: Calculate the gradient :

∇αf (αt−1) = ΘTΘαt−1 −ΘT x

4: Undergo shrinkage operator:

βt = Sλη (αt−1 − η∇αf (αt−1)) .

5: Compute the coefficient:

κt =1 +

√1 + 4κ2

t−12

6: Update with momentum:

αt = βt +κt−1 − 1

κt(βt − βt−1).

7: end for8: return βT .

undergo the shrinkage operator, which introduces the sparsityconstraint as (13)

βt = Sλclassη (αt−1 − η∇αf (αt−1)) , (13)

where Sχ(α) is the shrinkage operator of threshold χ and willupdate each entry of α by

Sχ(α)[j] = (|α[j]| − χ)+ sgn(α[j]),

j = 1, 2, . . . , dclass (14)

(ξ)+ = max(ξ, 0). (15)

Asλ becomes larger, there will be more zero entries inβt, whichcan directly be observed via (14) and (15). Finally, the sparsecoefficients are obtained by the linear combination of currentshrunk vector and the previous shrunk vector as (16), which canbe viewed as a momentum acceleration method.

αt = βt +κt−1 − 1

κt(βt − βt−1), (16)

where κt =(1 +

√1 + 4κ2

t−1)/2 and κ0 = 1. After the fixed

iteration (T ) or termination threshold (ζ) we will obtain thesparse coefficients, namely, αclass. Multiplying αclass by W T

and choosing the bigger probability, we can classify x. To thebest of the author’s knowledge, it is the first time TDDL isutilized for the compressed learning.

Here we state the complexity result of FISTA in our setting byfollowing the theorem in [25] and encourage interested readersto check the detail. We first take the derivative of ∇αf(α) in(12) and show that it is L-smooth, where L is the maximumeigenvalue of ΘTΘ as below

∇2αf(α) = ΘTΘ � LI, (17)

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 6: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

CHOU et al.: LOW-COMPLEXITY ON-DEMAND RECONSTRUCTION FOR COMPRESSIVELY SENSED PROBLEMATIC SIGNALS 4099

Fig. 3. The sparse coding coefficients of αrec and αrec.

Fig. 4. The original AF signal and reconstructed AF signals via sparse coefficients from (a) classification stage, (b) sparse transform stage, and (c) CAR stage.

where ΘTΘ � LI suggests LI −ΘTΘ is positive semi-definite. Then, let {αt} and {βt} be generated by FISTA inAlgorithm 2 and choose stepsize η = 1/L. We have the follow-ing result

F (βt)− F (β�) ≤ 2L(t+ 1)2

‖β0 − β�‖22. (18)

In other words, if we want to get an ζ-optimal solution β′ suchthat F (β′)− F (β�) ≤ ζ, we need at most C iterations, where

C =

⌈√2Lζ‖β0 − β�‖2 − 1

⌉. (19)

There are two direct observations from (19): 1) we know thecomplexity result is O(1/T 2) or O(1/√ζ) and 2) A goodinitialization point can reduce the error by a constant factor. Theobservation 2) motivates us to think if we can reuse the informa-tion in the classification stage to accelerate the reconstructionspeed. We can consider the αclass from the classification stageas coarse sparse coefficients. With an effective transform, it willdefinitely serve as a good initialization point for the refinedsparse coefficients in the reconstruction stage. We propose thesparse transform in Section III-B1 and use Figs. 3 and 4 todemonstrate its efficacy in Section III-B2.

In fact, there are many other algorithms can be utilized for�1-optimization, such as OMP and SP. However, in our case,FISTA outperforms them in both classification accuracy andreconstruction quality, which will be further compared in Sec-tion IV-C. In addition, FISTA has good hardware feasibility andhas similar architecture for both classification and reconstruc-tion, which is elaborated more in Section V-A. Therefore, we

utilize FISTA to acquire SC in both training and inference forclassification and reconstruction. The FISTA for �1-optimizationis summarized in Algorithm 2.

B. Reconstruction Stage: On-Demand Classification-AidedReconstruction

In this subsection, we consider the reconstruction of the com-pressed signals, such as compressed sensing ECG signals. Tostart with, we learn the reconstruction dictionary in the Nyquistdomain. In addition, we reuse the information obtained at theclassification stage, namely, the predictive sparse coefficients(αclass). To be more specific, we proposed a sparse transform,so the sparse coefficients (αclass) based on the classificationdictionary (Dclass) are transformed to the reconstruction dictio-nary. As for the on-line stage, we will execute the reconstructiononly if the target signal is classified as problematic signal. Wecan reuse the information in the classification stage via thelearned sparse transform, so the reconstruction stage has a goodinitial point. This is the reason why it is named classification-aided reconstruction (CAR). CAR can largely accelerate thereconstruction speed, which will be further illustrated inSection IV-B2. The red region in Fig. 2 illustrates the connectionbetween off-line training and on-line inference. In addition, thetraining and inference process are summarized in Algorithm 3and Algorithm 4, respectively.

1) Reconstruction Dictionary and Sparse Transform Learn-ing: In the proposed TS-CAR, we only focus on theon-demand reconstruction, which means we only want to re-construct the interested signals, for example AF signals in ECGmonitoring. Therefore, considering we have nAF AF signals,

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 7: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

4100 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020

Algorithm 3: Two-Stage Classification-Aided Reconstruc-tion (Off-Line).

RequireData: X ∈ RN×n, Label: y ∈ Rn, Initial Dictionaries:Dc0 ∈ Dc, Dr0 ∈ Dr, Initial Weight: W 0 ∈ W ,�1-penalty: λclass, λrec, The number of iterations:Tclass, Trec, Regularization: ν, Step size: η

1: Compute classification dictionary:

Dclass,W←MB-TDDL(X,y,Dc0,W 0, λclass, Tclass, ν, η).

2: Compute reconstruction dictionary:

Drec ← ODL(XAF ,Dr0, λrec, Trec).

3: Compute sparse transform:

Aclass ← argminA∈Rdclass×nAF

L(XAF ,Dclass, λclass),

Arec ← argminA∈Rdrec×nAF

L(XAF ,Drec, λrec),

T ← argminT∈Rdrec×dclass

‖Arec − T ·Aclass‖F .

4: return Dclass, W , Drec, T

XAF = [x′1 · · ·x′nAF] ∈ RN×nAF ⊂X , we can obtain the re-

construction dictionary via following optimization

Drec = argminD∈Dr

fr(D, λrec), (20)

where Dr � {D = [d1 · · ·ddrec] ∈ RN×drec s.t. ‖dj‖2 ≤

1, ∀j = 1, 2, . . . , drec}, λrec is the sparsity penalty used inreconstruction, drec is the number of atoms of atoms in thereconstruction dictionary and

fr(D, λrec) �1

nAF

nAF∑

i=1

�r(x′i,D, λrec), (21)

�r(x′i,D, λrec) =

1

2‖x′i −Dα′i‖22 + λrec‖α′i‖1, (22)

with α′i = α(x′i,D, λrec) defined in (4). We can obtain thereconstruction dictionary via online dictionary learning (ODL),which shows the state-of-the-art performance in both speed andoptimization [28].

The training of reconstruction dictionary and classificationdictionary share with the same building block, namely, sparsecoding computation, but the �1 penalty should be carefully cho-sen. For classification dictionary, λclass should better be chosennot too small since strong �1 penalty implies high sparsity.With high sparsity, the classifier (W ) can be designed withlow complexity [38]. On the other hand, for reconstructiondictionary, λrec should better be chosen not too big. This isdue to the fact that we want to more focus on the reconstruction�2 loss.

Inspired by only different �1 penalty choice, we considerthe two sparse coefficients of classification dictionary and re-construction dictionary may be connected by some transform.

Algorithm 4: Two-Stage Classification-Aided Reconstruc-tion (On-Line).

RequireData: x ∈ RM , Classification Dictionary:Dclass ∈ Dc, Reconstruction Dictionary: Drec ∈ Dr,

Weight: W ∈ W , Sparse Transform: T ∈ Rdrec×dclass ,Sensing Matrix: Φ ∈ RM×N , �1-penalty forclassification: λclass, �1-penalty for reconstruction: λrec

1: Θclass ← Φ ·Dclass, Θrec ← Φ ·Drec

Stage One: Classification2: Compute sparse coding for classification (e.g. FISTA):

αclass ← argminα∈Rdclass

1

2‖x−Θclassα‖22 + λclass‖α‖1.

3: Classify the signal:

p = W Tαclass.

4: if p[1] > p[2] then5: yest ← 0.6: return yest.7: else8: yest ← 1.9: Conduct classification-aided reconstruction:Stage Two: Classification-Aided Reconstruction

10: Sparse transform:

αrec = Tαclass.

11: Compute sparse coding for reconstruction withinitial αrec (e.g. FISTA):

αrec ← argminα∈Rdrec

1

2‖x−Θrecα‖22 + λrec‖α‖1.

12: Compute reconstructed signal:

xest = Drecαrec.

13: return yest, xest.14: end if

Some similar issues are concerned in computer vision society.For example, the sparse coefficients of the signal in the low-resolution dictionary can be used to well reconstruct its pairedsignal in the high-resolution dictionary [39]. In addition, thecross-view action recognition can relate pictures in one camerato other cameras by cross-domain transform [40]. Therefore, weproposed a linear sparse transform to map αclass to the sparsecoefficients on Drec. This sparse transform is learned by thefollowing

T = argminT∈Rdrec×dclass

‖Arec − T ·Aclass‖F , (23)

where Aclass and Arec are the sparse coefficients of XAF onDclass and Drec, respectively.

2) Classification-Aided Reconstruction: Consider the samecompressed signal in classification stage, x, if x is classified to

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 8: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

CHOU et al.: LOW-COMPLEXITY ON-DEMAND RECONSTRUCTION FOR COMPRESSIVELY SENSED PROBLEMATIC SIGNALS 4101

be problematic, we will execute the on-demand reconstruction.Because Drec is learned in the Nyquist domain, we also needto transform the learned dictionary according to the Φ via

Θrec = Φ ·Drec. (24)

After this step, we can start classification-aided reconstruction(CAR).

We first use sparse transform to map the sparse coefficientsof the classification dictionary to the reconstruction dictionary,namely,

αrec = Tαclass. (25)

Taking αrec as an initial point, we then utilize the similaroperations in (12), (13), (16), or Algorithm 2, to solve thefollowing optimization

αrec = argminα∈Rdrec

F (α; x,Θrec, λrec) (26)

with λclass, dclass and Θclass are replaced by λrec, drec andΘrec, respectively. Fig. 3 shows an example of proposed sparsetransform. It is observed that αrec is very closed to αrec. There-fore, it is reasonable that the information reuse can accelerate thereconstruction speed. The results showed αrec can have half asmany iterations required by the zero vector initial point, whichwill further be presented in Section IV-B2.

Fig. 4 compares the reconstruction results based on αclass in(11), αrec in (25) and αrec in (26), where αrec and αrec arealso depicted by blue and red stem plots in Fig. 3. It is observedαclass can roughly reconstruct x and catch the major peaks.After sparse transform, αrec can mitigate the fluctuation of smallamplitude region. Finally, the αrec can perfectly reconstruct xwith less than 6% percent root mean square difference (PRD),where PRD is defined as

PRD � ‖x− xrec‖2‖x‖2 × 100% (27)

and xrec is the reconstructed signal. Fig. 4 also implies Dclass

can be viewed as a low-resolution dictionary, possibly dueto the fact of large �1 penalty, while Drec can be viewedas a high-resolution dictionary. Therefore, the learned sparsetransform served as an effective but efficient linear mappingfrom low-resolution to high-resolution. To the best of the au-thor’s knowledge, it is the first time the cross-domain trans-form is applied to different tasks, namely, from classification toreconstruction.

IV. NUMERICAL EXPERIMENTS AND ANALYSIS OF

COMPUTATIONAL COST

A. Experimental Settings

We use a case study of AF detection to validate the benefits ofour proposed algorithm. Raw ECG signals were recorded fromthe intensive care unit (ICU) of stroke in National Taiwan Uni-versity Hospital (NTUH). There are total 231 non-AF recordsand 58 AF records labeled by doctors, where each record spansten minutes at sample frequency of 512 Hz. The original ECGsignals are segmented into several N = 512-point samples. Werandomly sample 3500 seconds from non-AF and AF records

TABLE IIPARAMETERS SETTING FOR LEARNING MODELS

Fig. 5. The accuracy under different number of atoms in the dictionary.

respectively and divide these samples into 5000 for trainingand 2000 for testing. Then, we apply random projection in CSon each sample to obtain the compressed sensing signal withmeasured dimension, M = 128, using Bernoulli matrix, whoseentries are randomly chosen from i.i.d.Bernoulli(0.5). Thesimulation is run with 100 trails to ensure the robustness underrandom measurement matrix.

We compare our system with other state-of-the-art algorithms.In Section IV-B1, we compare the classification accuracy andcomputational cost, which is the required number of multiplica-tions, of proposed compressed learning by TDDL (CL-TDDL)with residual learning by DLSI [19] (ResL-DLSI), residuallearning by LRSDL [21] (ResL-LRSDL), compressed learningby SVM (CL-SVM) and compressed learning by CNN (CL-CNN). On the other hand, we contrast the reconstruction qualityand computational cost of proposed classification-aided recon-struction (CAR) with FISTA of zero vector initialization (REC)and ResL in Section IV-B2. The hyper-parameters of all thesealgorithms are determined by the validation and the detailedsearch pool is summarized in Table II. Finally, in Section IV-B3,we draw the overall comparison between four classification plusreconstruction frameworks, namely, ResL, CL-CNN + REC, theproposed CL-TDDL + REC and the proposed TS-CAR.

B. Evaluation of TS-CAR Framework

1) Evaluation of CL-TDDL: Fig. 5 shows the accuracy underdifferent number of atoms in the dictionary. It is observed CL-SVM and CL-CNN have 94% and 93% accuracy, respectively.

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 9: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

4102 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020

TABLE IIICOMPUTATIONAL COST OF DIFFERENT CLASSIFICATION ALGORITHMS

Fig. 6. The PRD under different number of atoms in the dictionary.

To have the similar accuracy as CL-SVM and CL-CNN, theproposed CL-TDDL needs 100 atoms; however, other DL-basedalgorithms, such as ResL-DLSI and ResL-LRSDL, need 200atoms to reach about 90% accuracy.

The computational cost of classification algorithms is sum-marized in Table III. In DL-based algorithms, such as ResL andthe proposed CL-TDDL, the complexity is dominated by thegradient computation times the number of iterations in FISTA.To reach above 90% accuracy, we choose dclass = 100 forCL-TDDL, dtotal = 200 for ResL and iterSC = 25 for all threealgorithms. It is observed the proposed CL-TDDL has the lowestcomputational cost because of utilizing the sparsity in naturesignals, which serves as an efficient classifier. Therefore, theproposed CL-TDDL only has low overhead, making it the mostappropriate classifier for compressed inference.

2) Evaluation of CAR: Fig. 6 shows the PRD under differentnumber of atoms in the dictionary. Since the proposed CAR andREC use the same dictionary, we plot the curve of REC and CARtogether. It is observed REC or CAR needs only 100 atoms tohave PRD lower than 9%, while ResL needs about 200 atoms.Compared to Fig. 5, ResL has better accuracy and lower PRD asthe number of atoms increasing. In other words, the performanceof ResL relies on the full reconstruction, so the dictionary is notspecialized as our proposed TS-CAR. Therefore, ResL requiresmore computation to reach similar PRD as proposed TS-CAR.

Fig. 7 shows the PRD under different iterations in FISTA. Theproposed CAR evidently accelerates the reconstruction, whichcan be seen from the initial PRD and the convergence speed. The

Fig. 7. The PRD under different number of iterations in the sparse coding.

TABLE IVCOMPUTATIONAL COST OF DIFFERENT RECONSTRUCTION ALGORITHMS

initial PRD demonstrates that the proposed sparse transform wellreuses the information from the classification stage, and thusprovides a better starting point. As for the convergence speed,we choose 7% PRD as our comparison criterion since ResL canonly reach about 7% PRD no matter how many the iterations are.It is observed that the proposed CAR only needs 25 iterations toreach about 7% PRD, while REC requires 40 iterations.

The computational cost of reconstruction algorithms is sum-marized in Table IV. To reach about 7% PRD, we choosedtotal = 200 for ResL, drec = 150 for REC and the proposedCAR, iterrec = 40 for REC and iterCAR = 25 for the proposedCAR. For ResL, after classification, we only need to pay theoverhead computation by N × dtotal. Nonetheless, we want tohighlight again ResL conducts the full reconstruction in theclassification stage, so ResL doesn’t possess the property ofon-demand reconstruction. For REC and the proposed CAR,we need to compute the sparse coding again. However, with in-formation reuse, we only need to pay the negligible computationof sparse transform to evidently reduce the number of iterations,and thus achieve less computational cost.

3) Analysis of Computational Cost of Different Frameworks:To compare different classification plus reconstruction frame-works, we summarizes the accuracy, PRD and computationalcost in Table V. The proposed TS-CAR has the lightest classifier,while the sparse transform can further reduce the computationalcost in reconstruction. Therefore, the proposed TS-CAR onlyhas the computational cost equals to the reconstruction stage ofCL-CNN + REC or ResL (ResL conducts full reconstruction inthe classification stage). In other words, we don’t pay any addi-tional overhead, but we can realize on-demand reconstruction.

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 10: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

CHOU et al.: LOW-COMPLEXITY ON-DEMAND RECONSTRUCTION FOR COMPRESSIVELY SENSED PROBLEMATIC SIGNALS 4103

TABLE VPERFORMANCE AND COMPLEXITY OF DIFFERENT CLASSIFICATION PLUS

RECONSTRUCTION FRAMEWORKS

Fig. 8. The required number of multiplications versus the percentage of AFsignals for different on-demand CS reconstruction frameworks.

Fig. 8 shows the total number of multiplications versus dif-ferent percentages of AF signals. The number of multiplicationsis calculated by the number of computations in classificationadded by the number of computations in reconstruction timesthe percentage of AF signals. It is observed ResL-DLSI hashigh number of multiplications under all percentages of AFsignals because ResL-DLSI needs the full reconstruction inclassification stage.

When the percentage of AF signals is low, the number ofmultiplications is dominated by the classification stage. Ourproposed CL-TDDL takes the advantage of sparsity, makingthe classifier have the lowest computational cost and outper-form ResL-DLSI and CL-CNN. On the other hand, when thepercentage of AF signals is high, the computation requirementfor reconstruction cannot be neglected. We proposed the CARto reuse the information obtained in the classification stage andlargely reduces the computational cost. In summary, the averagecomputational cost of on-demand reconstruction required by theproposed TS-CAR is 2.37× fewer compared to ResL and 2.25×fewer compared to CL-REC when the percentage of AF signalsis among 10% to 50%.

C. Comparison of Sparse Coding Computation Algorithms

We compare FISTA with other SC computation algorithms,such as OMP and SP. The ACC, PRD and the hyper-parameters

TABLE VIPERFORMANCE OF CS RECONSTRUCTION ALGORITHMS

are summarized in Table VI, where the dictionary size is deter-mined by previous sections and the termination criteria is basedon the specific threshold, ζ. It is observed FISTA noticeablyoutperforms OMP and SP in reconstruction quality since weonly consider PRD < 9% as acceptable reconstruction [42]. Asfor classification performance, FISTA is the best. Although SPmay have good accuracy, it requires known sparsity, k; however,k is used to be unavailable in practical situations.

V. ARCHITECTURE DESIGN AND VLSI IMPLEMENTATION

In this section, based on the on-line stage of TS-CAR algo-rithm (Algorithm 4) and sparse coding computation using FISTA(Algorithm 2), we propose an intelligent CS reconstructionengine. The proposed engine can classify compressed signalsand conduct the reconstruction only for the signals which areidentified as interested (or problematic) ones, saving abundantenergy and time to reconstruct uninterested (or normal) signalswhich are far more than high-risk signals.

As discussed in the previous section, N = 512, M =128, dclass = 100, drec = 150, λclass = 0.1, λrec = 0.005,iterSC = 25 for both classification and on-demand reconstruc-tion, and nc = 2 for binary classification are used in the im-plementation of the proposed TS-CAR on ECG AF detectiontask.

A. Hardware Implementation Reasons for Choosing FISTA

Matching pursuit (MP) based sparse coding algorithms, whichapplied in prior CS reconstruction engines [6]–[8], update thesupport set and compute the corresponding sparse coefficientsin each iteration. Different from MP-based algorithms, FISTAis based on proximal gradient descent and thus has severaladvantages in hardware design.

The support set in MP-based algorithm will lead to the trade-off between reconstruction performance and implementationcost. The sparsity is different in each nature signal. Hence,the design should support the maximum sparsity in the signals;otherwise, the performance will degrade. However, the hardwarecost and the computational overhead will also increase whenthe size of the support set increases. On the contrary, FISTA ismore flexible without support set, and can be more universalto the nature signals with arbitrary sparsity. In addition, thenature of gradient descent in FISTA avoids complicated matrixdecomposition design, which usually sacrifices flexibility inupdating support set in order to prevent the high-complexitycomputation of the pseudo-inverse in least-square [8]. On theother hand, other than dictionary choice, the only difference

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 11: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

4104 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020

Fig. 9. Proposed hardware sharing architecture for TS-CAR.

between sparse coding computation in the classification andreconstruction is the �1 penalty. Therefore, applying FISTA canallow TS-CAR to easily switch between two stages by adjustingthe threshold in the shrinkage operator without any extra controlstrategy.

B. Hardware Sharing Architecture

The overall architecture of the proposed engine is shown inFig. 9. It consists of five memory banks, six main memorybuffers, and four main arithmetic units. The entries of the matrix,such as ΘT , ΘTΘ, W T , and T , will be memorized in thememory banks. The memory buffers include x, ΘT x, αt, βt,the data from the memory banks, and the temporal buffers. Thearithmetic units includes the inner product unit consisting par-allel multipliers and adder-tree, similar to the structure depictedin [7], the adder unit, the multiplier unit, and the shrinkageoperator.

There are two levels of hardware sharing design in the pro-posed engine. First of all, the computation in the classificationand reconstruction stage is almost the same, so we can use asingle hardware to conduct both tasks. Second, different stepsin TS-CAR engine, such as gradient calculation in FISTA in(12), weighted sum of predictive sparse coefficients and sparsetransform in (25), all require the same matrix-vector multipli-cation. Furthermore, those processes are computed in sequence.Thus, these steps share the inner product and adder unit.

C. Data Path of Each Step in Execution Flow

The execution flow of the proposed engine is shown in Fig. 10.In the configuring stage, the entries of the matrices are stored inthe memory banks. The classification and reconstruction processshare the same operation, namely FISTA algorithm for sparsecoding, which is the left cycle in the execution flow chart. First,the compressed signal is classified by computing predictivesparse coding and weighted sum. If the signal is uninterested, theTS-CAR engine will classify next target signal. On the contrary,if the signal is interested, the TS-CAR engine will conduct sparsetransform and compute reconstructive sparse coding. At last, the

Fig. 10. Execution flow of the proposed TS-CAR engine.

TS-CAR engine will output the reconstruction result and turn tothe next compressed signal.

Each step in the execution flow chart has its own color, andarithmetic units in the architecture with corresponding colorrepresent the activated data path in that step. The detail datapath of each step is as follows.

1) Gradient calculation (blue): Since ΘT x is unchangedamong iterations, ΘT x can be calculated and stored be-fore starting iterative optimization in FISTA. The com-pressed signal is shifted into the x buffers and will beinputted into the inner product unit with ΘT

class or ΘTrec

from the memory banks. In our design, the inner productunit can only process 10 entries of data vector at a time.Therefore, we partition the entire matrix vector productinto 13 times sub-matrix vector product. In each time, ittakes 100 cycles and 150 cycles to calculate the entiresub-matrix vector product for each column in classifica-tion and reconstruction dictionary, respectively. The resultof calculation from the inner product unit will be addedwith the value in the temporal buffers and stored back tothe temporal buffers. At the end of calculating ΘT x, thevalue in the temporal buffers will be copied and stored intothe ΘT x buffers.In each iteration, t, the gradient is updated by only cal-culation of ΘTΘαt−1. The sparse coefficients in the αt

buffers and ΘTclassΘclass or ΘT

recΘrec from the memorybanks are inputted into the inner product unit. The initialvalue inside the temporal buffers is ΘT x. Similar toprevious, it takes 10 × 100 cycles and 15 × 150 cyclesto calculate the matrix vector product in classification andreconstruction, respectively.

2) Undergoing shrinkage operator (red): After gradient cal-culation, the updated sparse coefficients undergo the

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 12: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

CHOU et al.: LOW-COMPLEXITY ON-DEMAND RECONSTRUCTION FOR COMPRESSIVELY SENSED PROBLEMATIC SIGNALS 4105

shrinkage operator, which introduces the sparsity con-straint as (13). ∇αf(αt−1) in the temporal buffers willbe added to the sparse coefficients in the αt buffers, andbe inputted to the shrinkage operator. The new shrunkterm βt from the shrinkage operator is stored to thetemporal buffers and the αt buffers. The threshold in theshrinkage operator is ηλclass and ηλrec in classificationand reconstruction, respectively.

3) Nesterov acceleration with momentum (green and brown):The new sparse coefficients αt are obtained by the linearcombination of new shrunk term and the previous shrunkterm as (16). In momentum calculation stage, the pre-vious shrunk term from the βt buffers will go throughan inverted and be added to the new shrunk term in thetemporal buffers. After that, the results will be multipliedby the momentum coefficients which can be pre-computedoff-line and stored in the circuits, and the momentum willbe then stored into the βt buffers.In stage Nesterov combination, the momentum in theβt buffers will be added to the new shrunk term in thetemporal buffers, and the new sparse coefficients αt willbe stored into the αt buffers. At the end of this stage, thenumber of iterations in FISTA will add one. In order toprepare for the gradient calculation in the next iteration,the new shrunk term in the temporal buffers will be copiedand stored into theβt buffers, and the pre-computed partialgradient in the ΘT x buffers will be copied and stored intothe temporal buffers.

4) Weighted sum in classification (blue and gray): The prob-ability is obtained from multiplying the predictive sparsecoefficients αclass by W T . αclass in the αt buffers andW T from the memory banks are inputted into the innerproduct unit. The initial value inside the temporal buffersis the bias of the linear model. Similar to previous, ittakes 10× 2 cycles to calculate the matrix vector product.After calculating the probability, the estimated label isdetermined by choosing the bigger probability.

5) Sparse transform in reconstruction (blue): The predictivesparse coefficients which are identified as problematicones will be transformed to the estimated reconstructivesparse coefficients αrec as (25). αclass in the αt buffersand T from the memory banks are inputted into the innerproduct unit. Similar to previous, it takes 10× 150 cyclesto calculate the matrix vector product. At the end ofcalculating αrec, the value in the temporal buffers willbe copied and stored into the αt and the βt buffers as theinitial point of the reconstructive sparse coding.

D. Fixed-Point Analysis

The data bits used in the arithmetic units, memory buffers,and memory banks determine the hardware efficiency of theproposed TS-CAR. To make a fair comparison with state-of-the-art CS engine [6], the data bits of the arithmetic units andthe memory buffers in the TS-CAR engine is set to be 32-bitfixed-point format. For memory banks, such as ΘT , ΘTΘ,W T , and T , we conduct fixed-point analysis to decide the

Fig. 11. (a) Accuracy and (b) PRD under different fractional bits.

Fig. 12. Layout of the proposed on-demand CS reconstruction engine.

TABLE VIICOMPARISON OF THE ML, CS, AND TS-CAR ENGINES

*Technology scaling to 40 nm: delay∼ 1/S, power ∼ 1/U2, where S = L/40

nm and U = VDD/0.9V . †Without on-chip memory for sensing matrix.

minimal required fractional bits by evaluating the classificationperformance and the reconstruction quality as the integer bits isset to be 3. As shown in Fig. 11, the proposed TS-CAR algorithmhas almost the same performance as float-point counterpart whenwe use 15 fractional bits. Hence, each entry stored in memorybanks is set to be signed 19-bit word length consisting of 1 signbit, 3 integer bits, and 15 fractional bits.

E. Implementation Results and Comparisons

The proposed TS-CAR engine is implemented in TSMC40 nm technology by using Cadence Innovus tool. Fig. 12 showsthe IP layout of the intelligent CS reconstruction engine. Thecore area is 0.41 mm2 at 135 MHz operation frequency.

Table VII shows a comparison of the proposed TS-CARengine, which can achieve both classification and reconstruction,

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 13: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

4106 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 68, 2020

with the state-of-the-art ASIC designs of ML and CS recon-struction engines. To make a fair comparison, we normalizedthe area and energy consumption of the classification engine ac-cording to the technology and the CNN model size described inTable II. Since the design goal in the low-complexity on-demandCS reconstruction is area- and energy-efficient, we defined thearea-energy-efficiency (AEE, in Sample/(μJ × mm2)) as

AEE =1

Energy Efficiency× Area, (28)

which describes the computation ability provided by energyconsumption and silicon cost in order to identify the problematiccompressed signals and conduct reconstruction.

In the CL-CNN + REC framework, a compressed signal isfirst processed by the ML chip. If the signal is problematic, itwill be reconstructed by the CS chip. Therefore, the AEE of theCL-REC framework is 1/((1.32 + 0.39)× (4.64 + 5.13)) = 0.06,which is quite low because of double hardware and computationoverhead. On the contrary, the proposed TS-CAR engine sharesthe hardware for compressed inference and on-demand recon-struction. Referring to the execution flow in Fig. 10, it takes33751 cycles to classify, and 70780 cycles to further conductsparse transform and full reconstruction of AF signal. Therefore,the AEE of the proposed engine is 1/(5.88 × 0.41) = 0.41. Inthe ResL framework, larger dictionary is required (dtotal = 200in Section IV-B). Therefore, the core area is normalized to0.48 (mm2), the power is normalized to 9.56 (mW), and the cyclecount is 117330 according to Fig. 10. Hence, the AEE of ResLframework is 1/(8.31 × 0.48) = 0.25. Nonetheless, we wantto emphasize again ResL wastes lots of efforts to reconstructuninterested (or normal) signals, and the computational cost isanalyzed in Section IV-B3. In summary, the AEE of the proposed2-in-1 intelligent CS reconstruction chip based on TS-CARalgorithm is 1.64× compared to ResL and 6.83× compared toCL-REC.

VI. CONCLUSION

In this work, to eliminate the abundant cost of recoveringirrelevant data, we proposed a novel two-stage CS on-demandreconstruction framework. The proposed TS-CAR frameworkcan not only directly classify the compressed signals, but alsoaccelerate the reconstruction speed by using the information inclassification. The average computational cost of on-demandreconstruction required by the proposed TS-CAR is 2.25× fewercompared with prior related CS frameworks when the percentageof interested signals is among 10% to 50%. In ASIC design,the post-layout results show that the proposed intelligent CSreconstruction engine based on TS-CAR framework can providea competitive area- and energy-efficiency compared to state-of-the-art CS and ML engines because of the hardware sharingdesign between classification and reconstruction.

REFERENCES

[1] F. Touati and R. Tabish, “U-healthcare system: State-of-the-art review andchallenges,” J. Med. Syst., vol. 37, no. 3, 2013, Art. no. 9949.

[2] C. Wen, M.-F. Yeh, K.-C. Chang, and R.-G. Lee, “Real-time ECG tele-monitoring system design with mobile phone platform,” Measurement,vol. 41, no. 4, pp. 463–470, 2008.

[3] H. Mamaghanian, N. Khaled, D. Atienza, and P. Vandergheynst, “Com-pressed sensing for real-time energy-efficient ECG compression on wire-less body sensor nodes,” IEEE Trans. Biomed. Eng., vol. 58, no. 9,pp. 2456–2466, Sep. 2011.

[4] D. L. Donoho, “Compressed sensing,” IEEE Trans. Inf. Theory, vol. 52,no. 4, pp. 1289–1306, Apr. 2006.

[5] A. M. R. Dixon, E. G. Allstot, D. Gangopadhyay, and D. J. Allstot,“Compressed sensing system considerations for ECG and EMG wirelessbiosensors,” IEEE Trans. Biomed. Circuits Syst., vol. 6, no. 2, pp. 156–166,Apr. 2012.

[6] F. Ren and D. Markovic, “A configurable 12–237 kS/s 12.8 mW sparse-approximation engine for mobile data aggregation of compressively sam-pled physiological signals,” IEEE J. Solid-State Circuits, vol. 51, no. 1,pp. 68–78, Jan. 2016.

[7] Y. Lin, Y. Chen, N. Huang, and A. Wu, “Low-complexity stochasticgradient pursuit algorithm and architecture for robust compressive sensingreconstruction,” IEEE Trans. Signal Process., vol. 65, no. 3, pp. 638–650,Feb. 2017.

[8] T. Chen, H. Kuo, and A. Wu, “A 232–1996-kS/s robust compressivesensing reconstruction engine for real-time physiological signals moni-toring,” IEEE J. Solid-State Circuits, vol. 54, no. 1, pp. 307–317, Jan.2019.

[9] M. A. T. Figueiredo, R. D. Nowak, and S. J. Wright, “Gradient projectionfor sparse reconstruction: Application to compressed sensing and otherinverse problems,” IEEE J. Sel. Topics Signal Process., vol. 1, no. 4,pp. 586–597, Dec. 2007.

[10] M. Zoni-Berisso, F. Lercari, T. Carazza, and S. Domenicucci, “Epidemi-ology of atrial fibrillation: European perspective,” Clin. Epidemiol., vol. 6,pp. 213–220, Jun. 2014.

[11] R. Calderbank, S. Jafarpour, and R. Schapire, “Compressed learning: Uni-versal sparse dimensionality reduction and learning in the measurementdomain,” preprint, Tech. Rep., 2009.

[12] O. Maillard and R. Munos, “Compressed least-squares regression,” inProc. Advances Neural Inf. Process. Syst., 2009, pp. 1213–1221.

[13] M. Shoaib, N. K. Jha, and N. Verma, “Signal processing with directcomputations on compressively sensed data,” IEEE Trans. Very LargeScale Integr. (VLSI) Syst., vol. 23, no. 1, pp. 30–43, Jan. 2015.

[14] K. Kulkarni and P. Turaga, “Reconstruction-free action inference fromcompressive imagers,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 38,no. 4, pp. 772–784, Apr. 2016.

[15] S. Lohit, K. Kulkarni, and P. Turaga, “Direct inference on compressivemeasurements using convolutional neural networks,” in Proc. IEEE Int.Conf. Image Process., 2016, pp. 1913–1917.

[16] Y. Chen, T. Krishna, J. S. Emer, and V. Sze, “Eyeriss: An energy-efficient reconfigurable accelerator for deep convolutional neural net-works,” IEEE J. Solid-State Circuits, vol. 52, no. 1, pp. 127–138, Jan.2017.

[17] S. Yin, “A high energy efficient reconfigurable hybrid neural networkprocessor for deep learning applications,” IEEE J. Solid-State Circuits,vol. 53, no. 4, pp. 968–982, Apr. 2018.

[18] H. Kuo, Y. Lin, and A. Wu, “Compressive sensing based ECG monitoringwith effective AF detection,” in Proc. IEEE Int. Conf. Acoust., SpeechSignal Process., 2017, pp. 1008–1012.

[19] I. Ramirez, P. Sprechmann, and G. Sapiro, “Classification and clusteringvia dictionary learning with structured incoherence and shared features,”in Proc. IEEE Comput. Soc. Conf. Comput. Vision Pattern Recognit., 2010,pp. 3501–3508.

[20] S. Kong and D. Wang, “A dictionary learning approach for classification:Separating the particularity and the commonality,” in Proc. Eur. Conf.Comput. Vision, 2012, pp. 186–199.

[21] T. H. Vu and V. Monga, “Fast low-rank shared dictionary learningfor image classification,” IEEE Trans. Image Process., vol. 26, no. 11,pp. 5160–5175, Nov. 2017.

[22] E. J. Candes and T. Tao, “Near-optimal signal recovery from randomprojections: Universal encoding strategies?” IEEE Trans. Inf. Theory,vol. 52, no. 12, pp. 5406–5425, Dec. 2006.

[23] J. A. Tropp and A. C. Gilbert, “Signal recovery from random measurementsvia orthogonal matching pursuit,” IEEE Trans. Inf. Theory, vol. 53, no. 12,pp. 4655–4666, Dec. 2007.

[24] W. Dai and O. Milenkovic, “Subspace pursuit for compressive sensing sig-nal reconstruction,” IEEE Trans. Inf. Theory, vol. 55, no. 5, pp. 2230–2249,May 2009.

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.

Page 14: Low-Complexity On-Demand Reconstruction for Compressively ...access.ee.ntu.edu.tw/Publications/Journal/(2020)Low... · CS reconstruction, we propose a two-stage classification-aided

CHOU et al.: LOW-COMPLEXITY ON-DEMAND RECONSTRUCTION FOR COMPRESSIVELY SENSED PROBLEMATIC SIGNALS 4107

[25] A. Beck and M. Teboulle, “A fast iterative shrinkage-thresholding al-gorithm for linear inverse problems,” SIAM J. Imag. Sci., vol. 2, no. 1,pp. 183–202, 2009.

[26] K. Engan, S. O. Aase, and J. H. Husoy, “Frame based signal compressionusing method of optimal directions (mod),” in Proc. IEEE Int. Symp.Circuits Syst. VLSI, 1999, vol. 4, pp. 1–4.

[27] M. Aharon, M. Elad, and A. Bruckstein, “K-SVD: An algorithm for de-signing overcomplete dictionaries for sparse representation,” IEEE Trans.Signal Process., vol. 54, no. 11, pp. 4311–4322, Nov. 2006.

[28] J. Mairal, F. Bach, J. Ponce, and G. Sapiro, “Online learning for matrixfactorization and sparse coding,” J. Mach. Learn. Res., vol. 11, pp. 19–60,2010.

[29] J. Mairal, F. Bach, and J. Ponce, “Task-driven dictionary learning,” IEEETrans. Pattern Anal. Mach. Intell., vol. 34, no. 4, pp. 791–804, Apr. 2012.

[30] K. Hsu, B. Cho, C. Chou, and A. Wu, “Low-complexity compressedanalysis in eigenspace with limited labeled data for real-time electrocardio-graphy telemonitoring,” in Proc. IEEE Global Conf. Signal Inf. Process.,2018, pp. 459–463.

[31] F. Giovanneschi, K. V. Mishra, M. A. González-Huici, Y. C. Eldar, and J.H. G. Ender, “Dictionary learning for adaptive GPR target classification,”2018, arXiv:1806.04599.

[32] S. Dasgupta and A. Gupta, “An elementary proof of a theorem of Johnsonand Lindenstrauss,” Random Struct. Algorithms, vol. 22, no. 1, pp. 60–65,2003.

[33] R. Durrant and A. Kabán, “Sharp generalization error bounds forrandomly-projected classifiers,” in Proc. Int. Conf. Mach. Learn., 2013,pp. 693–701.

[34] A. Kabán, “New bounds on compressive linear least squares regression,”Artif. Intell. Statist., pp. 448–456, 2014.

[35] M. Tsai, C. Chou, and A. Wu, “Robust compressed analysis usingsubspace-based dictionary for ECG telemonitoring systems,” in Proc.IEEE Int. Workshop Signal Process. Syst., 2017, pp. 1–5.

[36] C. Chou, E. Chang, H. Li, and A. Wu, “Low-complexity privacy-preservingcompressive analysis using subspace-based dictionary for ECG telemon-itoring system,” IEEE Trans. Biomed. Circuits Syst., vol. 12, no. 4,pp. 801–811, Aug. 2018.

[37] C. Chou and A. Wu, “Low-complexity compressive analysis in sub-eigenspace for ECG telemonitoring system,” in Proc. IEEE Int. Conf.Acoust., Speech Signal Process., 2019, pp. 7575–7579.

[38] B. Krishnapuram, L. Carin, M. A. T. Figueiredo, and A. J. Hartemink,“Sparse multinomial logistic regression: Fast algorithms and generaliza-tion bounds,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 27, no. 6,pp. 957–968, Jun. 2005.

[39] J. Yang, Z. Wang, Z. Lin, S. Cohen, and T. Huang, “Coupled dictionarytraining for image super-resolution,” IEEE Trans. Image Process., vol. 21,no. 8, pp. 3467–3478, Aug. 2012.

[40] D.-A. Huang and Y.-C. Frank Wang, “Coupled dictionary and feature spacelearning with applications to cross-domain image synthesis and recogni-tion,” in Proc. IEEE Int. Conf. Comput. Vision, 2013, pp. 2496–2503.

[41] C.-C. Chang and C.-J. Lin, “LIBSVM: A library for support vectormachines,” ACM Trans. Intell. Syst. Technol., vol. 2, no. 3, Art. no. 27.[Online]. Available: http://www.csie.ntu.edu.tw/∼cjlin/libsvm

[42] Y. Zigel, A. Cohen, and A. Katz, “The weighted diagnostic distortion(WDD) measure for ECG signal compression,” IEEE Trans. Biomed. Eng.,vol. 47, no. 11, pp. 1422–1430, Nov. 2000.

Ching-Yao Chou (Student Member, IEEE) receivedthe B.S. degree in electrical engineering from Na-tional Taiwan University (NTU), Taipei, Taiwan, in2014, and the Ph.D. degree in the Graduate Instituteof Electronics Engineering from National TaiwanUniversity in 2019. His research interests includethe areas of compressed sensing, machine learningin compressed domain, and VLSI implementation ofDSP algorithms.

Kai-Chieh Hsu (Student Member, IEEE) receivedthe B.S. degree in electrical engineering from Na-tional Taiwan University, Taipei, Taiwan, in 2019.He is currently working toward the Ph.D. degreein the electrical engineering, Princeton University.His research interests include machine learning, con-trol theory, reinforcement learning and compressedsensing.

Bo-Hong Cho (Student Member, IEEE) received theB.S. degree in electrical engineering from NationalTaiwan University, Taipei, Taiwan, in 2019. He iscurrently pursuing the M.S. degree in the electrical &computer engineering, from the University of Michi-gan. His research interests are mainly in the realmsof biomedical signal processing, compressed sensing,and natural language processing.

Kuan-Chun Chen received the B.S. degree in elec-tronic engineering from National Chiao Tung Uni-versity, Hsinchu, Taiwan, in 2018. He is currentlyworking toward the M.S. degree in the GraduateInstitute of Electronics Engineering, National TaiwanUniversity. His research interests include the areasof biomedical signal processing, compressed sensing,and VLSI implementation of DSP algorithms.

An-Yeu (Andy) Wu (Fellow, IEEE) received the B.S.degree in electrical engineering from National TaiwanUniversity (NTU), Taipei, Taiwan, in 1987, and theM.S. and Ph.D. degrees in electrical engineering fromthe University of Maryland of College Park, MD,USA, in 1992 and 1995, respectively.

In 2000, he joined the Department of ElectricalEngineering and the Graduate Institute of ElectronicsEngineering (GIEE), NTU, as a Faculty Member,where he is currently a Distinguished Professor andalso has been the Director of GIEE since 2016. His

research interests include VLSI architectures for signal processing and commu-nications, and adaptive/multirate signal processing. He has published more than250 refereed journal and conference papers in above research areas, togetherwith five book chapters and 20 granted US patents. From 2007 to 2009, hewas on leave from NTU and served as the Deputy General Director of theSoC Technology Center (STC), Industrial Technology Research Institute (ITRI),Hsinchu, Taiwan.

In 2015, Prof. Wu was elevated to IEEE Fellow for his contributions to “DSPalgorithms and VLSI designs for communication IC/SoC.” He serves as a Boardof Governor (BoG) Member in IEEE Circuits and Systems Society (CASS) fortwo terms (2016–2018, 2019–2021). From August 2016 to July 2019, he servedas the Director of Graduate Institute of Electronics Engineering (GIEE), NationalTaiwan University. He now serves as Editor-in-Chief (EiC) of IEEE Journal onEmerging and Selected Topics in Circuits and Systems (JETCAS).

Authorized licensed use limited to: National Taiwan University. Downloaded on August 10,2020 at 16:40:48 UTC from IEEE Xplore. Restrictions apply.