1
Aacking Split Manufacturing from a Deep Learning Perspective Haocheng Li 1 , Satwik Patnaik 2 , Abhrajit Sengupta 2 , Haoyu Yang 1 , Johann Knechtel 3 , Bei Yu 1 , Evangeline F. Y. Young 1 , Ozgur Sinanoglu 3 1 The Chinese University of Hong Kong 2 New York University 3 New York University Abu Dhabi Split Manufacturing The notion of integrated circuit split manufacturing which delegates the front-end-of-line (FEOL) and back-end-of-line (BEOL) parts to dierent foundries [McCants 2011], is to prevent overproduction, piracy of the intellectual property (IP) [Shamsi et al. 2019], or targeted insertion of hardware Trojans [Li et al. 2018] by adversaries in the FEOL facility. M1 V1 M2 V2 M3 V3 M4 V4 M5 V5 M6 V6 M7 V7 M8 V8 M9 V9 M10 Front-end-of-line (FEOL) Back-end-of-line (BEOL) We challenge the security promise of split manufacturing by formulat- ing various layout-level placement and routing hints as vector-based and image-based features. We construct a sophisticated deep neural net- work which can infer the missing BEOL connections with high accuracy. Compared with the network-flow aack [Wang et al. 2018] for the same set of ISCAS-85 designs, we achieve 1.21× accuracy when spliing on M1 and 1.12× accuracy when spliing on M3 with less than 1% running time. Vector- and Image-based Feature Extraction A!acking Designs Training Designs Network Training A!acking Threat Model Available FEOL design, cell library, database of layouts generated in a similar manner. Via in FEOL layer Virtual pin in split layer M1 FEOL wiring fragment Mapping for virtual pin pairs M1 M2 M1 M3 M2 M3 M3 M3 M1 M1 M2 M2 M1 Objective correct connection rate: CCR = m i =1 c i x i m i =1 c i , (1) where m is the number of sink fragments, c 1 , c 2 ,..., c m are the numbers of sinks in every fragment, x i = 1 when a positive virtual pin pair (VPP) is selected for the i -th sink fragment, x i = 0 when a negative VPP is selected for the i -th sink fragment. Feature Extraction Vector-based Features I Distances for VPPs along the preferred and the non-preferred routing direction. I Maximum capacitance of the driver and pin capacitance of the sinks. I Number of sinks connected within the sink fragment. I Wirelength and via contribution in each FEOL metal layer. I Driver delay based on the underlying timing paths. Image-based Features We represent the routing layout of the local regions centering the virtual pin as gray-scale layout images. We consider three dierent scales with the same image shape but dierent precisions. Feature Image 1 Feature Image 2 Feature Image 3 Each image is 99 pixels wide and high, representing 99 × 99 consecutive regions. Since wires closer to the BEOL carry more information about the connection, those in higher metal layers are encoded in more signif- icant bits. ’100’ ’111’ ’001’ ’000’ ’011’ ’001’ Metal 3 Metal 2 Metal 1 Sample Selection We select n candidate VPPs for each sink in training and testing based on three criteria. Direction Criterion For a VPP (p, q ), if q is on the opposite side of one of the wire segments directly connected to p, we then say the virtual pin p prefers virtual pin q . Metal 3 Via 3 Source A Source B Sink A Sink B A VPP is not considered as a candidate in case both source and sink pins do not prefer each other. Sk Sc Sk Prefers Sc Sc Prefers Sk Direction Criterion A A A B B A B B Non-duplication Criterion If a sink fragment or source fragment have multiple virtual pins, for each pair of sink fragment and source fragment, only the VPP with the short- est distance apart in the non-preferred routing direction of the split layer is considered as candidate. Distance Criterion If the number of VPPs remaining is greater than n, the VPPs with shorter distance in the non-preferred routing direction of the split layer have higher priority to be selected. Model Architecture Input: I a batch of features corresponding to a sink fragment including the vector-based features of n selected VPPs with the sink fragment; I the image-based features of n source fragments in the related VPPs; I the image-based features of the sink fragment itself. Output: I scores for every VPP in the batch. To handle vector-based and image-based features in the same network, the proposed neural network first extracts underlying features from het- erogeneous input by processing vector-based features (shown in the up- per le) and image-based features (shown in the upper middle) individ- ually, and then processing them together (shown in the lower le) aer concatenating the output of the vector and image part together. shared network n x 256 n x 256 input vector features input source images input sink image 3x3 conv1-1, 16 3x3 conv1-2, 16 3x3 conv1-3, 16 3x3 conv2-1, 32, /3 3x3 conv2-2, 32 3x3 conv2-3, 32 3x3 conv3-1, 64, /3 3x3 conv3-2, 64 3x3 conv3-3, 64 3x3 conv4-1, 128, /3 3x3 conv4-2, 128 3x3 conv4-3, 128 fc3, 256 fc4, 128 fc5-1, 128 fc1, 128 n x 128 n x 128 n x 128 1 x 128 fc5-2, 128 fc6, 32 fc7, 1 output scores res, 128 res, 128 res, 128 res, 128 res, 128 res, 128 res, 128 fc2, 128 fc2, 128 fc2, 128 ResNet Block For the image part of the network, note that the image-based features of the sink fragment are the same in the batch, so we only process them once to save runtime and its output is distributed to the output of every source images. Besides, all the image-based features go through the same shared network. Part Layer Parameter Output Vector fc1 27 × 128 n × 128 part fc2 [128 × 128] × 12 n × 128 conv1 [3 × 3, 16] × 3 (n + 1) × 99 × 99 × 16 conv2 [3 × 3, 32] × 3 (n + 1) × 33 × 33 × 32 Image conv3 [3 × 3, 64] × 3 (n + 1) × 11 × 11 × 64 conv4 [3 × 3, 128] × 3 (n + 1) × 4 × 4 × 128 part fc3 128 × 256 (n + 1) × 256 fc4 256 × 128 (n + 1) × 128 fc5 256 × 128 n × 128 fc5 256 × 128 n × 128 Merged fc2 [128 × 128] × 9 n × 128 part fc6 128 × 32 n × 32 fc7 32 × 1 n × 1 Both fully connected layers and convolutional layers are followed by a leaky rectified linear unit (LReLU) y = max(0.01x , x ), (2) as activation, where x is the input and y is the output [Maas, Hannun, and Ng 2013]. Somax Regression Loss Conventional Approach The loss of the two-class classification is l r = - 1 n log e s + t e s - t + e s + t + X j 6=t log e s - j e s - j + e s + j , (3) whose partial derivative is l r s + j = - l r s - j = - e s - j n e s - j + e s + j if j = t , e s + j n e s - j + e s + j otherwise. (4) The partial derivative in the last FC layer is l r w + i = - l r w - i = 1 n n X j =1 e s + j x i ,j e s - j + e s + j - x i ,t . (5) Misprediction of one VPP, which significantly influences CCR, barely aects the average loss. It also has a serious imbalance problem as it can easily gain a high accuracy by simply classifying all VPPs as negative, which is meaningless. Our Method We propose the following somax regression loss l c = - log e s t n j =1 e s j , (6) whose partial derivative is l c s j = e s j n j =1 e s j - 1 if j = t , e s j n j =1 e s j otherwise. (7) The partial derivative in the last FC layer is l c w i = n j =1 e s j x i ,j n j =1 e s j - x i ,t . (8) The source fragment with higher score contributes much more signifi- cantly in the gradient with an exponential factor. The summation of the coeicients in the positive part equals to that of the negative part, so there is no imbalance issue. Experimental Results Comparison between Ours and Wang (TVLSI’18) 0 0.5 1 1.5 M1 CCR M3 CCR M1 Time M3 Time 1.21 1.12 1 · 10 -3 2 · 10 -3 1 1 1 1 Average Ratio Wang Ours 0 5 10 15 b7 b11 b13 b14 b15_1 b17_1 b18 c432 c880 c1355 c1908 c2670 c3540 c5315 c6288 c7552 M1 CCR (%) Comparison between dierent seings 50 55 60 65 CCR (%) 0 20 40 Time (s) Two-class Vec Vec & Img Conclusion I Demonstrate vector-based and image-based features. I Process heterogeneous features simultaneously in a neural network. I Propose a somax regression loss that directly reflects on the accuracy for the virtual pin pair matching problem of split manufacturing. References Li, Meng, Bei Yu, Yibo Lin, Xiaoqing Xu, Wuxi Li, and David Z. Pan (2018). “A Practical Split Manufacturing Framework for Trojan Prevention via Simultaneous Wire Liing and Cell Insertion”. In: IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems (TCAD). Maas, Andrew L, Awni Y Hannun, and Andrew Y Ng (2013). “Rectifier non- linearities improve neural network acoustic models”. In: International Con- ference on Machine Learning (ICML). McCants, C (2011). “Trusted integrated chips (TIC)”. In: Intelligence Ad- vanced Research Projects Activity (IARPA), Tech. Rep. Shamsi, Kaveh, Travis Meade, Meng Li, David Z. Pan, and Yier Jin (2019). “On the approximation resiliency of logic locking and IC camouflaging schemes”. In: IEEE Transactions on Information Forensics and Security 14.2, pp. 347–359. Wang, Yujie, Pu Chen, Jiang Hu, Guofeng Li, and Jeyavijayan Rajendran (2018). “The cat and mouse in split manufacturing”. In: IEEE Transactions on Very Large Scale Integration Systems (TVLSI) 26.5, pp. 805–817. Haocheng Li et al. Aacking Split Manufacturing from a Deep Learning Perspective [email protected]

Attacking Split Manufacturing from a Deep Learning ...byu/papers/C81-DAC2019-AttackSPM...A˛acking Split Manufacturing from a Deep Learning Perspective Haocheng Li1, Satwik Patnaik2,

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Attacking Split Manufacturing from a Deep Learning ...byu/papers/C81-DAC2019-AttackSPM...A˛acking Split Manufacturing from a Deep Learning Perspective Haocheng Li1, Satwik Patnaik2,

A�acking Split Manufacturing from a Deep Learning PerspectiveHaocheng Li1, Satwik Patnaik2, Abhrajit Sengupta2, Haoyu Yang1,

Johann Knechtel3, Bei Yu1, Evangeline F. Y. Young1, Ozgur Sinanoglu3

1The Chinese University of Hong Kong2New York University

3New York University Abu Dhabi

Split Manufacturing

The notion of integrated circuit split manufacturing which delegates thefront-end-of-line (FEOL) and back-end-of-line (BEOL) parts to di�erentfoundries [McCants 2011], is to prevent overproduction, piracy of theintellectual property (IP) [Shamsi et al. 2019], or targeted insertion ofhardware Trojans [Li et al. 2018] by adversaries in the FEOL facility.

M1V1M2V2M3V3M4V4M5V5M6V6M7V7M8V8M9V9

M10

Front-end-of-line (FEOL)Back-end-of-line (BEOL)

We challenge the security promise of split manufacturing by formulat-ing various layout-level placement and routing hints as vector-based andimage-based features. We construct a sophisticated deep neural net-work which can infer the missing BEOL connections with high accuracy.Compared with the network-flow a�ack [Wang et al. 2018] for the sameset of ISCAS-85 designs, we achieve 1.21× accuracy when spli�ing onM1 and 1.12× accuracy when spli�ing on M3 with less than 1% runningtime.

Vector- and Image-based Feature ExtractionA!acking DesignsTraining Designs

Network Training A!acking

Threat Model

Available FEOL design, cell library, database of layouts generated in asimilar manner.

Via in FEOL layer

Virtual pin in split layer

M1 FEOL wiring fragment

Mapping for virtual pin pairs

M1

M2

M1M3M2

M3

M3

M3

M1

M1

M2 M2 M1

Objective correct connection rate:

CCR =

∑mi=1 cixi∑mi=1 ci

, (1)

where m is the number of sink fragments, c1, c2, . . . , cm are the numbersof sinks in every fragment, xi = 1 when a positive virtual pin pair (VPP)is selected for the i-th sink fragment, xi = 0 when a negative VPP isselected for the i-th sink fragment.

Feature Extraction

Vector-based FeaturesI Distances for VPPs along the preferred and the non-preferred routing

direction.I Maximum capacitance of the driver and pin capacitance of the sinks.I Number of sinks connected within the sink fragment.I Wirelength and via contribution in each FEOL metal layer.I Driver delay based on the underlying timing paths.

Image-based FeaturesWe represent the routing layout of the local regions centering the virtualpin as gray-scale layout images. We consider three di�erent scales withthe same image shape but di�erent precisions.

Feature Image 1

Feature Image 2

Feature Image 3

Each image is 99 pixels wide and high, representing 99× 99 consecutiveregions. Since wires closer to the BEOL carry more information aboutthe connection, those in higher metal layers are encoded in more signif-icant bits.

’100’

’111’

’001’

’000’

’011’

’001’

Metal 3

Metal 2

Metal 1

Sample Selection

We select n candidate VPPs for each sink in training and testing basedon three criteria.

Direction CriterionFor a VPP (p, q), if q is on the opposite side of one of the wire segmentsdirectly connected to p, we then say the virtual pin p prefers virtual pinq.

Metal 3

Via 3Source A Source B

Sink A

Sink B

A VPP is not considered as a candidate in case both source and sink pinsdo not prefer each other.

Sk Sc Sk Prefers Sc Sc Prefers Sk Direction CriterionA A 3 7 3

A B 3 3 3

B A 7 7 7

B B 3 3 3

Non-duplication CriterionIf a sink fragment or source fragment have multiple virtual pins, for eachpair of sink fragment and source fragment, only the VPP with the short-est distance apart in the non-preferred routing direction of the split layeris considered as candidate.

Distance CriterionIf the number of VPPs remaining is greater than n, the VPPs with shorterdistance in the non-preferred routing direction of the split layer havehigher priority to be selected.

Model Architecture

Input:I a batch of features corresponding to a sink fragment including the

vector-based features of n selected VPPs with the sink fragment;I the image-based features of n source fragments in the related VPPs;I the image-based features of the sink fragment itself.Output:I scores for every VPP in the batch.To handle vector-based and image-based features in the same network,the proposed neural network first extracts underlying features from het-erogeneous input by processing vector-based features (shown in the up-per le�) and image-based features (shown in the upper middle) individ-ually, and then processing them together (shown in the lower le�) a�erconcatenating the output of the vector and image part together.

shared network

n x 256

n x 256

input vector features input source images input sink image3x3 conv1-1, 16

3x3 conv1-2, 16

3x3 conv1-3, 16

3x3 conv2-1, 32, /3

3x3 conv2-2, 32

3x3 conv2-3, 32

3x3 conv3-1, 64, /3

3x3 conv3-2, 64

3x3 conv3-3, 64

3x3 conv4-1, 128, /3

3x3 conv4-2, 128

3x3 conv4-3, 128

fc3, 256

fc4, 128

fc5-1, 128

fc1, 128

n x 128 n x 128

n x 128 1 x 128

fc5-2, 128

fc6, 32

fc7, 1

output scores

res, 128

res, 128

res, 128

res, 128

res, 128

res, 128

res, 128 fc2, 128

fc2, 128

fc2, 128

ResNet Block

For the image part of the network, note that the image-based featuresof the sink fragment are the same in the batch, so we only process themonce to save runtime and its output is distributed to the output of everysource images. Besides, all the image-based features go through thesame shared network.

Part Layer Parameter OutputVector fc1 27× 128 n× 128part fc2 [128× 128]× 12 n× 128

conv1 [3× 3, 16]× 3 (n + 1)× 99× 99× 16conv2 [3× 3, 32]× 3 (n + 1)× 33× 33× 32

Image conv3 [3× 3, 64]× 3 (n + 1)× 11× 11× 64conv4 [3× 3, 128]× 3 (n + 1)× 4× 4× 128

part fc3 128× 256 (n + 1)× 256fc4 256× 128 (n + 1)× 128fc5 256× 128 n× 128fc5 256× 128 n× 128

Merged fc2 [128× 128]× 9 n× 128part fc6 128× 32 n× 32

fc7 32× 1 n× 1

Both fully connected layers and convolutional layers are followed by aleaky rectified linear unit (LReLU)

y = max(0.01x, x), (2)

as activation, where x is the input and y is the output [Maas, Hannun,and Ng 2013].

So�max Regression Loss

Conventional ApproachThe loss of the two-class classification is

lr = −1n

loges+t

es−t + es+t+∑j 6=t

loges−j

es−j + es+j

, (3)

whose partial derivative is

∂lr∂s+j

= − ∂lr∂s−j

=

− es−j

n(

es−j + es+j) if j = t,

es+j

n(

es−j + es+j) otherwise.

(4)

The partial derivative in the last FC layer is

∂lr∂w+

i= − ∂lr

∂w−i=

1n

n∑j=1

es+j xi,j

es−j + es+j− xi,t

. (5)

Misprediction of one VPP, which significantly influences CCR, barelya�ects the average loss. It also has a serious imbalance problem as it caneasily gain a high accuracy by simply classifying all VPPs as negative,which is meaningless.

Our MethodWe propose the following so�max regression loss

lc = − logest∑nj=1 esj

, (6)

whose partial derivative is

∂lc∂sj

=

esj∑nj=1 esj

− 1 if j = t,

esj∑nj=1 esj

otherwise.(7)

The partial derivative in the last FC layer is

∂lc∂wi

=

∑nj=1 esjxi,j∑n

j=1 esj− xi,t. (8)

The source fragment with higher score contributes much more signifi-cantly in the gradient with an exponential factor. The summation of thecoe�icients in the positive part equals to that of the negative part, sothere is no imbalance issue.

Experimental Results

Comparison between Ours and Wang (TVLSI’18)

0 0.5 1 1.5

M1 CCR

M3 CCR

M1 Time

M3 Time

1.21

1.12

1 · 10−3

2 · 10−3

1

1

1

1

Average Ratio

WangOurs

0 5 10 15

b7b11b13b14

b15_1b17_1

b18c432c880

c1355c1908c2670c3540c5315c6288c7552

M1 CCR (%)

Comparison between di�erent se�ings

50

55

60

65

CC

R(%

)

0

20

40

Tim

e(s

) Two-classVec

Vec & Img

Conclusion

I Demonstrate vector-based and image-based features.I Process heterogeneous features simultaneously in a neural network.I Propose a so�max regression loss that directly reflects on the

accuracy for the virtual pin pair matching problem of splitmanufacturing.

References

Li, Meng, Bei Yu, Yibo Lin, Xiaoqing Xu, Wuxi Li, and David Z. Pan (2018).“A Practical Split Manufacturing Framework for Trojan Prevention viaSimultaneous Wire Li�ing and Cell Insertion”. In: IEEE Transactions onComputer-Aided Design of Integrated Circuits and Systems (TCAD).

Maas, Andrew L, Awni Y Hannun, and Andrew Y Ng (2013). “Rectifier non-linearities improve neural network acoustic models”. In: International Con-ference on Machine Learning (ICML).

McCants, C (2011). “Trusted integrated chips (TIC)”. In: Intelligence Ad-vanced Research Projects Activity (IARPA), Tech. Rep.

Shamsi, Kaveh, Travis Meade, Meng Li, David Z. Pan, and Yier Jin (2019).“On the approximation resiliency of logic locking and IC camouflagingschemes”. In: IEEE Transactions on Information Forensics and Security 14.2,pp. 347–359.

Wang, Yujie, Pu Chen, Jiang Hu, Guofeng Li, and Jeyavijayan Rajendran(2018). “The cat and mouse in split manufacturing”. In: IEEE Transactionson Very Large Scale Integration Systems (TVLSI) 26.5, pp. 805–817.

Haocheng Li et al. A�acking Split Manufacturing from a Deep Learning Perspective [email protected]