9
Separate and Aend in Personal Email Search Yu Meng 1, Maryam Karimzadehgan 2 , Honglei Zhuang 2 , Donald Metzler 2 1 Department of Computer Science, University of Illinois at Urbana-Champaign, IL, USA 2 Google LLC, Mountain View, CA, USA 1 [email protected] 2 {maryamk, hlz, metzler}@google.com ABSTRACT In personal email search, user queries often impose different re- quirements on different aspects of the retrieved emails. For exam- ple, the query “my recent flight to the US” requires emails to be ranked based on both textual contents and recency of the email documents, while other queries such as “medical history” do not impose any constraints on the recency of the email. Recent deep learning-to-rank models for personal email search often directly concatenate dense numerical features 1 (e.g., document age) with embedded sparse features (e.g., n-gram embeddings). In this paper, we first show with a set of experiments on synthetic datasets that direct concatenation of dense and sparse features does not lead to the optimal search performance of deep neural ranking models. To effectively incorporate both sparse and dense email features into personal email search ranking, we propose a novel neural model, SepAn. SepAn first builds two separate neural models to learn from sparse and dense features respectively, and then applies an attention mechanism at the prediction level to derive the final pre- diction from these two models. We conduct a comprehensive set of experiments on a large-scale email search dataset, and demonstrate that our SepAn model consistently improves the search quality over the baseline models. CCS CONCEPTS Information systems Learning to rank; Computing methodologies Neural networks; Regularization. KEYWORDS Learning-to-Rank, Email Search, Neural Attention Model ACM Reference Format: Yu Meng 1, Maryam Karimzadehgan 2 , Honglei Zhuang 2 , Donald Metzler 2 . 2020. Separate and Attend in Personal Email Search. In The Thirteenth ACM International Conference on Web Search and Data Mining (WSDM’20), February 3–7, 2020, Houston, TX, USA. ACM, New York, NY, USA, 9 pages. https://doi.org/10.1145/XXXXXX.XXXXXX 1 Throughout the paper, we use dense features to refer to dense numerical features that cannot be derived from sparse features; not to be confused with embedded sparse features. Work done while Yu Meng was at Google. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. WSDM ’20, February 3–7, 2020, Houston, TX, USA © 2020 Association for Computing Machinery. ACM ISBN 978-1-4503-6822-3/20/02. . . $15.00 https://doi.org/10.1145/XXXXXX.XXXXXX 1 INTRODUCTION Email has long been an important means of daily communication. Personal email search, which helps users to quickly retrieve the emails they are looking for from their own corpora, has been an intriguing research topic in information retrieval (IR) for years. Email search is formulated as a learning-to-rank problem, which has been tackled with different learning models, such as boosted trees [8], SVM-based linear models [10, 12, 23], and shallow neural networks [9, 11]. Recently, deep neural networks (DNNs) have shown great suc- cess in learning-to-rank tasks. They significantly improve the per- formance of search engines in the presence of large-scale query logs in both web search [19] and email settings [39, 45, 51]. The ad- vantages of DNNs over traditional models are mainly two-fold: (1) DNNs have strong power to learn embedded representations from sparse features, including words [32] and characters [6]. This allows effective and accurate matching of textual features between queries and documents. (2) DNNs are proved to have universal approxi- mation capability [21], allowing to capture high-order interactions between query and document features. In the personal email search scenario, user queries impose dif- ferent requirements on different aspects of email documents to be retrieved. For example, the query “my recent flight to the US” requires the email search system to focus on both the textual con- tents and the recency of email documents, while queries such as “medical history” expect emails to be retrieved regardless of the recency. In email search models, different properties of email docu- ments are reflected by different types of features, including dense numerical ones (e.g., document age) and sparse categorical ones (e.g., n-grams). However, there have been few efforts that study how to effectively combine dense features with sparse features in the learning-to-rank setting, probably because a natural approach exists—simply concatenating dense features with embedded sparse features and feeding them into the DNNs. Indeed, many previous deep neural email search models use direct concatenation of dense features with embedded sparse features [15, 38, 39, 45]. In this paper, we first begin with a set of empirical findings and analyses to show that direct concatenation of dense with embedded sparse features does not lead to the optimal performance of DNN- based ranking models. As a result, we propose the SepAn model as an effective way to incorporate both dense and sparse features into the personal email search model. More specifically, SepAn model consists of three major modules: (1) Separate DNN models to learn from sparse and dense features, respectively. (2) An attention mechanism that aggregates the outputs from the sparse feature and dense feature DNN models. (3) A regularizer that enables the joint learning of the sparse and dense feature DNN models. The main advantage of SepAn over the simple concatenation approach is

Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

Separate and Attend in Personal Email SearchYu Meng1∗, Maryam Karimzadehgan2, Honglei Zhuang2, Donald Metzler21Department of Computer Science, University of Illinois at Urbana-Champaign, IL, USA

2Google LLC, Mountain View, CA, [email protected] 2{maryamk, hlz, metzler}@google.com

ABSTRACTIn personal email search, user queries often impose different re-quirements on different aspects of the retrieved emails. For exam-ple, the query “my recent flight to the US” requires emails to beranked based on both textual contents and recency of the emaildocuments, while other queries such as “medical history” do notimpose any constraints on the recency of the email. Recent deeplearning-to-rank models for personal email search often directlyconcatenate dense numerical features1 (e.g., document age) withembedded sparse features (e.g., n-gram embeddings). In this paper,we first show with a set of experiments on synthetic datasets thatdirect concatenation of dense and sparse features does not lead tothe optimal search performance of deep neural ranking models. Toeffectively incorporate both sparse and dense email features intopersonal email search ranking, we propose a novel neural model,SepAttn. SepAttn first builds two separate neural models to learnfrom sparse and dense features respectively, and then applies anattention mechanism at the prediction level to derive the final pre-diction from these two models. We conduct a comprehensive set ofexperiments on a large-scale email search dataset, and demonstratethat our SepAttn model consistently improves the search qualityover the baseline models.

CCS CONCEPTS• Information systems → Learning to rank; • Computingmethodologies→ Neural networks; Regularization.

KEYWORDSLearning-to-Rank, Email Search, Neural Attention ModelACM Reference Format:Yu Meng1∗, Maryam Karimzadehgan2, Honglei Zhuang2, Donald Metzler2.2020. Separate and Attend in Personal Email Search. In The ThirteenthACM International Conference on Web Search and Data Mining (WSDM’20),February 3–7, 2020, Houston, TX, USA. ACM, New York, NY, USA, 9 pages.https://doi.org/10.1145/XXXXXX.XXXXXX

1Throughout the paper, we use dense features to refer to dense numerical featuresthat cannot be derived from sparse features; not to be confused with embedded sparsefeatures.

∗Work done while Yu Meng was at Google.

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full citationon the first page. Copyrights for components of this work owned by others than ACMmust be honored. Abstracting with credit is permitted. To copy otherwise, or republish,to post on servers or to redistribute to lists, requires prior specific permission and/or afee. Request permissions from [email protected] ’20, February 3–7, 2020, Houston, TX, USA© 2020 Association for Computing Machinery.ACM ISBN 978-1-4503-6822-3/20/02. . . $15.00https://doi.org/10.1145/XXXXXX.XXXXXX

1 INTRODUCTIONEmail has long been an important means of daily communication.Personal email search, which helps users to quickly retrieve theemails they are looking for from their own corpora, has been anintriguing research topic in information retrieval (IR) for years.Email search is formulated as a learning-to-rank problem, whichhas been tackled with different learning models, such as boostedtrees [8], SVM-based linear models [10, 12, 23], and shallow neuralnetworks [9, 11].

Recently, deep neural networks (DNNs) have shown great suc-cess in learning-to-rank tasks. They significantly improve the per-formance of search engines in the presence of large-scale querylogs in both web search [19] and email settings [39, 45, 51]. The ad-vantages of DNNs over traditional models are mainly two-fold: (1)DNNs have strong power to learn embedded representations fromsparse features, including words [32] and characters [6]. This allowseffective and accurate matching of textual features between queriesand documents. (2) DNNs are proved to have universal approxi-mation capability [21], allowing to capture high-order interactionsbetween query and document features.

In the personal email search scenario, user queries impose dif-ferent requirements on different aspects of email documents tobe retrieved. For example, the query “my recent flight to the US”requires the email search system to focus on both the textual con-tents and the recency of email documents, while queries such as“medical history” expect emails to be retrieved regardless of therecency. In email search models, different properties of email docu-ments are reflected by different types of features, including densenumerical ones (e.g., document age) and sparse categorical ones(e.g., n-grams). However, there have been few efforts that studyhow to effectively combine dense features with sparse features inthe learning-to-rank setting, probably because a natural approachexists—simply concatenating dense features with embedded sparsefeatures and feeding them into the DNNs. Indeed, many previousdeep neural email search models use direct concatenation of densefeatures with embedded sparse features [15, 38, 39, 45].

In this paper, we first begin with a set of empirical findings andanalyses to show that direct concatenation of dense with embeddedsparse features does not lead to the optimal performance of DNN-based ranking models. As a result, we propose the SepAttn modelas an effective way to incorporate both dense and sparse featuresinto the personal email search model. More specifically, SepAttnmodel consists of three major modules: (1) Separate DNNmodels tolearn from sparse and dense features, respectively. (2) An attentionmechanism that aggregates the outputs from the sparse feature anddense feature DNN models. (3) A regularizer that enables the jointlearning of the sparse and dense feature DNN models. The mainadvantage of SepAttn over the simple concatenation approach is

Page 2: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

that SepAttn separates dense from sparse features, and automati-cally learns to explicitly focus on the important feature set, whileignore the unimportant ones for different query types.

In summary, the followings are the contributions of this paper:• We empirically show that simply concatenating embeddedsparse features with dense numerical features is sub-optimalfor DNN-based neural ranking models.

• To effectively leverage both sparse and dense features inneural ranking models, we propose to learn two separatemodels where their outputs are aggregated via an attentionmechanism to derive the final output.

• We propose a regularization method to train the sparse fea-ture model and dense feature model in a collaborative man-ner.

• We conduct a comprehensive set of experiments on syntheticdatasets and a large-scale real-world email search dataset todemonstrate the advantage of our proposed method.

The rest of the paper is organized as follows. Section 2 reviewsrelated work. Section 3 motivates the problem through a set of ex-periments on synthetic datasets. Section 4 introduces our proposedSepAttn model. Section 5 presents and analyzes the performancesof SepAttn on the synthetic datasets. In Section 6, we report anddiscuss the experimental results on a real-world large-scale emailsearch dataset. Section 7 concludes the paper and discusses futuredirections.

2 RELATEDWORKIn this section, we review related works on learning-to-rank, emailsearch models and state-of-the-art neural attention models.

2.1 Learning-to-RankLearning-to-rank refers to building ranking models with machinelearning algorithms. In early years, learning-to-rank has been stud-ied with different models, such as boosted trees [8], SVM-basedlinear models [10, 12, 23] and shallow neural networks [9, 11].Recent years have witnessed great success of applying DNNs tolearning-to-rank, such as [7, 15, 35, 38]. For a complete literaturereview on neural ranking models for information retrieval, pleaserefer to a survey by Mitra and Craswell [33].

2.2 Email SearchThere have been several studies in the IR community focusing on thetask of email search. The Enterprise tracks of TREC 2005 [40] andTREC 2006 [41] provide public datasets containing email data andsummarize some early explorations [14, 31, 34]. A typical trade-offin email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]proposed an email search framework with a learning-to-rank re-ranking module combining freshness with relevance signals ofemails as well as other features such as user actions. Alternatively,Carmel et al. [13] studied to present users with both the relevance-ranked results as well as the time-ranked results in two separatelists for better user experience. A number of studies specificallyfocus on improving the content-based relevance signals in emailsearch. Kuzi et al. [27] explored several methods to expand theusually short and sparse queries by finding more related terms to

improve the relevance results. Li et al. [29] studied a more specificsynonym expansion problem to improve email search performance.

User interaction data such as clicks is another important signalfor learning-to-rank models in email search. Bendersky et al. [4]leveraged user interactions by attribute parameterization. Wang etal. [48] mitigated the position bias in click data for better trainingof the model. In addition, Zamani et al. [51] showed that contextssuch as search request time and location of users were helpful foremail search quality.

There are also studies on understanding and leveraging queryintent information in email search. Ai et al. [2] conducted a thor-ough survey of search intent by analyzing user logs of email search.Shen et al. [39] categorized email search queries into different clus-ters before adding the query cluster information to improve emailranking.

To the best of our knowledge, there is no previous work onemail search that studied how to effectively combine dense numer-ical features with embedded sparse features in DNN-based rankingmodels. In the previous works, dense features were directly concate-nated with embedded sparse features, which led to the suboptimalperformances of DNN ranking models, as we will show later.

2.3 Attention ModelsRecently, neural attention mechanisms have demonstrated enor-mous power on sequence modeling. They derived the optimal se-quence representation by learning to focus on the important to-kens in the sequence and down-weighting unimportant ones fordownstream tasks. The attention mechanism was first proposed byBahdanau et al. [3] in machine translation, where attention wasused on top of RNN encoders for input-output alignment. Later,the attention mechanism has been adapted to a wide range ofcompelling sequence modeling tasks, including image caption gen-eration [49], text classification [50] and natural language questionanswering [20, 26, 43].

The above studies employ attention mechanism in conjunctionwith RNN or CNN models. Vaswani et al. [46] proposed the Trans-former, which used self-attention along with positional encoding.Later, Devlin et al. [16] proposed a deep bidirectional Transformerstructure, BERT, which becomes one of the state-of-the-art pre-trained language models benefiting many downstream tasks withfine-tuning. The attention mechanism has also been generalized toattend to a group of structurally adjacent items instead of singleones, as studied by Li et al. [30].

Regardless of the manner that attention mechanisms are used inthe previous works, the common purpose was to derive sequencerepresentations as a weighted average of token representations.Therefore, the attention mechanisms were applied at the featurelevel, as a step of feature learning. In our work, however, the atten-tion is used to aggregate the outputs from two models—the sparseand the dense feature models, to derive the final scoring outputs ofthe ranking model. To the best of our knowledge, this is the firstwork that applies attention mechanism at the prediction level.

Page 3: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

0 25 50 75 100 125 150 175 200Steps (thousands)

0.7500.7750.8000.8250.8500.8750.9000.9250.950

Accu

racy

ConcatSparseSepAttn

(a) First Synthetic dataset (Sparse only)

0 25 50 75 100 125 150 175 200Steps (thousands)

0.90

0.92

0.94

0.96

0.98

1.00

Accu

racy

ConcatDenseSepAttn

(b) Second Synthetic dataset (Dense only)

0 25 50 75 100 125 150 175 200Steps (thousands)

0.70

0.75

0.80

0.85

0.90

0.95

Accu

racy

ConcatSparseDenseSepAttn

(c) Third Synthetic dataset (Sparse and Dense Combined)

Figure 1: Training and testing performance of three models, i.e., Sparse only, Dense only and Sparse + Dense models on threesynthetic datasets. The dotted lines represent training accuracy and the solid lines represent testing accuracy. For complete-ness, we also show the performance of our SepAttnmodel in these figures. “Accuracy” is the ratio of correctly classified docu-ments.

3 MOTIVATIONIn this section, we provide motivation for our proposed “Separateand Attend” model, i.e., SepAttn, and show that simply concatenat-ing different feature types (i.e., dense and sparse features) does notlead to the optimal performance. Different queries in email searchimpose different requirements. For example, a query such as “medi-cal history” does not indicate requirements on dense features (e.g.,recency of an email document). As a result, simply concatenatingboth dense and sparse features and feeding them into the modelmight lead to suboptimal performance for such queries. In such acase, dense features are purely noise and could potentially distractthe model from concentrating on useful features.

In order to simulate the personal email search scenario, we createthree synthetic datasets and consider a simple binary classificationproblem, where a document d is classified to be relevant (y = 1)or irrelevant (y = 0) for a given query q. We use 100-dimensionalGloVe pretrained embeddings [37] as embedded sparse feature rep-resentations for queries and documents.

The query features q and document sparse features dsparse areobtained by randomly sampling word embeddings from the pre-trained embedding vocabulary. The document dense featuresddenseare generated from a uniform distribution in the interval [−0.5, 0.5].We set the dimension of dsparse (after embedding) and ddense tobe both 100. The ground-truth labels are generated differently foreach of these synthetic datasets which will be explained later inthis section.

We compare the performance of these models: (1) Themodel thatlearns fromq anddsparse only; (2) The model that learns fromq andddense only; (3) The model that learns fromq and the concatenationof dsparse and ddense. All models are built based on a feed-forwardDNN [28] with a hidden layer of 50 dimensions; the only differencebetween the models is that they learn from different feature sets.We generate 20, 000 samples for both training and testing sets.

For thefirst dataset, the following rule describes how the ground-truth data is generated:

y =

{1, if cos(dsparse,q) > 00, else

which simulates the scenario where document and query matchingis based only on sparse features, and dense features are unimpor-tant/unused.

Similarly, we create the second synthetic dataset by generatingground-truth labels only based on dense features as below:

y =

{1, ifq ∈ Q and

∑ddense < 0

0, else

where Q is a manually selected vocabulary of words that imposerequirements on the recency of the documents, such as “recent”,“latest”, “newest”, “today”.

We show the performance of the aforementioned models on thefirst synthetic dataset and second synthetic dataset in Figure 1(a)and Figure 1(b), respectively. In these figures, dotted lines repre-sent training curves, and solid lines represent testing curves. Wealso show the performance of our proposed method SepAttn forcompleteness in these figures and postpone the discussions aboutits performance to Section 5.

On the first synthetic dataset, sparse only model is the bench-mark because it learns only from sparse features and inherentlyavoids unimportant dense features. We observe from Figure 1(a)that the concatenation model overfits the training set—when itstraining accuracy goes up, its testing accuracy goes down. This in-dicates that the concatenation model learns noisy signals from thetraining set that cannot be generalized to the testing set, demon-strating its ineffectiveness of discriminating the useful featuresfrom unimportant ones.

On the second synthetic dataset, dense only model is the bench-mark since it avoids the unimportant sparse features. Figure 1(b)shows that the testing accuracy of concatenation model falls be-hind that of the dense only model. This again demonstrates thatthe concatenation model is negatively influenced by unimportantfeatures.

The third synthetic dataset is generated based on combinationof both sparse and dense features and the ground-truth labels aredefined as follows:

Page 4: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

y =

{1, if

(cos(dsparse,q) > 0

)∧

(q ∈ Q and

∑ddense < 0

)0, else

where the document-query matching requires both sparse anddense feature matching.

The performance of the above models using this dataset is shownin Figure 1(c). Although the concatenation model outperformssparse only and dense only models, it still suffers from overfit-ting (testing accuracy increases while training accuracy decreases).This demonstrates that DNNs are ineffective to learn from concate-nated sparse and dense features. We explain the reason as below:ddense and dsparse come from different feature space—dense numer-ical feature values (e.g., document age) have practical meanings(i.e., a higher value of document age means the email was receivedearlier), while embedding feature values (e.g., n-gram embedding)do not have practical meanings [42] (i.e., a higher value in n-gramembedding dimensions is not interpretable). If ddense and dsparseare directly concatenated and fed into the DNNs, operations appliedbetween ddense and dsparse are meaningless (e.g., it does not makesense to add/multiply document age with n-gram embeddings) andcan lead to ineffectiveness and overfitting of DNNs.

In the next section, we describe the details of our model, i.e.,SepAttn.

4 METHODOLOGYWe first formulate the problem and define the notations. Then wedescribe our SepAttn model.

4.1 Problem FormulationThe inputs for personal email search problem are tuples (q,D)

where q is a user query string, and D = {d1, . . . ,dn } is a list ofcandidate documents. The ground truth labels y ∈ {0, 1}n are userclick-through data indicating whether the corresponding documentis clicked. The goal of personal email search model is to rank D sothat the clicked document is ranked as high as possible.

The query features q are n-grams and character n-grams ex-tracted from the original user query string q, the document featuresd include sparse categorical features dsparse (e.g., n-grams) anddense numerical features ddense (e.g., document ages).

To preserve privacy, the query-document inputs are anonymizedbased on k-anonymity approach [44], and only query and documentn-grams that are frequent in the entire corpus are retained. Wesummarize the features used in the model in Table 1.

Given the input features X = {q,d1, . . . ,dn }, the email searchmodel learns a per-item scoring function h : (q,di ) → R, such that

h(X) =[h(q,d1) · · · h(q,dn )

]⊤ (1)

induces a permutation π and h(X)��π −1(r ) monotonically decreases

with increasing rank r . The per-item scoring function h could belearned by gradient boosted trees [18], support vectormachines [24]or neural networks [9]. In this paper, we aim to design a DNN-based model that effectively incorporates both dsparse and ddensefor improving personal email search performance.

With the input features (q,di ) described above, ground truthclick-through data yi and a per-item scoring function h, we follow

Table 1: Features used in personal email search.

Type Features

Query Feature (q) n-gramscharacter n-grams

Document Sparse Feature (dsparse)n-grams

character n-grams

Document Dense Feature (ddense)number of attachmentsnumber of recipients

document age

the Softmax Cross-Entropy listwise loss setting in [36] to train themodel. Specifically, the loss is defined as

Llistwise = −

n∑i=1

yi log(

exp (h(q,di ))∑nj=1 exp

(h(q,dj )

) ) . (2)

4.2 Separate and Attend MethodIn this section, we introduce our “Separate and Attend model”, i.e.,SepAttnmodel, to address the observed DNN learning issues whensparse and dense features are directly concatenated, as described inSection 3. The model structure of SepAttn is presented in Figure 2.

Often times different queries impose different requirements re-garding an email document to be retrieved, and different emaildocument properties are reflected by the sparse and dense featuresextracted. In this section, we first build separate models that learnfrom each sparse and dense features individually, and then developan attention mechanism at the prediction stage that enables themodel to explicitly learn to focus on the important features fordifferent queries.

Specifically, we first build two models to learn two per-itemscoring functions (one model for sparse and one model for densefeatures):

hsparse : (q,di :sparse) → R, hdense : (q,di :dense) → R,

which score a document given a query based on sparse and densedocument features, respectively.

Then the scores of a list of n documents given by sparse anddense feature models are:

hsparse(X) =[hsparse(q,d1:sparse) · · · hsparse(q,dn:sparse)

]⊤,

hdense(X) =[hdense(q,d1:dense) · · · hdense(q,dn:dense)

]⊤.

Next, we introduce an attention mechanism that aggregateshsparse(X) and hdense(X) to derive the final scores for the docu-ment list. Specifically, we first feed hsparse(X) and hdense(X) to aone-layer feed-forward neural network [28] to derive the hiddenrepresentation of the document scores, and then compute theirattention weights based on the hidden representations as well asa context vector. Finally, hsparse(X) and hdense(X) are weighted-averaged according to their attention weights to derive the finalscores h(X) for all the documents. Mathematically, the attention

Page 5: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

d1:dense<latexit sha1_base64="GSC0fnxVu7TeFK8ql13wMBTZL5E=">AAACCHicbVA9SwNBEN2LXzF+RS0tPAyCVbiLAcUqYGMZwXxAEsLe3iRZsrd37M6J4bjSxr9iY6GIrT/Bzn/jXpJCEx8s+3hvhpl5XiS4Rsf5tnIrq2vrG/nNwtb2zu5ecf+gqcNYMWiwUISq7VENgktoIEcB7UgBDTwBLW98nfmte1Cah/IOJxH0AjqUfMAZRSP1i8ddLxS+ngTmS/y0n7hXXYQHTHyQGtK0Xyw5ZWcKe5m4c1Iic9T7xa+uH7I4AIlMUK07rhNhL6EKOROQFrqxhoiyMR1Cx1BJA9C9ZHpIap8axbcHoTJPoj1Vf3ckNNDZrqYyoDjSi14m/ud1Yhxc9hIuoxhBstmgQSxsDO0sFdvnChiKiSGUKW52tdmIKsrQZFcwIbiLJy+TZqXsnpcrt9VSrTqPI0+OyAk5Iy65IDVyQ+qkQRh5JM/klbxZT9aL9W59zEpz1rznkPyB9fkDNc2asA==</latexit>

d1:dense<latexit sha1_base64="GSC0fnxVu7TeFK8ql13wMBTZL5E=">AAACCHicbVA9SwNBEN2LXzF+RS0tPAyCVbiLAcUqYGMZwXxAEsLe3iRZsrd37M6J4bjSxr9iY6GIrT/Bzn/jXpJCEx8s+3hvhpl5XiS4Rsf5tnIrq2vrG/nNwtb2zu5ecf+gqcNYMWiwUISq7VENgktoIEcB7UgBDTwBLW98nfmte1Cah/IOJxH0AjqUfMAZRSP1i8ddLxS+ngTmS/y0n7hXXYQHTHyQGtK0Xyw5ZWcKe5m4c1Iic9T7xa+uH7I4AIlMUK07rhNhL6EKOROQFrqxhoiyMR1Cx1BJA9C9ZHpIap8axbcHoTJPoj1Vf3ckNNDZrqYyoDjSi14m/ud1Yhxc9hIuoxhBstmgQSxsDO0sFdvnChiKiSGUKW52tdmIKsrQZFcwIbiLJy+TZqXsnpcrt9VSrTqPI0+OyAk5Iy65IDVyQ+qkQRh5JM/klbxZT9aL9W59zEpz1rznkPyB9fkDNc2asA==</latexit>

· · ·<latexit sha1_base64="knvjzPn9Dm3sfwmmLM2CwU7glWY=">AAAB7XicbVBNS8NAEJ3Ur1q/qh69BIvgqSRV0GNBDx4r2A9oQ9lsNu3azW7YnQil9D948aCIV/+PN/+N2zYHbX0w8Hhvhpl5YSq4Qc/7dgpr6xubW8Xt0s7u3v5B+fCoZVSmKWtSJZTuhMQwwSVrIkfBOqlmJAkFa4ejm5nffmLacCUfcJyyICEDyWNOCVqp1aORQtMvV7yqN4e7SvycVCBHo1/+6kWKZgmTSAUxput7KQYTopFTwaalXmZYSuiIDFjXUkkSZoLJ/Nqpe2aVyI2VtiXRnau/JyYkMWachLYzITg0y95M/M/rZhhfBxMu0wyZpItFcSZcVO7sdTfimlEUY0sI1dze6tIh0YSiDahkQ/CXX14lrVrVv6jW7i8r9ds8jiKcwCmcgw9XUIc7aEATKDzCM7zCm6OcF+fd+Vi0Fpx85hj+wPn8AbA/jzU=</latexit>

dn:dense<latexit sha1_base64="c8Hsqwd81lwRuT6Zyw6iDmA7oa4=">AAACCHicbVA9SwNBEN2LXzF+RS0tPAyCVbiLAcUqYGMZwXxAEsLe3iRZsrd37M6J4bjSxr9iY6GIrT/Bzn/jXpJCEx8s+3hvhpl5XiS4Rsf5tnIrq2vrG/nNwtb2zu5ecf+gqcNYMWiwUISq7VENgktoIEcB7UgBDTwBLW98nfmte1Cah/IOJxH0AjqUfMAZRSP1i8ddLxS+ngTmS/y0n8irLsIDJj5IDWnaL5acsjOFvUzcOSmROer94lfXD1kcgEQmqNYd14mwl1CFnAlIC91YQ0TZmA6hY6ikAeheMj0ktU+N4tuDUJkn0Z6qvzsSGuhsV1MZUBzpRS8T//M6MQ4uewmXUYwg2WzQIBY2hnaWiu1zBQzFxBDKFDe72mxEFWVosiuYENzFk5dJs1J2z8uV22qpVp3HkSdH5IScEZdckBq5IXXSIIw8kmfySt6sJ+vFerc+ZqU5a95zSP7A+vwBlZea7Q==</latexit>

dn:dense<latexit sha1_base64="c8Hsqwd81lwRuT6Zyw6iDmA7oa4=">AAACCHicbVA9SwNBEN2LXzF+RS0tPAyCVbiLAcUqYGMZwXxAEsLe3iRZsrd37M6J4bjSxr9iY6GIrT/Bzn/jXpJCEx8s+3hvhpl5XiS4Rsf5tnIrq2vrG/nNwtb2zu5ecf+gqcNYMWiwUISq7VENgktoIEcB7UgBDTwBLW98nfmte1Cah/IOJxH0AjqUfMAZRSP1i8ddLxS+ngTmS/y0n8irLsIDJj5IDWnaL5acsjOFvUzcOSmROer94lfXD1kcgEQmqNYd14mwl1CFnAlIC91YQ0TZmA6hY6ikAeheMj0ktU+N4tuDUJkn0Z6qvzsSGuhsV1MZUBzpRS8T//M6MQ4uewmXUYwg2WzQIBY2hnaWiu1zBQzFxBDKFDe72mxEFWVosiuYENzFk5dJs1J2z8uV22qpVp3HkSdH5IScEZdckBq5IXXSIIw8kmfySt6sJ+vFerc+ZqU5a95zSP7A+vwBlZea7Q==</latexit>· · ·<latexit sha1_base64="knvjzPn9Dm3sfwmmLM2CwU7glWY=">AAAB7XicbVBNS8NAEJ3Ur1q/qh69BIvgqSRV0GNBDx4r2A9oQ9lsNu3azW7YnQil9D948aCIV/+PN/+N2zYHbX0w8Hhvhpl5YSq4Qc/7dgpr6xubW8Xt0s7u3v5B+fCoZVSmKWtSJZTuhMQwwSVrIkfBOqlmJAkFa4ejm5nffmLacCUfcJyyICEDyWNOCVqp1aORQtMvV7yqN4e7SvycVCBHo1/+6kWKZgmTSAUxput7KQYTopFTwaalXmZYSuiIDFjXUkkSZoLJ/Nqpe2aVyI2VtiXRnau/JyYkMWachLYzITg0y95M/M/rZhhfBxMu0wyZpItFcSZcVO7sdTfimlEUY0sI1dze6tIh0YSiDahkQ/CXX14lrVrVv6jW7i8r9ds8jiKcwCmcgw9XUIc7aEATKDzCM7zCm6OcF+fd+Vi0Fpx85hj+wPn8AbA/jzU=</latexit>

pdense<latexit sha1_base64="QtadI/TWSj5ydbhWgcQE6tZDjhI=">AAACBnicbVBNS8NAEN34WetX1aMIwSJ4Kkkt6LHgxWMF+wFNKZvNtF262YTdiVhCTl78K148KOLV3+DNf+O2zUFbHyz7eG+GmXl+LLhGx/m2VlbX1jc2C1vF7Z3dvf3SwWFLR4li0GSRiFTHpxoEl9BEjgI6sQIa+gLa/vh66rfvQWkeyTucxNAL6VDyAWcUjdQvnXh+JAI9Cc2Xxlk/9RAeMA1AasiyfqnsVJwZ7GXi5qRMcjT6pS8viFgSgkQmqNZd14mxl1KFnAnIil6iIaZsTIfQNVTSEHQvnZ2R2WdGCexBpMyTaM/U3x0pDfV0U1MZUhzpRW8q/ud1Exxc9VIu4wRBsvmgQSJsjOxpJnbAFTAUE0MoU9zsarMRVZShSa5oQnAXT14mrWrFvahUb2vlei2Po0COySk5Jy65JHVyQxqkSRh5JM/klbxZT9aL9W59zEtXrLzniPyB9fkDSNiaPQ==</latexit>

pdense<latexit sha1_base64="QtadI/TWSj5ydbhWgcQE6tZDjhI=">AAACBnicbVBNS8NAEN34WetX1aMIwSJ4Kkkt6LHgxWMF+wFNKZvNtF262YTdiVhCTl78K148KOLV3+DNf+O2zUFbHyz7eG+GmXl+LLhGx/m2VlbX1jc2C1vF7Z3dvf3SwWFLR4li0GSRiFTHpxoEl9BEjgI6sQIa+gLa/vh66rfvQWkeyTucxNAL6VDyAWcUjdQvnXh+JAI9Cc2Xxlk/9RAeMA1AasiyfqnsVJwZ7GXi5qRMcjT6pS8viFgSgkQmqNZd14mxl1KFnAnIil6iIaZsTIfQNVTSEHQvnZ2R2WdGCexBpMyTaM/U3x0pDfV0U1MZUhzpRW8q/ud1Exxc9VIu4wRBsvmgQSJsjOxpJnbAFTAUE0MoU9zsarMRVZShSa5oQnAXT14mrWrFvahUb2vlei2Po0COySk5Jy65JHVyQxqkSRh5JM/klbxZT9aL9W59zEtXrLzniPyB9fkDSNiaPQ==</latexit>

pfinal<latexit sha1_base64="EGxtgnVmMOUBwsYfQQTBmTph4Os=">AAACBnicbVDLSsNAFJ34rPUVdSlCsAiuSlILuiy4cVnBPqAJYTKdtEMnkzBzI5aQlRt/xY0LRdz6De78GydtFtp6YJjDOfdy7z1BwpkC2/42VlbX1jc2K1vV7Z3dvX3z4LCr4lQS2iExj2U/wIpyJmgHGHDaTyTFUcBpL5hcF37vnkrFYnEH04R6ER4JFjKCQUu+eeIGMR+qaaS/LMn9zAX6AFnIBOZ57ps1u27PYC0TpyQ1VKLtm1/uMCZpRAUQjpUaOHYCXoYlMMJpXnVTRRNMJnhEB5oKHFHlZbMzcutMK0MrjKV+AqyZ+rsjw5EqNtWVEYaxWvQK8T9vkEJ45WVMJClQQeaDwpRbEFtFJtaQSUqATzXBRDK9q0XGWGICOrmqDsFZPHmZdBt156LeuG3WWs0yjgo6RqfoHDnoErXQDWqjDiLoET2jV/RmPBkvxrvxMS9dMcqeI/QHxucPQTyaOA==</latexit>

pfinal<latexit sha1_base64="EGxtgnVmMOUBwsYfQQTBmTph4Os=">AAACBnicbVDLSsNAFJ34rPUVdSlCsAiuSlILuiy4cVnBPqAJYTKdtEMnkzBzI5aQlRt/xY0LRdz6De78GydtFtp6YJjDOfdy7z1BwpkC2/42VlbX1jc2K1vV7Z3dvX3z4LCr4lQS2iExj2U/wIpyJmgHGHDaTyTFUcBpL5hcF37vnkrFYnEH04R6ER4JFjKCQUu+eeIGMR+qaaS/LMn9zAX6AFnIBOZ57ps1u27PYC0TpyQ1VKLtm1/uMCZpRAUQjpUaOHYCXoYlMMJpXnVTRRNMJnhEB5oKHFHlZbMzcutMK0MrjKV+AqyZ+rsjw5EqNtWVEYaxWvQK8T9vkEJ45WVMJClQQeaDwpRbEFtFJtaQSUqATzXBRDK9q0XGWGICOrmqDsFZPHmZdBt156LeuG3WWs0yjgo6RqfoHDnoErXQDWqjDiLoET2jV/RmPBkvxrvxMS9dMcqeI/QHxucPQTyaOA==</latexit>

· · ·<latexit sha1_base64="knvjzPn9Dm3sfwmmLM2CwU7glWY=">AAAB7XicbVBNS8NAEJ3Ur1q/qh69BIvgqSRV0GNBDx4r2A9oQ9lsNu3azW7YnQil9D948aCIV/+PN/+N2zYHbX0w8Hhvhpl5YSq4Qc/7dgpr6xubW8Xt0s7u3v5B+fCoZVSmKWtSJZTuhMQwwSVrIkfBOqlmJAkFa4ejm5nffmLacCUfcJyyICEDyWNOCVqp1aORQtMvV7yqN4e7SvycVCBHo1/+6kWKZgmTSAUxput7KQYTopFTwaalXmZYSuiIDFjXUkkSZoLJ/Nqpe2aVyI2VtiXRnau/JyYkMWachLYzITg0y95M/M/rZhhfBxMu0wyZpItFcSZcVO7sdTfimlEUY0sI1dze6tIh0YSiDahkQ/CXX14lrVrVv6jW7i8r9ds8jiKcwCmcgw9XUIc7aEATKDzCM7zCm6OcF+fd+Vi0Fpx85hj+wPn8AbA/jzU=</latexit>

d1:sparse<latexit sha1_base64="XO4x69x5ACjHwA/uMJ57wf7eCe0=">AAACCXicbVC7SgNBFJ31GeMramkzGASrsBsDilXAxjKCeUCyhNnZm2TI7IOZu2JYtrXxV2wsFLH1D+z8GyfJFpp4YJjDOfdy7z1eLIVG2/62VlbX1jc2C1vF7Z3dvf3SwWFLR4ni0OSRjFTHYxqkCKGJAiV0YgUs8CS0vfH11G/fg9IiCu9wEoMbsGEoBoIzNFK/RHteJH09CcyX+lk/da56CA+Y6pgpDVnWL5Xtij0DXSZOTsokR6Nf+ur5EU8CCJFLpnXXsWN0U6ZQcAlZsZdoiBkfsyF0DQ1ZANpNZ5dk9NQoPh1EyrwQ6Uz93ZGyQE+XNZUBw5Fe9Kbif143wcGlm4owThBCPh80SCTFiE5job5QwFFODGFcCbMr5SOmGEcTXtGE4CyevExa1YpzXqne1sr1Wh5HgRyTE3JGHHJB6uSGNEiTcPJInskrebOerBfr3fqYl65Yec8R+QPr8wco0Js5</latexit>

d1:sparse<latexit sha1_base64="XO4x69x5ACjHwA/uMJ57wf7eCe0=">AAACCXicbVC7SgNBFJ31GeMramkzGASrsBsDilXAxjKCeUCyhNnZm2TI7IOZu2JYtrXxV2wsFLH1D+z8GyfJFpp4YJjDOfdy7z1eLIVG2/62VlbX1jc2C1vF7Z3dvf3SwWFLR4ni0OSRjFTHYxqkCKGJAiV0YgUs8CS0vfH11G/fg9IiCu9wEoMbsGEoBoIzNFK/RHteJH09CcyX+lk/da56CA+Y6pgpDVnWL5Xtij0DXSZOTsokR6Nf+ur5EU8CCJFLpnXXsWN0U6ZQcAlZsZdoiBkfsyF0DQ1ZANpNZ5dk9NQoPh1EyrwQ6Uz93ZGyQE+XNZUBw5Fe9Kbif143wcGlm4owThBCPh80SCTFiE5job5QwFFODGFcCbMr5SOmGEcTXtGE4CyevExa1YpzXqne1sr1Wh5HgRyTE3JGHHJB6uSGNEiTcPJInskrebOerBfr3fqYl65Yec8R+QPr8wco0Js5</latexit>

· · ·<latexit sha1_base64="knvjzPn9Dm3sfwmmLM2CwU7glWY=">AAAB7XicbVBNS8NAEJ3Ur1q/qh69BIvgqSRV0GNBDx4r2A9oQ9lsNu3azW7YnQil9D948aCIV/+PN/+N2zYHbX0w8Hhvhpl5YSq4Qc/7dgpr6xubW8Xt0s7u3v5B+fCoZVSmKWtSJZTuhMQwwSVrIkfBOqlmJAkFa4ejm5nffmLacCUfcJyyICEDyWNOCVqp1aORQtMvV7yqN4e7SvycVCBHo1/+6kWKZgmTSAUxput7KQYTopFTwaalXmZYSuiIDFjXUkkSZoLJ/Nqpe2aVyI2VtiXRnau/JyYkMWachLYzITg0y95M/M/rZhhfBxMu0wyZpItFcSZcVO7sdTfimlEUY0sI1dze6tIh0YSiDahkQ/CXX14lrVrVv6jW7i8r9ds8jiKcwCmcgw9XUIc7aEATKDzCM7zCm6OcF+fd+Vi0Fpx85hj+wPn8AbA/jzU=</latexit>

dn:sparse<latexit sha1_base64="s9AeAEq4rR+Q4MYSJWs32alv87g=">AAACCXicbVC7SgNBFJ31GeMramkzGASrsBsDilXAxjKCeUASwuzsTTJkdnaZuSuGZVsbf8XGQhFb/8DOv3HyKDTxwDCHc+7l3nv8WAqDrvvtrKyurW9s5rby2zu7e/uFg8OGiRLNoc4jGemWzwxIoaCOAiW0Yg0s9CU0/dH1xG/egzYiUnc4jqEbsoESfcEZWqlXoB0/koEZh/ZLg6yXqqsOwgOmJmbaQJb1CkW35E5Bl4k3J0UyR61X+OoEEU9CUMglM6btuTF2U6ZRcAlZvpMYiBkfsQG0LVUsBNNNp5dk9NQqAe1H2j6FdKr+7khZaCbL2sqQ4dAsehPxP6+dYP+ymwoVJwiKzwb1E0kxopNYaCA0cJRjSxjXwu5K+ZBpxtGGl7cheIsnL5NGueSdl8q3lWK1Mo8jR47JCTkjHrkgVXJDaqROOHkkz+SVvDlPzovz7nzMSlecec8R+QPn8weI15t2</latexit>

dn:sparse<latexit sha1_base64="s9AeAEq4rR+Q4MYSJWs32alv87g=">AAACCXicbVC7SgNBFJ31GeMramkzGASrsBsDilXAxjKCeUASwuzsTTJkdnaZuSuGZVsbf8XGQhFb/8DOv3HyKDTxwDCHc+7l3nv8WAqDrvvtrKyurW9s5rby2zu7e/uFg8OGiRLNoc4jGemWzwxIoaCOAiW0Yg0s9CU0/dH1xG/egzYiUnc4jqEbsoESfcEZWqlXoB0/koEZh/ZLg6yXqqsOwgOmJmbaQJb1CkW35E5Bl4k3J0UyR61X+OoEEU9CUMglM6btuTF2U6ZRcAlZvpMYiBkfsQG0LVUsBNNNp5dk9NQqAe1H2j6FdKr+7khZaCbL2sqQ4dAsehPxP6+dYP+ymwoVJwiKzwb1E0kxopNYaCA0cJRjSxjXwu5K+ZBpxtGGl7cheIsnL5NGueSdl8q3lWK1Mo8jR47JCTkjHrkgVXJDaqROOHkkz+SVvDlPzovz7nzMSlecec8R+QPn8weI15t2</latexit>

· · ·<latexit sha1_base64="knvjzPn9Dm3sfwmmLM2CwU7glWY=">AAAB7XicbVBNS8NAEJ3Ur1q/qh69BIvgqSRV0GNBDx4r2A9oQ9lsNu3azW7YnQil9D948aCIV/+PN/+N2zYHbX0w8Hhvhpl5YSq4Qc/7dgpr6xubW8Xt0s7u3v5B+fCoZVSmKWtSJZTuhMQwwSVrIkfBOqlmJAkFa4ejm5nffmLacCUfcJyyICEDyWNOCVqp1aORQtMvV7yqN4e7SvycVCBHo1/+6kWKZgmTSAUxput7KQYTopFTwaalXmZYSuiIDFjXUkkSZoLJ/Nqpe2aVyI2VtiXRnau/JyYkMWachLYzITg0y95M/M/rZhhfBxMu0wyZpItFcSZcVO7sdTfimlEUY0sI1dze6tIh0YSiDahkQ/CXX14lrVrVv6jW7i8r9ds8jiKcwCmcgw9XUIc7aEATKDzCM7zCm6OcF+fd+Vi0Fpx85hj+wPn8AbA/jzU=</latexit>

psparse<latexit sha1_base64="b7fxbSD7IzvYeLroPQSLdO6uDLM=">AAACB3icbVDLSsNAFJ34rPUVdSlIsAiuSlILuiy4cVnBPqAJZTK9bYdOHszciCVk58ZfceNCEbf+gjv/xkmbhbYeGOZwzr3ce48fC67Qtr+NldW19Y3N0lZ5e2d3b988OGyrKJEMWiwSkez6VIHgIbSQo4BuLIEGvoCOP7nO/c49SMWj8A6nMXgBHYV8yBlFLfXNE9ePxEBNA/2lcdZPXYQHTFVMpYIs65sVu2rPYC0TpyAVUqDZN7/cQcSSAEJkgirVc+wYvZRK5ExAVnYTBTFlEzqCnqYhDUB56eyOzDrTysAaRlK/EK2Z+rsjpYHKV9WVAcWxWvRy8T+vl+Dwykt5GCcIIZsPGibCwsjKQ7EGXAJDMdWEMsn1rhYbU0kZ6ujKOgRn8eRl0q5VnYtq7bZeadSLOErkmJySc+KQS9IgN6RJWoSRR/JMXsmb8WS8GO/Gx7x0xSh6jsgfGJ8/O3iaxg==</latexit>

psparse<latexit sha1_base64="b7fxbSD7IzvYeLroPQSLdO6uDLM=">AAACB3icbVDLSsNAFJ34rPUVdSlIsAiuSlILuiy4cVnBPqAJZTK9bYdOHszciCVk58ZfceNCEbf+gjv/xkmbhbYeGOZwzr3ce48fC67Qtr+NldW19Y3N0lZ5e2d3b988OGyrKJEMWiwSkez6VIHgIbSQo4BuLIEGvoCOP7nO/c49SMWj8A6nMXgBHYV8yBlFLfXNE9ePxEBNA/2lcdZPXYQHTFVMpYIs65sVu2rPYC0TpyAVUqDZN7/cQcSSAEJkgirVc+wYvZRK5ExAVnYTBTFlEzqCnqYhDUB56eyOzDrTysAaRlK/EK2Z+rsjpYHKV9WVAcWxWvRy8T+vl+Dwykt5GCcIIZsPGibCwsjKQ7EGXAJDMdWEMsn1rhYbU0kZ6ujKOgRn8eRl0q5VnYtq7bZeadSLOErkmJySc+KQS9IgN6RJWoSRR/JMXsmb8WS8GO/Gx7x0xSh6jsgfGJ8/O3iaxg==</latexit>

· · ·<latexit sha1_base64="knvjzPn9Dm3sfwmmLM2CwU7glWY=">AAAB7XicbVBNS8NAEJ3Ur1q/qh69BIvgqSRV0GNBDx4r2A9oQ9lsNu3azW7YnQil9D948aCIV/+PN/+N2zYHbX0w8Hhvhpl5YSq4Qc/7dgpr6xubW8Xt0s7u3v5B+fCoZVSmKWtSJZTuhMQwwSVrIkfBOqlmJAkFa4ejm5nffmLacCUfcJyyICEDyWNOCVqp1aORQtMvV7yqN4e7SvycVCBHo1/+6kWKZgmTSAUxput7KQYTopFTwaalXmZYSuiIDFjXUkkSZoLJ/Nqpe2aVyI2VtiXRnau/JyYkMWachLYzITg0y95M/M/rZhhfBxMu0wyZpItFcSZcVO7sdTfimlEUY0sI1dze6tIh0YSiDahkQ/CXX14lrVrVv6jW7i8r9ds8jiKcwCmcgw9XUIc7aEATKDzCM7zCm6OcF+fd+Vi0Fpx85hj+wPn8AbA/jzU=</latexit>

Lreg<latexit sha1_base64="oRd5XKXUpQnKaF11a3LQnZFTLIE=">AAACAXicbVA9SwNBEN2LXzF+ndoINodBsAp3MaBlwMbCIoL5gOQIe5tJsmTvg905MRxn41+xsVDE1n9h579xL7lCEx8MPN6bYWaeFwmu0La/jcLK6tr6RnGztLW9s7tn7h+0VBhLBk0WilB2PKpA8ACayFFAJ5JAfU9A25tcZX77HqTiYXCH0whcn44CPuSMopb65lHPpzhmVCQ3aT/pITxgImGUpn2zbFfsGaxl4uSkTHI0+uZXbxCy2IcAmaBKdR07QjehEjkTkJZ6sYKIsgkdQVfTgPqg3GT2QWqdamVgDUOpK0Brpv6eSKiv1NT3dGd2r1r0MvE/rxvj8NJNeBDFCAGbLxrGwsLQyuKwBlwCQzHVhDLJ9a0WG1NJGerQSjoEZ/HlZdKqVpzzSvW2Vq7X8jiK5JickDPikAtSJ9ekQZqEkUfyTF7Jm/FkvBjvxse8tWDkM4fkD4zPH8/ml7k=</latexit>

Lreg<latexit sha1_base64="oRd5XKXUpQnKaF11a3LQnZFTLIE=">AAACAXicbVA9SwNBEN2LXzF+ndoINodBsAp3MaBlwMbCIoL5gOQIe5tJsmTvg905MRxn41+xsVDE1n9h579xL7lCEx8MPN6bYWaeFwmu0La/jcLK6tr6RnGztLW9s7tn7h+0VBhLBk0WilB2PKpA8ACayFFAJ5JAfU9A25tcZX77HqTiYXCH0whcn44CPuSMopb65lHPpzhmVCQ3aT/pITxgImGUpn2zbFfsGaxl4uSkTHI0+uZXbxCy2IcAmaBKdR07QjehEjkTkJZ6sYKIsgkdQVfTgPqg3GT2QWqdamVgDUOpK0Brpv6eSKiv1NT3dGd2r1r0MvE/rxvj8NJNeBDFCAGbLxrGwsLQyuKwBlwCQzHVhDLJ9a0WG1NJGerQSjoEZ/HlZdKqVpzzSvW2Vq7X8jiK5JickDPikAtSJ9ekQZqEkUfyTF7Jm/FkvBjvxse8tWDkM4fkD4zPH8/ml7k=</latexit>

SoftmaxSoftmax

Prediction from dense features

Prediction from sparse features Llistwise

<latexit sha1_base64="5M5IkUmDMe1InSqlxibwqyVkXT0=">AAACBnicbVA9SwNBEN3zM8avU0sRDoNgFe5iQMuAjYVFBPMByRH2NpNkyd4Hu3NqOK6y8a/YWChi62+w89+4l1yhiQ8GHu/NMDPPiwRXaNvfxtLyyuraemGjuLm1vbNr7u03VRhLBg0WilC2PapA8AAayFFAO5JAfU9AyxtfZn7rDqTiYXCLkwhcnw4DPuCMopZ65lHXpzhiVCTXaS/pIjxgkq295wrStGeW7LI9hbVInJyUSI56z/zq9kMW+xAgE1SpjmNH6CZUImcC0mI3VhBRNqZD6GgaUB+Um0zfSK0TrfStQSh1BWhN1d8TCfWVmvie7syOVvNeJv7ndWIcXLgJD6IYIWCzRYNYWBhaWSZWn0tgKCaaUCa5vtViIyopQ51cUYfgzL+8SJqVsnNWrtxUS7VqHkeBHJJjckocck5q5IrUSYMw8kieySt5M56MF+Pd+Ji1Lhn5zAH5A+PzBxsImiE=</latexit>

Llistwise<latexit sha1_base64="5M5IkUmDMe1InSqlxibwqyVkXT0=">AAACBnicbVA9SwNBEN3zM8avU0sRDoNgFe5iQMuAjYVFBPMByRH2NpNkyd4Hu3NqOK6y8a/YWChi62+w89+4l1yhiQ8GHu/NMDPPiwRXaNvfxtLyyuraemGjuLm1vbNr7u03VRhLBg0WilC2PapA8AAayFFAO5JAfU9AyxtfZn7rDqTiYXCLkwhcnw4DPuCMopZ65lHXpzhiVCTXaS/pIjxgkq295wrStGeW7LI9hbVInJyUSI56z/zq9kMW+xAgE1SpjmNH6CZUImcC0mI3VhBRNqZD6GgaUB+Um0zfSK0TrfStQSh1BWhN1d8TCfWVmvie7syOVvNeJv7ndWIcXLgJD6IYIWCzRYNYWBhaWSZWn0tgKCaaUCa5vtViIyopQ51cUYfgzL+8SJqVsnNWrtxUS7VqHkeBHJJjckocck5q5IrUSYMw8kieySt5M56MF+Pd+Ji1Lhn5zAH5A+PzBxsImiE=</latexit>

hdense(X )<latexit sha1_base64="HpiNKma4t2s/1k6kciAjFI4xbNk=">AAACE3icbVBNS8NAEN34WetX1aOXYBHUQ0mqoEfBi8cK9gOaEjabiV3cbMLuRCwh/8GLf8WLB0W8evHmv3HT9qDWgWUf780wb16QCq7Rcb6sufmFxaXlykp1dW19Y7O2td3RSaYYtFkiEtULqAbBJbSRo4BeqoDGgYBucHtR6t07UJon8hpHKQxieiN5xBlFQ/m1Iy9IRKhHsfnyYeHnHsI95iFIDUVx4MUUh4yKvFcc+rW603DGZc8CdwrqZFotv/bphQnLYpDIBNW67zopDnKqkDMBRdXLNKSU3dIb6BsoaQx6kI9vKux9w4R2lCjzJNpj9udETmNd2jadpUf9VyvJ/7R+htHZIOcyzRAkmyyKMmFjYpcB2SFXwFCMDKBMcePVZkOqKEMTY9WE4P49eRZ0mg33uNG8Oqmfn0zjqJBdskcOiEtOyTm5JC3SJow8kCfyQl6tR+vZerPeJ61z1nRmh/wq6+MbNYufjg==</latexit>

hdense(X )<latexit sha1_base64="HpiNKma4t2s/1k6kciAjFI4xbNk=">AAACE3icbVBNS8NAEN34WetX1aOXYBHUQ0mqoEfBi8cK9gOaEjabiV3cbMLuRCwh/8GLf8WLB0W8evHmv3HT9qDWgWUf780wb16QCq7Rcb6sufmFxaXlykp1dW19Y7O2td3RSaYYtFkiEtULqAbBJbSRo4BeqoDGgYBucHtR6t07UJon8hpHKQxieiN5xBlFQ/m1Iy9IRKhHsfnyYeHnHsI95iFIDUVx4MUUh4yKvFcc+rW603DGZc8CdwrqZFotv/bphQnLYpDIBNW67zopDnKqkDMBRdXLNKSU3dIb6BsoaQx6kI9vKux9w4R2lCjzJNpj9udETmNd2jadpUf9VyvJ/7R+htHZIOcyzRAkmyyKMmFjYpcB2SFXwFCMDKBMcePVZkOqKEMTY9WE4P49eRZ0mg33uNG8Oqmfn0zjqJBdskcOiEtOyTm5JC3SJow8kCfyQl6tR+vZerPeJ61z1nRmh/wq6+MbNYufjg==</latexit>

hsparse(X )<latexit sha1_base64="ORdBkn6ndwkrRQ9hxLxSbuxO2dY=">AAACFHicbVDLSsNAFJ34rPVVdekmWISKUJJa0GXBjcsK9gFNCJPpbTt08mDmRiwhH+HGX3HjQhG3Ltz5N07aLrT1wjCHc+7lnnv8WHCFlvVtrKyurW9sFraK2zu7e/ulg8O2ihLJoMUiEcmuTxUIHkILOQroxhJo4Avo+OPrXO/cg1Q8Cu9wEoMb0GHIB5xR1JRXOnf8SPTVJNBfOsq81EF4wFTFVCrIsooTUBwxKtJuduaVylbVmpa5DOw5KJN5Nb3Sl9OPWBJAiExQpXq2FaObUomcCciKTqIgpmxMh9DTMKQBKDedHpWZp5rpm4NI6heiOWV/T6Q0ULlv3Zl7VItaTv6n9RIcXLkpD+MEIWSzRYNEmBiZeUJmn0tgKCYaUCa59mqyEZWUoc6xqEOwF09eBu1a1b6o1m7r5UZ9HkeBHJMTUiE2uSQNckOapEUYeSTP5JW8GU/Gi/FufMxaV4z5zBH5U8bnDy6woBc=</latexit>

hsparse(X )<latexit sha1_base64="ORdBkn6ndwkrRQ9hxLxSbuxO2dY=">AAACFHicbVDLSsNAFJ34rPVVdekmWISKUJJa0GXBjcsK9gFNCJPpbTt08mDmRiwhH+HGX3HjQhG3Ltz5N07aLrT1wjCHc+7lnnv8WHCFlvVtrKyurW9sFraK2zu7e/ulg8O2ihLJoMUiEcmuTxUIHkILOQroxhJo4Avo+OPrXO/cg1Q8Cu9wEoMb0GHIB5xR1JRXOnf8SPTVJNBfOsq81EF4wFTFVCrIsooTUBwxKtJuduaVylbVmpa5DOw5KJN5Nb3Sl9OPWBJAiExQpXq2FaObUomcCciKTqIgpmxMh9DTMKQBKDedHpWZp5rpm4NI6heiOWV/T6Q0ULlv3Zl7VItaTv6n9RIcXLkpD+MEIWSzRYNEmBiZeUJmn0tgKCYaUCa59mqyEZWUoc6xqEOwF09eBu1a1b6o1m7r5UZ9HkeBHJMTUiE2uSQNckOapEUYeSTP5JW8GU/Gi/FufMxaV4z5zBH5U8bnDy6woBc=</latexit>

h(X )<latexit sha1_base64="ZpGd2JjwhskqVOmvyWfxkSUvu0g=">AAACBHicbVC7TsMwFHXKq5RXgLGLRYVUlioplWCsxMJYJPqQ2qhyHKe16tiR7SBVUQYWfoWFAYRY+Qg2/ganzQAtV7J8dM69uuceP2ZUacf5tkobm1vbO+Xdyt7+weGRfXzSUyKRmHSxYEIOfKQIo5x0NdWMDGJJUOQz0vdnN7nefyBSUcHv9TwmXoQmnIYUI22osV0d+YIFah6ZL51m9VGE9BQjlg6yi7FdcxrOouA6cAtQA0V1xvbXKBA4iQjXmCGlhq4Tay9FUlPMSFYZJYrECM/QhAwN5CgiyksXR2Tw3DABDIU0j2u4YH9PpChSuU/TmXtUq1pO/qcNEx1eeynlcaIJx8tFYcKgFjBPBAZUEqzZ3ACEJTVeIZ4iibA2uVVMCO7qyeug12y4l43mXavWbhVxlEEVnIE6cMEVaINb0AFdgMEjeAav4M16sl6sd+tj2VqyiplT8Keszx9TBJh5</latexit>

h(X )<latexit sha1_base64="ZpGd2JjwhskqVOmvyWfxkSUvu0g=">AAACBHicbVC7TsMwFHXKq5RXgLGLRYVUlioplWCsxMJYJPqQ2qhyHKe16tiR7SBVUQYWfoWFAYRY+Qg2/ganzQAtV7J8dM69uuceP2ZUacf5tkobm1vbO+Xdyt7+weGRfXzSUyKRmHSxYEIOfKQIo5x0NdWMDGJJUOQz0vdnN7nefyBSUcHv9TwmXoQmnIYUI22osV0d+YIFah6ZL51m9VGE9BQjlg6yi7FdcxrOouA6cAtQA0V1xvbXKBA4iQjXmCGlhq4Tay9FUlPMSFYZJYrECM/QhAwN5CgiyksXR2Tw3DABDIU0j2u4YH9PpChSuU/TmXtUq1pO/qcNEx1eeynlcaIJx8tFYcKgFjBPBAZUEqzZ3ACEJTVeIZ4iibA2uVVMCO7qyeug12y4l43mXavWbhVxlEEVnIE6cMEVaINb0AFdgMEjeAav4M16sl6sd+tj2VqyiplT8Keszx9TBJh5</latexit>

Softmax

↵dense<latexit sha1_base64="SrlS1gXhGZfEH0fKkmoqqxGyBO0=">AAACCnicbVBNS8NAEN34WetX1aOXaBE8laQW9Fjw4rGC/YAmlM122i7dbMLuRCwhZy/+FS8eFPHqL/Dmv3Hb5qCtDwYe783szrwgFlyj43xbK6tr6xubha3i9s7u3n7p4LClo0QxaLJIRKoTUA2CS2giRwGdWAENAwHtYHw99dv3oDSP5B1OYvBDOpR8wBlFI/VKJ97sjTQQCWQeFfGI9lIP4QHTPkgNWdYrlZ2KM4O9TNyclEmORq/05fUjloQgkQmqddd1YvRTqpAzAVnRSzTElI3pELqGShqC9tPZGpl9ZpS+PYiUKYn2TP09kdJQ60kYmM6Q4kgvelPxP6+b4ODKT7mMEwTJ5h8NEmFjZE9zsftcAUMxMYQyxc2uNhtRRRma9IomBHfx5GXSqlbci0r1tlau1/I4CuSYnJJz4pJLUic3pEGahJFH8kxeyZv1ZL1Y79bHvHXFymeOyB9Ynz9X55vn</latexit>

↵sparse<latexit sha1_base64="d7upo7RAdS3/6W6uxejGzjM4R7g=">AAACC3icbVDJSgNBEO1xjXGLevQyJAiewkwM6DHgxWMEs0AmhJpOJWnSs9BdI4Zh7l78FS8eFPHqD3jzb+wsB018UPB4r6q76vmxFJoc59taW9/Y3NrO7eR39/YPDgtHx00dJYpjg0cyUm0fNEoRYoMESWzHCiHwJbb88fXUb92j0iIK72gSYzeAYSgGggMZqVcoerM3Ul8mmHkg4xH0Uo/wgVIdg9KYZb1CySk7M9irxF2QElug3it8ef2IJwGGxCVo3XGdmLopKBJcYpb3Eo0x8DEMsWNoCAHqbjrbI7PPjNK3B5EyFZI9U39PpBBoPQl80xkAjfSyNxX/8zoJDa66qQjjhDDk848GibQpsqfB2H2hkJOcGAJcCbOrzUeggJOJL29CcJdPXiXNStm9KFduq6VadRFHjp2yIjtnLrtkNXbD6qzBOHtkz+yVvVlP1ov1bn3MW9esxcwJ+wPr8wdMEZxw</latexit>

↵dense<latexit sha1_base64="SrlS1gXhGZfEH0fKkmoqqxGyBO0=">AAACCnicbVBNS8NAEN34WetX1aOXaBE8laQW9Fjw4rGC/YAmlM122i7dbMLuRCwhZy/+FS8eFPHqL/Dmv3Hb5qCtDwYe783szrwgFlyj43xbK6tr6xubha3i9s7u3n7p4LClo0QxaLJIRKoTUA2CS2giRwGdWAENAwHtYHw99dv3oDSP5B1OYvBDOpR8wBlFI/VKJ97sjTQQCWQeFfGI9lIP4QHTPkgNWdYrlZ2KM4O9TNyclEmORq/05fUjloQgkQmqddd1YvRTqpAzAVnRSzTElI3pELqGShqC9tPZGpl9ZpS+PYiUKYn2TP09kdJQ60kYmM6Q4kgvelPxP6+b4ODKT7mMEwTJ5h8NEmFjZE9zsftcAUMxMYQyxc2uNhtRRRma9IomBHfx5GXSqlbci0r1tlau1/I4CuSYnJJz4pJLUic3pEGahJFH8kxeyZv1ZL1Y79bHvHXFymeOyB9Ynz9X55vn</latexit>

↵sparse<latexit sha1_base64="d7upo7RAdS3/6W6uxejGzjM4R7g=">AAACC3icbVDJSgNBEO1xjXGLevQyJAiewkwM6DHgxWMEs0AmhJpOJWnSs9BdI4Zh7l78FS8eFPHqD3jzb+wsB018UPB4r6q76vmxFJoc59taW9/Y3NrO7eR39/YPDgtHx00dJYpjg0cyUm0fNEoRYoMESWzHCiHwJbb88fXUb92j0iIK72gSYzeAYSgGggMZqVcoerM3Ul8mmHkg4xH0Uo/wgVIdg9KYZb1CySk7M9irxF2QElug3it8ef2IJwGGxCVo3XGdmLopKBJcYpb3Eo0x8DEMsWNoCAHqbjrbI7PPjNK3B5EyFZI9U39PpBBoPQl80xkAjfSyNxX/8zoJDa66qQjjhDDk848GibQpsqfB2H2hkJOcGAJcCbOrzUeggJOJL29CcJdPXiXNStm9KFduq6VadRFHjp2yIjtnLrtkNXbD6qzBOHtkz+yVvVlP1ov1bn3MW9esxcwJ+wPr8wdMEZxw</latexit>

Ldense<latexit sha1_base64="CAcekDSYY5hhWDodSwQ/qDynZio=">AAACA3icbVA9SwNBEN2LXzF+Re20OQyCVbiLAS0DNhYWEcwHJEfY20ySJXt7x+6cGI4DG/+KjYUitv4JO/+Nm+QKTXww8Hhvhpl5fiS4Rsf5tnIrq2vrG/nNwtb2zu5ecf+gqcNYMWiwUISq7VMNgktoIEcB7UgBDXwBLX98NfVb96A0D+UdTiLwAjqUfMAZRSP1ikfdgOKIUZHcpL2ki/CASR+khjTtFUtO2ZnBXiZuRkokQ71X/Or2QxYHIJEJqnXHdSL0EqqQMwFpoRtriCgb0yF0DJU0AO0lsx9S+9QofXsQKlMS7Zn6eyKhgdaTwDed04v1ojcV//M6MQ4uvYTLKEaQbL5oEAsbQ3saiN3nChiKiSGUKW5utdmIKsrQxFYwIbiLLy+TZqXsnpcrt9VSrZrFkSfH5IScEZdckBq5JnXSIIw8kmfySt6sJ+vFerc+5q05K5s5JH9gff4Aa3OYng==</latexit>

Ldense<latexit sha1_base64="CAcekDSYY5hhWDodSwQ/qDynZio=">AAACA3icbVA9SwNBEN2LXzF+Re20OQyCVbiLAS0DNhYWEcwHJEfY20ySJXt7x+6cGI4DG/+KjYUitv4JO/+Nm+QKTXww8Hhvhpl5fiS4Rsf5tnIrq2vrG/nNwtb2zu5ecf+gqcNYMWiwUISq7VMNgktoIEcB7UgBDXwBLX98NfVb96A0D+UdTiLwAjqUfMAZRSP1ikfdgOKIUZHcpL2ki/CASR+khjTtFUtO2ZnBXiZuRkokQ71X/Or2QxYHIJEJqnXHdSL0EqqQMwFpoRtriCgb0yF0DJU0AO0lsx9S+9QofXsQKlMS7Zn6eyKhgdaTwDed04v1ojcV//M6MQ4uvYTLKEaQbL5oEAsbQ3saiN3nChiKiSGUKW5utdmIKsrQxFYwIbiLLy+TZqXsnpcrt9VSrZrFkSfH5IScEZdckBq5JnXSIIw8kmfySt6sJ+vFerc+5q05K5s5JH9gff4Aa3OYng==</latexit>

Lsparse<latexit sha1_base64="4hB6brZjT/AdRv5k1QiuK4aBN24=">AAACBHicbVA9SwNBEN2LXzF+nVqmWQyCVbiLAS0DNhYWEcwHJCHsbSbJkr0PdufEcFxh41+xsVDE1h9h579xk1yhiQ8GHu/NMDPPi6TQ6DjfVm5tfWNzK79d2Nnd2z+wD4+aOowVhwYPZajaHtMgRQANFCihHSlgvieh5U2uZn7rHpQWYXCH0wh6PhsFYig4QyP17WLXZzjmTCY3aT/pIjxgoiOmNKRp3y45ZWcOukrcjJRIhnrf/uoOQh77ECCXTOuO60TYS5hCwSWkhW6sIWJ8wkbQMTRgPuheMn8ipadGGdBhqEwFSOfq74mE+VpPfc90zk7Wy95M/M/rxDi87CUiiGKEgC8WDWNJMaSzROhAKOAop4YwroS5lfIxU4yjya1gQnCXX14lzUrZPS9XbqulWjWLI0+K5IScEZdckBq5JnXSIJw8kmfySt6sJ+vFerc+Fq05K5s5Jn9gff4AXIyZJw==</latexit>

Lsparse<latexit sha1_base64="4hB6brZjT/AdRv5k1QiuK4aBN24=">AAACBHicbVA9SwNBEN2LXzF+nVqmWQyCVbiLAS0DNhYWEcwHJCHsbSbJkr0PdufEcFxh41+xsVDE1h9h579xk1yhiQ8GHu/NMDPPi6TQ6DjfVm5tfWNzK79d2Nnd2z+wD4+aOowVhwYPZajaHtMgRQANFCihHSlgvieh5U2uZn7rHpQWYXCH0wh6PhsFYig4QyP17WLXZzjmTCY3aT/pIjxgoiOmNKRp3y45ZWcOukrcjJRIhnrf/uoOQh77ECCXTOuO60TYS5hCwSWkhW6sIWJ8wkbQMTRgPuheMn8ipadGGdBhqEwFSOfq74mE+VpPfc90zk7Wy95M/M/rxDi87CUiiGKEgC8WDWNJMaSzROhAKOAop4YwroS5lfIxU4yjya1gQnCXX14lzUrZPS9XbqulWjWLI0+K5IScEZdckBq5JnXSIJw8kmfySt6sJ+vFerc+Fq05K5s5Jn9gff4AXIyZJw==</latexit>

Figure 2: The overview of SepAttn model. SepAttn builds separate models to learn from sparse and dense features, and thenaggregates the two models’ outputs via an attention mechanism. SepAttn also employs a regularization term to enable jointtraining of the two models, and the final loss function is a combination of the listwise loss with the regularization loss. Forsimplicity, we omit the query features which are concatenated with the document dense/sparse features before fed into theDNN models.

mechanism is presented as:

usparse = tanh(Whsparse(X) + b),

udense = tanh(Whdense(X) + b),

αsparse =exp(u⊤sparsev)

exp(u⊤sparsev) + exp(u⊤densev), (3)

αdense =exp(u⊤densev)

exp(u⊤sparsev) + exp(u⊤densev), (4)

h(X) = αsparsehsparse(X) + αdensehdense(X), (5)

whereW ∈ Rn×n and b ∈ Rn are trainable weights and bias vectorrespectively;v ∈ Rn is the context vector, randomly initialized andtrained together with other parameters.

Here αsparse and αdense are attention weights assigned to thesparse feature model and the dense feature model, respectively. Asshown in Equation (5), the higher the attention weight a modelgets, the more dominant role it plays in the final result.

We explain why the above attention mechanism enables themodel to focus on important features for different queries. Thehidden representations usparse and udense can be interpreted asencoding the importance of hsparse(X) and hdense(X) respectively.For example, if dense features are noisy features (i.e., not importantfor a given query which might mislead the model) with respect to a

specific query, hdense(X) will be random across all documents, andudense will encode that this is an unimportant result. The contextvectorv can be seen as a representation of important results, andtherefore a hidden representation similar to v will be assignedhigher weights. In this way, the unimportant/noisy features will bedown-weighted by the attention mechanism, and predictions fromthe important features will dominate the final score.

In the following section, we describe the model learning withregularization.

4.3 Joint Learning with RegularizationUntil now, the dense and sparse feature models are trained withoutany interactions, i.e., the outputs of dense feature model will notdirectly influence the training of the sparse feature model, and viceversa. To enable direct interactions between the two models, weintroduce a regularization term into SepAttn, motivated by the ideaof co-training [5] that when two learning models capture differentand complementary feature sets of the same instances, the twomodels can mutually enhance each other during training.

To encourage the dense and sparse feature models to output con-sistent results given the same set of documents, a straightforwardway is to minimize the difference betweenhsparse(X) andhdense(X).However, since the two sets of scores are computed from differentfeature sets, they will be in different scales, making it ineffective

Page 6: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

to regularize directly on hsparse(X) and hdense(X). Therefore, weinstead regularize on the probability distributions after normalizinghsparse(X) and hdense(X) with Softmax, i.e.,

psparse =

[exp

(hsparse(q,d1:sparse)

)∑nj=1 exp

(hsparse(q,dj :sparse)

) · · ·exp(hsparse(q,dn:sparse))∑nj=1 exp

(hsparse(q,dj :sparse)

) ]⊤ ,

pdense =

[exp(hdense(q,d1:dense))∑nj=1 exp

(hdense(q,dj :dense)

) · · ·exp(hdense(q,dn:dense))∑nj=1 exp

(hdense(q,dj :dense)

) ]⊤ .

We then minimize the KL divergence of psparse and pdense withrespect to the final prediction pfinal:

Lsparse = DKL(pfinal ∥ psparse) =n∑i=1

pi :final log(pi :finalpi :sparse

),

Ldense = DKL(pfinal ∥ pdense) =n∑i=1

pi :final log(pi :finalpi :dense

),

where pfinal is SepAttn’s final prediction after normalizing h(X)

in Equation (5) via Softmax.The regularization loss becomes the weighted average of Lsparse

and Ldense:

Lreg = αsparseLsparse + αdenseLdense, (6)

where αsparse and αdense directly come from Equations (3) and (4),respectively, and are automatically learned during training.

The purpose of including the attention weights αsparse and αdenseagain in the regularization term is to prevent noisy/unimportantfeatures from misleading the regularization and disturbing the pre-dictions learned from important features. For example, when densefeatures are noisy features with respect to a specific query, pdense isa noisy distribution which is not meaningful to be regularized. Sincethe attention mechanism automatically down-weights unimportantfeatures, the regularization term will encourage the interactionbetween the two models only when both feature sets are important.

Finally, the SepAttn model is trained via a combination of thelistwise loss (Equation (2)) with the regularization loss:

L = Llistwise + λLreg, (7)

where λ > 0 is a hyperparameter that controls the importanceof regularization. We will study its effect in model training in theexperiment section.

5 RESULTS ON SYNTHETIC DATASETSIn this section, we describe the results of our method on syntheticdatasets described in Section 3. From Figures 1(a) and 1(b), weobserve that SepAttn achieves comparable performance on thetesting set to Sparse-only (benchmark on the first synthetic dataset)and Dense-only (benchmark on the second synthetic dataset) mod-els, respectively, which shows the robustness of our model againstnoisy/unimportant features. To understand how SepAttn is ableto focus on important features and ignore noisy ones, we visualizethe attention weights, i.e., αsparse and αdense in Equations (3) and(4) on the first and second synthetic datasets. We randomly select20 samples. It can be observed from Figure 3(a) that on the firstsynthetic dataset, SepAttn assigns almost zero attention weightsto the dense features, and effectively focuses on the sparse features.This explains why SepAttn can achieve similar performance with

the Sparse-only model. Similarly, as shown in Figure 3(b), Sep-Attn focuses mainly on the dense features for the second syntheticdataset, and does not pay attention to the sparse features.

In Figure 1(c), SepAttn performs the best on the testing set,demonstrating its effectiveness on learning from the combinationof sparse and dense features. We visualize the attention weightsin Figure 3(c), where we use 0/1 values to denote the classificationoutputs from sparse and dense feature models. For example, thefirst sample has a 0 in the sparse row, and a 1 in the dense row. Thismeans that for the first sample, the sparse featuremodel predictsy =0 and the dense feature model predicts y = 1. Recall that the thirdsynthetic dataset is generated from both sparse and dense features,i.e., only when both dense and sparse document features matchwith the query features, the ground-truth label is 1. In this case, theSepAttn model mainly focuses on the 0 predictions as shown inFigure 3(c), which explains why SepAttn works as expected—thefinal prediction is dominated by negative predictions from eitherdense or sparse features.

6 EXPERIMENTSIn this section, we begin with a description of the datasets we usein our experiments. Then, we evaluate our proposed technique andbaselines using the evaluation metrics that we will define. Finally,we discuss the hyperparameter sensitivity of our method.

6.1 DatasetDue to the private and sensitive nature of personal email data, thereis no publicly available large-scale email search dataset. Therefore,the data we use comes from the search click logs of Gmail searchengine. The training set contains around 317 million queries, andthe testing set contains around 41million queries. All queries in thetesting set are issued strictly later than all queries in the training set.Each query has six candidate documents (i.e., n = 6 in Equation (1)),one of which is clicked2. The goal is to rank the six documents toincrease the likelihood of a higher ranked document being clicked.

6.2 Model EvaluationIn this subsection, we describe the evaluation metrics. We denotethe evaluation set as Q , and the ranking position of the clickeddocument for query q as r∗q .

• Mean reciprocal rank (MRR) of the clicked document:

MRR = 1|Q |

∑q∈Q

1r∗q.

• Weighted mean reciprocal rank (WMRR) [47] of the clickeddocument:

WMRR = 1∑q∈ |Q | wq

∑q∈Q

wq

r∗q,

wherewq is the bias correction weight, and is inversely pro-portional to the probability of observing a click at the clicked

2These six candidate documents are presented in the dropdown menu of the Gmailsearch box while users type their queries but before they click the search bottom.Whenusers find the target email, the system will direct them to the exact email, generatingexactly one click.

Page 7: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

Table 2: Evaluations on testing set with percentage of improvements over the best baseline. Metrics with an upper arrow (↑)indicate higher is better; metrics with a down arrow (↓) indicate lower is better. Improvements with an asterisk (*) denotesstatistical significance relative to the best performing baseline (Concatenation), according to the two-tailed paired t-test.

Methods MRR (↑) WMRR (↑) ARP (↓) WARP (↓) DCG (↑)Dense-only method 0.650 0.563 2.143 2.474 0.759Sparse-only method 0.665 0.578 2.067 2.412 0.767

Concatenation method 0.682 0.589 2.002 2.366 0.770SepAttn 0.686 0.595 1.992 2.350 0.781

∆(%) +0.59∗ +1.02∗ −0.50∗ −0.68∗ +1.43∗

position of q. We set those weights using result randomiza-tion, as described in [47]. This serves as our main evaluationmetric.

• Average relevance value position (ARP) [52]:

ARP = 1|Q |

∑q∈Q

r∗q .

• Weighted average relevance value position (WARP), whichis the weighted version of the above ARP metric:

WARP = 1∑q∈ |Q | wq

∑q∈Q

wqr∗q ,

wherewq is the bias correction weight as in WMRR.• Discounted Cumulative Gain (DCG) [22]:

DCG = 1|Q |

∑q∈Q

1log2(1 + r∗q )

.

6.3 Results and DiscussionsIn this section, we compare our proposed approach, i.e., SepAttnmodel with the following baseline models.

• Dense-only method: The DNN model that only uses densedocument features.

• Sparse-onlymethod: TheDNNmodel that only uses sparsedocument features.

• Concatenation method: The DNN model that learns fromconcatenated dense and embedded sparse features.

The above models are implemented using TF-ranking [36] which isa scalable open-source learning-to-rank Tensorflow library [1]. Dueto the data anonymization process (i.e., bag of frequent n-grams),the sequential information in the raw email documents is lost, andthus sequence-aware models like CNNs [38] and RNNs [35] cannotbe applied to our case. The configuration of the DNN model isdescribed below: The DNN model has three hidden layers whosedimensions are 256, 128, and 64, respectively. The n-grams andcharacter n-grams are first passed through an embedding layerof size 20 followed by an average pooling layer before fed intothe DNN. We use Adagrad [17] with 0.1 learning rate and batchsize 100 to train the model. The above hyperparameters are theoptimal settings of the Concatenation method, obtained fromtuning the model in the following hyperparameter ranges: Layerdimension in {64, 128, 256, 512, 768, 1024}; embedding dimension

in {20, 30, 40, 50, 100}; learning rate in {0.05, 0.08, 0.1, 0.15, 0.3};optimization algorithm in {Adam [25],Adagrad [17]}.

We set the regularization weight to be 1 in our SepAttn model,and will study model sensitivity to this hyperparameter in Sec-tion 6.4.

The results are presented in Table 2. From the table, we can seethat across all metrics, the concatenation of dense features withembedded sparse features (i.e., Concatenation method) consistentlyleads to better results than both Sparse-only and Dense-only mod-els. An interesting finding is that even without any sparse features(Dense-only), the model achieves reasonably good performance,probably because emails that are more recent or with more attach-ments are more likely to be the user-desired ones.

Moreover, our SepAttn method achieves statistically significantimprovements (using the two-tailed paired t-test with 99% confi-dence level) over the Concatenation method.

We note that an improvement of 1% is considered to be highlysignificant for our email search ranking system. In Table 2, wesee that the improvement of WMRR metric for SepAttn methodis 1.02% over the best performing baseline model (Concatenationmethod).

The results demonstrate that dense features indeed play a criticalrole in email search tasks, and they need to be incorporated into themodel effectively, i.e., simply concatenating them with embeddedsparse features is suboptimal; SepAttn provides a more effectiveway of learning from both sparse and dense features.

6.4 Sensitivity AnalysisIn this section, we discuss the sensitivity of SepAttn method tothe regularization parameter λ in Equation (7). We vary the regu-larization parameter λ in the range [0, 2.5] and report SepAttn’sperformance (with respect to WMRR metric) in Figure 4.

We observe that when λ > 0, the performance is significantlybetter than that when λ = 0. This shows that jointly training thesparse and dense feature models by encouraging them to makeconsistent prediction is indeed beneficial for improving the rankingperformance. When λ continues to grow larger, the performanceof the model becomes stable, which is a favorable property becauseλ can be safely set within a relatively wide range of values.

Page 8: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

Sparse

Dense

(a) First synthetic dataset

Sparse

Dense

(b) Second synthetic dataset

Sparse

Dense

0 0 1 1 0 0 1 0 0 1 0 0 1 0 0 1 0 0 1 1

1 1 1 0 1 1 0 1 1 1 1 1 0 1 1 0 0 1 1 1

(c) Third synthetic dataset

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20Samples

Sparse

Dense

(d) Real email search dataset

0.0

0.2

0.4

0.6

0.8

1.0

Figure 3: Attention weights visualization (αsparse and αdense from Equations (3) and (4)) over 20 randomly selected samples onfour datasets: (a) First synthetic dataset; (b) Second synthetic dataset; (c) Third synthetic dataset; (d) Real email search dataset.The 0/1 values on (c) indicate classification outputs from sparse and dense feature models.

0.0 0.5 1.0 1.5 2.0 2.5Regularization Weight

0.591

0.592

0.593

0.594

0.595

WM

RR

Figure 4: Sensitivity of regularization parameter λ onWMRR metric.

6.5 Attention Weights VisualizationTo understand how SepAttnworks for different email search queries,we visualize the attention weights (αsparse and αdense in Equa-tions (3) and (4)) of SepAttn model on our email search datasetin Figure 3(d). We randomly select 20 queries. We observe the fol-lowing: (1) Both sparse and dense features are important for emailsearch tasks, and sparse features generally obtain higher weightsthan dense features, which corresponds to our intuition that textual

features are more important than numerical features in email search.(2) Different queries have different attention weight distributionon sparse and dense features, which verifies the necessity to buildan attentive ranking model that learns to focus on different fea-ture sets according to different queries types. For example, the 2nd,13th, 17th and 19th samples in Figure 3(d) have higher attentionweights on dense features, while for other queries, sparse featuresare more important. SepAttn automatically learns to assign appro-priate weights to dense and sparse features in order to achieve thebest performance.

7 CONCLUSIONS AND FUTUREWORKIn this paper, we studied how to improve email search ranking byeffectively learning from a combination of dense and sparse docu-ment features using DNNs. We first showed on a set of syntheticdatasets that simply concatenating dense features with embeddedsparse features leads to suboptimal performance of DNN rankingmodels, mainly because these features are from different featurespace. Motivated by this drawback, we proposed the SepAttnmodelwhich automatically learns to focus on important features for differ-ent queries through the attention mechanism. We evaluated ourSepAttn model on a large-scale email search dataset and showedthat it significantly outperforms the baseline approach—direct con-catenation of dense features with sparse features.

In the future, we would like to extend our study to other IR taskssuch as web search and recommendation. Our proposed methodcan be easily generalized to other scenarios where dense numerical

Page 9: Separate and Attend in Personal Email Search...in email search system is to balance the importance of content-based relevance and other features, e.g. freshness. Carmel et al. [12]

features and sparse categorical features both play a role. Further-more, the attention mechanism at the prediction level introduced inthis paper may facilitate further research on ensembling multiplelearning models.

ACKNOWLEDGMENTSWe thank Zhongliang Li for his help on the model implementa-tion. We thank anonymous reviewers for valuable and insightfulfeedback.

REFERENCES[1] Martin Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, and Jeffrey

Dean. 2016. Tensorflow: a system for large-scale machine learning. OperatingSystems Design and Implementation 16 (2016), 265–283.

[2] Qingyao Ai, Susan T Dumais, Nick Craswell, and Dan Liebling. 2017. Character-izing email search using large-scale behavioral logs and surveys. In WWW.

[3] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural MachineTranslation by Jointly Learning to Align and Translate. In ICLR.

[4] Michael Bendersky, Xuanhui Wang, Donald Metzler, and Marc Najork. 2017.Learning from user interactions in personal search via attribute parameterization.In WSDM.

[5] Avrim Blum and Tom M. Mitchell. 1998. Combining Labeled and Unlabeled Datawith Co-Training. In COLT.

[6] Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2016. En-riching Word Vectors with Subword Information. Transactions of the Associationfor Computational Linguistics 5 (2016), 135–146.

[7] Alexey Borisov, Ilya Markov, Maarten de Rijke, and Pavel Serdyukov. 2016. ANeural Click Model for Web Search. In WWW.

[8] Christopher J. C. Burges. 2010. From RankNet to LambdaRank to LambdaMART:An Overview.

[9] Christopher J. C. Burges, Tal Shaked, Erin Renshaw, Ari Lazier, Matt Deeds, NicoleHamilton, and Gregory N. Hullender. 2005. Learning to rank using gradientdescent. In ICML.

[10] Yunbo Cao, Jun Xu, T. M. Liu, Hang Li, Yalou Huang, and Hsiao-Wuen Hon. 2006.Adapting ranking SVM to document retrieval. In SIGIR.

[11] Zhe Cao, Tao Qin, T. M. Liu, Ming-Feng Tsai, and Hang Li. 2007. Learning torank: from pairwise approach to listwise approach. In ICML.

[12] David Carmel, Guy Halawi, Liane Lewin-Eytan, Yoelle Maarek, and Ariel Raviv.2015. Rank by Time or Relevance?: Revisiting Email Search. In CIKM.

[13] David Carmel, Liane Lewin-Eytan, Alex Libov, Yoelle Maarek, and Ariel Raviv.2017. Promoting relevant results in time-ranked mail search. In WWW.

[14] Nick Craswell, Hugo Zaragoza, and Stephen Robertson. 2005. Microsoft cam-bridge at TREC-14: Enterprise track. In TREC.

[15] Mostafa Dehghani, Hamed Zamani, Aliaksei Severyn, Jaap Kamps, andWilliam Bruce Croft. 2017. Neural Ranking Models with Weak Supervision.In SIGIR.

[16] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT:Pre-training of Deep Bidirectional Transformers for Language Understanding. InHLT-NAACL.

[17] John C. Duchi, Elad Hazan, and Yoram Singer. 2010. Adaptive SubgradientMethods for Online Learning and Stochastic Optimization. J. Mach. Learn. Res.12 (2010), 2121–2159.

[18] Jerome H. Friedman. 2001. Greedy function approximation: A gradient boostingmachine. Annals of statistics (2001).

[19] Jiafeng Guo, Yixing Fan, Qingyao Ai, and W Bruce Croft. 2016. A deep RelevanceMatching Model For Ad-hoc Retrieval. In CIKM.

[20] Karl Moritz Hermann, Tomás Kociský, Edward Grefenstette, Lasse Espeholt, WillKay, Mustafa Suleyman, and Phil Blunsom. 2015. Teaching Machines to Readand Comprehend. In NIPS.

[21] Kurt Hornik. 1991. Approximation capabilities of multilayer feedforward net-works. Neural Networks 4 (1991), 251–257.

[22] Kalervo Järvelin and Jaana Kekäläinen. 2002. Cumulated gain-based evaluationof IR techniques. ACM Trans. Inf. Syst. 20 (2002), 422–446.

[23] Thorsten Joachims. 2002. Optimizing search engines using clickthrough data. InKDD.

[24] Thorsten Joachims. 2006. Training linear SVMs in linear time. In KDD.[25] Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Opti-

mization. In ICLR.[26] Ankit Kumar, Ozan Irsoy, Jonathan Su, James Bradbury, Robert English, Brian

Pierce, Peter Ondruska, Ishaan Gulrajani, and Richard Socher. 2015. Ask MeAnything: Dynamic Memory Networks for Natural Language Processing. InICML.

[27] Saar Kuzi, David Carmel, Alex Libov, and Ariel Raviv. 2017. Query expansion foremail search. In SIGIR.

[28] Yann LeCum, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep Learning. Nature521 (2015), 436–444.

[29] Cheng Li, Mingyang Zhang, Michael Bendersky, Hongbo Deng, Donald Metzler,andMarc Najork. 2019. Multi-view Embedding-based Synonyms for Email Search.In SIGIR.

[30] Yang Li, Lukasz Kaiser, Samy Bengio, and Si Si. 2019. Area attention. In ICML.[31] Craig Macdonald and Iadh Ounis. 2006. Combining fields in known-item email

search. In SIGIR.[32] Tomas Mikolov, Ilya Sutskever, Kai Chen, Gregory S. Corrado, and Jeffrey Dean.

2013. Distributed Representations of Words and Phrases and their Composition-ality. In NIPS.

[33] Bhaskar Mitra and Nick Craswell. 2017. Neural Models for Information Retrieval.ArXiv abs/1705.01509 (2017).

[34] Paul Ogilvie and Jamie Callan. 2005. Experiments with Language Models forKnown-Item Finding of E-mail Messages. In TREC.

[35] Liang Pang, Yanyan Lan, Jiafeng Guo, Jun Xu, Jingfang Xu, and Xueqi Cheng.2017. DeepRank: A NewDeep Architecture for Relevance Ranking in InformationRetrieval. In CIKM.

[36] Rama Kumar Pasumarthi, Sebastian Bruch, Xuanhui Wang, Carlos Pignataro,Michael Bendersky, Marc Najork, Jan Pfeifer, Nadav Golbandi, Rohan Anil, andStephan Wolf. 2019. TF-Ranking: Scalable TensorFlow Library for Learning-to-Rank. In KDD.

[37] Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. Glove:Global Vectors for Word Representation. In EMNLP.

[38] Aliaksei Severyn and Alessandro Moschitti. 2015. Learning to Rank Short TextPairs with Convolutional Deep Neural Networks. In SIGIR.

[39] Jiaming Shen, Maryam Karimzadehgan, Michael Bendersky, Zhen Qin, and Don-ald Metzler. 2018. Multi-Task Learning for Email Search Ranking With AuxiliaryQuery Clustering. In CIKM.

[40] Ian Soboroff, Arjen P de Vries, and Nick Craswell. 2005. Overview of the TREC2005 Enterprise Track. In TREC.

[41] Ian Soboroff, Arjen P de Vries, and Nick Craswell. 2006. Overview of the TREC2006 Enterprise Track. In TREC.

[42] Anant Subramanian, Danish Pruthi, Harsh Jhamtani, Taylor Berg-Kirkpatrick,and Eduard H. Hovy. 2017. SPINE: SParse Interpretable Neural Embeddings. InAAAI.

[43] Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston, and Rob Fergus. 2015. End-To-End Memory Networks. In NIPS.

[44] Latanya Sweeney. 2002. k-Anonymity: A Model for Protecting Privacy. Interna-tional Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 10 (2002),557–570.

[45] Brandon Tran, Maryam Karimzadehgan, Rama Kumar Pasumarthi, Michael Ben-dersky, and Donald Metzler. 2019. Domain Adaptation for Enterprise EmailSearch. In SIGIR.

[46] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Lawrence Jones,Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention Is All YouNeed. In NIPS.

[47] Xuanhui Wang, Michael Bendersky, Donald Metzler, and Marc Najork. 2016.Learning to Rank with Selection Bias in Personal Search. In SIGIR.

[48] Xuanhui Wang, Nadav Golbandi, Michael Bendersky, Donald Metzler, and MarcNajork. 2018. Position bias estimation for unbiased learning to rank in personalsearch. In WSDM.

[49] Kelvin Xu, Jimmy Ba, Jamie Ryan Kiros, Kyunghyun Cho, Aaron C. Courville,Ruslan Salakhutdinov, Richard S. Zemel, and Yoshua Bengio. 2015. Show, Attendand Tell: Neural Image Caption Generation with Visual Attention. In ICML.

[50] Zichao Yang, Diyi Yang, Chris Dyer, Xiaodong He, Alexander J. Smola, and Ed-uardH. Hovy. 2016. Hierarchical AttentionNetworks for Document Classification.In HLT-NAACL.

[51] Hamed Zamani, Michael Bendersky, Xuanhui Wang, and Mingyang Zhang. 2017.Situational Context for Ranking in Personal Search. In WWW.

[52] Mu Zhu. 2004. Recall, Precision and Average Precision.