https://arxiv.org/abs/1701.07795v1 Match-Tensor: a Deep Relevance Model for Search

The application of Deep Neural Networks for ranking in search engines may obviate the need for the extensive feature engineering common to current learning-to-rank methods. However, we show that combining simple relevance matching features like BM25 with existing Deep Neural Net models often substantially improves the accuracy of these models, indicating that they do not capture essential local relevance matching signals. We describe a novel deep Recurrent Neural Net-based model that we call Match-Tensor. The architecture of the Match-Tensor model simultaneously accounts for both local relevance matching and global topicality signals allowing for a rich interplay between them when computing the relevance of a document to a query. On a large held-out test set consisting of social media documents, we demonstrate not only that Match-Tensor outperforms BM25 and other classes of DNNs but also that it largely subsumes signals present in these models.

https://arxiv.org/abs/1707.00189v1 An Approach for Weakly-Supervised Deep Information Retrieval

We present an approach for generating weak supervision training data for use in a neural IR model. Specifically, we use a news corpus with article headlines acting as pseudo-queries and article content as pseudo-documents, and we propose a measure of interaction similarity to filter these pseudo-documents

https://arxiv.org/abs/1705.03556v2 Relevance-based Word Embedding

The primary objective in various IR tasks is to capture relevance instead of term proximity, syntactic, or even semantic similarity. This is the motivation for developing unsupervised relevance-based word embedding models that learn word representations based on query-document relevance information. In this paper, we propose two learning models with different objective functions; one learns a relevance distribution over the vocabulary set for each query, and the other classifies each term as belonging to the relevant or non-relevant class for each query.

https://arxiv.org/abs/1706.04922v2 DSRIM: A Deep Neural Information Retrieval Model Enhanced by a Knowledge Resource Driven Representation of Documents

Guided by the intuition that the relational semantics might improve the effectiveness of deep neural approaches, we propose the Deep Semantic Resource Inference Model (DSRIM) that relies on: 1) a representation of raw-data that models the relational semantics of text by jointly considering objects and relations expressed in a knowledge resource, and 2) an end-to-end neural architecture that learns the query-document relevance by leveraging the distributional and relational semantics of documents and queries. The experimental evaluation carried out on two TREC datasets from TREC Terabyte and TREC CDS tracks relying respectively on WordNet and MeSH resources, indicates that our model outperforms state-of-the-art semantic and deep neural IR models.

https://arxiv.org/pdf/1701.07810.pdf Learning to Effectively Select Topics For Information Retrieval Test Collections

https://arxiv.org/pdf/1702.02535.pdf Exploiting Domain Knowledge via Grouped Weight Sharing with Application to Text Categorization

we treat weight sharing as a flexible mechanism for incorporating prior knowledge into neural models. We show that this approach consistently yields improved performance on classifi- cation tasks compared to baseline strategies that do not exploit weight sharing.

https://arxiv.org/pdf/1708.01713.pdf Automatic Question-Answering Using A Deep Similarity Neural Network

https://arxiv.org/pdf/1708.02702v1.pdf Neural Vector Spaces for Unsupervised Information Retrieval

https://arxiv.org/abs/1704.04572 Task-Oriented Query Reformulation with Reinforcement Learning

In this work, we introduce a query reformulation system based on a neural network that rewrites a query to maximize the number of relevant documents returned. We train this neural network with reinforcement learning. The actions correspond to selecting terms to build a reformulated query, and the reward is the document recall. We evaluate our approach on three datasets against strong baselines and show a relative improvement of 5-20% in terms of recall. https://github.com/nyu-dl/QueryReformulator

https://arxiv.org/abs/1106.1925 Ranking via Sinkhorn Propagation

https://arxiv.org/pdf/1711.08726.pdf Modelling Domain Relationships for Transfer Learning on Retrieval-based Question Answering Systems in E-commerce

https://arxiv.org/pdf/1705.01509.pdf Neural Models for Information Retrieval

https://arxiv.org/abs/1807.02299 On the Equilibrium of Query Reformulation and Document Retrieval

https://github.com/hamed-zamani/snrm Standalone Neural Ranking Model (SNRM)

https://github.com/williamleif/graphqembed Embedding Logical Queries on Knowledge Graphs https://arxiv.org/pdf/1806.01445.pdf

Here we introduce a framework to efficiently make predictions about conjunctive logical queries—a flexible but tractable subset of first-order logic—on incomplete knowledge graphs. In our approach, we embed graph nodes in a low-dimensional space and represent logical operators as learned geometric operations (e.g., translation, rotation) in this embedding space. By performing logical operations within a low-dimensional embedding space, our approach achieves a time complexity that is linear in the number of query variables, compared to the exponential complexity required by a naive enumeration-based approach. We demonstrate the utility of this framework in two application studies on real-world datasets with millions of relations: predicting logical relationships in a network of drug-gene-disease interactions and in a graph-based representation of social interactions derived from a popular web forum.

https://arxiv.org/abs/1809.05679v1 Graph Convolutional Networks for Text Classification

We build a single text graph for a corpus based on word co-occurrence and document word relations, then learn a Text Graph Convolutional Network (Text GCN) for the corpus. Our Text GCN is initialized with one-hot representation for word and document, it then jointly learns the embeddings for both words and documents, as supervised by the known class labels for documents.

https://arxiv.org/abs/1803.01707v2 Neural Architectures for Open-Type Relation Argument Extraction

https://arxiv.org/abs/1810.09591 Applying Deep Learning To Airbnb Search

The application to search ranking is one of the biggest machine learning success stories at Airbnb. Much of the initial gains were driven by a gradient boosted decision tree model. The gains, however, plateaued over time. This paper discusses the work done in applying neural networks in an attempt to break out of that plateau. We present our perspective not with the intention of pushing the frontier of new modeling techniques. Instead, ours is a story of the elements we found useful in applying neural networks to a real life product. Deep learning was steep learning for us. To other teams embarking on similar journeys, we hope an account of our struggles and triumphs will provide some useful pointers. Bon voyage!

https://ai.googleblog.com/2018/12/tf-ranking-scalable-tensorflow-library.html