{"id":231283,"date":"2016-05-12T11:52:39","date_gmt":"2016-05-12T18:52:39","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&p=231283"},"modified":"2025-08-06T12:01:09","modified_gmt":"2025-08-06T19:01:09","slug":"neuir2016","status":"publish","type":"msr-event","link":"https:\/\/www.microsoft.com\/en-us\/research\/event\/neuir2016\/","title":{"rendered":"Neu-IR: The SIGIR 2016 Workshop on Neural Information Retrieval"},"content":{"rendered":"\n\n
<<\u00a0Final Workshop Report (opens in new tab)<\/span><\/a>\u00a0>><\/strong><\/p>\n Submission Deadline<\/strong>: May 16 The first international Neu-IR (pronounced “new<\/em> IR”) workshop on neural information retrieval will be hosted at SIGIR 2016 (opens in new tab)<\/span><\/a>in Pisa, Tuscany, Italy on 21 July, 2016.Opens in a new tab<\/span><\/p>\n (The final report on the workshop is available here (opens in new tab)<\/span><\/a>.)<\/p>\n In recent years, deep neural networks have yielded significant performance improvements on speech recognition and computer vision tasks, as well as led to exciting breakthroughs in novel application areas such as automatic voice translation, image captioning, and conversational agents. Despite demonstrating good performance on natural language processing (NLP) tasks, the performance of deep neural networks on IR tasks has had relatively less scrutiny.<\/p>\n The lack of many positive results in the area of information retrieval is partially due to the fact that IR tasks such as ranking are fundamentally different from NLP tasks, but also because the IR and neural network communities are only beginning to focus on the application of these techniques to core information retrieval problems. Given that deep learning has made such a big impact, first on speech processing and computer vision and now, increasingly, also on computational linguistics, it seems clear that deep learning will have a major impact on information retrieval and that this is an ideal time for a workshop in this area. Our focus is on the applicability of deep neural networks to information retrieval: demonstrating performance improvements on public or private information retrieval datasets, identifying key modelling challenges and best practices, and thinking about what insights deep neural network architectures give us about information retrieval problems.<\/p>\n Neu-IR 2016 <\/strong>will be a highly interactive full day workshop that will provide a forum for academic and industrial researchers working at the intersection of IR and neural networks. The purpose is to provide an opportunity for people to present new work and early results, compare notes on neural network toolkits, share best practices, and discuss the main challenges facing this line of research.<\/p>\n Please use the tabs above to navigate to see the program, the accepted papers and other details of this workshop.<\/p>\n Opens in a new tab<\/span><\/p>\n Neu-IR will be a highly interactive full day workshop, featuring a mix of presentation and interaction formats. The full schedule is presented below.<\/p>\n Morning Session I<\/strong> Welcome and opening announcements [slides (opens in new tab)<\/span><\/a>] Keynote: Recurrent Networks and Beyond [slides (opens in new tab)<\/span><\/a>] Paper: Query Expansion with Locally-Trained Word Embeddings [slides (opens in new tab)<\/span><\/a>] Paper: Uncertainty in Neural Network Word Embedding Exploration of Potential Threshold [slides (opens in new tab)<\/span><\/a>] Coffee Break<\/strong> Morning Session II<\/strong> Lessons from the Trenches [slides (opens in new tab)<\/span><\/a>] Poster presentations Lunch Break<\/strong> Afternoon Session I<\/strong> Keynote: Does IR Need Deep Learning? [slides (opens in new tab)<\/span><\/a>] Paper: Modelling User Preferences using Word Embeddings for Context-Aware Venue Recommendation [slides (opens in new tab)<\/span><\/a>] Paper: A Study of MatchPyramid Models on Ad-hoc Retrieval [slides (opens in new tab)<\/span><\/a>] Paper: Emulating Human Conversations using Convolutional Neural Network-based IR [slides (opens in new tab)<\/span><\/a>] Coffee Break<\/strong> Afternoon Session II<\/strong> Breakout session Breakout session retrospective Concluding remarks Opens in a new tab<\/span><\/p>\n Tomas Mikolov, Facebook AI Research<\/p>\n Bio: Tomas Mikolov (opens in new tab)<\/span><\/a>\u00a0is a research scientist at Facebook AI Research since May 2014. Previously\u00a0he has been a member of the Google Brain team, where\u00a0he developed and implemented efficient algorithms for computing distributed representations of words (word2vec project). He obtained his PhD from Brno University of Technology (Czech Republic) for\u00a0his work on recurrent neural network based language models (RNNLM).\u00a0His long term research goal is to develop intelligent machines capable of learning and communication with people using natural language.<\/p>\n <\/p>\n Hang Li, Huawei Technologies<\/p>\n In this talk, I will argue that, if we take a broad view on IR, then we arrive at a conclusion that deep learning can indeed greatly boost IR. Actually it has been observed that deep learning can make great improvements on some hard problems in IR such as question answering from knowledge base, image retrieval, etc; on the other hand, for some traditional IR tasks, in some sense easy tasks, such as document retrieval, the improvements might not be so notable. I will introduce some of the work on deep learning for IR conducted at Huawei Noah\u2019s Ark Lab, to support my claim. I will also make discussions on the strength and limitation of deep learning, IR problems on which deep learning can potentially make significant contributions, as well as future directions of research on IR.<\/p>\n Bio: Hang Li (opens in new tab)<\/span><\/a> is director of the Noah\u2019s Ark Lab of Huawei Technologies, adjunct professors of Peking University and Nanjing University. He is ACM Distinguished Scientist. His research areas include information retrieval, natural language processing, statistical machine learning, and data mining. Hang graduated from Kyoto University in 1988 and earned his PhD from the University of Tokyo in 1998. He worked at the NEC lab as researcher during 1991 and 2001, and Microsoft Research Asia as senior researcher and research manager during 2001 and 2012. He joined Huawei Technologies in 2012. Hang has published three technical books, and more than 120 technical papers at top international conferences including SIGIR, WWW, WSDM, ACL, EMNLP, ICML, NIPS, SIGKDD, AAAI, IJCAI, and top international journals including CL, NLE, JMLR, TOIS, IRJ, IPM, TKDE, TWEB, TIST. He and his colleagues\u2019 papers received the SIGKDD\u201908 best application paper award, the SIGIR\u201908 best student paper award, the ACL\u201912 best student paper award. Hang worked on the development of several products such as Microsoft SQL Server 2005, Office 2007, Live Search 2008, Bing 2009, Office 2010, Bing 2010, Office 2012, Huawei Smartphones 2014. He has 42 granted US patents. Hang is also very active in the research communities and has served or is serving top international conferences as PC chair, Senior PC member, or PC member, including SIGIR, WWW, WSDM, ACL, NACL, EMNLP, NIPS, SIGKDD, ICDM, IJCAI, ACML, and top international journals as associate editor or editorial board member, including CL, IRJ, TIST, JASIST, JCST.<\/p>\n Opens in a new tab<\/span><\/p>\n We had 27 submissions (excluding three incomplete submissions). Every paper was reviewed by at least two members of the program committee and finally 19 submission were accepted\u00a0(acceptance rate of 73%). Among the accepted papers, there were a few popular themes. 8 papers were related to learning and applications of word embeddings. 10 papers focused on applications of deep neural networks for different IR tasks. The accepted papers also covered a broad range\u00a0of tasks, including question\/answering, proactive IR, knowledge-based IR,\u00a0conversational models and\u00a0text-to-image, but document ranking was a popular choice with 7 papers using it as the evaluation task.\u00a0The word cloud summary (generated using http:\/\/www.wordle.net (opens in new tab)<\/span><\/a>)\u00a0of the abstracts of the accepted papers highlights additional themes across all the submissions.<\/p>\n Geographically, the accepted papers (based on the first author) ranged from\u00a09 countries and 3 continents (FR: 4, IN: 4, CN: 2, DK: 2, UK: 2, US: 2, AT: 1, FI: 1 and IT: 1). Based on the first author’s affiliation, 2 of the accepted papers came from the industry and the rest from academia.<\/p>\n <\/p>\n The full list of accepted papers is below:<\/p>\n An empirical study on large scale text classification with skip-gram embeddings\u00a0 Deep Feature Fusion Network for Answer Quality Prediction in Community Question Answering Selective Term Proximity Scoring Via BP-ANN\u00a0 Adaptability of Neural Networks on Varying Granularity IR Tasks\u00a0 Emulating Human Conversations using Convolutional Neural Network-based IR\u00a0 A Study of MatchPyramid Models on Ad-hoc Retrieval\u00a0 Learning text representation using recurrent convolutional neural network with highway layers\u00a0 Toward Word Embedding for Personalized Information Retrieval\u00a0 Toward a Deep Neural Approach for Knowledge-Based IR\u00a0 Query Expansion with Locally-Trained Word Embeddings\u00a0 LSTM-Based Predictions for Proactive Information Retrieval\u00a0 Picture It In Your Mind: Generating High Level Visual Representations From Textual Descriptions Uncertainty in Neural Network Word Embedding Exploration of Potential Threshold\u00a0 Deep Learning Relevance: Creating Relevant Information (as Opposed to Retrieving it) Learning Dynamic Classes of Events using Stacked Multilayer Perceptron Networks\u00a0 Representing Documents and Queries as Sets of Word Embedded Vectors for Information Retrieval\u00a0 Using Word Embeddings for Automatic Query Expansion\u00a0 Modelling User Preferences using Word Embeddings for Context-Aware Venue Recommendation\u00a0 Using Word Embeddings in Twitter Election Classification\u00a0 Opens in a new tab<\/span><\/p>\n The Lessons from the Trenches will be a series of “lightning talks” by\u00a0researchers\u00a0who are actively working in the intersection of information retrieval and neural networks who\u00a0want to share their personal insights and learning with the broader community. In particular, we are hoping to hear about,<\/p>\n <\/p>\n The following people have signed-up to present at this session.<\/p>\n Opens in a new tab<\/span><\/p>\n We solicit submission of papers of two to six pages (excluding references), representing reports of original research, preliminary research results, proposals for new work, descriptions of neural network based toolkits tailored for IR, and position papers. Papers presented at the workshop will be required to be uploaded to arXiv.org but will be considered non-archival<\/strong>, and may be submitted elsewhere (modified or not), although the workshop site will maintain a link to the arXiv versions. This makes the workshop a forum for the presentation and discussion of current work, without preventing the work from being published elsewhere.<\/p>\n We are interested in submissions relevant to the following main themes:<\/p>\n All papers will be peer reviewed (single-blind) by the program committee and judged by their relevance to the workshop, especially to the main themes identified above, and their potential to generate discussion. All submissions must be formatted according to the ACM SIG proceedings template. Please note that at least one of the authors of each accepted paper must register for the workshop and present the paper in-person.<\/p>\n Submission url: https:\/\/easychair.org\/conferences\/?conf=neuir2016 (opens in new tab)<\/span><\/a>Opens in a new tab<\/span><\/p>\n Organizers<\/strong><\/p>\n Nick Craswell (opens in new tab)<\/span><\/a>, Microsoft, Bellevue, US <\/p>\n Program Committee<\/strong><\/p>\n Carsten Eickhoff (opens in new tab)<\/span><\/a>, ETH Zurich The first international Neu-IR (pronounced “new IR”) workshop on neural information retrieval will be hosted at SIGIR 2016 in Pisa, Tuscany, Italy on 21 July, 2016.<\/p>\n","protected":false},"featured_media":241994,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_startdate":"2016-07-21","msr_enddate":"2016-07-21","msr_location":"Pisa, Italy","msr_expirationdate":"","msr_event_recording_link":"","msr_event_link":"","msr_event_link_redirect":false,"msr_event_time":"","msr_hide_region":false,"msr_private_event":false,"msr_hide_image_in_river":0,"footnotes":""},"research-area":[13555],"msr-region":[239178],"msr-event-type":[197941],"msr-video-type":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[],"class_list":["post-231283","msr-event","type-msr-event","status-publish","has-post-thumbnail","hentry","msr-research-area-search-information-retrieval","msr-region-europe","msr-event-type-conferences","msr-locale-en_us"],"msr_about":"\n\n <<\u00a0Final Workshop Report (opens in new tab)<\/span><\/a>\u00a0>><\/strong><\/p>\n Submission Deadline<\/strong>: May 16 The first international Neu-IR (pronounced “new<\/em> IR”) workshop on neural information retrieval will be hosted at SIGIR 2016 (opens in new tab)<\/span><\/a>in Pisa, Tuscany, Italy on 21 July, 2016.Opens in a new tab<\/span><\/p>\n (The final report on the workshop is available here (opens in new tab)<\/span><\/a>.)<\/p>\n
\nAcceptance Notifications<\/strong>: June 6
\nCamera-ready Deadline<\/strong>: June 17
\nWorkshop<\/strong>: July 21<\/p>\n
\n09:00 \u2013 10:30<\/span><\/p>\n
\nBhaskar Mitra
\n15 mins<\/span><\/p>\n
\nTomas Mikolov, Facebook AI Research
\n45 mins<\/span><\/p>\n
\nFernando Diaz, Bhaskar Mitra and Nick Craswell
\n15 mins<\/span><\/p>\n
\nNavid Rekabsaz, Mihai Lupu and Allan Hanbury
\n15 mins<\/span><\/p>\n
\n10:30 \u2013 11:00<\/span><\/p>\n
\n11:00 \u2013 12:30<\/span><\/p>\n
\n45 mins<\/span><\/p>\n
\n45 mins<\/span><\/p>\n
\n12:30 \u2013 14:00<\/span><\/p>\n
\n14:00 \u2013 15:30<\/span><\/p>\n
\nHang Li, Huawei Technologies
\n45 mins<\/span><\/p>\n
\nJarana Manotumruksa, Craig Macdonald and Iadh Ounis
\n15 mins<\/span><\/p>\n
\nLiang Pang, Yanyan Lan, Jiafeng Guo, Jun Xu and Xueqi Cheng
\n15 mins<\/span><\/p>\n
\nAbhay Prakash, Chris Brockett and Puneet Agrawal
\n15 mins<\/span><\/p>\n
\n15:30 \u2013 16:00<\/span><\/p>\n
\n16:00 \u2013 17:45<\/span><\/p>\n
\n45 mins<\/span><\/p>\n
\n45 mins<\/span><\/p>\n
\n15 mins<\/span><\/p>\nRecurrent Networks and Beyond<\/h3>\n
(opens in new tab)<\/span><\/a>Abstract: In this talk, I will give a brief overview of recurrent networks and their applications. I will then present several extensions that aim to help these powerful models to learn more patterns from training data. This will include a simple modification of the architecture that allows to capture longer context information, and an architecture that allows to learn complex algorithmic patterns. The talk will be concluded with a discussion of a long term research plan on how to advance machine learning techniques towards development of artificial intelligence.<\/p>\nDoes IR Need Deep Learning?<\/h3>\n
(opens in new tab)<\/span><\/a>Abstract: In recent years, deep learning has become the key technology of state-of-the-art systems in many areas of computer science, such as computer vision, speech processing, and natural language processing. A question naturally arises, that is, can deep learning also bring breakthrough into IR (information retrieval)? In fact, there has been a large amount of effort made to address the question and significant progress has been achieved. Yet there is still doubt about whether it is the case.<\/p>\n
<\/p>\n
(opens in new tab)<\/span><\/a>
\nGeorgios Balikas and Massih-Reza Amini<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nSai Praneeth Suggu, Kushwanth N. Goutham T, Manoj K. Chinnakotla and Manish Shrivastava<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nJu Yang, Rebecca Stones, Gang Wang and Xiaoguang Liu<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nDaniel Cohen, Qingyao Ai and W. Bruce Croft<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nAbhay Prakash, Chris Brockett and Puneet Agrawal<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nLiang Pang, Yanyan Lan, Jiafeng Guo, Jun Xu and Xueqi Cheng<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nYing Wen, Weinan Zhang, Rui Luo and Jun Wang<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nNawal Ould Amer, Philippe Mulhem and Mathias G\u00e9ry<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nGia-Hung Nguyen, Lynda Tamine, Laure Soulier and Nathalie Bricon-Souf<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nFernando Diaz, Bhaskar Mitra and Nick Craswell<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nPetri Luukkonen, Markus Koskela and Patrik Flor\u00e9en<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nFabio Carrara, Andrea Esuli, Tiziano Fagni, Fabrizio Falchi and Alejandro Moreo Fern\u00e1ndez<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nNavid Rekabsaz, Mihai Lupu and Allan Hanbury<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nChristina Lioma, Birger Larsen, Casper Petersen and Jakob Grue Simonsen<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nNattiya Kanhabua, Huamin Ren and Thomas B. Moeslund<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nDebasis Ganguly, Dwaipayan Roy, Mandar Mitra and Gareth Jones<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nDwaipayan Roy, Debjyoti Paul and Mandar Mitra<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nJarana Manotumruksa, Craig Macdonald and Iadh Ounis<\/span><\/p>\n
(opens in new tab)<\/span><\/a>
\nXiao Yang, Craig Macdonald and Iadh Ounis<\/span><\/p>\n\n
\n
\n
\n
\n
\n
\nW. Bruce Croft (opens in new tab)<\/span><\/a>, University of Massachusetts, Amherst, US
\nJiafeng Guo (opens in new tab)<\/span><\/a>, Chinese Academy of Sciences, Beijing, China
\nBhaskar Mitra (opens in new tab)<\/span><\/a>, Microsoft, Cambridge, UK
\nMaarten de Rijke (opens in new tab)<\/span><\/a>, University of Amsterdam, Amsterdam, The Netherlands<\/p>\n
\nDebasis Ganguly (opens in new tab)<\/span><\/a>, Dublin City University
\nKatja Hoffman (opens in new tab)<\/span><\/a>, Microsoft Research
\nHang Li (opens in new tab)<\/span><\/a>, Huawei Technologies
\nPiotr Mirowski (opens in new tab)<\/span><\/a>, Google DeepMind
\nAlessandro Moschitti (opens in new tab)<\/span><\/a>, Qatar Computing Research Institute, HKBU
\nPavel Serdyukov (opens in new tab)<\/span><\/a>, Yandex
\nFabrizio Silvestri (opens in new tab)<\/span><\/a>, Yahoo Labs
\nAlessandro Sordoni (opens in new tab)<\/span><\/a>, University of MontrealOpens in a new tab<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"
\nAcceptance Notifications<\/strong>: June 6
\nCamera-ready Deadline<\/strong>: June 17
\nWorkshop<\/strong>: July 21<\/p>\n