default search action
RepL4NLP@ACL 2018 Melbourne, Australia
- Isabelle Augenstein, Kris Cao, He He, Felix Hill, Spandana Gella, Jamie Kiros, Hongyuan Mei, Dipendra Misra:
Proceedings of The Third Workshop on Representation Learning for NLP, Rep4NLP@ACL 2018, Melbourne, Australia, July 20, 2018. Association for Computational Linguistics 2018, ISBN 978-1-948087-43-8 - Edgar Altszyler, Mariano Sigman, Diego Fernández Slezak:
Corpus Specificity in LSA and Word2vec: The Role of Out-of-Domain Documents. 1-10 - Shang Gao, Arvind Ramanathan, Georgia D. Tourassi:
Hierarchical Convolutional Attention Networks for Text Classification. 11-23 - Hwiyeol Jo, Stanley Jungkyu Choi:
Extrofitting: Enriching Word Representation and its Vector Space with Semantic Lexicons. 24-29 - Dennis Singh Moirangthem, Minho Lee:
Chat Discrimination for Intelligent Conversational Agents with a Hybrid CNN-LMTGRU Network. 30-40 - Amr Rekaby Salama, Özge Alaçam, Wolfgang Menzel:
Text Completion using Context-Integrated Dependency Parsing. 41-49 - Qiuchi Li, Sagar Uprety, Benyou Wang, Dawei Song:
Quantum-Inspired Complex Word Embedding. 50-57 - Kosuke Nishida, Kyosuke Nishida, Hisako Asano, Junji Tomita:
Natural Language Inference with Definition Embedding Considering Context On the Fly. 58-63 - Lidia Pivovarova, Roman Yangarber:
Comparison of Representations of Named Entities for Document Classification. 64-68 - Shuai Tang, Hailin Jin, Chen Fang, Zhaowen Wang, Virginia R. de Sa:
Speeding up Context-based Sentence Representation Learning with Non-autoregressive Convolutional Decoding. 69-78 - Gil Levi:
Connecting Supervised and Unsupervised Sentence Embeddings. 79-83 - Weiguo Sheng:
A Hybrid Learning Scheme for Chinese Word Embedding. 84-90 - Kawin Ethayarajh:
Unsupervised Random Walk Sentence Embeddings: A Strong but Simple Baseline. 91-100 - Yadollah Yaghoobzadeh, Katharina Kann, Hinrich Schütze:
Evaluating Word Embeddings in Multi-label Classification Using Fine-Grained Name Typing. 101-106 - Ade Romadhony, Alfan Farizki Wicaksono, Ayu Purwarianti, Dwi Hendratmo Widyantoro:
A Dense Vector Representation for Open-Domain Relation Tuples. 107-112 - Jilei Wang, Shiying Luo, Weiyan Shi, Tao Dai, Shu-Tao Xia:
Exploiting Common Characters in Chinese and Japanese to Learn Cross-Lingual Word Embeddings via Matrix Factorization. 113-121 - Chakaveh Saedi, António Branco, João António Rodrigues, João Silva:
WordNet Embeddings. 122-131 - Yanrong Wu, Zhichun Wang:
Knowledge Graph Embedding with Numeric Attributes of Entities. 132-136 - Ivan Vulic:
Injecting Lexical Contrast into Word Vectors by Guiding Vector Space Specialisation. 137-143 - Ahmet Üstün, Murathan Kurfali, Burcu Can:
Characters or Morphemes: How to Represent Words? 144-153 - Athul Paul Jacob, Zhouhan Lin, Alessandro Sordoni, Yoshua Bengio:
Learning Hierarchical Structures On-The-Fly with a Recurrent-Recursive Model for Sequences. 154-158 - Mareike Hartmann, Anders Søgaard:
Limitations of Cross-Lingual Learning from Image Search. 159-163 - Yinfei Yang, Steve Yuan, Daniel Cer, Sheng-yi Kong, Noah Constant, Petr Pilar, Heming Ge, Yun-Hsuan Sung, Brian Strope, Ray Kurzweil:
Learning Semantic Textual Similarity from Conversations. 164-174 - Katherine Yu, Haoran Li, Barlas Oguz:
Multilingual Seq2seq Training with Similarity Loss for Cross-Lingual Document Classification. 175-179 - Nelson F. Liu, Omer Levy, Roy Schwartz, Chenhao Tan, Noah A. Smith:
LSTMs Exploit Linguistic Attributes of Data. 180-186 - Samuel Broscheit:
Learning Distributional Token Representations from Visual Features. 187-194 - Denis Newman-Griffis, Albert M. Lai, Eric Fosler-Lussier:
Jointly Embedding Entities and Text with Distant Supervision. 195-206 - Angel Daza, Anette Frank:
A Sequence-to-Sequence Model for Semantic Role Labeling. 207-216 - Nikola Ljubesic, Darja Fiser, Anita Peti-Stantic:
Predicting Concreteness and Imageability of Words Within and Across Languages via Word Embeddings. 217-222
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.