Lis Kanashiro Pereira

リズ カナシロ ペレイラ

I’m a Lecturer working on Natural Language Processing (NLP) @ Ochanomizu University, Japan.

I'm currently working on representation learning methods for NLP.

Contact Info: x@ocha.ac.jp where x = kanashiro.pereira

Address: Advanced Sciences, Graduate School of Humanities and Sciences, Ochanomizu University 2-1-1 Ohtsuka Bunkyo-ku Tokyo 112-8610 Japan

RECENT activities

**New**: Our paper OCHADAI at SemEval-2022 Task 2: Adversarial Training for Multilingual Idiomaticity Detection has been accepted to SEMEVAL 2022.

**New**: Our paper Attention is All you Need for Robust Temporal Reasoning has been accepted to LREC 2022.

**New**: Gave a talk at Microsoft Redmond Campus.

Our paper ALICE++: Adversarial Training for Robust and Effective Temporal Reasoning has been accepted to PACLIC 2021.

Our paper Multi-Layer Random Perturbation Training for Improving Model Generalization has been accepted to BlackBoxNLP@EMNLP 2021.

Our paper Towards a Language Model for Temporal Commonsense Reasoning has been accepted to the RANLP 2021 Student Workshop.

Our paper OCHADAI-KYODAI at SemEval-2021 Task 1: Enhancing Model Generalization and Robustness for Lexical Complexity Prediction won an Honorable Mention at SemEval-2021! --> https://semeval.github.io/SemEval2021/awards

Our recent paper Targeted Adversarial Training for Natural Language Understanding has been accepted to NAACL-2021. #1 on XNLI dataset

Our recent paper Posterior Differential Regularization with f-divergence for Improving Model Robustness has been accepted to NAACL-2021.

Our recent paper OCHADAI-KYODAI at SemEval-2021 Task 1: Enhancing Model Generalization and Robustness for Lexical Complexity Prediction has been accepted to be presented at SemEval 2021.

Our recent paper OCHADAI at SMM4H-2021 Task 5: Classifying self-reporting tweets on potential cases of COVID-19 by ensembling pre-trained language models has been accepted to be presented at SMM4H 2021@NAACL.

Our system ranked among the top-10 systems on both subtasks of SemEval-2021 Task 1 (Lexical Complexity Prediction).

Our recent paper Posterior Differential Regularization with f-divergence for Improving Model Robustness has been uploaded to arxiv.

Our model (ALICE) ranked 1st on the Allen Institute for AI's Leaderboard on the Temporal Commonsense Comprehension Task. (As of June 11th, 2020), outperforming the T5 model.

Our paper "Adversarial Training for Commonsense Inference" got accepted at the ACL 2020 Workshop on Representation Learning for NLP (RepL4NLP-2020).

Our model (ALICE) ranked 2nd on the Allen Institute for AI's Leaderboard on the SciTail dataset (Natural Language Inference dataset). (As of May 15th, 2020)

Our model (ALICE) ranked 1st on the Allen Institute for AI's Leaderboard on the CosmosQA dataset (Machine Reading Comprehension with Contextual Commonsense Reasoning). (As of March 23rd, 2020)

We ranked 1st on the Allen Institute for AI's Leaderboard on the Temporal Commonsense Comprehension Task. (As of February 27th, 2020)



Journal Reviewer: CALICO Journal

Conference Paper Reviewer: BEA 2022, BlackBoxNLP 2022, ACL 2022, BlackBoxNLP 2021, EMNLP 2021, ACL 2021, NAACL 2021, AAAI 2021, WSDM 2021, AAAI 2020, ACL 2020, SIGIR 2020, EMNLP 2020, AACL 2020, EMNLP 2019

Conference Program Committee member: Workshop on Automatic Speech Recognition for spontaneous and prepared speech & Speech Emotion Recognition in Portuguese (S&ER 2022) @ PROPOR 2022

Education

Professional Experience

  • Lecturer (Apr. 2021- Present)

  • Assistant Professor (Dec. 2019- Mar. 2021)

  • NLP Engineer (2019)

    • A.I. Squared Inc., Japan.

  • NLP Engineer (2017-2019)

    • Weathernews Inc., Japan.

  • Postdoctoral Fellow (2017)

    • City University of Hong Kong, Hong Kong.

    • Supervisor: John Lee

  • Information Technology Analyst, Federal Government Employee (2008- 2010​)

    • Universidade Federal do Pará (UFPA), Information Technology and Communication Center (CTIC), Pará, Brazil

AWARDS

  • Distinguished Student (2007), Universidade Federal do Pará (UFPA), link

  • Monbukagakusho Scholarship Recipient (2010-2016)

  • SemEval 2021, Honorable Mention, link

TALKS

Microsoft Redmond Campus, January 2022. Title: Robust and Generalizable Language Model Fine-Tuning.

Publications

Lis Kanashiro Pereira, Kevin Duh, Fei Cheng, Masayuki Asahara, Ichiro Kobayashi . Attention-Focused Adversarial Training for Robust Temporal Reasoning. LREC 2022. (To appear)


Lis Kanashiro Pereira, Ichiro Kobayashi. OCHADAI at SemEval-2022 Task 2: Adversarial Training for Multilingual Idiomaticity Detection. SemEval 2022.

Lis Pereira, Fei Cheng, Masayuki Asahara, and Ichiro Kobayashi. ALICE++: Adversarial Training for Robust and Effective Temporal Reasoning. PACLIC 2021.

Lis Pereira, Yuki Taya, , and Ichiro Kobayashi. Multi-Layer Random Perturbation Training for Improving Model Generalization. BlackBoxNLP@EMNLP 2021. (to appear)


Mayuko Kimura, Lis Kanashiro Pereira and Ichiro Kobayashi. Towards a Language Model for Temporal Commonsense Reasoning. Recent Advances in Natural Language Processing (RANLP 2021) Student Workshop.


Lis Pereira*, Xiaodong Liu*, Hao Cheng, Hoifung Poon, Jianfeng Gao and Ichiro Kobayashi. Targeted Adversarial Training for Natural Language Understanding. NAACL-2021 (*equal contribution). [pdf] #1 on the XNLI dataset

Hao Cheng, Xiaodong Liu, Lis Pereira, Yaoliang Yu, Jianfeng Gao. Posterior Differential Regularization with f-divergence for Improving Model Robustness. NAACL-2021. [pdf]

Yuki Taya, Lis Pereira, Fei Cheng, Ichiro Kobayashi. OCHADAI-KYODAI at SemEval-2021 Task 1: Enhancing Model Generalization and Robustness for Lexical Complexity Prediction. SemEval 2021@ACL. (honorable mention) [pdf]

Ying Luo, Lis Pereira, Ichiro Kobayashi. OCHADAI at SMM4H-2021 Task 5: Classifying self-reporting tweets on potential cases of COVID-19 by ensembling pre-trained language models. SMM4H 2021@NAACL.

Murayama, Y., Pereira, L. and Kobayashi, I. Dialogue over Context and Structured Knowledge using a Neural Network Model with External Memories. KNLP@AACL-IJCNLP 2020. [pdf]

Pereira, L., Liu, X., Cheng, F., Asahara, M. and Kobayashi, I. 2020. Adversarial Training for Commonsense Inference. ACL 2020 Workshop on Representation Learning for NLP (Rep4NLP@ACL2020). [pdf] #1 on the CosmosQA and on the MC-TACO datasets

Pereira, L., Liu, X., and Lee, J. 2017. Lexical Simplification using Deep Semantic Structured Models. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 430-435, Taipei, Taiwan.

Lis Pereira and Yuji Matsumoto. Leveraging a Learner Corpus for Automated Collocation Suggestion for Learners of Japanese as a Second Language. CALICO Journal. vol 33.3 2016 311–333, doi : 10.1558/cj.v33i3.26444

Lis Pereira and Yuji Matsumoto. 2015. Collocational Aid for Learners of Japanese as a Second Language. In Proceedings of the ACL 2015 Workshop on Natural Language Processing Techniques for Educational Applications (NLP-TEA-2), Beijing, China.

Lis W.K. Pereira, Elga Strafella, Kevin Duh and Yuji Matsumoto. 2014. Identifying Collocations using Cross-lingual Association Measures. In Proceedings of the EACL 2014 Workshop on Multiword Expressions, Gothenburg, Sweden.

Lis W.K. Pereira, Elga Strafella and Yuji Matsumoto. 2014. Collocation or Free Combination? – Applying Machine Translation Techniques to identify collocations in Japanese. Proceedings of the 9th Language Resources and Evaluation Conference (LREC 2014), Reykjavik, Iceland.

Lis W.K. Pereira, Erlyn Manguilimotan and Yuji Matsumoto. 2013. Automated Collocation Suggestion for Japanese Second Language Learners. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL), Student Research Workshop, pp.52-58, Sofia, Bulgaria.

Keisuke Sakaguchi, Yuta Hayashibe, Shuhei Kondo, Lis Kanashiro, Tomoya Mizumoto, Mamoru Komachi and Yuji Matsumoto. NAIST at the HOO 2012 Shared Task. The Seventh Workshop on Building Educational Applications Using NLP, pp.281-288, Montreal, Canada, 2012 June 7th.

Domestic Conferences

木村麻友子, Lis Kanashiro Pereira (お茶大), 浅原正幸 (国語研), Fei Cheng (京大), 越智綾子 (国語研), 小林一郎 (お茶大). 時間的常識理解へ向けた効果的なマスク言語モデルの検証. 言語処理学会第28回年次大会(NLP2022)

Kimura, M., Pereira, L. and Kobayashi, I. Towards a Temporal Commonsense Aware Language Model. In Proceedings of the Twenty-Seventh Annual Meeting of the Association for Natural Language Processing (NLP-2021). (In Japanese)


深層強化学習モデルの内部挙動の言語化による制御手法構築へ向けて
圓田彩乃、Lis Kanashiro Pereira、小林一郎, Domestic, 2021.06, 第35回人工知能学会全国大会, オンライン

特性を顕在化する言語の意味を反映した画像生成
渡邊清子、Lis Kanashiro Pereira、小林一郎, Domestic, 2021.06, 第35回人工知能学会全国大会, オンライン


Lis W.K. Pereira, Erlyn Manguilimotan and Yuji Matsumoto. 2013. Data Coverage vs. Data Size: A comparison of two large-scale corpora in Collocation Suggestion for Japanese Second Language Learners. In Proceedings of the Nineteenth Annual Meeting of the Association for Natural Language Processing (NLP-2013), pp.74-76, Nagoya, Japan, March 2013.

Lis W.K.Pereira, Erlyn Manguilimotan and Yuji Matsumoto. 2013. Collocation Suggestion for Japanese Second Language Learners, 情報処理学会研究報告 第210回自然言語処理研究、Vol.2013-NL-210、No.3、pp.1-5, January 2013.

Pereira, Lis W.K., Chagas, Larissa F., Souza, Jñane (2009) , Melhoria de Processo de Software no CTIC-UFPa: Um relato de experiência, III Workshop de Tecnologia de Informação das IFES, Brazil.

Mota, Marcelle; Pereira, Lis W.K. & Favero, Eloi (2008), JavaTool, uma ferramenta para ensino de programação, XXVIII Congress of the Brazilian Computer Society (SBC 2008), Brazil.