Publications

2025

  1. TactfulToM: Do LLMs have the Theory of Mind ability to understand White Lies?
    Yiwei Liu, Emma Jane Pretty, Jiahao Huang, and 1 more author
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, Nov 2025
  2. Are Checklists Really Useful for Automatic Evaluation of Generative Tasks?
    Momoka Furuhashi, Kouta Nakayama, Takashi Kodama, and 1 more author
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, Nov 2025
  3. Specification-Aware Machine Translation and Evaluation for Purpose Alignment
    Yoko Kayano and Saku Sugawara
    In Proceedings of the Tenth Conference on Machine Translation, Nov 2025
  4. MCQFormatBench: Robustness Tests for Multiple-Choice Questions
    Hiroo Takizawa, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the Fourth Workshop on Generation, Evaluation and Metrics (GEM\texttwosuperior), Jul 2025
  5. Development of Numerical Error Detection Tasks to Analyze the Numerical Capabilities of Language Models
    Taku Sakamoto, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the 31st International Conference on Computational Linguistics, Jan 2025
  6. Quality Text, Robust Vision: The Role of Language in Enhancing Visual Robustness of Vision-Language Models
    Futa Waseda, Saku Sugawara, and Isao Echizen
    In Proceedings of the 33rd ACM International Conference on Multimedia, Dublin, Ireland, 2025

2024

  1. Rationale-Aware Answer Verification by Pairwise Self-Evaluation
    Akira Kawabata and Saku Sugawara
    In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Nov 2024
  2. Can Language Models Induce Grammatical Knowledge from Indirect Evidence?
    Miyu Oba, Yohei Oseki, Akiyo Fukatsu, and 4 more authors
    In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Nov 2024
  3. Modeling Overregularization in Children with Small Language Models
    Akari Haga, Saku Sugawara, Akiyo Fukatsu, and 4 more authors
    In Findings of the Association for Computational Linguistics: ACL 2024, Aug 2024
  4. What Makes Language Models Good-enough?
    Daiki Asami and Saku Sugawara
    In Findings of the Association for Computational Linguistics: ACL 2024, Aug 2024
  5. AUTOMATIC FEEDBACK GENERATION FOR SHORT ANSWER QUESTIONS USING ANSWER DIAGNOSTIC GRAPHS
    M. Furuhashi, H. Funayama, Y. Iwase, and 5 more authors
    In EDULEARN24 Proceedings, Palma, Spain, 1-3 july, 2024 2024

2023

  1. PROPRES: Investigating the Projectivity of Presupposition with Various Triggers and Environments
    Daiki Asami and Saku Sugawara
    In Proceedings of the 27th Conference on Computational Natural Language Learning (CoNLL), Dec 2023
  2. Evaluating the Rationale Understanding of Critical Reasoning in Logical Reading Comprehension
    Akira Kawabata and Saku Sugawara
    In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, Dec 2023
  3. On Degrees of Freedom in Defining and Testing Natural Language Understanding
    Saku Sugawara and Shun Tsugita
    In Findings of the Association for Computational Linguistics: ACL 2023, Jul 2023
  4. Probing Physical Reasoning with Counter-Commonsense Context
    Kazushi Kondo, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Jul 2023
  5. Which Shortcut Solution Do Question Answering Models Prefer to Learn?
    Kazutoshi Shinoda, Saku Sugawara, and Akiko Aizawa
    Proceedings of the AAAI Conference on Artificial Intelligence, Jun 2023
  6. Analyzing the Effectiveness of the Underlying Reasoning Tasks in Multi-hop Question Answering
    Xanh Ho, Anh-Khoa Duong Nguyen, Saku Sugawara, and 1 more author
    In Findings of the Association for Computational Linguistics: EACL 2023, May 2023
  7. A Survey on Measuring and Mitigating Reasoning Shortcuts in Machine Reading Comprehension
    Xanh Ho, Johannes Mario Meissner, Saku Sugawara, and 1 more author
    2023

2022

  1. Cross-Modal Similarity-Based Curriculum Learning for Image Captioning
    Hongkuan Zhang, Saku Sugawara, Akiko Aizawa, and 3 more authors
    In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Dec 2022
  2. Debiasing Masks: A New Framework for Shortcut Mitigation in NLU
    Johannes Mario Meissner, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Dec 2022
  3. Look to the Right: Mitigating Relative Position Bias in Extractive Question Answering
    Kazutoshi Shinoda, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the Fifth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, Dec 2022
  4. How Well Do Multi-hop Reading Comprehension Models Understand Date Information?
    Xanh Ho, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), Nov 2022
  5. Possible Stories: Evaluating Situated Commonsense Reasoning under Multiple Possible Scenarios
    Mana Ashida and Saku Sugawara
    In Proceedings of the 29th International Conference on Computational Linguistics, Oct 2022
  6. What Makes Reading Comprehension Questions Difficult?
    Saku Sugawara, Nikita Nangia, Alex Warstadt, and 1 more author
    In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), May 2022
  7. Penalizing Confident Predictions on Largely Perturbed Inputs Does Not Improve Out-of-Distribution Generalization in Question Answering
    Kazutoshi Shinoda, Saku Sugawara, and Akiko Aizawa
    2022

2021

  1. Can Question Generation Debias Question Answering Models? A Case Study on Question–Context Lexical Overlap
    Kazutoshi Shinoda, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the 3rd Workshop on Machine Reading for Question Answering, Nov 2021
  2. Embracing Ambiguity: Shifting the Training Target of NLI Models
    Johannes Mario Meissner, Napat Thumwanit, Saku Sugawara, and 1 more author
    In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), Aug 2021
  3. What Ingredients Make for an Effective Crowdsourcing Protocol for Difficult NLU Data Collection Tasks?
    Nikita Nangia, Saku Sugawara, Harsh Trivedi, and 3 more authors
    In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Aug 2021
  4. Improving the Robustness of QA Models to Challenge Sets with Variational Question-Answer Pair Generation
    Kazutoshi Shinoda, Saku Sugawara, and Akiko Aizawa
    In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: Student Research Workshop, Aug 2021
  5. Benchmarking Machine Reading Comprehension: A Psychological Perspective
    Saku Sugawara, Pontus Stenetorp, and Akiko Aizawa
    In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Apr 2021

2020

  1. Constructing A Multi-hop QA Dataset for Comprehensive Evaluation of Reasoning Steps
    Xanh Ho, Anh-Khoa Duong Nguyen, Saku Sugawara, and 1 more author
    In Proceedings of the 28th International Conference on Computational Linguistics, Dec 2020
  2. Constructing A Multi-hop QA Dataset for Comprehensive Evaluation of Reasoning Steps
    Xanh Ho, Anh-Khoa Duong Nguyen, Saku Sugawara, and 1 more author
    In Proceedings of the 28th International Conference on Computational Linguistics, Dec 2020
  3. Assessing the Benchmarking Capacity of Machine Reading Comprehension Datasets
    Saku Sugawara, Pontus Stenetorp, Kentaro Inui, and 1 more author
    In AAAI, 2020

2018

  1. What Makes Reading Comprehension Questions Easier?
    Saku Sugawara, Kentaro Inui, Satoshi Sekine, and 1 more author
    In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Oct 2018

2017

  1. Evaluation Metrics for Machine Reading Comprehension: Prerequisite Skills and Readability
    Saku Sugawara, Yusuke Kido, Hikaru Yokono, and 1 more author
    In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Jul 2017
  2. Prerequisite Skills for Reading Comprehension: Multi-Perspective Analysis of MCTest Datasets and Systems
    Saku Sugawara, Hikaru Yokono, and Akiko Aizawa
    Proceedings of the AAAI Conference on Artificial Intelligence, Feb 2017

2016

  1. Annotation and Analysis of Discourse Relations, Temporal Relations and Multi-Layered Situational Relations in Japanese Texts
    Kimi Kaneko, Saku Sugawara, Koji Mineshima, and 1 more author
    In Proceedings of the 12th Workshop on Asian Language Resources (ALR12), Dec 2016
  2. An Analysis of Prerequisite Skills for Reading Comprehension
    Saku Sugawara and Akiko Aizawa
    In Proceedings of the Workshop on Uphill Battles in Language Processing: Scaling Early Achievements to Robust Methods, Nov 2016