Program for CoNLL 2024

Day 1 (Friday, Nov 15, 2024):

09:00 – 09:10: Opening Remarks
09:10 – 10:30: Keynote 1: Lorna Quandt
10:30 – 11:00: Coffee Break
11:00 – 12:30: Oral Session 1: Psycholinguistic Session (chair: Libby Barak)

  • Leveraging a Cognitive Model to Measure Subjective Similarity of Human and GPT-4 Written Content
    Tyler Malloy, Maria José Ferreira, Fei Fang, Cleotilde Gonzalez
  • SPAWNing Structural Priming Predictions from a Cognitively Motivated Parser
    Grusha Prasad, Tal Linzen
  • Lossy Context Surprisal Predicts Task-Dependent Patterns in Relative Clause Processing
    Kate McCurdy, Michael Hahn
  • Multimodal Large Language Models “Foresee” Objects Based on Verb Information But Not Gender
    Shuqi Wang, Xufeng Duan, Zhenguang Cai

12:30 – 13:45: Lunch
13:45 – 15:30: Poster Session 1
15:30 – 16:00: Coffee Break
16:00 – 17:30: Oral Session 2: Syntax and Structure Session (chair: Omri Abend)

  • Is Structure Dependence Shaped for Efficient Communication?: A Case Study on Coordination
    Kohei Kajikawa, Yusuke Kubota, Yohei Oseki
  • NeLLCom-X: A Comprehensive Neural-Agent Framework to Simulate Language Learning and Group Communication
    Yuchen Lian, Tessa Verhoef, Arianna Bisazza
  • Solving the Challenge Set without Solving the Task: On Winograd Schemas as a Test of Pronominal Coreference Resolution
    Ian Porada, Jackie CK Cheung
  • Global Learning with Triplet Relations in Abstractive Summarization
    Jiaxin Duan, Fengyu Lu, Junfei Liu 
     

Day 2 (Saturday, Nov 16, 2024):

09:00 – 09:10: Best Paper Awards
09:10 – 10:30: Keynote 2: Thamar Solorio
10:30 – 10:45: Coffee Break
10:45 – 12:15: Oral Session 3: LLM Session (Chair: Malihe Alikhani)

  • Global-Pruner: A Stable and Efficient Pruner for Retraining-Free Pruning of Encoder-Based Language Models
    Guangzhen Yao, Sandong Zhu, Long Zhang, MiaoQI
  • Investigating large language models for their competence in extracting grammatically sound sentences from transcribed noisy utterances
    Alina Wróblewska
  • The Effect of Word Predictability on Reading Times in Information Seeking and Repeated Reading
    Keren Gruteke Klein, Yoav Meiri, Omer Shubi, Yevgeni Berzak
  • Multi-Cultural Norm Base: Frame-based Norm Discovery in Multi-Cultural Settings
    Viet Thanh Pham, SHILIN QU, Farhad Moghimifar, Suraj Sharma, Yuan-Fang Li, Weiqing Wang, Reza Haf

12:15 – 13:45: Lunch
13:45 – 15:00: Poster Session 2
15:00 – 15:30: BabyLM Challenge (oral session)
15:30 – 16:00: Coffee Break
16:00 – 17:20: BabyLM Challenge (poster session)
17:20 – 17:30: Closing Remarks

  • All times are given in local time, which is in Miami, Florida, Eastern Time (ET).

 

Poster Sessions:

  • Text2Afford: Probing Object Affordance Prediction abilities of Language Models solely from Text

    Sayantan Adak, Daivik Agrawal, Animesh Mukherjee, Somak Aditya 

  • Transformer verbatim in-context retrieval across time and scale

    Kristijan Armeni, Marko Pranjić, Senja Pollak 

  • Of Models and Men: Probing Neural Networks for Agreement Attraction with Psycholinguistic Data

    Maxim Bazhukov, Ekaterina Voloshina, Sergey Pletenev, Arseny Anisimov, Oleg Serikov, Svetlana Toldova 

  • How Are Metaphors Processed by Language Models? The Case of Analogies

    Joanne Boisson 

  • AIStorySimilarity: Quantifying Story Similarity Using Narrative for Search, IP Infringement, and Guided Creativity

    Jon Chun 

  • Explaining the Hardest Errors of Contextual Embedding Based Classifiers

    Claudio Moisés Valiense de Andrade, Washington Cunha, Guilherme Fonseca, Ana Clara Souza Pagano, Luana de Castro Santos, Adriana Silvina Pagano, Leonardo Chaves Dutra da Rocha, Marcos André Gonçalves 

  • EditEval: An Instruction-Based Benchmark for Text Improvements

    Jane Dwivedi-Yu, Timo Schick, Zhengbao Jiang, Maria Lomeli, Patrick Lewis, Gautier Izacard, Edouard Grave, Sebastian Riedel, Fabio Petroni 

  • Advancing Arabic Sentiment Analysis: ArSen Benchmark and the Improved Fuzzy Deep Hybrid Network

    Yang Fang, Cheng Xu, Shuhao Guan, Nan Yan, Yuke Mei 

  • Critical Questions Generation: Motivation and Challenges

    Blanca Calvo Figueras, Rodrigo Agerri 

  • Generalizations across filler-gap dependencies in neural language models

    Katherine Howitt, Sathvik Nair, Allison Dods, Robert Melvin Hopkins 

  • Continuous Attentive Multimodal Prompt Tuning for Few-Shot Multimodal Sarcasm Detection

    Soumyadeep Jana, Animesh Dey, Ranbir Singh Sanasam 

  • Aligning Alignments: Do Colexification and Distributional Similarity Align as Measures of cross-lingual Lexical Alignment?

    Taelin Karidi, Eitan Grossman, Omri Abend 

  • On Functional Competence of LLMs for Linguistic Disambiguation

    Raihan Kibria, Sheikh Intiser Uddin Dipta, Muhammad Abdullah Adnan 

  • TpT-ADE: Transformer Based Two-Phase ADE Extraction

    Suryamukhi Kuchibhotla, Manish Singh 

  • PRACT: Optimizing Principled Reasoning and Acting of LLM Agent

    Zhiwei Liu, Weiran Yao, Jianguo Zhang, Zuxin Liu, Liangwei Yang, Rithesh R N, Tian Lan, Ming Zhu, Juntao Tan, Shirley Kokane, Thai Quoc Hoang, Juan Carlos Niebles, Shelby Heinecke, Huan Wang, Silvio Savarese, Caiming Xiong 

  • Mitigating Bias in Language Model Evaluators: A Causal ATE Approach

    Rahul Madhavan, Kahini Wadhawan 

  • Words That Stick: Using Keyword Cohesion to Improve Text Segmentation

    Amit Maraj, Miguel Vargas Martin, Masoud Makrehchi 

  • An Empirical Comparison of Vocabulary Expansion and Initialization Approaches For Language Models

    Nandini Mundra, Aditya Nanda Kishore Khandavally, Raj Dabre, Ratish Puduppully, Anoop Kunchukuttan, Mitesh M Khapra 

  • Revisiting Hierarchical Text Classification: Inference and Metrics

    Roman Plaud, Matthieu Labeau, Antoine Saillenfest, Thomas Bonald 

  • Image-conditioned human language comprehension and psychometric benchmarking of visual language models

    Subha Nawer Pushpita, Roger P. Levy 

  • Large Language Model Recall Uncertainty is Modulated by the Fan Effect

    Jesse Roberts, Kyle Moore, Douglas Fisher, Oseremhen Ewaleifoh, Thao Pham 

  • Self-supervised speech representations display some human-like cross-linguistic perceptual abilities

    Joselyn Rodriguez, Kamala Sreepada, Ruolan Leslie Famularo, Sharon Goldwater, Naomi Feldman 

  • One-Vs-Rest Neural Network English Grapheme Segmentation: A Linguistic Perspective

    Samuel Rose, Nina Dethlefs, C. Kambhampati 

  • CrowdCounter: A benchmark type-specific multi-target counterspeech dataset

    Punyajoy Saha, Abhilash Datta, Abhik Jana, Animesh Mukherjee 

  • Translating Across Cultures: LLMs for Intralingual Cultural Adaptation

    Pushpdeep Singh, Mayur Patidar, Lovekesh Vig 

  • Making Distilled Language Models Even Smaller: Lightweight Reconstruction of Rare Token Embeddings

    Kohki Tamura, Naoki Yoshinaga, Masato Neishi 

  • A Novel Instruction Tuning Method for Vietnamese Math Reasoning using Trainable Open-Source Large Language Models

    Nguyen Quang Vinh, Thanh-Do Nguyen, Vinh Van Nguyen, Nam Khac-Hoai Bui 

  • Information Association for Language Model Updating by Mitigating LM-Logical Discrepancy

    Pengfei Yu, Heng Ji 

 

Webmaster: Jens Lemmens