University of California, San Diego | Nov 22, 2024
Welcome
The goal of the Southern California Natural Language Processing Symposium is to gather researchers from the southern California region with broad expertise in machine learning and natural language processing. The symposium will provide a participatory environment where attendees from a variety of fields can share and discuss their latest findings.
Important Dates
Submission deadline: Oct 26 (Saturday), 2024, 11:59 PM Anywhere On Earth [OpenReview Submission Portal] Acceptance notification: Nov. 1, 2024 Registration deadline: Nov 10, 2024 [Eventbrite Registration Link] Symposium date: Nov 22, 2024 (Friday) Submission format:
If the paper is published in a recent conference or journal, you can directly submit the paper without modification. Please indicate where the paper is published in the Abstract when submitting the paper.
For papers that have not been previously published, we recommend submitting an extended summary spanning no more than two pages, formatted according to the ACL guidelines, However, submissions of a longer length will also be considered.
All submissions are single-blind reviewed.
Venue
Date: Nov 22, 2024
Locations: There are two venues involved in this event. They are close to each other.
Breakfast, Lunch, and Poster: Great Hall
Talk: Institute of the America's Hojel Auditorium
Parking information:
Campus parking is very limited and is not recommended. You can check available options here. The closest parking lot for visitors is located on the top floor of Hopkins Parking Structure. You can pay the fee in pay station or pay-by-phone. Note that most visitor spots are limited to two hours. Even though the app allows you to pay for longer periods, you will get a ticket after that time.
There is free parking at Torrey Pines Gliderport, if you do not mind a 30-minute walk to the venue. The view from the Gliderport is spectacular.
View from the Torrey Pines Gliderport. Photo credit: Gina Trapani (CC BY-NC-SA 2.0 Deed)
Schedule
Great Hall
09:00am - 09:45am Breakfast, Registration & Social
Abstract: Large language models (LLMs) and vision-language models (VLMs) have demonstrated remarkable performance on knowledge reasoning tasks, owing to their implicit knowledge derived from extensive pretraining data. However, their inherent knowledge bases often suffer from disorganization and illusion, bias towards common entities, and rapid obsolescence. Consequently, LLMs frequently make up untruthful information, exhibit resistance to updating outdated knowledge, or struggle with generalizing across multiple languages. In this talk I will discuss several research directions that aim to make foundation models’ knowledge more accurate, organized, up-to-date and fair: (1) Where and How is Knowledge Stored in LLM? (2) How to Control LLM’s Knowledge? (3) How to Update LLM’s Dynamic Knowledge? (4) How to Bridge the Knowledge Gap between Natural Language and Unnatural Language?
Bio: Heng Ji is a professor at Siebel School of Computing and Data Science, and an affiliated faculty member at Electrical and Computer Engineering Department, Coordinated Science Laboratory, and Carl R. Woese Institute for Genomic Biology of University of Illinois Urbana-Champaign. She is an Amazon Scholar. She is the Founding Director of Amazon-Illinois Center on AI for Interactive Conversational Experiences (AICE). She received her B.A. and M. A. in Computational Linguistics from Tsinghua University, and her M.S. and Ph.D. in Computer Science from New York University. Her research interests focus on Natural Language Processing, especially on Multimedia Multilingual Information Extraction, Knowledge-enhanced Large Language Models and Vision-Language Models, and AI for Science. The awards she received include Outstanding Paper Award at ACL2024, two Outstanding Paper Awards at NAACL2024, "Young Scientist" by the World Laureates Association in 2023 and 2024, "Young Scientist" and a member of the Global Future Council on the Future of Computing by the World Economic Forum in 2016 and 2017, "Women Leaders of Conversational AI" (Class of 2023) by Project Voice, "AI's 10 to Watch" Award by IEEE Intelligent Systems in 2013, NSF CAREER award in 2009, PACLIC2012 Best paper runner-up, "Best of ICDM2013" paper award, "Best of SDM2013" paper award, ACL2018 Best Demo paper nomination, ACL2020 Best Demo Paper Award, NAACL2021 Best Demo Paper Award, Google Research Award in 2009 and 2014, IBM Watson Faculty Award in 2012 and 2014 and Bosch Research Award in 2014-2018. She served as the associate editor for IEEE/ACM Transaction on Audio, Speech, and Language Processing, and the Program Committee Co-Chair of many conferences including NAACL-HLT2018 and AACL-IJCNLP2022. She was elected as the North American Chapter of the Association for Computational Linguistics (NAACL) secretary 2020-2023.
Eunsol Choi
Title: Equipping LLMs for Complex Knowledge Scenarios: Interaction and Retrieval
Abstract: Language models are increasingly used as an interface to gather information. Yet trusting the answers generated from LMs is risky, as they often contain incorrect or misleading information. Why is this happening? We identify two key issues: (1) ambiguous and underspecified user questions and (2) imperfect knowledge in LMs, especially for long tail or recent events. To address the first issue, we propose a system that can interact with users to clarify their intent before answering. By simulating their expected outcomes in the future turns, we reward LMs for generating clarifying questions and not just answering immediately. In the second part of the talk, I will discuss the state of retrieval augmentation, which is often lauded as the path to provide up-to-date, relevant knowledge to LMs. While their success is evident in scenarios where there exists a single gold document, incorporating information from a diverse set of documents remains challenging for both retrievers and LMs. Together, the talk highlights key research directions for building reliable LMs to answer information seeking questions.
Bio: Eunsol Choi is an assistant professor of computer science and data science at New York University. Her research spans natural language processing and machine learning, with a focus on interpreting and reasoning about text in dynamic real-world contexts. Prior to joining NYU, she was an assistant professor at the University of Texas at Austin. She also spent a year at Google AI as a visiting researcher. She holds a Ph.D. in computer science and engineering from the University of Washington. She is a recipient of a Facebook research fellowship, Google faculty research award, Sony faculty award, and an outstanding paper award at EMNLP.
Dawei Huang
Title: Enabling extremely fast inference and training performance using dataflow and custom chip
Abstract: As the pursuit of larger language models continues to push the boundaries of computational demands, the traditional silicon chip is facing a daunting memory/power wall. In this talk, we present a novel chip design, SN40L, to tackle this challenge. This chip combines a reconfigurable data-flow architecture with a tightly coupled 3-tier memory hierarchy to enable efficient compute-intensive training and memory-bound inference workloads for a wide variety of neural network architectures. We discuss the various advantages of this chip via case studies. In our first case study, we discuss how the dataflow architecture coupled with on-chip SRAM and HBM empowers operation fusion capabilities enabling 1000+ tokens/second inference performance without sacrificing on precision. In the second case study we look at the training performance of various model architectures and compare their performance against traditional kernel by kernel execution-based architectures. We show how dataflow architecture can help accelerate LLM training for traditional dense, sparse and novel state space models, while allowing one to train extremely large models on a smaller footprint. In the third case study we will discuss how you can use the strongly coupled DRAM, HBM and SRAM to develop new neural network architecture that can help scale to 100+ billion parameters efficiently. This new architecture uses a modular and coarse-grained approach to a mixture of experts that allows for incremental updates to models for new capabilities and knowledge and smaller footprint execution during inference.
Bio: Dawei Huang received BS from Tsinghua University and MS from UCSD. He has 25+ years experience in Computer/AI Systems and has 47 issued patents. Currently, he is Senior Director of Engineering at Sambanova Systems. His current responsibilities include ML Model Performance and Customer Solution.
Taylor Berg-Kirkpatrick
Title: Large Language Models and Cybersecurity: New Threats and Safeguards
Abstract: The rapid development of large-language models has created new attack surfaces for adversaries, but at the same time has enabled new mitigations for human-driven fraud and social engineering attacks. This talk will consider two new interfaces between NLP and computer security: (1) the potential for new types of attacks as LLMs are increasingly treated like operating systems and (2) the possibility of using LLMs to combat and track social engineering attacks at scale via LLM-driven honeypots.
Bio: Taylor Berg-Kirkpatrick is an Associate Professor in the Computer Science and Engineering Department at the University of California San Diego. Taylor's research focuses on using machine learning to understand structured human data, including language but also sources like music, document images, and other complex artifacts.
Accepted Work
Non-archival, randomly ordered
Poster Session #1 (11:30am - 12:30pm)
CLIMB: A Benchmark of Clinical Bias in Large Language Models Yubo Zhang, Shudi Hou, Mingyu Derek Ma, Wei Wang, Muhao Chen, Jieyu Zhao
Exploring the Design Space of Diffusion Bridge Models via Stochasticity Control Shaorong Zhang, Yuanbin Cheng, Xianghao Kong, Greg Ver Steeg
A Novel Multi-Document Retrieval Benchmark Grounded on Journalist Source-Selection in Newswriting Alexander Spangher, Tenghao Huang, Yiqin Huang, Liheng Lai, Lucas Spangher, Sewon Min, Mark Dredze
Chain-of-Thought Augmentation with Logit Contrast for Enhanced Reasoning in Language Models Jay Shim, Grant Kruttschnitt, Alyssa Ma, Daniel Kim, Benjamin Chek, Athul Anand, Kevin Zhu, Sean O'Brien
Improving Inference on Theory of Mind with Evidence Chain Qiutong Tony Yi, Wang Zhu
An Audio checkup for Language Models Arjun Prasaath Anbazhagan, Parteek Kumar, Souritra Kar, Ujjwal Kaur
Compare without Despair: Reliable Preference Evaluation with Generation Separability Sayan Ghosh, Tejas Srinivasan, Swabha Swayamdipta
Towards Interventions for Suicide Prevention with LLM Assistants in Social Work Jaspreet Ranjit, Hyundong Justin Cho, Myles Phung, John R. Blosnich, Swabha Swayamdipta
Verify with Caution: The Pitfalls of Relying on Imperfect Factuality Metrics Ameya Godbole, Robin Jia
The Dark Side of Instagram: A Large Dataset for Identifying Persian Harmful Comments Hadi Davardoust, Hadi Zare, Hossein Rafiee zade
Eliciting Better Multilingual Structured Reasoning from LLMs through Code Bryan Li, Tamer Alkhouli, Daniele Bonadiman, Nikolaos Pappas, Saab Mansour
Unlocking Decoding-Time Controllability: Gradient-Free Multi-Objective Alignment with Contrastive Prompts Tingchen Fu, Yupeng Hou, Julian McAuley, Rui Yan
AgentGrow: LLMs as Scalable, Customizable General-Purpose Simulators For Language Agent Training Yiming Wang, Yuedong Cui, Da Yin, Zongyu Lin, Di Wu, Xueqing Wu, Chenchen Ye, Kai-Wei Chang
SWE-Search: Enhancing Software Agents with Monte Carlo Tree Search and Iterative Refinement Antonis Antoniades, Albert Örwall, Kexun Zhang, Yuxi Xie, Anirudh Goyal, William Yang Wang
INTERINTENT: Investigating Social Intelligence of LLMs via Intention Understanding in an Interactive Game Context Ziyi Liu, Abhishek Anand, Pei Zhou, Jen-tse Huang, Jieyu Zhao
Contrastive Instruction Tuning Tianyi Lorena Yan, Fei Wang, James Y. Huang, Wenxuan Zhou, Fan Yin, Aram Galstyan, Wenpeng Yin, Muhao Chen
Embodied Agent Interface: Benchmarking LLMs for Embodied Decision Making Manling Li, Shiyu Zhao, Qineng Wang, Kangrui Wang, Yu Zhou, Sanjana Srivastava, Cem Gokmen, Tony Lee, Li Erran Li, Ruohan Zhang, Weiyu Liu, Percy Liang, Li Fei-Fei, Jiayuan Mao, Jiajun Wu
Teaching Models to Understand but Not Generate Ryan Yixiang Wang, Matthew Finlayson, Luca Soldaini, Robin Jia, Swabha Swayamdipta
IsoBench: Benchmarking Multimodal Foundation Models on Isomorphic Representations Deqing Fu, Ruohao Guo, Ghazal Khalighinejad, Ollie Liu, Bhuwan Dhingra, Dani Yogatama, Robin Jia, Willie Neiswanger
IllusionVQA: A Challenging Optical Illusion Dataset for Vision Language Models Haz Sameen Shahgir, Khondker Salman Sayeed, Abhik Bhattacharjee, Wasi Uddin Ahmad, Yue Dong, Rifat Shahriyar
Stress-Testing Long-Context Language Models with Lifelong ICL and Task Haystack Xiaoyue Xu, Qinyuan Ye, Xiang Ren
Can Textual Unlearning Solve Cross-Modality Safety Alignment? Trishna Chakraborty, Erfan Shayegani, Zikui Cai, Nael Abu-Ghazaleh, M. Salman Asif, Yue Dong, Amit Roy-Chowdhury, Chengyu Song
Evaluating Grounding Gaps in LLMs via News Interviews Michael Lu, Hyundong Justin Cho, Weiyan Shi, Jonathan May, Alexander Spangher
LogogramNLP: Comparing Visual and Textual Representations of Ancient Logographic Writing Systems for NLP Danlu Chen, Freda Shi, Aditi Agarwal, Jacobo Myerston, Taylor Berg-Kirkpatrick
Question-Analysis Prompting Improves LLM Performance in Reasoning Tasks Dharunish Yugeswardeenoo, Kevin Zhu, Sean O'Brien
Answer is All You Need: Instruction-following Text Embedding via Answering the Question Letian Peng, Yuwei Zhang, Zilong Wang, Jayanth Srinivasa, Gaowen Liu, Zihan Wang, Jingbo Shang
Socratic Mind: Scalable Oral Assessment Powered By AI Jui-Tse Hung, Christopher Zhang Cui, Diana M. Popescu, Saurabh Chatterjee, Thad Starner
Will Trump Win in 2024? Predicting the US Presidential Election via Multi-step Reasoning with Large Language Models Chenxiao Yu, Zhaotian Weng, Xiyang Hu, Yue Zhao
Faithful Persona-based Conversational Dataset Generation with Large Language Models Pegah Jandaghi, Xianghai Sheng, Xinyi Bai, Jay Pujara, Hakim Sidahmed
From Bias to Balance: Detecting Facial Expression Recognition Biases in Multimodal Foundation Models Kaylee Chhua, Zhoujinyi Wen, Vedant Hathalia, Kevin Zhu, Sean O'Brien
AAVENUE: Detecting LLM Biases on NLU Tasks in AAVE via a Novel Benchmark Abhay Gupta, Ece Yurtseven, Philip Meng, Sean O'Brien, Kevin Zhu
Language Models Implicitly Learn a Unified Representation Space Zhaofeng Wu, Xinyan Velocity Yu, Dani Yogatama, Jiasen Lu, Yoon Kim
Scaling LLM Inference with Optimized Sample Compute Allocation Kexun Zhang, Shang Zhou, Danqing Wang, William Yang Wang, Lei Li
Evaluating Human Alignment and Model Faithfulness of LLM Rationale Mohsen Fayyaz, Fan Yin, Jiao Sun, Nanyun Peng
Guiding Through Complexity: What Makes Good Supervision for Hard Reasoning Tasks? Xuan He, Da Yin, Nanyun Peng
A Little Human Data Goes A Long Way Dhananjay Ashok, Jonathan May
Enhancing Depression Diagnosis with Chain-of-Thought Prompting Elysia Shi, Adithri Manda, London Chowdhury, Runeema Arun, Kevin Zhu, Michael Lam
ChunkRAG: Novel LLM-Chunk Filtering Method for RAG Systems Ishneet Sukhvinder Singh, Ritvik Aggarwal, Ibrahim Allahverdiyev, Muhammad Taha, Aslihan Akalin, Kevin Zhu, Sean O'Brien
Reversing the Forget-Retain Objectives: An Efficient LLM Unlearning Framework from Logit Difference Jiabao Ji, Yujian Liu, Yang Zhang, Gaowen Liu, Ramana Rao Kompella, Sijia Liu, Shiyu Chang
DeLLMa: Decision Making Under Uncertainty with Large Language Models Ollie Liu, Deqing Fu, Dani Yogatama, Willie Neiswanger
Apathetic or Empathetic? Evaluating LLMs' Emotional Alignments with Humans Jen-tse Huang, Man Ho LAM, Eric John Li, Shujie Ren, Wenxuan Wang, Wenxiang Jiao, Zhaopeng Tu, Michael Lyu
Interleaved Multimodal Decision Transformer for Embodied Agents Bosung Kim
DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks Kaijie Zhu, Jiaao Chen, Jindong Wang, Neil Zhenqiang Gong, Diyi Yang, Xing Xie
Learn from Failure: Fine-Tuning LLMs with Trial-and-Error Data for Intuitionistic Propositional Logic Proving Chenyang An, Zhibo Chen, Qihao Ye, Emily First, Letian Peng, Jiayun Zhang, Zihan Wang, Sorin Lerner, Jingbo Shang
Control Large Language Models via Divide and Conquer Bingxuan Li, Yiwei Wang, Tao Meng, Kai-Wei Chang, Nanyun Peng
On the Resilience of Multi-Agent Systems with Malicious Agents Jen-tse Huang, Jiaxu Zhou, Tailin Jin, Xuhui Zhou, Zixi Chen, Wenxuan Wang, Youliang Yuan, Maarten Sap, Michael Lyu
ClimaQA: An Automated Evaluation Framework for Climate Foundation Models Veeramakali Vignesh Manivannan, Yasaman Jafari, Srikar Eranky, Spencer Ho, Rose Yu, Duncan Watson-Parris, Yian Ma, Leon Bergen, Taylor Berg-Kirkpatrick
Q-Debias: Quantum-Enhanced ActAdd-Guided Bias Reduction in LLMs Shardul Kulkarni, Rishi Gandhe, Ryunosuke Tatlock, Dylan De Schryver, Daniel Cho, Jonathan Lu, Kevin Zhu
Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective Zhaotian Weng, Zijun Gao, Jerone Andrews, Jieyu Zhao
Explain it to me like I’m five: Are LLM Explanations Effective for All Users? Brihi Joshi, Swabha Swayamdipta, Xiang Ren
Poster Session #2 (1:30pm - 2:30pm)
MultiAgent Collaboration Attack: Investigating Adversarial Attacks in Large Language Model Collaborations via Debate Alfonso Amayuelas
The African Languages Lab: Advancing NLP for Underrepresented African Languages through Global Collaboration Sheriff Issaka
Autonomous agents from automatic reward modeling and planning Rui Sun
Clustering Running Titles to Understand the Printing of Early Modern Books Nikolai Vogler, Kartik Goyal, Christopher Warren, Max G'Sell, Taylor Berg-Kirkpatrick
Evaluating K-Fold Cross Validation for Transformer Based Symbolic Regression Models Kaustubh Kislay, Shlok Singh, Rohan Dutta, Soham Joshi, Jay Shim, George Flint, Kevin Zhu
Optimizing LLMs with K-Shot Prompting for Synthetic Data Generation Vivaan Sehgal, Prahalad Anand, Brian Lin, Shridhar Garg, George Flint, Kevin Zhu
Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting Tim Knappe, Ryan Luo Li, Ayush Chauhan, Kaylee Chhua, Kevin Zhu, Sean O'Brien
Q*Agent: Optimizing Language Agents with Q-Guided Exploration Zongyu Lin, Yao Tang, Da Yin, Xingcheng Yao, Ziniu Hu, Yizhou Sun, Kai-Wei Chang
3D-LLM: Injecting the 3D World into Large Language Models Yining Hong
ContextRef: Evaluating referenceless metrics for image description generation Elisa Kreiss, Eric Zelikman, Christopher Potts, Nick Haber
A Debate-Driven Experiment on LLM Hallucinations and Accuracy Ruikun Li, Tanishka Bagade, Kevin Martinez, Flora Yasmin, Grant Ayala, Michael Lam, Kevin Zhu
Can Large Language Models Successfully Understand Errors? Jason Li, Lauren Gabrielle Yraola, Kevin Zhu, Sean O'Brien
Promoting Fairness in Link Prediction with Graph Enhancement Yezi Liu, Hanning Chen, Mohsen Imani
Stronger Random Baselines for In-Context Learning Gregory Yauney, David Mimno
Challenges in Cross-Lingual Transfer Learning for Offensive Language Detection: Insights from 30 Languages Elnaz Rahmati, Alireza Salkhordeh Ziabari, Morteza Dehghani
Beyond LLMs: A Linguistic Approach to Causal Graph Generation from Narrative Texts Zehan Li, Ruhua Pan, Xinyu Pi
BPO: Staying Close to the Behavior LLM Creates Better Online LLM Alignment Wenda Xu, Jiachen Li, William Yang Wang, Lei Li
Safety Alignment in NLP Tasks: Weakly Aligned Summarization as an In-Context Attack Yu Fu, Yufei Li, Wen Xiao, Cong Liu, Yue Dong
Are Large-Language Models Graph Algorithmic Reasoners? Alexander K Taylor, Anthony Cuturrufo, Vishal Yathish, Mingyu Derek Ma, Wei Wang
Can Language Models Learn to Skip Steps? Tengxiao Liu, Qipeng Guo, Xiangkun Hu, Cheng Jiayang, Yue Zhang, Xipeng Qiu, Zheng Zhang
Open-world Multi-label Text Classification with Extremely Weak Supervision Xintong Li, Jinya Jiang, Ria Dharmani, Jayanth Srinivasa, Gaowen Liu, Jingbo Shang
The Factuality Tax of Diversity-Intervened Text-to-Image Generation: Benchmark and Fact-Augmented Intervention Yixin Wan, Di Wu, Haoran Wang, Kai-Wei Chang
Enhancing Transparency in RLHF: An Explainable Reward Framework for Aligning Large Language Models Yiran Shen, Aditya Emmanuel Arokiaraj John, Brandon Fain
Dialect Bias in Affective Computing: African American English and Text-Based Emotion Classification Rebecca Dorn, Christina A Chance, Casandra Rusti
Cross-Task Generalization Abilities of Large Language Models Qinyuan Ye
Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling Hritik Bansal, Arian Hosseini, Rishabh Agarwal, Vinh Q. Tran, Mehran Kazemi
Machine Learning Research Transparency through Positionality Statements Christina A Chance, Rebecca Pattichis
Comparing Bad Apples to Good Oranges: Aligning Large Language Models via Joint Preference Optimization Hritik Bansal, Ashima Suvarna, Gantavya Bhatt, Nanyun Peng, Kai-Wei Chang, Aditya Grover
Matryoshka Query Transformer for Large Vision-Language Models Wenbo Hu, Zi-Yi Dou, Liunian Harold Li, Amita Kamath, Nanyun Peng, Kai-Wei Chang
Dynamic Rewarding with Prompt Optimization Enables Tuning-free Self-Alignment of Language Models Somanshu Singla, Zhen Wang, Tianyang Liu, Abdullah Ashfaq, Zhiting Hu, Eric P. Xing
Sufficient Context: A New Lens on Retrieval Augmented Generation Systems Hailey Joren, Jianyi Zhang, Chun-Sung Ferng, Da-Cheng Juan, Ankur Taly, Cyrus Rashtchian
LLM-Coordination: Evaluating and Analyzing Multi-agent Coordination Abilities in Large Language Models Saaket Agashe, Yue Fan, Anthony Reyna, Xin Eric Wang
Speechworthy Instruction-tuned Language Models Hyundong Justin Cho, Nicolaas Paul Jedema, Leonardo F. R. Ribeiro, Karishma Sharma, Pedro Szekely, Alessandro Moschitti, Ruben Janssen, Jonathan May
EditRoom: LLM-parameterized Graph Diffusion for Composable 3D Room Layout Editing Kaizhi Zheng, Xiaotong Chen, Xuehai He, Jing Gu, Linjie Li, Zhengyuan Yang, Kevin Lin, Jianfeng Wang, Lijuan Wang, Xin Eric Wang
Using Linguistic Entrainment to Evaluate Large Language Models for use in Cognitive Behavioral Therapy Mina Kian, Kaleen Shrestha, Katrin Fischer, Xiaoyuan Zhu, Jonathan Ong, Aryan Trehan, Jessica Wang, Gloria Chang, Séb Arnold, Maja Mataric
VSP: Assessing the dual challenges of perception and reasoning in spatial planning tasks for MLLMs Qiucheng Wu, Handong Zhao, Michael Saxon, Trung Bui, William Yang Wang, Yang Zhang, Shiyu Chang
Quantifying Generalization Complexity for Large Language Models Zhenting Qi, Hongyin Luo, Xuliang Huang, Zhuokai Zhao, Yibo Jiang, Xiangjun Fan, Himabindu Lakkaraju, James R. Glass
LlamaTales: Studying the Effects of Training Data on Small Language Models Ivan Lee, Taylor Berg-Kirkpatrick
A Comparative Study of Translation Bias and Accuracy in Multilingual Large Language Models for Cross-Language Claim Verification Aryan Singhal, Veronica Shao, Gary Sun, Ryan Ding, Jonathan Lu, Kevin Zhu
Town Hall Debate Prompting: Enhancing Logical Reasoning in LLMs through Multi-Persona Interaction Bhav Jain, Vivaan Sandwar, Ishaan Garg, Rishan Thangaraj, Michael Lam, Kevin Zhu
StyleRemix: Interpretable Authorship Obfuscation via Distillation and Perturbation of Style Elements Jillian Fisher, Skyler Hallinan, Ximing Lu, Mitchell L Gordon, Zaid Harchaoui, Yejin Choi
Goal-Driven Explainable Clustering via Language Descriptions Zihan Wang, Jingbo Shang, Ruiqi Zhong
Inferring from Logits: Exploring Best Practices for Decoding-Free Generative Candidate Selection Mingyu Derek Ma, Yanna Ding, Zijie Huang, Jianxi Gao, Yizhou Sun, Wei Wang
Follow My Instruction and Spill the Beans: Scalable Data Extraction from Retrieval-Augmented Generation Systems Zhenting Qi, Hanlin Zhang, Eric P. Xing, Sham M. Kakade, Himabindu Lakkaraju
Generative Verifiers: Reward Modeling as Next-Token Prediction Lunjun Zhang, Arian Hosseini, Hritik Bansal, Mehran Kazemi, Aviral Kumar, Rishabh Agarwal
Leveraging Language Models to Detect Greenwashing Margaret Capetz, Christina A Chance, Rebecca Pattichis, Reshmi Ghosh, Avalon Vinella
Asymmetric Bias in Text-to-Image Generation with Adversarial Attacks Haz Sameen Shahgir, Xianghao Kong, Greg Ver Steeg, Yue Dong
Enhancing Knowledge Distillation for LLMs with Response-Priming Prompting Vijay Goyal, Mustafa Khan, Aprameya Tirupati, Harveer Saini, Michael Lam, Kevin Zhu
Learning a Decision Tree Algorithm with Transformers Yufan Zhuang, Liyuan Liu, Chandan Singh, Jingbo Shang, Jianfeng Gao
Evaluating Creativity and Deception in Large Language Models: A Simulation Framework for Multi-Agent Balderdash Parsa Hejabi, Elnaz Rahmati, Alireza Salkhordeh Ziabari, Preni Golazizian, Jesse Thomason, Morteza Dehghani
Planning as Inpainting: A Generative Framework for Realistic Embodied Path Planning Cheng-Fu Yang, Kai-Wei Chang
Multi-step Problem Solving Through a Verifier: An Empirical Analysis on Model-induced Process Supervision Zihan Wang, Yunxuan Li, Yuexin Wu, Liangchen Luo, Le Hou, Hongkun Yu, Jingbo Shang
Poster Session #3 (2:40pm - 3:40pm)
Pre-trained Large Language Models Use Fourier Features to Compute Addition Tianyi Zhou, Deqing Fu, Vatsal Sharan, Robin Jia
NusaMT-7B: Machine Translation for Low-Resource Indonesian Languages with Large Language Models William Tan, Kevin Zhu
Adaptive In-conversation Team Building for Language Model Agents Linxin Song, Jiale Liu, Jieyu Zhang, Shaokun Zhang, Ao Luo, Shijian Wang, Qingyun Wu, Chi Wang
PatentEdits: Framing Patent Novelty as Textual Entailment Ryan Lee, Alexander Spangher, Xuezhe Ma
Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue Jia-Chen Gu, Hao-Xiang Xu, Jun-Yu Ma, Pan Lu, Zhen-Hua Ling, Kai-Wei Chang, Nanyun Peng
Fictitious Synthetic Data Can Improve LLM Factuality via Prerequisite Learning Yujian Liu, Shiyu Chang, Tommi Jaakkola, Yang Zhang
Adversarial Attacks on Parts of Speech: An Empirical Study in Text-to-Image Generation G M Shahariar, Jia Chen, Jiachen Li, Yue Dong
SoMeR: A Multi-View Social Media User Representation Learning Framework Keith Burghardt, Siyi Guo, Valeria Pantè, Kristina Lerman
Communicate to Play: Pragmatic Reasoning for Efficient Cross-Cultural Communication Isadora White, Sashrika Pandey, Michelle Pan
Weak-to-Strong Reasoning Yuqing Yang, Yan Ma, Pengfei Liu
MARVEL: Multidimensional Abstraction and Reasoning through Visual Evaluation and Learning Yifan Jiang, Jiarui Zhang, Kexuan Sun, Zhivar Sourati, Kian Ahrabian, Kaixin Ma, Filip Ilievski, Jay Pujara
STAR: A Simple Training-free Approach for Recommendations using Large Language Models Dong-Ho Lee, Adam Kraft, Long Jin, Nikhil Mehta, Taibai Xu, Lichan Hong, Ed H. Chi, Xinyang Yi
Ladder Residual: Redefining Tensor Parallelism in Transformers for Accelerated Inference Muru Zhang, Mayank Mishra, Zhongzhu Zhou, William Brandon, Jue WANG, Yoon Kim, Jonathan Ragan-Kelley, Shuaiwen Leon Song, Ben Athiwaratkun, Tri Dao
CLEAR: Contrastive Learning with Experts and Amateurs for Reasoning Andrew Rufail, Abenezer Tessema, Buse Toksoz, Chizoba Okoli, Jonathan Pei, Kevin Zhu
Layer Swapping for Zero-Shot Cross-Lingual Transfer in Large Language Models Lucas Bandarkar
MORL-Prompt: An Empirical Analysis of Multi-Objective Reinforcement Learning for Discrete Prompt Optimization Yasaman Jafari, Dheeraj Mekala, Rose Yu, Taylor Berg-Kirkpatrick
Evaluating Cultural and Social Awareness of LLM Web Agents Haoyi Qiu, Alexander Fabbri, Divyansh Agarwal, Kung-Hsiang Huang, Sarah Tan, Nanyun Peng, Chien-Sheng Wu
VideoPhy: Evaluating Physical Commonsense for Video Generation Hritik Bansal, Zongyu Lin, Tianyi Xie, Zeshun Zong, Michal Yarom, Yonatan Bitton, Chenfanfu Jiang, Yizhou Sun, Kai-Wei Chang, Aditya Grover
COMMUNITY-CROSS-INSTRUCT: Unsupervised Instruction Generation for Aligning Large Language Models to Online Communities Zihao He, Minh Duc Chu, Rebecca Dorn, Siyi Guo, Kristina Lerman
Exploring Synthetic Datasets for Large Language Model Unlearning Xiaoyuan Zhu, Ollie Liu, Muru Zhang, Willie Neiswanger
MAPO: Momentum-Aided Gradient Descent Prompt Optimization Anthony Cui, Pranav Nandyalam, Kevin Zhu
Out-of-Distribution Detection through Soft Clustering with Non-Negative Kernel Regression Aryan Gulati, Xingjian Dong, Carlos Hurtado, Sarath Shekkizhar, Swabha Swayamdipta, Antonio Ortega
CoLLAP: Contrastive Long-form Language-Audio Pretraining with Musical Temporal Structure Augmentation Junda Wu, Warren Li, Zachary Novack, Amit Namburi, Carol Chen, Julian McAuley
Data Contamination Can Cross Language Barriers Feng Yao, Yufan Zhuang, Zihao Sun, Sunan Xu, Animesh Kumar, Jingbo Shang
Quantifying and Optimizing Global Faithfulness in Persona-driven Role-playing Letian Peng, Jingbo Shang
DiversityMedQA: A Benchmark for Assessing Demographic Biases in Medical Diagnosis using Large Language Models Rajat Rawat, Hudson McBride, Rajarshi Ghosh, Dhiyaan Chakkresh Nirmal, Jong Moon, Dhruv Karthik Alamuri, Sean O'Brien, Kevin Zhu
Language Models Can Infer Action Semantics for Symbolic Planners from Environment Feedback Wang Zhu, Ishika Singh, Robin Jia, Jesse Thomason
A Probabilistic Framework for LLM Hallucination Detection via Belief Tree Propagation Bairu Hou, Yang Zhang, Jacob Andreas, Shiyu Chang
GPT is Not an Annotator: The Necessity of Human Annotation in Fairness Benchmark Construction Virginia K. Felkner, Jennifer A. Thompson, Jonathan May
Adaptable Logical Control for Large Language Models Honghua Zhang, Po-Nien Kung, Masahiro Yoshida, Guy Van den Broeck, Nanyun Peng
Strategic Moves: Negotiation Agents with Dynamic Adaptation to Opponent Behavior for Optimizing Counter-Offers Deuksin Kwon, Jiwon Hae, Gale Lucas
Solving for X and Beyond: Can Large Language Models Solve Complex Math Problems with More-Than-Two Unknowns? Kuei-Chun Kao, Ruochen Wang, Cho-Jui Hsieh
Capturing the Essence of a Phrase: Extracting Physical and Sensory Information from Text Abhinav Gupta
Efficient LLM Scheduling by Learning to Rank Yichao Fu, Siqi Zhu, Runlong Su, Aurick Qiao, Ion Stoica, Hao Zhang
Improving LLM Abilities in Idiomatic Translation Maximilian Spencer, Sundesh Donthi, Om Patel, Joon Young Doh, Eid Rodan
Large Language Models Can Be Contextual Privacy Protection Learners Yijia Xiao, Yiqiao Jin, Yushi Bai, Yue Wu, Xianjun Yang, Xiao Luo, Wenchao Yu, Xujiang Zhao, Yanchi Liu, Quanquan Gu, Haifeng Chen, Wei Wang, Wei Cheng
TrimLLM: Progressive Layer Dropping for Efficient LLM Fine-tuning and Inference Lanxiang Hu, Tajana Rosing, Hao Zhang
Revisiting Representation Collapse in Language Models Atharva Kulkarni, Swabha Swayamdipta
Logits of API-Protected LLMs Leak Proprietary Information Matthew Finlayson, Swabha Swayamdipta, Xiang Ren
Grammar-Aligned Decoding Kanghee Park, Jiayu Wang, Taylor Berg-Kirkpatrick, Nadia Polikarpova, Loris D'Antoni
Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented Generation Di Wu, Jia-Chen Gu, Fan Yin, Nanyun Peng, Kai-Wei Chang
Who Evaluates the Evaluations? Objectively Scoring Text-to-Image Prompt Coherence Metrics with T2IScoreScore Michael Saxon, Fatima Jahara, Mahsa Khoshnoodi, Yujie Lu, Aditya Sharma, William Yang Wang
When Parts Are Greater Than Sums: Individual LLM Components Can Outperform Full Models Ting-Yun Chang, Jesse Thomason, Robin Jia
Evaluating LLM Chatbot-Generated Answers to Patient Questions Wang Zhu, Tian-qi Chen, Ruishan Liu, Robin Jia
The Strong Pull of Prior Knowledge in Large Language Models and Its Impact on Emotion Recognition Georgios Chochlakis, Alexandros Potamianos, Kristina Lerman, Shrikanth Narayanan
Firm’s AI Exposure without Labor Data: The Expected Impact of AI Jiacheng Liu
Fine-Tuning Language Models for Ethical Ambiguity: A Comparative Study of Alignment with Human Responses Pranav Senthilkumar, Visshwa Balasubramanian, Prisha Jain, Aneesa Maity, Jonathan Lu, Kevin Zhu
SPEED++: A Multilingual Event Extraction Framework for Epidemic Prediction and Preparedness Tanmay Parekh, Jeffrey Kwan, Jiarui Yu, Sparsh Johri, Hyosang Ahn, Sreya Muppalla, Kai-Wei Chang, Wei Wang, Nanyun Peng
Stackelberg Games for Persuasive Advocacy in LLM Simulations Xinyue Cui, Swabha Swayamdipta
Style-Compress: An LLM-Based Prompt Compression Framework Considering Task-Specific Styles Xiao Pu, Tianxing He, Xiaojun Wan
BIASDETECTOR: Multi-Agent Synergy for Comprehensive Bias Detection in Structural Data Haoxuan Li, Mingyu Derek Ma, Jen-tse Huang, Wei Wang, Jieyu Zhao
mDPO: Conditional Preference Optimization for Multimodal Large Language Models Fei Wang, Wenxuan Zhou, James Y. Huang, Nan Xu, Sheng Zhang, Hoifung Poon, Muhao Chen
Linear Layer Extrapolation for Fine-Grained Emotion Classification Mayukh Sharma, Sean O'Brien, Julian McAuley
NewsHomepages: Homepage Layouts Capture Information Prioritization Decisions Arda Kaz, Alexander Spangher, Michael Vu, Naitian Zhou, Ben Welsh
Predictive Modeling for Early Alzheimer's Disease Using Natural Language Processing Azra Emekci