Welcome


The goal of the Southern California Natural Language Processing Symposium is to gather researchers from the southern California region with broad expertise in machine learning and natural language processing. The symposium will provide a participatory environment where attendees from a variety of fields can share and discuss their latest findings.

Important Dates

Submission deadline: Oct 26 (Saturday), 2024, 11:59 PM Anywhere On Earth [OpenReview Submission Portal]
Acceptance notification: Nov. 1, 2024
Registration deadline: Nov 10, 2024 [Eventbrite Registration Link]
Symposium date: Nov 22, 2024 (Friday)
Submission format:

  • If the paper is published in a recent conference or journal, you can directly submit the paper without modification. Please indicate where the paper is published in the Abstract when submitting the paper.
  • For papers that have not been previously published, we recommend submitting an extended summary spanning no more than two pages, formatted according to the ACL guidelines, However, submissions of a longer length will also be considered.
  • All submissions are single-blind reviewed.


Venue


Date: Nov 22, 2024

Locations: There are two venues involved in this event. They are close to each other.

Breakfast, Lunch, and Poster: Great Hall



Talk: Institute of the America's Hojel Auditorium



Parking information: Campus parking is very limited and is not recommended. You can check available options here. The closest parking lot for visitors is located on the top floor of Hopkins Parking Structure. You can pay the fee in pay station or pay-by-phone. Note that most visitor spots are limited to two hours. Even though the app allows you to pay for longer periods, you will get a ticket after that time.


There is free parking at Torrey Pines Gliderport, if you do not mind a 30-minute walk to the venue. The view from the Gliderport is spectacular.
View from the Torrey Pines Gliderport. Photo credit: Gina Trapani (CC BY-NC-SA 2.0 Deed)


Schedule


Great Hall
09:00am - 09:45am Breakfast, Registration & Social

Institute of the America's Hojel Auditorium
09:45am - 10:00am Opening Remarks
10:00am - 10:40am Heng Ji (UIUC)
Towards Knowledgeable Foundation Models
10:40pm - 11:20am Eunsol Choi (NYU)
Equipping LLMs for Complex Knowledge Scenarios: Interaction and Retrieval
11:20am - 11:30am Mini Break

Great Hall
11:30am - 12:30pm Poster Session 1
12:00pm - 02:00pm Lunch Reception & AIX-sponsored Social Hours
01:30pm - 02:30pm Poster Session 2
02:30pm - 02:40pm Mini Break
02:40pm - 03:40pm Poster Session 3

Institute of the America's Hojel Auditorium
04:00pm - 04:40pm Dawei Huang (SambaNova.ai)
Enabling extremely fast inference and training performance using dataflow and custom chip
04:40pm - 05:20pm Taylor Berg-Kirkpatrick (UCSD)
Large Language Models and Cybersecurity: New Threats and Safeguards
05:20pm - 05:30pm Closing Remarks & Awards


Invited Speakers


Heng Ji

Professor

Computer Science

UIUC

Eunsol Choi

Assistant Professor

Computer Science and Data Science

NYU

Dawei Huang

Senior Director of Engineering

SambaNova Systems

Taylor Berg-Kirkpatrick

Associate Professor

Computer Science

UCSD



Heng Ji
Title: Towards Knowledgeable Foundation Models

Abstract: Large language models (LLMs) and vision-language models (VLMs) have demonstrated remarkable performance on knowledge reasoning tasks, owing to their implicit knowledge derived from extensive pretraining data. However, their inherent knowledge bases often suffer from disorganization and illusion, bias towards common entities, and rapid obsolescence. Consequently, LLMs frequently make up untruthful information, exhibit resistance to updating outdated knowledge, or struggle with generalizing across multiple languages. In this talk I will discuss several research directions that aim to make foundation models’ knowledge more accurate, organized, up-to-date and fair: (1) Where and How is Knowledge Stored in LLM? (2) How to Control LLM’s Knowledge? (3) How to Update LLM’s Dynamic Knowledge? (4) How to Bridge the Knowledge Gap between Natural Language and Unnatural Language?

Bio: Heng Ji is a professor at Siebel School of Computing and Data Science, and an affiliated faculty member at Electrical and Computer Engineering Department, Coordinated Science Laboratory, and Carl R. Woese Institute for Genomic Biology of University of Illinois Urbana-Champaign. She is an Amazon Scholar. She is the Founding Director of Amazon-Illinois Center on AI for Interactive Conversational Experiences (AICE). She received her B.A. and M. A. in Computational Linguistics from Tsinghua University, and her M.S. and Ph.D. in Computer Science from New York University. Her research interests focus on Natural Language Processing, especially on Multimedia Multilingual Information Extraction, Knowledge-enhanced Large Language Models and Vision-Language Models, and AI for Science. The awards she received include Outstanding Paper Award at ACL2024, two Outstanding Paper Awards at NAACL2024, "Young Scientist" by the World Laureates Association in 2023 and 2024, "Young Scientist" and a member of the Global Future Council on the Future of Computing by the World Economic Forum in 2016 and 2017, "Women Leaders of Conversational AI" (Class of 2023) by Project Voice, "AI's 10 to Watch" Award by IEEE Intelligent Systems in 2013, NSF CAREER award in 2009, PACLIC2012 Best paper runner-up, "Best of ICDM2013" paper award, "Best of SDM2013" paper award, ACL2018 Best Demo paper nomination, ACL2020 Best Demo Paper Award, NAACL2021 Best Demo Paper Award, Google Research Award in 2009 and 2014, IBM Watson Faculty Award in 2012 and 2014 and Bosch Research Award in 2014-2018. She served as the associate editor for IEEE/ACM Transaction on Audio, Speech, and Language Processing, and the Program Committee Co-Chair of many conferences including NAACL-HLT2018 and AACL-IJCNLP2022. She was elected as the North American Chapter of the Association for Computational Linguistics (NAACL) secretary 2020-2023.


Eunsol Choi
Title: Equipping LLMs for Complex Knowledge Scenarios: Interaction and Retrieval

Abstract: Language models are increasingly used as an interface to gather information. Yet trusting the answers generated from LMs is risky, as they often contain incorrect or misleading information. Why is this happening? We identify two key issues: (1) ambiguous and underspecified user questions and (2) imperfect knowledge in LMs, especially for long tail or recent events. To address the first issue, we propose a system that can interact with users to clarify their intent before answering. By simulating their expected outcomes in the future turns, we reward LMs for generating clarifying questions and not just answering immediately. In the second part of the talk, I will discuss the state of retrieval augmentation, which is often lauded as the path to provide up-to-date, relevant knowledge to LMs. While their success is evident in scenarios where there exists a single gold document, incorporating information from a diverse set of documents remains challenging for both retrievers and LMs. Together, the talk highlights key research directions for building reliable LMs to answer information seeking questions.

Bio: Eunsol Choi is an assistant professor of computer science and data science at New York University. Her research spans natural language processing and machine learning, with a focus on interpreting and reasoning about text in dynamic real-world contexts. Prior to joining NYU, she was an assistant professor at the University of Texas at Austin. She also spent a year at Google AI as a visiting researcher. She holds a Ph.D. in computer science and engineering from the University of Washington. She is a recipient of a Facebook research fellowship, Google faculty research award, Sony faculty award, and an outstanding paper award at EMNLP.


Dawei Huang
Title: Enabling extremely fast inference and training performance using dataflow and custom chip

Abstract: As the pursuit of larger language models continues to push the boundaries of computational demands, the traditional silicon chip is facing a daunting memory/power wall. In this talk, we present a novel chip design, SN40L, to tackle this challenge. This chip combines a reconfigurable data-flow architecture with a tightly coupled 3-tier memory hierarchy to enable efficient compute-intensive training and memory-bound inference workloads for a wide variety of neural network architectures. We discuss the various advantages of this chip via case studies. In our first case study, we discuss how the dataflow architecture coupled with on-chip SRAM and HBM empowers operation fusion capabilities enabling 1000+ tokens/second inference performance without sacrificing on precision. In the second case study we look at the training performance of various model architectures and compare their performance against traditional kernel by kernel execution-based architectures. We show how dataflow architecture can help accelerate LLM training for traditional dense, sparse and novel state space models, while allowing one to train extremely large models on a smaller footprint. In the third case study we will discuss how you can use the strongly coupled DRAM, HBM and SRAM to develop new neural network architecture that can help scale to 100+ billion parameters efficiently. This new architecture uses a modular and coarse-grained approach to a mixture of experts that allows for incremental updates to models for new capabilities and knowledge and smaller footprint execution during inference.

Bio: Dawei Huang received BS from Tsinghua University and MS from UCSD. He has 25+ years experience in Computer/AI Systems and has 47 issued patents. Currently, he is Senior Director of Engineering at Sambanova Systems. His current responsibilities include ML Model Performance and Customer Solution.


Taylor Berg-Kirkpatrick
Title: Large Language Models and Cybersecurity: New Threats and Safeguards

Abstract: The rapid development of large-language models has created new attack surfaces for adversaries, but at the same time has enabled new mitigations for human-driven fraud and social engineering attacks. This talk will consider two new interfaces between NLP and computer security: (1) the potential for new types of attacks as LLMs are increasingly treated like operating systems and (2) the possibility of using LLMs to combat and track social engineering attacks at scale via LLM-driven honeypots.

Bio: Taylor Berg-Kirkpatrick is an Associate Professor in the Computer Science and Engineering Department at the University of California San Diego. Taylor's research focuses on using machine learning to understand structured human data, including language but also sources like music, document images, and other complex artifacts.



Accepted Work

Non-archival, randomly ordered


Poster Session #1 (11:30am - 12:30pm)

  1. CLIMB: A Benchmark of Clinical Bias in Large Language Models
    Yubo Zhang, Shudi Hou, Mingyu Derek Ma, Wei Wang, Muhao Chen, Jieyu Zhao
  2. Exploring the Design Space of Diffusion Bridge Models via Stochasticity Control
    Shaorong Zhang, Yuanbin Cheng, Xianghao Kong, Greg Ver Steeg
  3. A Novel Multi-Document Retrieval Benchmark Grounded on Journalist Source-Selection in Newswriting
    Alexander Spangher, Tenghao Huang, Yiqin Huang, Liheng Lai, Lucas Spangher, Sewon Min, Mark Dredze
  4. Chain-of-Thought Augmentation with Logit Contrast for Enhanced Reasoning in Language Models
    Jay Shim, Grant Kruttschnitt, Alyssa Ma, Daniel Kim, Benjamin Chek, Athul Anand, Kevin Zhu, Sean O'Brien
  5. Improving Inference on Theory of Mind with Evidence Chain
    Qiutong Tony Yi, Wang Zhu
  6. An Audio checkup for Language Models
    Arjun Prasaath Anbazhagan, Parteek Kumar, Souritra Kar, Ujjwal Kaur
  7. REFFLY: Melody-Constrained Lyrics Editing Model
    Songyan Zhao, Bingxuan Li, Yufei Tian, Nanyun Peng
  8. Compare without Despair: Reliable Preference Evaluation with Generation Separability
    Sayan Ghosh, Tejas Srinivasan, Swabha Swayamdipta
  9. Towards Interventions for Suicide Prevention with LLM Assistants in Social Work
    Jaspreet Ranjit, Hyundong Justin Cho, Myles Phung, John R. Blosnich, Swabha Swayamdipta
  10. Verify with Caution: The Pitfalls of Relying on Imperfect Factuality Metrics
    Ameya Godbole, Robin Jia
  11. The Dark Side of Instagram: A Large Dataset for Identifying Persian Harmful Comments
    Hadi Davardoust, Hadi Zare, Hossein Rafiee zade
  12. Eliciting Better Multilingual Structured Reasoning from LLMs through Code
    Bryan Li, Tamer Alkhouli, Daniele Bonadiman, Nikolaos Pappas, Saab Mansour
  13. Unlocking Decoding-Time Controllability: Gradient-Free Multi-Objective Alignment with Contrastive Prompts
    Tingchen Fu, Yupeng Hou, Julian McAuley, Rui Yan
  14. AgentGrow: LLMs as Scalable, Customizable General-Purpose Simulators For Language Agent Training
    Yiming Wang, Yuedong Cui, Da Yin, Zongyu Lin, Di Wu, Xueqing Wu, Chenchen Ye, Kai-Wei Chang
  15. SWE-Search: Enhancing Software Agents with Monte Carlo Tree Search and Iterative Refinement
    Antonis Antoniades, Albert Örwall, Kexun Zhang, Yuxi Xie, Anirudh Goyal, William Yang Wang
  16. INTERINTENT: Investigating Social Intelligence of LLMs via Intention Understanding in an Interactive Game Context
    Ziyi Liu, Abhishek Anand, Pei Zhou, Jen-tse Huang, Jieyu Zhao
  17. Contrastive Instruction Tuning
    Tianyi Lorena Yan, Fei Wang, James Y. Huang, Wenxuan Zhou, Fan Yin, Aram Galstyan, Wenpeng Yin, Muhao Chen
  18. Embodied Agent Interface: Benchmarking LLMs for Embodied Decision Making
    Manling Li, Shiyu Zhao, Qineng Wang, Kangrui Wang, Yu Zhou, Sanjana Srivastava, Cem Gokmen, Tony Lee, Li Erran Li, Ruohan Zhang, Weiyu Liu, Percy Liang, Li Fei-Fei, Jiayuan Mao, Jiajun Wu
  19. Teaching Models to Understand but Not Generate
    Ryan Yixiang Wang, Matthew Finlayson, Luca Soldaini, Robin Jia, Swabha Swayamdipta
  20. IsoBench: Benchmarking Multimodal Foundation Models on Isomorphic Representations
    Deqing Fu, Ruohao Guo, Ghazal Khalighinejad, Ollie Liu, Bhuwan Dhingra, Dani Yogatama, Robin Jia, Willie Neiswanger
  21. IllusionVQA: A Challenging Optical Illusion Dataset for Vision Language Models
    Haz Sameen Shahgir, Khondker Salman Sayeed, Abhik Bhattacharjee, Wasi Uddin Ahmad, Yue Dong, Rifat Shahriyar
  22. Stress-Testing Long-Context Language Models with Lifelong ICL and Task Haystack
    Xiaoyue Xu, Qinyuan Ye, Xiang Ren
  23. Can Textual Unlearning Solve Cross-Modality Safety Alignment?
    Trishna Chakraborty, Erfan Shayegani, Zikui Cai, Nael Abu-Ghazaleh, M. Salman Asif, Yue Dong, Amit Roy-Chowdhury, Chengyu Song
  24. VDebugger: Harnessing Execution Feedback for Debugging Visual Programs
    Xueqing Wu, Zongyu Lin, Songyan Zhao, Te-Lin Wu, Pan Lu, Nanyun Peng, Kai-Wei Chang
  25. Evaluating Grounding Gaps in LLMs via News Interviews
    Michael Lu, Hyundong Justin Cho, Weiyan Shi, Jonathan May, Alexander Spangher
  26. LogogramNLP: Comparing Visual and Textual Representations of Ancient Logographic Writing Systems for NLP
    Danlu Chen, Freda Shi, Aditi Agarwal, Jacobo Myerston, Taylor Berg-Kirkpatrick
  27. Question-Analysis Prompting Improves LLM Performance in Reasoning Tasks
    Dharunish Yugeswardeenoo, Kevin Zhu, Sean O'Brien
  28. AgentReview: Exploring Peer Review Dynamics with LLM Agents
    Yiqiao Jin, Qinlin Zhao, Yiyang Wang, Hao Chen, Kaijie Zhu, Yijia Xiao, Jindong Wang
  29. Answer is All You Need: Instruction-following Text Embedding via Answering the Question
    Letian Peng, Yuwei Zhang, Zilong Wang, Jayanth Srinivasa, Gaowen Liu, Zihan Wang, Jingbo Shang
  30. Socratic Mind: Scalable Oral Assessment Powered By AI
    Jui-Tse Hung, Christopher Zhang Cui, Diana M. Popescu, Saurabh Chatterjee, Thad Starner
  31. Will Trump Win in 2024? Predicting the US Presidential Election via Multi-step Reasoning with Large Language Models
    Chenxiao Yu, Zhaotian Weng, Xiyang Hu, Yue Zhao
  32. Faithful Persona-based Conversational Dataset Generation with Large Language Models
    Pegah Jandaghi, Xianghai Sheng, Xinyi Bai, Jay Pujara, Hakim Sidahmed
  33. Exploring Scientific Hypothesis Generation with Mamba
    Miaosen Chai, Emily Herron, Erick Cervantes, Tirthankar Ghosal
  34. From Bias to Balance: Detecting Facial Expression Recognition Biases in Multimodal Foundation Models
    Kaylee Chhua, Zhoujinyi Wen, Vedant Hathalia, Kevin Zhu, Sean O'Brien
  35. AAVENUE: Detecting LLM Biases on NLU Tasks in AAVE via a Novel Benchmark
    Abhay Gupta, Ece Yurtseven, Philip Meng, Sean O'Brien, Kevin Zhu
  36. Language Models Implicitly Learn a Unified Representation Space
    Zhaofeng Wu, Xinyan Velocity Yu, Dani Yogatama, Jiasen Lu, Yoon Kim
  37. Scaling LLM Inference with Optimized Sample Compute Allocation
    Kexun Zhang, Shang Zhou, Danqing Wang, William Yang Wang, Lei Li
  38. Evaluating Human Alignment and Model Faithfulness of LLM Rationale
    Mohsen Fayyaz, Fan Yin, Jiao Sun, Nanyun Peng
  39. Guiding Through Complexity: What Makes Good Supervision for Hard Reasoning Tasks?
    Xuan He, Da Yin, Nanyun Peng
  40. A Little Human Data Goes A Long Way
    Dhananjay Ashok, Jonathan May
  41. Enhancing Depression Diagnosis with Chain-of-Thought Prompting
    Elysia Shi, Adithri Manda, London Chowdhury, Runeema Arun, Kevin Zhu, Michael Lam
  42. ChunkRAG: Novel LLM-Chunk Filtering Method for RAG Systems
    Ishneet Sukhvinder Singh, Ritvik Aggarwal, Ibrahim Allahverdiyev, Muhammad Taha, Aslihan Akalin, Kevin Zhu, Sean O'Brien
  43. Reversing the Forget-Retain Objectives: An Efficient LLM Unlearning Framework from Logit Difference
    Jiabao Ji, Yujian Liu, Yang Zhang, Gaowen Liu, Ramana Rao Kompella, Sijia Liu, Shiyu Chang
  44. DeLLMa: Decision Making Under Uncertainty with Large Language Models
    Ollie Liu, Deqing Fu, Dani Yogatama, Willie Neiswanger
  45. Apathetic or Empathetic? Evaluating LLMs' Emotional Alignments with Humans
    Jen-tse Huang, Man Ho LAM, Eric John Li, Shujie Ren, Wenxuan Wang, Wenxiang Jiao, Zhaopeng Tu, Michael Lyu
  46. Interleaved Multimodal Decision Transformer for Embodied Agents
    Bosung Kim
  47. DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks
    Kaijie Zhu, Jiaao Chen, Jindong Wang, Neil Zhenqiang Gong, Diyi Yang, Xing Xie
  48. Learn from Failure: Fine-Tuning LLMs with Trial-and-Error Data for Intuitionistic Propositional Logic Proving
    Chenyang An, Zhibo Chen, Qihao Ye, Emily First, Letian Peng, Jiayun Zhang, Zihan Wang, Sorin Lerner, Jingbo Shang
  49. Control Large Language Models via Divide and Conquer
    Bingxuan Li, Yiwei Wang, Tao Meng, Kai-Wei Chang, Nanyun Peng
  50. On the Resilience of Multi-Agent Systems with Malicious Agents
    Jen-tse Huang, Jiaxu Zhou, Tailin Jin, Xuhui Zhou, Zixi Chen, Wenxuan Wang, Youliang Yuan, Maarten Sap, Michael Lyu
  51. Euclid: Supercharging Multimodal LLMs with Synthetic High-Fidelity Visual Descriptions
    Jiarui Zhang, Ollie Liu, Tianyu Yu, Jinyi Hu, Willie Neiswanger
  52. ClimaQA: An Automated Evaluation Framework for Climate Foundation Models
    Veeramakali Vignesh Manivannan, Yasaman Jafari, Srikar Eranky, Spencer Ho, Rose Yu, Duncan Watson-Parris, Yian Ma, Leon Bergen, Taylor Berg-Kirkpatrick
  53. Q-Debias: Quantum-Enhanced ActAdd-Guided Bias Reduction in LLMs
    Shardul Kulkarni, Rishi Gandhe, Ryunosuke Tatlock, Dylan De Schryver, Daniel Cho, Jonathan Lu, Kevin Zhu
  54. Images Speak Louder than Words: Understanding and Mitigating Bias in Vision-Language Model from a Causal Mediation Perspective
    Zhaotian Weng, Zijun Gao, Jerone Andrews, Jieyu Zhao
  55. Explain it to me like I’m five: Are LLM Explanations Effective for All Users?
    Brihi Joshi, Swabha Swayamdipta, Xiang Ren

Poster Session #2 (1:30pm - 2:30pm)

  1. MultiAgent Collaboration Attack: Investigating Adversarial Attacks in Large Language Model Collaborations via Debate
    Alfonso Amayuelas
  2. The African Languages Lab: Advancing NLP for Underrepresented African Languages through Global Collaboration
    Sheriff Issaka
  3. Autonomous agents from automatic reward modeling and planning
    Rui Sun
  4. Simplicity Bias of Transformers to Learn Low Sensitivity Functions
    Bhavya Vasudeva, Deqing Fu, Tianyi Zhou, Elliott Kau, Youqi Huang, Vatsal Sharan
  5. Clustering Running Titles to Understand the Printing of Early Modern Books
    Nikolai Vogler, Kartik Goyal, Christopher Warren, Max G'Sell, Taylor Berg-Kirkpatrick
  6. Evaluating K-Fold Cross Validation for Transformer Based Symbolic Regression Models
    Kaustubh Kislay, Shlok Singh, Rohan Dutta, Soham Joshi, Jay Shim, George Flint, Kevin Zhu
  7. Optimizing LLMs with K-Shot Prompting for Synthetic Data Generation
    Vivaan Sehgal, Prahalad Anand, Brian Lin, Shridhar Garg, George Flint, Kevin Zhu
  8. Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting
    Tim Knappe, Ryan Luo Li, Ayush Chauhan, Kaylee Chhua, Kevin Zhu, Sean O'Brien
  9. Q*Agent: Optimizing Language Agents with Q-Guided Exploration
    Zongyu Lin, Yao Tang, Da Yin, Xingcheng Yao, Ziniu Hu, Yizhou Sun, Kai-Wei Chang
  10. 3D-LLM: Injecting the 3D World into Large Language Models
    Yining Hong
  11. ContextRef: Evaluating referenceless metrics for image description generation
    Elisa Kreiss, Eric Zelikman, Christopher Potts, Nick Haber
  12. Detecting Machine-Generated Long-Form Content with Latent-Space Variables
    Yufei Tian, Zeyu Pan, Nanyun Peng
  13. A Debate-Driven Experiment on LLM Hallucinations and Accuracy
    Ruikun Li, Tanishka Bagade, Kevin Martinez, Flora Yasmin, Grant Ayala, Michael Lam, Kevin Zhu
  14. Can Large Language Models Successfully Understand Errors?
    Jason Li, Lauren Gabrielle Yraola, Kevin Zhu, Sean O'Brien
  15. Promoting Fairness in Link Prediction with Graph Enhancement
    Yezi Liu, Hanning Chen, Mohsen Imani
  16. Stronger Random Baselines for In-Context Learning
    Gregory Yauney, David Mimno
  17. Challenges in Cross-Lingual Transfer Learning for Offensive Language Detection: Insights from 30 Languages
    Elnaz Rahmati, Alireza Salkhordeh Ziabari, Morteza Dehghani
  18. Beyond LLMs: A Linguistic Approach to Causal Graph Generation from Narrative Texts
    Zehan Li, Ruhua Pan, Xinyu Pi
  19. Attributing Culture-Conditioned Generations to Pretraining Corpora
    Huihan Li, Arnav Goel, Keyu He, Xiang Ren
  20. BPO: Staying Close to the Behavior LLM Creates Better Online LLM Alignment
    Wenda Xu, Jiachen Li, William Yang Wang, Lei Li
  21. Safety Alignment in NLP Tasks: Weakly Aligned Summarization as an In-Context Attack
    Yu Fu, Yufei Li, Wen Xiao, Cong Liu, Yue Dong
  22. Are Large-Language Models Graph Algorithmic Reasoners?
    Alexander K Taylor, Anthony Cuturrufo, Vishal Yathish, Mingyu Derek Ma, Wei Wang
  23. Can Language Models Learn to Skip Steps?
    Tengxiao Liu, Qipeng Guo, Xiangkun Hu, Cheng Jiayang, Yue Zhang, Xipeng Qiu, Zheng Zhang
  24. Open-world Multi-label Text Classification with Extremely Weak Supervision
    Xintong Li, Jinya Jiang, Ria Dharmani, Jayanth Srinivasa, Gaowen Liu, Jingbo Shang
  25. The Factuality Tax of Diversity-Intervened Text-to-Image Generation: Benchmark and Fact-Augmented Intervention
    Yixin Wan, Di Wu, Haoran Wang, Kai-Wei Chang
  26. Enhancing Transparency in RLHF: An Explainable Reward Framework for Aligning Large Language Models
    Yiran Shen, Aditya Emmanuel Arokiaraj John, Brandon Fain
  27. Dialect Bias in Affective Computing: African American English and Text-Based Emotion Classification
    Rebecca Dorn, Christina A Chance, Casandra Rusti
  28. Cross-Task Generalization Abilities of Large Language Models
    Qinyuan Ye
  29. Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling
    Hritik Bansal, Arian Hosseini, Rishabh Agarwal, Vinh Q. Tran, Mehran Kazemi
  30. Machine Learning Research Transparency through Positionality Statements
    Christina A Chance, Rebecca Pattichis
  31. Comparing Bad Apples to Good Oranges: Aligning Large Language Models via Joint Preference Optimization
    Hritik Bansal, Ashima Suvarna, Gantavya Bhatt, Nanyun Peng, Kai-Wei Chang, Aditya Grover
  32. Matryoshka Query Transformer for Large Vision-Language Models
    Wenbo Hu, Zi-Yi Dou, Liunian Harold Li, Amita Kamath, Nanyun Peng, Kai-Wei Chang
  33. Dynamic Rewarding with Prompt Optimization Enables Tuning-free Self-Alignment of Language Models
    Somanshu Singla, Zhen Wang, Tianyang Liu, Abdullah Ashfaq, Zhiting Hu, Eric P. Xing
  34. Sufficient Context: A New Lens on Retrieval Augmented Generation Systems
    Hailey Joren, Jianyi Zhang, Chun-Sung Ferng, Da-Cheng Juan, Ankur Taly, Cyrus Rashtchian
  35. LLM-Coordination: Evaluating and Analyzing Multi-agent Coordination Abilities in Large Language Models
    Saaket Agashe, Yue Fan, Anthony Reyna, Xin Eric Wang
  36. Speechworthy Instruction-tuned Language Models
    Hyundong Justin Cho, Nicolaas Paul Jedema, Leonardo F. R. Ribeiro, Karishma Sharma, Pedro Szekely, Alessandro Moschitti, Ruben Janssen, Jonathan May
  37. EditRoom: LLM-parameterized Graph Diffusion for Composable 3D Room Layout Editing
    Kaizhi Zheng, Xiaotong Chen, Xuehai He, Jing Gu, Linjie Li, Zhengyuan Yang, Kevin Lin, Jianfeng Wang, Lijuan Wang, Xin Eric Wang
  38. Using Linguistic Entrainment to Evaluate Large Language Models for use in Cognitive Behavioral Therapy
    Mina Kian, Kaleen Shrestha, Katrin Fischer, Xiaoyuan Zhu, Jonathan Ong, Aryan Trehan, Jessica Wang, Gloria Chang, Séb Arnold, Maja Mataric
  39. VSP: Assessing the dual challenges of perception and reasoning in spatial planning tasks for MLLMs
    Qiucheng Wu, Handong Zhao, Michael Saxon, Trung Bui, William Yang Wang, Yang Zhang, Shiyu Chang
  40. Quantifying Generalization Complexity for Large Language Models
    Zhenting Qi, Hongyin Luo, Xuliang Huang, Zhuokai Zhao, Yibo Jiang, Xiangjun Fan, Himabindu Lakkaraju, James R. Glass
  41. LlamaTales: Studying the Effects of Training Data on Small Language Models
    Ivan Lee, Taylor Berg-Kirkpatrick
  42. A Comparative Study of Translation Bias and Accuracy in Multilingual Large Language Models for Cross-Language Claim Verification
    Aryan Singhal, Veronica Shao, Gary Sun, Ryan Ding, Jonathan Lu, Kevin Zhu
  43. Town Hall Debate Prompting: Enhancing Logical Reasoning in LLMs through Multi-Persona Interaction
    Bhav Jain, Vivaan Sandwar, Ishaan Garg, Rishan Thangaraj, Michael Lam, Kevin Zhu
  44. StyleRemix: Interpretable Authorship Obfuscation via Distillation and Perturbation of Style Elements
    Jillian Fisher, Skyler Hallinan, Ximing Lu, Mitchell L Gordon, Zaid Harchaoui, Yejin Choi
  45. Goal-Driven Explainable Clustering via Language Descriptions
    Zihan Wang, Jingbo Shang, Ruiqi Zhong
  46. Inferring from Logits: Exploring Best Practices for Decoding-Free Generative Candidate Selection
    Mingyu Derek Ma, Yanna Ding, Zijie Huang, Jianxi Gao, Yizhou Sun, Wei Wang
  47. Follow My Instruction and Spill the Beans: Scalable Data Extraction from Retrieval-Augmented Generation Systems
    Zhenting Qi, Hanlin Zhang, Eric P. Xing, Sham M. Kakade, Himabindu Lakkaraju
  48. Generative Verifiers: Reward Modeling as Next-Token Prediction
    Lunjun Zhang, Arian Hosseini, Hritik Bansal, Mehran Kazemi, Aviral Kumar, Rishabh Agarwal
  49. Leveraging Language Models to Detect Greenwashing
    Margaret Capetz, Christina A Chance, Rebecca Pattichis, Reshmi Ghosh, Avalon Vinella
  50. Asymmetric Bias in Text-to-Image Generation with Adversarial Attacks
    Haz Sameen Shahgir, Xianghao Kong, Greg Ver Steeg, Yue Dong
  51. Enhancing Knowledge Distillation for LLMs with Response-Priming Prompting
    Vijay Goyal, Mustafa Khan, Aprameya Tirupati, Harveer Saini, Michael Lam, Kevin Zhu
  52. Learning a Decision Tree Algorithm with Transformers
    Yufan Zhuang, Liyuan Liu, Chandan Singh, Jingbo Shang, Jianfeng Gao
  53. Evaluating Creativity and Deception in Large Language Models: A Simulation Framework for Multi-Agent Balderdash
    Parsa Hejabi, Elnaz Rahmati, Alireza Salkhordeh Ziabari, Preni Golazizian, Jesse Thomason, Morteza Dehghani
  54. Planning as Inpainting: A Generative Framework for Realistic Embodied Path Planning
    Cheng-Fu Yang, Kai-Wei Chang
  55. Multi-step Problem Solving Through a Verifier: An Empirical Analysis on Model-induced Process Supervision
    Zihan Wang, Yunxuan Li, Yuexin Wu, Liangchen Luo, Le Hou, Hongkun Yu, Jingbo Shang

Poster Session #3 (2:40pm - 3:40pm)

  1. Pre-trained Large Language Models Use Fourier Features to Compute Addition
    Tianyi Zhou, Deqing Fu, Vatsal Sharan, Robin Jia
  2. NusaMT-7B: Machine Translation for Low-Resource Indonesian Languages with Large Language Models
    William Tan, Kevin Zhu
  3. Adaptive In-conversation Team Building for Language Model Agents
    Linxin Song, Jiale Liu, Jieyu Zhang, Shaokun Zhang, Ao Luo, Shijian Wang, Qingyun Wu, Chi Wang
  4. PatentEdits: Framing Patent Novelty as Textual Entailment
    Ryan Lee, Alexander Spangher, Xuezhe Ma
  5. Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue
    Jia-Chen Gu, Hao-Xiang Xu, Jun-Yu Ma, Pan Lu, Zhen-Hua Ling, Kai-Wei Chang, Nanyun Peng
  6. Fictitious Synthetic Data Can Improve LLM Factuality via Prerequisite Learning
    Yujian Liu, Shiyu Chang, Tommi Jaakkola, Yang Zhang
  7. Adversarial Attacks on Parts of Speech: An Empirical Study in Text-to-Image Generation
    G M Shahariar, Jia Chen, Jiachen Li, Yue Dong
  8. SoMeR: A Multi-View Social Media User Representation Learning Framework
    Keith Burghardt, Siyi Guo, Valeria Pantè, Kristina Lerman
  9. Communicate to Play: Pragmatic Reasoning for Efficient Cross-Cultural Communication
    Isadora White, Sashrika Pandey, Michelle Pan
  10. Weak-to-Strong Reasoning
    Yuqing Yang, Yan Ma, Pengfei Liu
  11. MARVEL: Multidimensional Abstraction and Reasoning through Visual Evaluation and Learning
    Yifan Jiang, Jiarui Zhang, Kexuan Sun, Zhivar Sourati, Kian Ahrabian, Kaixin Ma, Filip Ilievski, Jay Pujara
  12. STAR: A Simple Training-free Approach for Recommendations using Large Language Models
    Dong-Ho Lee, Adam Kraft, Long Jin, Nikhil Mehta, Taibai Xu, Lichan Hong, Ed H. Chi, Xinyang Yi
  13. Ladder Residual: Redefining Tensor Parallelism in Transformers for Accelerated Inference
    Muru Zhang, Mayank Mishra, Zhongzhu Zhou, William Brandon, Jue WANG, Yoon Kim, Jonathan Ragan-Kelley, Shuaiwen Leon Song, Ben Athiwaratkun, Tri Dao
  14. CLEAR: Contrastive Learning with Experts and Amateurs for Reasoning
    Andrew Rufail, Abenezer Tessema, Buse Toksoz, Chizoba Okoli, Jonathan Pei, Kevin Zhu
  15. Layer Swapping for Zero-Shot Cross-Lingual Transfer in Large Language Models
    Lucas Bandarkar
  16. MORL-Prompt: An Empirical Analysis of Multi-Objective Reinforcement Learning for Discrete Prompt Optimization
    Yasaman Jafari, Dheeraj Mekala, Rose Yu, Taylor Berg-Kirkpatrick
  17. Evaluating Cultural and Social Awareness of LLM Web Agents
    Haoyi Qiu, Alexander Fabbri, Divyansh Agarwal, Kung-Hsiang Huang, Sarah Tan, Nanyun Peng, Chien-Sheng Wu
  18. VideoPhy: Evaluating Physical Commonsense for Video Generation
    Hritik Bansal, Zongyu Lin, Tianyi Xie, Zeshun Zong, Michal Yarom, Yonatan Bitton, Chenfanfu Jiang, Yizhou Sun, Kai-Wei Chang, Aditya Grover
  19. COMMUNITY-CROSS-INSTRUCT: Unsupervised Instruction Generation for Aligning Large Language Models to Online Communities
    Zihao He, Minh Duc Chu, Rebecca Dorn, Siyi Guo, Kristina Lerman
  20. Exploring Synthetic Datasets for Large Language Model Unlearning
    Xiaoyuan Zhu, Ollie Liu, Muru Zhang, Willie Neiswanger
  21. MAPO: Momentum-Aided Gradient Descent Prompt Optimization
    Anthony Cui, Pranav Nandyalam, Kevin Zhu
  22. Out-of-Distribution Detection through Soft Clustering with Non-Negative Kernel Regression
    Aryan Gulati, Xingjian Dong, Carlos Hurtado, Sarath Shekkizhar, Swabha Swayamdipta, Antonio Ortega
  23. CoLLAP: Contrastive Long-form Language-Audio Pretraining with Musical Temporal Structure Augmentation
    Junda Wu, Warren Li, Zachary Novack, Amit Namburi, Carol Chen, Julian McAuley
  24. Data Contamination Can Cross Language Barriers
    Feng Yao, Yufan Zhuang, Zihao Sun, Sunan Xu, Animesh Kumar, Jingbo Shang
  25. Quantifying and Optimizing Global Faithfulness in Persona-driven Role-playing
    Letian Peng, Jingbo Shang
  26. DiversityMedQA: A Benchmark for Assessing Demographic Biases in Medical Diagnosis using Large Language Models
    Rajat Rawat, Hudson McBride, Rajarshi Ghosh, Dhiyaan Chakkresh Nirmal, Jong Moon, Dhruv Karthik Alamuri, Sean O'Brien, Kevin Zhu
  27. Language Models Can Infer Action Semantics for Symbolic Planners from Environment Feedback
    Wang Zhu, Ishika Singh, Robin Jia, Jesse Thomason
  28. A Probabilistic Framework for LLM Hallucination Detection via Belief Tree Propagation
    Bairu Hou, Yang Zhang, Jacob Andreas, Shiyu Chang
  29. GPT is Not an Annotator: The Necessity of Human Annotation in Fairness Benchmark Construction
    Virginia K. Felkner, Jennifer A. Thompson, Jonathan May
  30. Adaptable Logical Control for Large Language Models
    Honghua Zhang, Po-Nien Kung, Masahiro Yoshida, Guy Van den Broeck, Nanyun Peng
  31. Strategic Moves: Negotiation Agents with Dynamic Adaptation to Opponent Behavior for Optimizing Counter-Offers
    Deuksin Kwon, Jiwon Hae, Gale Lucas
  32. Solving for X and Beyond: Can Large Language Models Solve Complex Math Problems with More-Than-Two Unknowns?
    Kuei-Chun Kao, Ruochen Wang, Cho-Jui Hsieh
  33. Capturing the Essence of a Phrase: Extracting Physical and Sensory Information from Text
    Abhinav Gupta
  34. Efficient LLM Scheduling by Learning to Rank
    Yichao Fu, Siqi Zhu, Runlong Su, Aurick Qiao, Ion Stoica, Hao Zhang
  35. Improving LLM Abilities in Idiomatic Translation
    Maximilian Spencer, Sundesh Donthi, Om Patel, Joon Young Doh, Eid Rodan
  36. Large Language Models Can Be Contextual Privacy Protection Learners
    Yijia Xiao, Yiqiao Jin, Yushi Bai, Yue Wu, Xianjun Yang, Xiao Luo, Wenchao Yu, Xujiang Zhao, Yanchi Liu, Quanquan Gu, Haifeng Chen, Wei Wang, Wei Cheng
  37. TrimLLM: Progressive Layer Dropping for Efficient LLM Fine-tuning and Inference
    Lanxiang Hu, Tajana Rosing, Hao Zhang
  38. Revisiting Representation Collapse in Language Models
    Atharva Kulkarni, Swabha Swayamdipta
  39. Logits of API-Protected LLMs Leak Proprietary Information
    Matthew Finlayson, Swabha Swayamdipta, Xiang Ren
  40. Grammar-Aligned Decoding
    Kanghee Park, Jiayu Wang, Taylor Berg-Kirkpatrick, Nadia Polikarpova, Loris D'Antoni
  41. Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented Generation
    Di Wu, Jia-Chen Gu, Fan Yin, Nanyun Peng, Kai-Wei Chang
  42. Who Evaluates the Evaluations? Objectively Scoring Text-to-Image Prompt Coherence Metrics with T2IScoreScore
    Michael Saxon, Fatima Jahara, Mahsa Khoshnoodi, Yujie Lu, Aditya Sharma, William Yang Wang
  43. LLM Self-contradictory Hallucination Mitigation via Retrieval-Based Post-Processing
    Zhaotian Weng, Xingjian Dong, Liyan Huang, Jingwen Qi
  44. When Parts Are Greater Than Sums: Individual LLM Components Can Outperform Full Models
    Ting-Yun Chang, Jesse Thomason, Robin Jia
  45. Evaluating LLM Chatbot-Generated Answers to Patient Questions
    Wang Zhu, Tian-qi Chen, Ruishan Liu, Robin Jia
  46. The Strong Pull of Prior Knowledge in Large Language Models and Its Impact on Emotion Recognition
    Georgios Chochlakis, Alexandros Potamianos, Kristina Lerman, Shrikanth Narayanan
  47. Firm’s AI Exposure without Labor Data: The Expected Impact of AI
    Jiacheng Liu
  48. Fine-Tuning Language Models for Ethical Ambiguity: A Comparative Study of Alignment with Human Responses
    Pranav Senthilkumar, Visshwa Balasubramanian, Prisha Jain, Aneesa Maity, Jonathan Lu, Kevin Zhu
  49. SPEED++: A Multilingual Event Extraction Framework for Epidemic Prediction and Preparedness
    Tanmay Parekh, Jeffrey Kwan, Jiarui Yu, Sparsh Johri, Hyosang Ahn, Sreya Muppalla, Kai-Wei Chang, Wei Wang, Nanyun Peng
  50. Stackelberg Games for Persuasive Advocacy in LLM Simulations
    Xinyue Cui, Swabha Swayamdipta
  51. Style-Compress: An LLM-Based Prompt Compression Framework Considering Task-Specific Styles
    Xiao Pu, Tianxing He, Xiaojun Wan
  52. BIASDETECTOR: Multi-Agent Synergy for Comprehensive Bias Detection in Structural Data
    Haoxuan Li, Mingyu Derek Ma, Jen-tse Huang, Wei Wang, Jieyu Zhao
  53. mDPO: Conditional Preference Optimization for Multimodal Large Language Models
    Fei Wang, Wenxuan Zhou, James Y. Huang, Nan Xu, Sheng Zhang, Hoifung Poon, Muhao Chen
  54. Linear Layer Extrapolation for Fine-Grained Emotion Classification
    Mayukh Sharma, Sean O'Brien, Julian McAuley
  55. NewsHomepages: Homepage Layouts Capture Information Prioritization Decisions
    Arda Kaz, Alexander Spangher, Michael Vu, Naitian Zhou, Ben Welsh
  56. Predictive Modeling for Early Alzheimer's Disease Using Natural Language Processing
    Azra Emekci


Organizers


General Chairs

Jingbo Shang

Associate Professor

CSE and HDSI

UCSD

Lianhui Qin

Assistant Professor

CSE

UCSD

Prithviraj (Raj) Ammanabrolu

Assistant Professor

CSE

UCSD

Hao Zhang

Assistant Professor

HDSI

UCSD



Program Chair

Zihan Wang

PhD student

CSE

UCSD



Reviewers

Xiaohan Fu, Yiran Shen, Ruiyi Wang, Junda Chen, Yufan Zhuang, Lanxiang Hu, Jiayun Zhang, Yupeng Hou, Xintong Li, Chenyang An, Matthew Ho, Feng Yao, Christopher Zhang Cui, Zihan Wang, Bosung Kim, Zilong Wang, Yuwei Zhang, Isadora White, Peiyuan Zhang, Animesh Kumar, Yi Gu, Yichao Fu, Letian Peng, and Weitang Liu.





Past Symposiums