{"title": "Proving Test Set Contamination in Black-Box Language Models", "url": "https://openreview.net/forum?id=KS8mIvetg2", "detail_url": "https://openreview.net/forum?id=KS8mIvetg2", "authors": "Yonatan Oren,Nicole Meister,Niladri S. Chatterji,Faisal Ladhak,Tatsunori Hashimoto", "tags": "ICLR 2024,Oral", "abstract": "Large language models are trained on vast amounts of internet data, prompting concerns that they have memorized public benchmarks. Detecting this type of contamination is challenging because the pretraining data used by proprietary models are often not publicly accessible.\n\nWe propose a procedure for detecting test set contamination of language models with exact false positive guarantees and without access to pretraining data or model weights. Our approach leverages the fact that when there is no data contamination, all orderings of an exchangeable benchmark should be equally likely. In contrast, the tendency for language models to memorize example order means that a contaminated language model will find certain canonical orderings to be much more likely than others. Our test flags potential contamination whenever the likelihood of a canonically ordered benchmark dataset is significantly higher than the likelihood after shuffling the examples.\n\nWe demonstrate that our procedure is sensitive enough to reliably detect contamination in challenging situations, including models as small as 1.4 billion parameters, on small test sets only 1000 examples, and datasets that appear only a few times in the pretraining corpus. Finally, we evaluate LLaMA-2 to apply our test in a realistic setting and find our results to be consistent with existing contamination evaluations.", "pdf": "https://openreview.net/pdf/cfd79aaab7bdcd4f7c032c57fe7e607058042c80.pdf"} {"title": "BooookScore: A systematic exploration of book-length summarization in the era of LLMs", "url": "https://openreview.net/forum?id=7Ttk3RzDeu", "detail_url": "https://openreview.net/forum?id=7Ttk3RzDeu", "authors": "Yapei Chang,Kyle Lo,Tanya Goyal,Mohit Iyyer", "tags": "ICLR 2024,Oral", "abstract": "Summarizing book-length documents ($>$100K tokens) that exceed the context window size of large language models (LLMs) requires first breaking the input document into smaller chunks and then prompting an LLM to merge, update, and compress chunk-level summaries. Despite the complexity and importance of this task, it has yet to be meaningfully studied due to the challenges of evaluation: existing book-length summarization datasets (e.g., BookSum) are in the pretraining data of most public LLMs, and existing evaluation methods struggle to capture errors made by modern LLM summarizers. In this paper, we present the first study of the coherence of LLM-based book-length summarizers implemented via two prompting workflows: (1) hierarchically merging chunk-level summaries, and (2) incrementally updating a running summary. We obtain 1193 fine-grained human annotations on GPT-4 generated summaries of 100 recently-published books and identify eight common types of coherence errors made by LLMs. Because human evaluation is expensive and time-consuming, we develop an automatic metric, BooookScore, that measures the proportion of sentences in a summary that do not contain any of the identified error types. BooookScore has high agreement with human annotations and allows us to systematically evaluate the impact of many other critical parameters (e.g., chunk size, base LLM) while saving \\$15K USD and 500 hours in human evaluation costs. We find that closed-source LLMs such as GPT-4 and Claude 2 produce summaries with higher BooookScore than those generated by open-source models. While LLaMA 2 falls behind other models, Mixtral achieves performance on par with GPT-3.5-Turbo. Incremental updating yields lower BooookScore but higher level of detail than hierarchical merging, a trade-off sometimes preferred by annotators. We release code and annotations to spur more principled research on book-length summarization.", "pdf": "https://openreview.net/pdf/975e393e430362eb39a2c1ceb2c750bd4bb80143.pdf"} {"title": "Generalization in diffusion models arises from geometry-adaptive harmonic representations", "url": "https://openreview.net/forum?id=ANvmVS2Yr0", "detail_url": "https://openreview.net/forum?id=ANvmVS2Yr0", "authors": "Zahra Kadkhodaie,Florentin Guth,Eero P Simoncelli,St\u00e9phane Mallat", "tags": "ICLR 2024,Oral", "abstract": "Deep neural networks (DNNs) trained for image denoising are able to generate high-quality samples with score-based reverse diffusion algorithms. These impressive capabilities seem to imply an escape from the curse of dimensionality, but recent reports of memorization of the training set raise the question of whether these networks are learning the \"true\" continuous density of the data. Here, we show that two DNNs trained on non-overlapping subsets of a dataset learn nearly the same score function, and thus the same density, when the number of training images is large enough. In this regime of strong generalization, diffusion-generated images are distinct from the training set, and are of high visual quality, suggesting that the inductive biases of the DNNs are well-aligned with the data density. We analyze the learned denoising functions and show that the inductive biases give rise to a shrinkage operation in a basis adapted to the underlying image. Examination of these bases reveals oscillating harmonic structures along contours and in homogeneous regions. We demonstrate that trained denoisers are inductively biased towards these geometry-adaptive harmonic bases since they arise not only when the network is trained on photographic images, but also when it is trained on image classes supported on low-dimensional manifolds for which the harmonic basis is suboptimal. Finally, we show that when trained on regular image classes for which the optimal basis is known to be geometry-adaptive and harmonic, the denoising performance of the networks is near-optimal.", "pdf": "https://openreview.net/pdf/84eb681ff8d070ce8c829cb2120dc133901594ce.pdf"} {"title": "Understanding In-Context Learning in Transformers and LLMs by Learning to Learn Discrete Functions", "url": "https://openreview.net/forum?id=ekeyCgeRfC", "detail_url": "https://openreview.net/forum?id=ekeyCgeRfC", "authors": "Satwik Bhattamishra,Arkil Patel,Phil Blunsom,Varun Kanade", "tags": "ICLR 2024,Oral", "abstract": "In order to understand the in-context learning phenomenon, recent works have adopted a stylized experimental framework and demonstrated that Transformers can match the performance of gradient-based learning algorithms for various classes of real-valued functions. However, the limitations of Transformers in implementing learning algorithms, and their ability to learn other forms of algorithms are not well understood. Additionally, the degree to which these capabilities are confined to attention-based models is unclear. Furthermore, it remains to be seen whether the insights derived from these stylized settings can be extrapolated to pretrained Large Language Models (LLMs). In this work, we take a step towards answering these questions by demonstrating the following: (a) On a test-bed with a variety of Boolean function classes, we find that Transformers can nearly match the optimal learning algorithm for 'simpler' tasks, while their performance deteriorates on more 'complex' tasks. Additionally, we find that certain attention-free models perform (almost) identically to Transformers on a range of tasks. (b) When provided a *teaching sequence*, i.e. a set of examples that uniquely identifies a function in a class, we show that Transformers learn more sample-efficiently. Interestingly, our results show that Transformers can learn to implement *two distinct* algorithms to solve a *single* task, and can adaptively select the more sample-efficient algorithm depending on the sequence of in-context examples. (c) Lastly, we show that extant LLMs, e.g. LLaMA-2, GPT-4, can compete with nearest-neighbor baselines on prediction tasks that are guaranteed to not be in their training set.", "pdf": "https://openreview.net/pdf/816f489eb70fe677c4ebc1cf159cf38b3062956b.pdf"} {"title": "The mechanistic basis of data dependence and abrupt learning in an in-context classification task", "url": "https://openreview.net/forum?id=aN4Jf6Cx69", "detail_url": "https://openreview.net/forum?id=aN4Jf6Cx69", "authors": "Gautam Reddy", "tags": "ICLR 2024,Oral", "abstract": "Transformer models exhibit in-context learning: the ability to accurately predict the response to a novel query based on illustrative examples in the input sequence, which contrasts with traditional in-weights learning of query-output relationships. What aspects of the training data distribution and architecture favor in-context vs in-weights learning? Recent work has shown that specific distributional properties inherent in language, such as burstiness, large dictionaries and skewed rank-frequency distributions, control the trade-off or simultaneous appearance of these two forms of learning. We first show that these results are recapitulated in a minimal attention-only network trained on a simplified dataset. In-context learning (ICL) is driven by the abrupt emergence of an induction head, which subsequently competes with in-weights learning. By identifying progress measures that precede in-context learning and targeted experiments, we construct a two-parameter model of an induction head which emulates the full data distributional dependencies displayed by the attention-based network. A phenomenological model of induction head formation traces its abrupt emergence to the sequential learning of three nested logits enabled by an intrinsic curriculum. We propose that the sharp transitions in attention-based networks arise due to a specific chain of multi-layer operations necessary to achieve ICL, which is implemented by nested nonlinearities sequentially learned during training.", "pdf": "https://openreview.net/pdf/4de2c24997e6d25adcda68f174ed540f41a217e8.pdf"} {"title": "Improved Techniques for Training Consistency Models", "url": "https://openreview.net/forum?id=WNzy9bRDvG", "detail_url": "https://openreview.net/forum?id=WNzy9bRDvG", "authors": "Yang Song,Prafulla Dhariwal", "tags": "ICLR 2024,Oral", "abstract": "Consistency models are a nascent family of generative models that can sample high quality data in one step without the need for adversarial training. Current consistency models achieve optimal sample quality by distilling from pre-trained diffusion models and employing learned metrics such as LPIPS. However, distillation limits the quality of consistency models to that of the pre-trained diffusion model, and LPIPS causes undesirable bias in evaluation. To tackle these challenges, we present improved techniques for consistency training, where consistency models learn directly from data without distillation. We delve into the theory behind consistency training and identify a previously overlooked flaw, which we address by eliminating Exponential Moving Average from the teacher consistency model. To replace learned metrics like LPIPS, we adopt Pseudo-Huber losses from robust statistics. Additionally, we introduce a lognormal noise schedule for the consistency training objective, and propose to double total discretization steps every set number of training iterations. Combined with better hyperparameter tuning, these modifications enable consistency models to achieve FID scores of 2.51 and 3.25 on CIFAR-10 and ImageNet $64\\times 64$ respectively in a single sampling step. These scores mark a 3.5$\\times$ and 4$\\times$ improvement compared to prior consistency training approaches. Through two-step sampling, we further reduce FID scores to 2.24 and 2.77 on these two datasets, surpassing those obtained via distillation in both one-step and two-step settings, while narrowing the gap between consistency models and other state-of-the-art generative models.", "pdf": "https://openreview.net/pdf/c40d76fe68ec3195a55ba242266828b01fdb06c5.pdf"} {"title": "Provable Compositional Generalization for Object-Centric Learning", "url": "https://openreview.net/forum?id=7VPTUWkiDQ", "detail_url": "https://openreview.net/forum?id=7VPTUWkiDQ", "authors": "Thadd\u00e4us Wiedemer,Jack Brady,Alexander Panfilov,Attila Juhos,Matthias Bethge,Wieland Brendel", "tags": "ICLR 2024,Oral", "abstract": "Learning representations that generalize to novel compositions of known concepts is crucial for bridging the gap between human and machine perception. One prominent effort is learning object-centric representations, which are widely conjectured to enable compositional generalization. Yet, it remains unclear when this conjecture will be true, as a principled theoretical or empirical understanding of compositional generalization is lacking. In this work, we investigate when compositional generalization is guaranteed for object-centric representations through the lens of identifiability theory. We show that autoencoders that satisfy structural assumptions on the decoder and enforce encoder-decoder consistency will learn object-centric representations that provably generalize compositionally. We validate our theoretical result and highlight the practical relevance of our assumptions through experiments on synthetic image data.", "pdf": "https://openreview.net/pdf/70cd6e52cd58ee0e0b07dfea409db6acc228b343.pdf"} {"title": "Predictive auxiliary objectives in deep RL mimic learning in the brain", "url": "https://openreview.net/forum?id=agPpmEgf8C", "detail_url": "https://openreview.net/forum?id=agPpmEgf8C", "authors": "Ching Fang,Kim Stachenfeld", "tags": "ICLR 2024,Oral", "abstract": "The ability to predict upcoming events has been hypothesized to comprise a key aspect of natural and machine cognition. This is supported by trends in deep reinforcement learning (RL), where self-supervised auxiliary objectives such as prediction are widely used to support representation learning and improve task performance. Here, we study the effects predictive auxiliary objectives have on representation learning across different modules of an RL system and how these mimic representational changes observed in the brain. We find that predictive objectives improve and stabilize learning particularly in resource-limited architectures, and we identify settings where longer predictive horizons better support representational transfer. Furthermore, we find that representational changes in this RL system bear a striking resemblance to changes in neural activity observed in the brain across various experiments. Specifically, we draw a connection between the auxiliary predictive model of the RL system and hippocampus, an area thought to learn a predictive model to support memory-guided behavior. We also connect the encoder network and the value learning network of the RL system to visual cortex and striatum in the brain, respectively. This work demonstrates how representation learning in deep RL systems can provide an interpretable framework for modeling multi-region interactions in the brain. The deep RL perspective taken here also suggests an additional role of the hippocampus in the brain-- that of an auxiliary learning system that benefits representation learning in other regions.", "pdf": "https://openreview.net/pdf/23365fd987e6b67de035adbd3b3bb679d36ddce7.pdf"} {"title": "Pre-Training Goal-based Models for Sample-Efficient Reinforcement Learning", "url": "https://openreview.net/forum?id=o2IEmeLL9r", "detail_url": "https://openreview.net/forum?id=o2IEmeLL9r", "authors": "Haoqi Yuan,Zhancun Mu,Feiyang Xie,Zongqing Lu", "tags": "ICLR 2024,Oral", "abstract": "Pre-training on task-agnostic large datasets is a promising approach for enhancing the sample efficiency of reinforcement learning (RL) in solving complex tasks. We present PTGM, a novel method that pre-trains goal-based models to augment RL by providing temporal abstractions and behavior regularization. PTGM involves pre-training a low-level, goal-conditioned policy and training a high-level policy to generate goals for subsequent RL tasks. To address the challenges posed by the high-dimensional goal space, while simultaneously maintaining the agent's capability to accomplish various skills, we propose clustering goals in the dataset to form a discrete high-level action space. Additionally, we introduce a pre-trained goal prior model to regularize the behavior of the high-level policy in RL, enhancing sample efficiency and learning stability. Experimental results in a robotic simulation environment and the challenging open-world environment of Minecraft demonstrate PTGM\u2019s superiority in sample efficiency and task performance compared to baselines. Moreover, PTGM exemplifies enhanced interpretability and generalization of the acquired low-level skills.", "pdf": "https://openreview.net/pdf/97ae12300fd1715ec484f1be154d49a619911fff.pdf"} {"title": "Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!", "url": "https://openreview.net/forum?id=hTEGyKf0dZ", "detail_url": "https://openreview.net/forum?id=hTEGyKf0dZ", "authors": "Xiangyu Qi,Yi Zeng,Tinghao Xie,Pin-Yu Chen,Ruoxi Jia,Prateek Mittal,Peter Henderson", "tags": "ICLR 2024,Oral", "abstract": "Optimizing large language models (LLMs) for downstream use cases often involves the customization of pre-trained LLMs through further fine-tuning. Meta's open-source release of Llama models and OpenAI's APIs for fine-tuning GPT-3.5 Turbo on customized datasets accelerate this trend. But, what are the safety costs associated with such customized fine-tuning? While existing safety alignment techniques restrict harmful behaviors of LLMs at inference time, they do not cover safety risks when fine-tuning privileges are extended to end-users. Our red teaming studies find that the safety alignment of LLMs can be compromised by fine-tuning with only a few adversarially designed training examples. For instance, we jailbreak GPT-3.5 Turbo's safety guardrails by fine-tuning it on only 10 such examples at a cost of less than $0.20 via OpenAI's APIs, making the model responsive to nearly any harmful instructions. Disconcertingly, our research also reveals that, even without malicious intent, simply fine-tuning with benign and commonly used datasets can also inadvertently degrade the safety alignment of LLMs, though to a lesser extent. These findings suggest that fine-tuning aligned LLMs introduces new safety risks that current safety infrastructures fall short of addressing --- even if a model's initial safety alignment is impeccable, how can it be maintained after customized fine-tuning? We outline and critically analyze potential mitigations and advocate for further research efforts toward reinforcing safety protocols for the customized fine-tuning of aligned LLMs. (This paper contains red-teaming data and model-generated content that can be offensive in nature.)", "pdf": "https://openreview.net/pdf/cf8a15c7b5a808ae67357cdde0c8f2bbd5c4b8ed.pdf"} {"title": "Never Train from Scratch: Fair Comparison of Long-Sequence Models Requires Data-Driven Priors", "url": "https://openreview.net/forum?id=PdaPky8MUn", "detail_url": "https://openreview.net/forum?id=PdaPky8MUn", "authors": "Ido Amos,Jonathan Berant,Ankit Gupta", "tags": "ICLR 2024,Oral", "abstract": "Modeling long-range dependencies across sequences is a longstanding goal in machine learning and has led to architectures, such as state space models, that dramatically outperform Transformers on long sequences. However, these impressive empirical gains have been by and large demonstrated on benchmarks (e.g. Long Range Arena), where models are randomly initialized and trained to predict a target label from an input sequence. In this work, we show that random initialization leads to gross overestimation of the differences between architectures and that pretraining with standard denoising objectives, *using only the downstream task data*, leads to dramatic gains across multiple architectures and to very small gaps between Transformers and state space models (SSMs). In stark contrast to prior works, we find vanilla Transformers to match the performance of S4 on Long Range Arena when properly pretrained, and we improve the best reported results of SSMs on the PathX-256 task by 20 absolute points. Subsequently, we analyze the utility of previously-proposed structured parameterizations for SSMs and show they become mostly redundant in the presence of data-driven initialization obtained through pretraining. Our work shows that, when evaluating different architectures on supervised tasks, incorporation of data-driven priors via pretraining is essential for reliable performance estimation, and can be done efficiently.", "pdf": "https://openreview.net/pdf/0f82cdb6beb87821d0a243ee526230c73d7ae798.pdf"} {"title": "LoftQ: LoRA-Fine-Tuning-aware Quantization for Large Language Models", "url": "https://openreview.net/forum?id=LzPWWPAdY4", "detail_url": "https://openreview.net/forum?id=LzPWWPAdY4", "authors": "Yixiao Li,Yifan Yu,Chen Liang,Nikos Karampatziakis,Pengcheng He,Weizhu Chen,Tuo Zhao", "tags": "ICLR 2024,Oral", "abstract": "Quantization is an indispensable technique for serving Large Language Models (LLMs) and has recently found its way into LoRA fine-tuning (Dettmers et al., 2023). In this work we focus on the scenario where quantization and LoRA fine- tuning are applied together on a pre-trained model. In such cases it is common to observe a consistent gap in the performance on downstream tasks between full fine-tuning and quantization plus LoRA fine-tuning approach. In response, we propose LoftQ (LoRA-Fine-Tuning-aware Quantization), a novel quantization framework that simultaneously quantizes an LLM and finds a proper low-rank initialization for LoRA fine-tuning. Such an initialization alleviates the discrep- ancy between the quantized and full-precision model and significantly improves the generalization in downstream tasks. We evaluate our method on natural lan- guage understanding, question answering, summarization, and natural language generation tasks. Experiments show that our method is highly effective and out- performs existing quantization methods, especially in the challenging 2-bit and 2/4-bit mixed precision regimes. We will release our code.", "pdf": "https://openreview.net/pdf/c8a3b2454c94e0374c1778862e8fca63e370ba5b.pdf"} {"title": "Graph Neural Networks for Learning Equivariant Representations of Neural Networks", "url": "https://openreview.net/forum?id=oO6FsMyDBt", "detail_url": "https://openreview.net/forum?id=oO6FsMyDBt", "authors": "Miltiadis Kofinas,Boris Knyazev,Yan Zhang,Yunlu Chen,Gertjan J. Burghouts,Efstratios Gavves,Cees G. M. Snoek,David W. Zhang", "tags": "ICLR 2024,Oral", "abstract": "Neural networks that process the parameters of other neural networks find applications in domains as diverse as classifying implicit neural representations, generating neural network weights, and predicting generalization errors. However, existing approaches either overlook the inherent permutation symmetry in the neural network or rely on intricate weight-sharing patterns to achieve equivariance, while ignoring the impact of the network architecture itself. In this work, we propose to represent neural networks as computational graphs of parameters, which allows us to harness powerful graph neural networks and transformers that preserve permutation symmetry. Consequently, our approach enables a single model to encode neural computational graphs with diverse architectures. We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations, predicting generalization performance, and learning to optimize, while consistently outperforming state-of-the-art methods. The source code is open-sourced at https://github.com/mkofinas/neural-graphs.", "pdf": "https://openreview.net/pdf/338609142f1f45e68ec5fc8b5d6c9a3c0247ee30.pdf"} {"title": "GNNCert: Deterministic Certification of Graph Neural Networks against Adversarial Perturbations", "url": "https://openreview.net/forum?id=IGzaH538fz", "detail_url": "https://openreview.net/forum?id=IGzaH538fz", "authors": "zaishuo xia,Han Yang,Binghui Wang,Jinyuan Jia", "tags": "ICLR 2024,Oral", "abstract": "Graph classification, which aims to predict a label for a graph, has many real-world applications such as malware detection, fraud detection, and healthcare. However, many studies show an attacker could carefully perturb the structure and/or node features in a graph such that a graph classifier misclassifies the perturbed graph. Such vulnerability impedes the deployment of graph classification in security/safety-critical applications. Existing empirical defenses lack formal robustness guarantees and could be broken by adaptive or unknown attacks. Existing provable defenses have the following limitations: 1) they achieve sub-optimal robustness guarantees for graph structure perturbation, 2) they cannot provide robustness guarantees for arbitrarily node feature perturbations, 3) their robustness guarantees are probabilistic, meaning they could be incorrect with a non-zero probability, and 4) they incur large computation costs. We aim to address those limitations in this work. We propose GNNCert, a certified defense against both graph structure and node feature perturbations for graph classification. Our GNNCert provably predicts the same label for a graph when the number of perturbed edges and the number of nodes with perturbed features are bounded. Our results on 8 benchmark datasets show that GNNCert outperforms three state-of-the-art methods.", "pdf": "https://openreview.net/pdf/03ea622e3c66547d24c4da2f725ddf1fe5db2233.pdf"} {"title": "Efficient Episodic Memory Utilization of Cooperative Multi-Agent Reinforcement Learning", "url": "https://openreview.net/forum?id=LjivA1SLZ6", "detail_url": "https://openreview.net/forum?id=LjivA1SLZ6", "authors": "Hyungho Na,Yunkyeong Seo,Il-chul Moon", "tags": "ICLR 2024,Oral", "abstract": "In cooperative multi-agent reinforcement learning (MARL), agents aim to achieve a common goal, such as defeating enemies or scoring a goal. Existing MARL algorithms are effective but still require significant learning time and often get trapped in local optima by complex tasks, subsequently failing to discover a goal-reaching policy. To address this, we introduce Efficient episodic Memory Utilization (EMU) for MARL, with two primary objectives: (a) accelerating reinforcement learning by leveraging semantically coherent memory from an episodic buffer and (b) selectively promoting desirable transitions to prevent local convergence. To achieve (a), EMU incorporates a trainable encoder/decoder structure alongside MARL, creating coherent memory embeddings that facilitate exploratory memory recall. To achieve (b), EMU introduces a novel reward structure called episodic incentive based on the desirability of states. This reward improves the TD target in Q-learning and acts as an additional incentive for desirable transitions. We provide theoretical support for the proposed incentive and demonstrate the effectiveness of EMU compared to conventional episodic control. The proposed method is evaluated in StarCraft II and Google Research Football, and empirical results indicate further performance improvement over state-of-the-art methods.", "pdf": "https://openreview.net/pdf/8b2d5ac5539754d00bf99458a60c63157c74fbdb.pdf"} {"title": "ClimODE: Climate and Weather Forecasting with Physics-informed Neural ODEs", "url": "https://openreview.net/forum?id=xuY33XhEGR", "detail_url": "https://openreview.net/forum?id=xuY33XhEGR", "authors": "Yogesh Verma,Markus Heinonen,Vikas Garg", "tags": "ICLR 2024,Oral", "abstract": "Climate and weather prediction traditionally relies on complex numerical simulations of atmospheric physics. Deep learning approaches, such as transformers, have recently challenged the simulation paradigm with complex network forecasts. However, they often act as data-driven black-box models that neglect the underlying physics and lack uncertainty quantification. We address these limitations with ClimODE, a spatiotemporal continuous-time process that implements a key principle of advection from statistical mechanics, namely, weather changes due to a spatial movement of quantities over time. ClimODE models precise weather evolution with value-conserving dynamics, learning global weather transport as a neural flow, which also enables estimating the uncertainty in predictions. Our approach outperforms existing data-driven methods in global and regional forecasting with an order of magnitude smaller parameterization, establishing a new state of the art.", "pdf": "https://openreview.net/pdf/d6e043c8dac8d842d6ba1816e2b687862e46f2bb.pdf"} {"title": "Mixed-Type Tabular Data Synthesis with Score-based Diffusion in Latent Space", "url": "https://openreview.net/forum?id=4Ay23yeuz0", "detail_url": "https://openreview.net/forum?id=4Ay23yeuz0", "authors": "Hengrui Zhang,Jiani Zhang,Zhengyuan Shen,Balasubramaniam Srinivasan,Xiao Qin,Christos Faloutsos,Huzefa Rangwala,George Karypis", "tags": "ICLR 2024,Oral", "abstract": "Recent advances in tabular data generation have greatly enhanced synthetic data quality. However, extending diffusion models to tabular data is challenging due to the intricately varied distributions and a blend of data types of tabular data. This paper introduces TabSyn, a methodology that synthesizes tabular data by leveraging a diffusion model within a variational autoencoder (VAE) crafted latent space. The key advantages of the proposed Tabsyn include (1) Generality: the ability to handle a broad spectrum of data types by converting them into a single unified space and explicitly capturing inter-column relations; (2) Quality: optimizing the distribution of latent embeddings to enhance the subsequent training of diffusion models, which helps generate high-quality synthetic data; (3) Speed: much fewer number of reverse steps and faster synthesis speed than existing diffusion-based methods. Extensive experiments on six datasets with five metrics demonstrate that Tabsyn outperforms existing methods. Specifically, it reduces the error rates by 86% and 67% for column-wise distribution and pair-wise column correlation estimations compared with the most competitive baselines. The code has been made available at https://github.com/amazon-science/tabsyn.", "pdf": "https://openreview.net/pdf/a916d9616f8be0fc9c47c323b6afe8398acf898d.pdf"} {"title": "Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement", "url": "https://openreview.net/forum?id=bNt7oajl2a", "detail_url": "https://openreview.net/forum?id=bNt7oajl2a", "authors": "Linlu Qiu,Liwei Jiang,Ximing Lu,Melanie Sclar,Valentina Pyatkin,Chandra Bhagavatula,Bailin Wang,Yoon Kim,Yejin Choi,Nouha Dziri,Xiang Ren", "tags": "ICLR 2024,Oral", "abstract": "The ability to derive underlying principles from a handful of observations and then generalize to novel situations---known as inductive reasoning---is central to human intelligence. Prior work suggests that language models (LMs) often fall short on inductive reasoning, despite achieving impressive success on research benchmarks. In this work, we conduct a systematic study of the inductive reasoning capabilities of LMs through $\\textit{iterative hypothesis refinement}$, a technique that more closely mirrors the human inductive process than standard input-output prompting. Iterative hypothesis refinement employs a three-step process: proposing, selecting, and refining hypotheses in the form of textual rules. By examining the intermediate rules, we observe that LMs are phenomenal $\\textit{hypothesis proposers}$ (i.e., generating candidate rules), and when coupled with a (task-specific) symbolic interpreter that is able to systematically filter the proposed set of rules, this hybrid approach achieves strong results across inductive reasoning benchmarks that require inducing causal relations, language-like instructions, and symbolic concepts. However, they also behave as puzzling $\\textit{inductive reasoners}$, showing notable performance gaps between rule induction (i.e., identifying plausible rules) and rule application (i.e., applying proposed rules to instances), suggesting that LMs are proposing hypotheses without being able to actually apply the rules. Through empirical and human analyses, we further reveal several discrepancies between the inductive reasoning processes of LMs and humans, shedding light on both the potentials and limitations of using LMs in inductive reasoning tasks.", "pdf": "https://openreview.net/pdf/4032df754ed3bcf600b7b70606e1de283e796547.pdf"} {"title": "Beyond Weisfeiler-Lehman: A Quantitative Framework for GNN Expressiveness", "url": "https://openreview.net/forum?id=HSKaGOi7Ar", "detail_url": "https://openreview.net/forum?id=HSKaGOi7Ar", "authors": "Bohang Zhang,Jingchu Gai,Yiheng Du,Qiwei Ye,Di He,Liwei Wang", "tags": "ICLR 2024,Oral", "abstract": "Designing expressive Graph Neural Networks (GNNs) is a fundamental topic in the graph learning community. So far, GNN expressiveness has been primarily assessed via the Weisfeiler-Lehman (WL) hierarchy. However, such an expressivity measure has notable limitations: it is inherently coarse, qualitative, and may not well reflect practical requirements (e.g., the ability to encode substructures). In this paper, we introduce a novel framework for quantitatively studying the expressiveness of GNN architectures, addressing all the above limitations. Specifically, we identify a fundamental expressivity measure termed homomorphism expressivity, which quantifies the ability of GNN models to count graphs under homomorphism. Homomorphism expressivity offers a complete and practical assessment tool: the completeness enables direct expressivity comparisons between GNN models, while the practicality allows for understanding concrete GNN abilities such as subgraph counting. By examining four classes of prominent GNNs as case studies, we derive simple, unified, and elegant descriptions of their homomorphism expressivity for both invariant and equivariant settings. Our results provide novel insights into a series of previous work, unify the landscape of different subareas in the community, and settle several open questions. Empirically, extensive experiments on both synthetic and real-world tasks verify our theory, showing that the practical performance of GNN models aligns well with the proposed metric.", "pdf": "https://openreview.net/pdf/1cdf9d7930ee08e1c02c2c2819a16e7a2cc56a4b.pdf"} {"title": "MathVista: Evaluating Mathematical Reasoning of Foundation Models in Visual Contexts", "url": "https://openreview.net/forum?id=KUNzEQMWU7", "detail_url": "https://openreview.net/forum?id=KUNzEQMWU7", "authors": "Pan Lu,Hritik Bansal,Tony Xia,Jiacheng Liu,Chunyuan Li,Hannaneh Hajishirzi,Hao Cheng,Kai-Wei Chang,Michel Galley,Jianfeng Gao", "tags": "ICLR 2024,Oral", "abstract": "Large Language Models (LLMs) and Large Multimodal Models (LMMs) exhibit impressive problem-solving skills in many tasks and domains, but their ability in mathematical reasoning in visual contexts has not been systematically studied. To bridge this gap, we present MathVista, a benchmark designed to combine challenges from diverse mathematical and visual tasks. It consists of 6,141 examples, derived from 28 existing multimodal datasets involving mathematics and 3 newly created datasets (i.e., IQTest, FunctionQA, and PaperQA). Completing these tasks requires fine-grained, deep visual understanding and compositional reasoning, which all state-of-the-art foundation models find challenging. With MathVista, we have conducted a comprehensive, quantitative evaluation of 12 prominent foundation models. The best-performing GPT-4V model achieves an overall accuracy of 49.9%, substantially outperforming Bard, the second-best performer, by 15.1%. Our in-depth analysis reveals that the superiority of GPT-4V is mainly attributed to its enhanced visual perception and mathematical reasoning. However, GPT-4V still falls short of human performance by 10.4%, as it often struggles to understand complex figures and perform rigorous reasoning. This significant gap underscores the critical role that MathVista will play in the development of general-purpose AI agents capable of tackling mathematically intensive and visually rich real-world tasks. We further explore the new ability of self-verification, the application of self-consistency, and the interactive chatbot capabilities of GPT-4V, highlighting its promising potential for future research. The project is available at https://mathvista.github.io/.", "pdf": "https://openreview.net/pdf/787a339a2bb6e601216540a43a659322ff3e4e9e.pdf"} {"title": "Protein Discovery with Discrete Walk-Jump Sampling", "url": "https://openreview.net/forum?id=zMPHKOmQNb", "detail_url": "https://openreview.net/forum?id=zMPHKOmQNb", "authors": "Nathan C. Frey,Dan Berenberg,Karina Zadorozhny,Joseph Kleinhenz,Julien Lafrance-Vanasse,Isidro Hotzel,Yan Wu,Stephen Ra,Richard Bonneau,Kyunghyun Cho,Andreas Loukas,Vladimir Gligorijevic,Saeed Saremi", "tags": "ICLR 2024,Oral", "abstract": "We resolve difficulties in training and sampling from a discrete generative model by learning a smoothed energy function, sampling from the smoothed data manifold with Langevin Markov chain Monte Carlo (MCMC), and projecting back to the true data manifold with one-step denoising. Our $\\textit{Discrete Walk-Jump Sampling}$ formalism combines the contrastive divergence training of an energy-based model and improved sample quality of a score-based model, while simplifying training and sampling by requiring only a single noise level. We evaluate the robustness of our approach on generative modeling of antibody proteins and introduce the $\\textit{distributional conformity score}$ to benchmark protein generative models. By optimizing and sampling from our models for the proposed distributional conformity score, 97-100\\% of generated samples are successfully expressed and purified and 70\\% of functional designs show equal or improved binding affinity compared to known functional antibodies on the first attempt in a single round of laboratory experiments. We also report the first demonstration of long-run fast-mixing MCMC chains where diverse antibody protein classes are visited in a single MCMC chain.", "pdf": "https://openreview.net/pdf/bd2adb2c58bf36a145a6eb40e827467a71d7aaf1.pdf"} {"title": "ExeDec: Execution Decomposition for Compositional Generalization in Neural Program Synthesis", "url": "https://openreview.net/forum?id=oTRwljRgiv", "detail_url": "https://openreview.net/forum?id=oTRwljRgiv", "authors": "Kensen Shi,Joey Hong,Yinlin Deng,Pengcheng Yin,Manzil Zaheer,Charles Sutton", "tags": "ICLR 2024,Oral", "abstract": "When writing programs, people have the ability to tackle a new complex task by decomposing it into smaller and more familiar subtasks. While it is difficult to measure whether neural program synthesis methods have similar capabilities, we can measure whether they compositionally generalize, that is, whether a model that has been trained on the simpler subtasks is subsequently able to solve more complex tasks. In this paper, we characterize several different forms of compositional generalization that are desirable in program synthesis, forming a meta-benchmark which we use to create generalization tasks for two popular datasets, RobustFill and DeepCoder. We then propose ExeDec, a novel decomposition-based synthesis strategy that predicts execution subgoals to solve problems step-by-step informed by program execution at each step. When used with Transformer models trained from scratch, ExeDec has better synthesis performance and greatly improved compositional generalization ability compared to baselines. Finally, we use our benchmarks to demonstrate that LLMs struggle to compositionally generalize when asked to do programming-by-example in a few-shot setting, but an ExeDec-style prompting approach can improve the generalization ability and overall performance.", "pdf": "https://openreview.net/pdf/a69b0e436a40cc8061344c5a3db100f446f53ee6.pdf"} {"title": "Batched Low-Rank Adaptation of Foundation Models", "url": "https://openreview.net/forum?id=w4abltTZ2f", "detail_url": "https://openreview.net/forum?id=w4abltTZ2f", "authors": "Yeming Wen,Swarat Chaudhuri", "tags": "ICLR 2024,Oral", "abstract": "Low-Rank Adaptation (LoRA) has recently gained attention for fine-tuning foundation models by incorporating trainable low-rank matrices, thereby reducing the number of trainable parameters. While \\lora/ offers numerous advantages, its applicability for real-time serving to a diverse and global user base \nis constrained by its incapability to handle multiple task-specific adapters efficiently. This imposes a performance bottleneck in scenarios requiring personalized, task-specific adaptations for each incoming request.\n\nTo address this, we introduce FLoRA (Fast LoRA), a framework in which each input example in a minibatch can be associated with its unique low-rank adaptation weights, allowing for efficient batching of heterogeneous requests. We empirically demonstrate that \\flora/ retains the performance merits of \\lora/, showcasing competitive results on the MultiPL-E code generation benchmark spanning over 8 languages and a multilingual speech recognition task across 6 languages.", "pdf": "https://openreview.net/pdf/49eea165a2219adfe98557e7d54b6ca13ebb7db9.pdf"} {"title": "Improved Active Learning via Dependent Leverage Score Sampling", "url": "https://openreview.net/forum?id=IYxDy2jDFL", "detail_url": "https://openreview.net/forum?id=IYxDy2jDFL", "authors": "Atsushi Shimizu,Xiaoou Cheng,Christopher Musco,Jonathan Weare", "tags": "ICLR 2024,Oral", "abstract": "We show how to obtain improved active learning methods in the agnostic (adversarial noise) setting by combining marginal leverage score sampling with non-independent sampling strategies that promote spatial coverage. In particular, we propose an easily implemented method based on the \\emph{pivotal sampling algorithm}, which we test on problems motivated by learning-based methods for parametric PDEs and uncertainty quantification. In comparison to independent sampling, our method reduces the number of samples needed to reach a given target accuracy by up to $50\\%$.\n\nWe support our findings with two theoretical results. First, we show that any non-independent leverage score sampling method that obeys a weak \\emph{one-sided $\\ell_{\\infty}$ independence condition} (which includes pivotal sampling) can actively learn $d$ dimensional linear functions with $O(d\\log d)$ samples, matching independent sampling. This result extends recent work on matrix Chernoff bounds under $\\ell_{\\infty}$ independence, and may be of interest for analyzing other sampling strategies beyond pivotal sampling. Second, we show that, for the important case of polynomial regression, our pivotal method obtains an improved bound of $O(d)$ samples.", "pdf": "https://openreview.net/pdf/98b84fd00d5f25df5c6927e10d5e51cde527543e.pdf"} {"title": "Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs", "url": "https://openreview.net/forum?id=uNrFpDPMyo", "detail_url": "https://openreview.net/forum?id=uNrFpDPMyo", "authors": "Suyu Ge,Yunan Zhang,Liyuan Liu,Minjia Zhang,Jiawei Han,Jianfeng Gao", "tags": "ICLR 2024,Oral", "abstract": "In this study, we introduce adaptive KV cache compression, a plug-and-play method that reduces the memory footprint of generative inference for Large Language Models (LLMs). Different from the conventional KV cache that retains key and value vectors for all context tokens, we conduct targeted profiling to discern the intrinsic structure of attention modules. Based on the recognized structure, we then construct the KV cache in an adaptive manner: evicting long-range contexts on attention heads emphasizing local contexts, discarding non-special tokens on attention heads centered on special tokens, and only employing the standard KV cache for attention heads that broadly attend to all tokens. Moreover, with the lightweight attention profiling used to guide the construction of the adaptive KV cache, FastGen can be deployed without resource-intensive fine-tuning or re-training. In our experiments across various asks, FastGen demonstrates substantial reduction on GPU memory consumption with negligible generation quality loss. We will release our code and the compatible CUDA kernel for reproducibility.", "pdf": "https://openreview.net/pdf/757a55aa24be0345fe1687e09fa5ca448934e52f.pdf"} {"title": "One-shot Empirical Privacy Estimation for Federated Learning", "url": "https://openreview.net/forum?id=0BqyZSWfzo", "detail_url": "https://openreview.net/forum?id=0BqyZSWfzo", "authors": "Galen Andrew,Peter Kairouz,Sewoong Oh,Alina Oprea,Hugh Brendan McMahan,Vinith Menon Suriyakumar", "tags": "ICLR 2024,Oral", "abstract": "Privacy estimation techniques for differentially private (DP) algorithms are useful for comparing against analytical bounds, or to empirically measure privacy loss in settings where known analytical bounds are not tight. However, existing privacy auditing techniques usually make strong assumptions on the adversary (e.g., knowledge of intermediate model iterates or the training data distribution), are tailored to specific tasks, model architectures, or DP algorithm, and/or require retraining the model many times (typically on the order of thousands). These shortcomings make deploying such techniques at scale difficult in practice, especially in federated settings where model training can take days or weeks. In this work, we present a novel \u201cone-shot\u201d approach that can systematically address these challenges, allowing efficient auditing or estimation of the privacy loss of a model during the same, single training run used to fit model parameters, and without requiring any a priori knowledge about the model architecture, task, or DP algorithm. We show that our method provides provably correct estimates for the privacy loss under the Gaussian mechanism, and we demonstrate its performance on a well-established FL benchmark dataset under several adversarial threat models.", "pdf": "https://openreview.net/pdf/7808a17938a3798b99894957cc00136bbf609c65.pdf"} {"title": "SWE-bench: Can Language Models Resolve Real-world Github Issues?", "url": "https://openreview.net/forum?id=VTF8yNQM66", "detail_url": "https://openreview.net/forum?id=VTF8yNQM66", "authors": "Carlos E Jimenez,John Yang,Alexander Wettig,Shunyu Yao,Kexin Pei,Ofir Press,Karthik R Narasimhan", "tags": "ICLR 2024,Oral", "abstract": "Language models have outpaced our ability to evaluate them effectively, but for their future development it is essential to study the frontier of their capabilities. We find real-world software engineering to be a rich, sustainable, and challenging testbed for evaluating the next generation of language models. To this end, we introduce SWE-bench, an evaluation framework consisting of 2,294 software engineering problems drawn from real GitHub issues and corresponding pull requests across 12 popular Python repositories. Given a codebase along with a description of an issue to be resolved, a language model is tasked with editing the codebase to address the issue. Resolving issues in SWE-bench frequently requires understanding and coordinating changes across multiple functions, classes, and even files simultaneously, calling for models to interact with execution environments, process extremely long contexts and perform complex reasoning that goes far beyond traditional code generation tasks. Our evaluations show that both state-of-the-art proprietary models and our fine-tuned model SWE-Llama can resolve only the simplest issues. The best-performing model, Claude 2, is able to solve a mere 1.96% of the issues. Advances on SWE-bench represent steps towards LMs that are more practical, intelligent, and autonomous.", "pdf": "https://openreview.net/pdf/c2a76eb44300a738cbd7cb95f5bc04df621f4d25.pdf"} {"title": "ReLU Strikes Back: Exploiting Activation Sparsity in Large Language Models", "url": "https://openreview.net/forum?id=osoWxY8q2E", "detail_url": "https://openreview.net/forum?id=osoWxY8q2E", "authors": "Seyed Iman Mirzadeh,Keivan Alizadeh-Vahid,Sachin Mehta,Carlo C del Mundo,Oncel Tuzel,Golnoosh Samei,Mohammad Rastegari,Mehrdad Farajtabar", "tags": "ICLR 2024,Oral", "abstract": "Large Language Models (LLMs) with billions of parameters have drastically transformed AI applications. However, their demanding computation during inference has raised significant challenges for deployment on resource-constrained devices. Despite recent trends favoring alternative activation functions such as GELU or SiLU, known for increased computation, this study strongly advocates for reinstating ReLU activation in LLMs. We demonstrate that using the ReLU activation function has a negligible impact on convergence and performance while significantly reducing computation and weight transfer. This reduction is particularly valuable during the memory-bound inference step, where efficiency is paramount. Exploring sparsity patterns in ReLU-based LLMs, we unveil the reutilization of activated neurons for generating new tokens and leveraging these insights, we propose practical strategies to substantially reduce LLM inference computation up to three times, using ReLU activations with minimal performance trade-offs.", "pdf": "https://openreview.net/pdf/a407324c94efa754d43a6c1718e24541d34e2f24.pdf"} {"title": "On the Joint Interaction of Models, Data, and Features", "url": "https://openreview.net/forum?id=ze7DOLi394", "detail_url": "https://openreview.net/forum?id=ze7DOLi394", "authors": "Yiding Jiang,Christina Baek,J Zico Kolter", "tags": "ICLR 2024,Oral", "abstract": "Learning features from data is one of the defining characteristics of deep learning,\nbut the theoretical understanding of the role features play in deep learning is still in\nearly development. To address this gap, we introduce a new tool, the interaction\ntensor, for empirically analyzing the interaction between data and model through\nfeatures. With the interaction tensor, we make several key observations about\nhow features are distributed in data and how models with different random seeds\nlearn different features. Based on these observations, we propose a conceptual\nframework for feature learning. Under this framework, the expected accuracy for a\nsingle hypothesis and agreement for a pair of hypotheses can both be derived in\nclosed form. We demonstrate that the proposed framework can explain empirically\nobserved phenomena, including the recently discovered Generalization Disagreement Equality (GDE) that allows for estimating the generalization error with only\nunlabeled data. Further, our theory also provides explicit construction of natural\ndata distributions that break the GDE. Thus, we believe this work provides valuable\nnew insight into our understanding of feature learning.", "pdf": "https://openreview.net/pdf/86a102e47488a58d90fc222cf560db16f68dc65d.pdf"} {"title": "Topological data analysis on noisy quantum computers", "url": "https://openreview.net/forum?id=dLrhRIMVmB", "detail_url": "https://openreview.net/forum?id=dLrhRIMVmB", "authors": "Ismail Yunus Akhalwaya,Shashanka Ubaru,Kenneth L. Clarkson,Mark S. Squillante,Vishnu Jejjala,Yang-Hui He,Kugendran Naidoo,Vasileios Kalantzis,Lior Horesh", "tags": "ICLR 2024,Oral", "abstract": "Topological data analysis (TDA) is a powerful technique for extracting complex and valuable shape-related summaries of high-dimensional data. However, the computational demands of classical algorithms for computing TDA are exorbitant, and quickly become impractical for high-order characteristics. Quantum computers offer the potential of achieving significant speedup for certain computational problems. Indeed, TDA has been purported to be one such problem, yet, quantum computing algorithms proposed for the problem, such as the original Quantum TDA (QTDA) formulation by Lloyd, Garnerone and Zanardi, require fault-tolerance qualifications that are currently unavailable. In this study, we present NISQ-TDA, a fully implemented end-to-end quantum machine learning algorithm needing only a short circuit-depth, that is applicable to high-dimensional classical data, and with provable asymptotic speedup for certain classes of problems. The algorithm neither suffers from the data-loading problem nor does it need to store the input data on the quantum computer explicitly. The algorithm was successfully executed on quantum computing devices, as well as on noisy quantum simulators, applied to small datasets. Preliminary empirical results suggest that the algorithm is robust to noise.", "pdf": "https://openreview.net/pdf/07776ae8b91f82e5061d6b246a4e9aacc7bddb41.pdf"} {"title": "Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection", "url": "https://openreview.net/forum?id=hSyW5go0v8", "detail_url": "https://openreview.net/forum?id=hSyW5go0v8", "authors": "Akari Asai,Zeqiu Wu,Yizhong Wang,Avirup Sil,Hannaneh Hajishirzi", "tags": "ICLR 2024,Oral", "abstract": "Despite their remarkable capabilities, large language models (LLMs) often produce responses containing factual inaccuracies due to their sole reliance on the parametric knowledge they encapsulate. Retrieval-Augmented Generation (RAG), an ad hoc approach that augments LMs with retrieval of relevant knowledge, decreases such issues. However, indiscriminately retrieving and incorporating a fixed number of retrieved passages, regardless of whether retrieval is necessary, or passages are relevant, diminishes LM versatility or can lead to unhelpful response generation. We introduce a new framework called **Self-Reflective Retrieval-Augmented Generation (Self-RAG)** that enhances an LM's quality and factuality through retrieval and self-reflection. \nOur framework trains a single arbitrary LM that adaptively retrieves passages on-demand, and generates and reflects on retrieved passages and its generations using special tokens, called {\\it reflection} tokens. Generating reflection tokens makes the LM controllable during the inference phase, enabling it to tailor its behavior to diverse task requirements. \nExperiments show that Self-RAG (7B and 13B parameters) significantly outperforms state-of-the-art LLMs and retrieval-augmented models on a diverse set of tasks. \nSpecifically, Self-RAG outperforms ChatGPT and retrieval-augmented Llama2-chat on Open-domain QA, reasoning, and fact verification tasks, and it shows significant gains in improving factuality and citation accuracy for long-form generations relative to these models. Our code and trained models are available at https://selfrag.github.io/", "pdf": "https://openreview.net/pdf/9a78cf641fab9032078e65ae2734293ae8e2f398.pdf"} {"title": "\"What Data Benefits My Classifier?\" Enhancing Model Performance and Interpretability through Influence-Based Data Selection", "url": "https://openreview.net/forum?id=HE9eUQlAvo", "detail_url": "https://openreview.net/forum?id=HE9eUQlAvo", "authors": "Anshuman Chhabra,Peizhao Li,Prasant Mohapatra,Hongfu Liu", "tags": "ICLR 2024,Oral", "abstract": "Classification models are ubiquitously deployed in society and necessitate high utility, fairness, and robustness performance. Current research efforts mainly focus on improving model architectures and learning algorithms on fixed datasets to achieve this goal. In contrast, in this paper, we address an orthogonal yet crucial problem: given a fixed convex learning model (or a convex surrogate for a non-convex model) and a function of interest, we assess what data benefits the model by interpreting the feature space, and then aim to improve performance as measured by this function. To this end, we propose the use of influence estimation models for interpreting the classifier's performance from the perspective of the data feature space. Additionally, we propose data selection approaches based on influence that enhance model utility, fairness, and robustness. Through extensive experiments on synthetic and real-world datasets, we validate and demonstrate the effectiveness of our approaches not only for conventional classification scenarios, but also under more challenging scenarios such as distribution shifts, fairness poisoning attacks, utility evasion attacks, online learning, and active learning.", "pdf": "https://openreview.net/pdf/c9c086d91e0480dcd349f7bb625a5031fabcc53a.pdf"} {"title": "Generative Modeling with Phase Stochastic Bridge", "url": "https://openreview.net/forum?id=tUtGjQEDd4", "detail_url": "https://openreview.net/forum?id=tUtGjQEDd4", "authors": "Tianrong Chen,Jiatao Gu,Laurent Dinh,Evangelos Theodorou,Joshua M. Susskind,Shuangfei Zhai", "tags": "ICLR 2024,Oral", "abstract": "Diffusion models (DMs) represent state-of-the-art generative models for continuous inputs. DMs work by constructing a Stochastic Differential Equation (SDE) in the input space (ie, position space), and using a neural network to reverse it. In this work, we introduce a novel generative modeling framework grounded in \\textbf{phase space dynamics}, where a phase space is defined as {an augmented space encompassing both position and velocity.} Leveraging insights from Stochastic Optimal Control, we construct a path measure in the phase space that enables efficient sampling. {In contrast to DMs, our framework demonstrates the capability to generate realistic data points at an early stage of dynamics propagation.} This early prediction sets the stage for efficient data generation by leveraging additional velocity information along the trajectory. On standard image generation benchmarks, our model yields favorable performance over baselines in the regime of small Number of Function Evaluations (NFEs). Furthermore, our approach rivals the performance of diffusion models equipped with efficient sampling techniques, underscoring its potential as a new tool generative modeling.", "pdf": "https://openreview.net/pdf/5d5ddf9cd03dbc97896ca72e62060b33d19f59e7.pdf"} {"title": "Zipformer: A faster and better encoder for automatic speech recognition", "url": "https://openreview.net/forum?id=9WD9KwssyT", "detail_url": "https://openreview.net/forum?id=9WD9KwssyT", "authors": "Zengwei Yao,Liyong Guo,Xiaoyu Yang,Wei Kang,Fangjun Kuang,Yifan Yang,Zengrui Jin,Long Lin,Daniel Povey", "tags": "ICLR 2024,Oral", "abstract": "The Conformer has become the most popular encoder model for automatic speech recognition (ASR). It adds convolution modules to a transformer to learn both local and global dependencies. In this work we describe a faster, more memory-efficient, and better-performing transformer, called Zipformer. Modeling changes include: 1) a U-Net-like encoder structure where middle stacks operate at lower frame rates; 2) reorganized block structure with more modules, within which we re-use attention weights for efficiency; 3) a modified form of LayerNorm called BiasNorm allows us to retain some length information; 4) new activation functions SwooshR and SwooshL work better than Swish. We also propose a new optimizer, called ScaledAdam, which scales the update by each tensor's current scale to keep the relative change about the same, and also explictly learns the parameter scale. It achieves faster converge and better performance than Adam. Extensive experiments on LibriSpeech, Aishell-1, and WenetSpeech datasets demonstrate the effectiveness of our proposed Zipformer over other state-of-the-art ASR models. Our code is publicly available at https://github.com/k2-fsa/icefall.", "pdf": "https://openreview.net/pdf/73f36dfc4a1fa9d3dd37fdb3cb11d5be19364046.pdf"} {"title": "MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework", "url": "https://openreview.net/forum?id=VtmBAGCN7o", "detail_url": "https://openreview.net/forum?id=VtmBAGCN7o", "authors": "Sirui Hong,Mingchen Zhuge,Jonathan Chen,Xiawu Zheng,Yuheng Cheng,Jinlin Wang,Ceyao Zhang,Zili Wang,Steven Ka Shing Yau,Zijuan Lin,Liyang Zhou,Chenyu Ran,Lingfeng Xiao,Chenglin Wu,J\u00fcrgen Schmidhuber", "tags": "ICLR 2024,Oral", "abstract": "Recently, remarkable progress has been made on automated problem solving through societies of agents based on large language models (LLMs). Previous LLM-based multi-agent systems can already solve simple dialogue tasks. More complex tasks, however, face challenges through logic inconsistencies due to cascading hallucinations caused by naively chaining LLMs. Here we introduce MetaGPT, an innovative meta-programming framework incorporating efficient human workflows into LLM-based multi-agent collaborations. MetaGPT encodes Standardized Operating Procedures (SOPs) into prompt sequences for more streamlined workflows, thus allowing agents with human-like domain expertise to verify intermediate results and reduce errors. MetaGPT utilizes an assembly line paradigm to assign diverse roles to various agents, efficiently breaking down complex tasks into subtasks involving many agents working together. On collaborative software engineering benchmarks, MetaGPT generates more coherent solutions than previous chat-based multi-agent systems.", "pdf": "https://openreview.net/pdf/474fc6dad3bd9bf7fdb97c7cd72b2cc0649a9647.pdf"} {"title": "ValUES: A Framework for Systematic Validation of Uncertainty Estimation in Semantic Segmentation", "url": "https://openreview.net/forum?id=yV6fD7LYkF", "detail_url": "https://openreview.net/forum?id=yV6fD7LYkF", "authors": "Kim-Celine Kahl,Carsten T. L\u00fcth,Maximilian Zenk,Klaus Maier-Hein,Paul F Jaeger", "tags": "ICLR 2024,Oral", "abstract": "Uncertainty estimation is an essential and heavily-studied component for the reliable application of semantic segmentation methods. While various studies exist claiming methodological advances on the one hand, and successful application on the other hand, the field is currently hampered by a gap between theory and practice leaving fundamental questions unanswered: Can data-related and model-related uncertainty really be separated in practice? Which components of an uncertainty method are essential for real-world performance? Which uncertainty method works well for which application? In this work, we link this research gap to a lack of systematic and comprehensive evaluation of uncertainty methods. Specifically, we identify three key pitfalls in current literature and present an evaluation framework that bridges the research gap by providing 1) a controlled environment for studying data ambiguities as well as distribution shifts, 2) systematic ablations of relevant method components, and 3) test-beds for the five predominant uncertainty applications: OoD-detection, active learning, failure detection, calibration, and ambiguity modeling. Empirical results on simulated as well as real-world data demonstrate how the proposed framework is able to answer the predominant questions in the field revealing for instance that 1) separation of uncertainty types works on simulated data but does not necessarily translate to real-world data, 2) aggregation of scores is a crucial but currently neglected component of uncertainty methods, 3) While ensembles are performing most robustly across the different downstream tasks and settings, test-time augmentation often constitutes a light-weight alternative. Code is at: https://github.com/IML-DKFZ/values", "pdf": "https://openreview.net/pdf/f1a6b968ddfb2f0ebdeb46499417239973e92e7e.pdf"} {"title": "Finetuning Text-to-Image Diffusion Models for Fairness", "url": "https://openreview.net/forum?id=hnrB5YHoYu", "detail_url": "https://openreview.net/forum?id=hnrB5YHoYu", "authors": "Xudong Shen,Chao Du,Tianyu Pang,Min Lin,Yongkang Wong,Mohan Kankanhalli", "tags": "ICLR 2024,Oral", "abstract": "The rapid adoption of text-to-image diffusion models in society underscores an urgent need to address their biases. Without interventions, these biases could propagate a skewed worldview and restrict opportunities for minority groups. In this work, we frame fairness as a distributional alignment problem. Our solution consists of two main technical contributions: (1) a distributional alignment loss that steers specific characteristics of the generated images towards a user-defined target distribution, and (2) adjusted direct finetuning of diffusion model's sampling process (adjusted DFT), which leverages an adjusted gradient to directly optimize losses defined on the generated images. Empirically, our method markedly reduces gender, racial, and their intersectional biases for occupational prompts. Gender bias is significantly reduced even when finetuning just five soft tokens. Crucially, our method supports diverse perspectives of fairness beyond absolute equality, which is demonstrated by controlling age to a 75% young and 25% old distribution while simultaneously debiasing gender and race. Finally, our method is scalable: it can debias multiple concepts at once by simply including these prompts in the finetuning data. We share code and various fair diffusion model adaptors at https://sail-sg.github.io/finetune-fair-diffusion/.", "pdf": "https://openreview.net/pdf/9fa6cd12f622fa7dffccbd1c62d26545e012eafa.pdf"} {"title": "Knowledge Card: Filling LLMs' Knowledge Gaps with Plug-in Specialized Language Models", "url": "https://openreview.net/forum?id=WbWtOYIzIK", "detail_url": "https://openreview.net/forum?id=WbWtOYIzIK", "authors": "Shangbin Feng,Weijia Shi,Yuyang Bai,Vidhisha Balachandran,Tianxing He,Yulia Tsvetkov", "tags": "ICLR 2024,Oral", "abstract": "By design, large language models (LLMs) are static general-purpose models, expensive to retrain or update frequently. As they are increasingly adopted for knowledge-intensive tasks, it becomes evident that these design choices lead to failures to generate factual, relevant, and up-to-date knowledge. To this end, we propose Knowledge Card, a modular framework to plug in new factual and relevant knowledge into general-purpose LLMs. We first introduce knowledge cards---specialized language models trained on corpora from specific domains and sources. Knowledge cards serve as parametric repositories that are selected at inference time to generate background knowledge for the base LLM. We then propose three content selectors to dynamically select and retain information in documents generated by knowledge cards, specifically controlling for relevance, brevity, and factuality of outputs. Finally, we propose two complementary integration approaches to augment the base LLM with the (relevant, factual) knowledge curated from the specialized LMs. Through extensive experiments, we demonstrate that Knowledge Card achieves state-of-the-art performance on six benchmark datasets. Ultimately, Knowledge Card framework enables dynamic synthesis and updates of knowledge from diverse domains. Its modularity will ensure that relevant knowledge can be continuously updated through the collective efforts of the research community.", "pdf": "https://openreview.net/pdf/93b8f30fd873a0887265f980d789959bfeb89e40.pdf"} {"title": "METRA: Scalable Unsupervised RL with Metric-Aware Abstraction", "url": "https://openreview.net/forum?id=c5pwL0Soay", "detail_url": "https://openreview.net/forum?id=c5pwL0Soay", "authors": "Seohong Park,Oleh Rybkin,Sergey Levine", "tags": "ICLR 2024,Oral", "abstract": "Unsupervised pre-training strategies have proven to be highly effective in natural language processing and computer vision. Likewise, unsupervised reinforcement learning (RL) holds the promise of discovering a variety of potentially useful behaviors that can accelerate the learning of a wide array of downstream tasks. Previous unsupervised RL approaches have mainly focused on pure exploration and mutual information skill learning. However, despite the previous attempts, making unsupervised RL truly scalable still remains a major open challenge: pure exploration approaches might struggle in complex environments with large state spaces, where covering every possible transition is infeasible, and mutual information skill learning approaches might completely fail to explore the environment due to the lack of incentives. To make unsupervised RL scalable to complex, high-dimensional environments, we propose a novel unsupervised RL objective, which we call Metric-Aware Abstraction (METRA). Our main idea is, instead of directly covering the entire state space, to only cover a compact latent space $\\mathcal{Z}$ that is metrically connected to the state space $\\mathcal{S}$ by temporal distances. By learning to move in every direction in the latent space, METRA obtains a tractable set of diverse behaviors that approximately cover the state space, being scalable to high-dimensional environments. Through our experiments in five locomotion and manipulation environments, we demonstrate that METRA can discover a variety of useful behaviors even in complex, pixel-based environments, being the first unsupervised RL method that discovers diverse locomotion behaviors in pixel-based Quadruped and Humanoid. Our code and videos are available at https://seohong.me/projects/metra/", "pdf": "https://openreview.net/pdf/957e22f4e911e7ad35fff291d142a0a622982c0a.pdf"} {"title": "Meta Continual Learning Revisited: Implicitly Enhancing Online Hessian Approximation via Variance Reduction", "url": "https://openreview.net/forum?id=TpD2aG1h0D", "detail_url": "https://openreview.net/forum?id=TpD2aG1h0D", "authors": "Yichen Wu,Long-Kai Huang,Renzhen Wang,Deyu Meng,Ying Wei", "tags": "ICLR 2024,Oral", "abstract": "Regularization-based methods have so far been among the *de facto* choices for continual learning. Recent theoretical studies have revealed that these methods all boil down to relying on the Hessian matrix approximation of model weights. \nHowever, these methods suffer from suboptimal trade-offs between knowledge transfer and forgetting due to fixed and unchanging Hessian estimations during training.\nAnother seemingly parallel strand of Meta-Continual Learning (Meta-CL) algorithms enforces alignment between gradients of previous tasks and that of the current task. \nIn this work we revisit Meta-CL and for the first time bridge it with regularization-based methods. Concretely, Meta-CL implicitly approximates Hessian in an online manner, which enjoys the benefits of timely adaptation but meantime suffers from high variance induced by random memory buffer sampling. \nWe are thus highly motivated to combine the best of both worlds, through the proposal of Variance Reduced Meta-CL (VR-MCL) to achieve both timely and accurate Hessian approximation.\nThrough comprehensive experiments across three datasets and various settings, we consistently observe that VR-MCL outperforms other SOTA methods, which further validates the effectiveness of VR-MCL.", "pdf": "https://openreview.net/pdf/28a552d86247251eb46610359a599b07e5b3e5eb.pdf"} {"title": "Improving Convergence and Generalization Using Parameter Symmetries", "url": "https://openreview.net/forum?id=L0r0GphlIL", "detail_url": "https://openreview.net/forum?id=L0r0GphlIL", "authors": "Bo Zhao,Robert M. Gower,Robin Walters,Rose Yu", "tags": "ICLR 2024,Oral", "abstract": "In many neural networks, different values of the parameters may result in the same loss value. Parameter space symmetries are loss-invariant transformations that change the model parameters. Teleportation applies such transformations to accelerate optimization. However, the exact mechanism behind this algorithm's success is not well understood. In this paper, we show that teleportation not only speeds up optimization in the short-term, but gives overall faster time to convergence. Additionally, teleporting to minima with different curvatures improves generalization, which suggests a connection between the curvature of the minimum and generalization ability. Finally, we show that integrating teleportation into a wide range of optimization algorithms and optimization-based meta-learning improves convergence. Our results showcase the versatility of teleportation and demonstrate the potential of incorporating symmetry in optimization.", "pdf": "https://openreview.net/pdf/5c8faf4be06ab48f03f7a0b88199632f8db72f7c.pdf"} {"title": "Flow Matching on General Geometries", "url": "https://openreview.net/forum?id=g7ohDlTITL", "detail_url": "https://openreview.net/forum?id=g7ohDlTITL", "authors": "Ricky T. Q. Chen,Yaron Lipman", "tags": "ICLR 2024,Oral", "abstract": "We propose Riemannian Flow Matching (RFM), a simple yet powerful framework for training continuous normalizing flows on manifolds. Existing methods for generative modeling on manifolds either require expensive simulation, are inherently unable to scale to high dimensions, or use approximations for limiting quantities that result in biased training objectives. Riemannian Flow Matching bypasses these limitations and offers several advantages over previous approaches: it is simulation-free on simple geometries, does not require divergence computation, and computes its target vector field in closed-form. The key ingredient behind RFM is the construction of a relatively simple premetric for defining target vector fields, which encompasses the existing Euclidean case. To extend to general geometries, we rely on the use of spectral decompositions to efficiently compute premetrics on the fly. Our method achieves state-of-the-art performance on real-world non-Euclidean datasets, and we demonstrate tractable training on general geometries, including triangular meshes with highly non-trivial curvature and boundaries.", "pdf": "https://openreview.net/pdf/00e980dec1d5ee17094141c71986553014f8a41a.pdf"} {"title": "Ghost on the Shell: An Expressive Representation of General 3D Shapes", "url": "https://openreview.net/forum?id=Ad87VjRqUw", "detail_url": "https://openreview.net/forum?id=Ad87VjRqUw", "authors": "Zhen Liu,Yao Feng,Yuliang Xiu,Weiyang Liu,Liam Paull,Michael J. Black,Bernhard Sch\u00f6lkopf", "tags": "ICLR 2024,Oral", "abstract": "The creation of photorealistic virtual worlds requires the accurate modeling of 3D surface geometry for a wide range of objects. For this, meshes are appealing since they enable 1) fast physics-based rendering with realistic material and lighting, 2) physical simulation, and 3) are memory-efficient for modern graphics pipelines. Recent work on reconstructing and statistically modeling 3D shape, however, has critiqued meshes as being topologically inflexible. To capture a wide range of object shapes, any 3D representation must be able to model solid, watertight, shapes as well as thin, open, surfaces. Recent work has focused on the former, and methods for reconstructing open surfaces do not support fast reconstruction with material and lighting or unconditional generative modelling. Inspired by the observation that open surfaces can be seen as islands floating on watertight surfaces, we parametrize open surfaces by defining a manifold signed distance field on watertight templates. With this parametrization, we further develop a grid-based and differentiable representation that parametrizes both watertight and non-watertight meshes of arbitrary topology. Our new representation, called Ghost-on-the-Shell (G-Shell), enables two important applications: differentiable rasterization-based reconstruction from multiview images and generative modelling of non-watertight meshes. We empirically demonstrate that G-Shell achieves state-of-the-art performance on non-watertight mesh reconstruction and generation tasks, while also performing effectively for watertight meshes.", "pdf": "https://openreview.net/pdf/97f11bc98d70c1fbee4e5f3325299c53225c6bfc.pdf"} {"title": "W\u00fcrstchen: An Efficient Architecture for Large-Scale Text-to-Image Diffusion Models", "url": "https://openreview.net/forum?id=gU58d5QeGv", "detail_url": "https://openreview.net/forum?id=gU58d5QeGv", "authors": "Pablo Pernias,Dominic Rampas,Mats Leon Richter,Christopher Pal,Marc Aubreville", "tags": "ICLR 2024,Oral", "abstract": "We introduce W\u00fcrstchen, a novel architecture for text-to-image synthesis that combines competitive performance with unprecedented cost-effectiveness for large-scale text-to-image diffusion models.\nA key contribution of our work is to develop a latent diffusion technique in which we learn a detailed but extremely compact semantic image representation used to guide the diffusion process. This highly compressed representation of an image provides much more detailed guidance compared to latent representations of language and this significantly reduces the computational requirements to achieve state-of-the-art results. Our approach also improves the quality of text-conditioned image generation based on our user preference study.\nThe training requirements of our approach consists of 24,602 A100-GPU hours - compared to Stable Diffusion 2.1's 200,000 GPU hours. \nOur approach also requires less training data to achieve these results. Furthermore, our compact latent representations allows us to perform inference over twice as fast, slashing the usual costs and carbon footprint of a state-of-the-art (SOTA) diffusion model significantly, without compromising the end performance. In a broader comparison against SOTA models our approach is substantially more efficient and compares favourably in terms of image quality.\nWe believe that this work motivates more emphasis on the prioritization of both performance and computational accessibility.", "pdf": "https://openreview.net/pdf/31506ae62c31613539a0623777d341cb424cf5b9.pdf"} {"title": "Unified Generative Modeling of 3D Molecules with Bayesian Flow Networks", "url": "https://openreview.net/forum?id=NSVtmmzeRB", "detail_url": "https://openreview.net/forum?id=NSVtmmzeRB", "authors": "Yuxuan Song,Jingjing Gong,Hao Zhou,Mingyue Zheng,Jingjing Liu,Wei-Ying Ma", "tags": "ICLR 2024,Oral", "abstract": "Advanced generative model (\\textit{e.g.}, diffusion model) derived from simplified continuity assumptions of data distribution, though showing promising progress, has been difficult to apply directly to geometry generation applications due to the \\textit{multi-modality} and \\textit{noise-sensitive} nature of molecule geometry. \nThis work introduces Geometric Bayesian Flow Networks (GeoBFN), which naturally fits molecule geometry by modeling diverse modalities in the differentiable parameter space of distributions. GeoBFN maintains the SE-(3) invariant density modeling property by incorporating equivariant inter-dependency modeling on parameters of distributions and unifying the probabilistic modeling of different modalities. \nThrough optimized training and sampling techniques, we demonstrate that GeoBFN achieves state-of-the-art performance on multiple 3D molecule generation benchmarks in terms of generation quality (90.87\\% molecule stability in QM9 and 85.6\\% atom stability in GEOM-DRUG\\footnote{The scores are reported at 1k sampling steps for fair comparison, and our scores could be further improved if sampling sufficiently longer steps.}). GeoBFN can also conduct sampling with any number of steps to reach an optimal trade-off between efficiency and quality (\\textit{e.g.}, 20$\\times$ speedup without sacrificing performance).", "pdf": "https://openreview.net/pdf/ddfe46bc639f9c1dc849398c8b3d978ffd171431.pdf"} {"title": "Small-scale proxies for large-scale Transformer training instabilities", "url": "https://openreview.net/forum?id=d8w0pmvXbZ", "detail_url": "https://openreview.net/forum?id=d8w0pmvXbZ", "authors": "Mitchell Wortsman,Peter J Liu,Lechao Xiao,Katie E Everett,Alexander A Alemi,Ben Adlam,John D Co-Reyes,Izzeddin Gur,Abhishek Kumar,Roman Novak,Jeffrey Pennington,Jascha Sohl-Dickstein,Kelvin Xu,Jaehoon Lee,Justin Gilmer,Simon Kornblith", "tags": "ICLR 2024,Oral", "abstract": "Teams that have trained large Transformer-based models have reported training instabilities at large scale that did not appear when training with the same hyperparameters at smaller scales. Although the causes of such instabilities are of scientific interest, the amount of resources required to reproduce them has made investigation difficult. In this work, we seek ways to reproduce and study training instability at smaller scales. First, we focus on two sources of training instability described in previous work: the growth of logits in attention layers (Dehghani et al., 2023) and divergence of the output logits from the log probabilities (Chowdhery et al., 2022). By measuring the relationship between learning rate and loss across scales, we show that these instabilities also appear in small models when training at high learning rates, and that mitigations previously employed at large scales are equally effective in this regime. This prompts us to investigate the extent to which other known optimizer and model interventions influence the sensitivity of the final loss to changes in the learning rate. To this end, we study methods such as warm-up, weight decay, and the MuParam (Yang et al., 2022), and combine techniques to train small models that achieve similar losses across orders of magnitude of learning rate variation. Finally, to conclude our exploration we study two cases where instabilities can be predicted before they emerge by examining the scaling behavior of model characteristics such as activation and gradient norms.", "pdf": "https://openreview.net/pdf/779db5974973fe74f026f4a70e3f08d16c11cadb.pdf"} {"title": "How I Warped Your Noise: a Temporally-Correlated Noise Prior for Diffusion Models", "url": "https://openreview.net/forum?id=pzElnMrgSD", "detail_url": "https://openreview.net/forum?id=pzElnMrgSD", "authors": "Pascal Chang,Jingwei Tang,Markus Gross,Vinicius C. Azevedo", "tags": "ICLR 2024,Oral", "abstract": "Video editing and generation methods often rely on pre-trained image-based diffusion models. During the diffusion process, however, the reliance on rudimentary noise sampling techniques that do not preserve correlations present in subsequent frames of a video is detrimental to the quality of the results. This either produces high-frequency flickering, or texture-sticking artifacts that are not amenable to post-processing. With this in mind, we propose a novel method for preserving temporal correlations in a sequence of noise samples. This approach is materialized by a novel noise representation, dubbed $\\int$-noise (integral noise), that reinterprets individual noise samples as a continuously integrated noise field: pixel values do not represent discrete values, but are rather the integral of an underlying infinite-resolution noise over the pixel area. Additionally, we propose a carefully tailored transport method that uses $\\int$-noise to accurately advect noise samples over a sequence of frames, maximizing the correlation between different frames while also preserving the noise properties. Our results demonstrate that the proposed $\\int$-noise can be used for a variety of tasks, such as video restoration, surrogate rendering, and conditional video generation.", "pdf": "https://openreview.net/pdf/c35a99656514c0312f7f69d2ecda8ffec1a632de.pdf"} {"title": "Vision Transformers Need Registers", "url": "https://openreview.net/forum?id=2dnO3LLiJ1", "detail_url": "https://openreview.net/forum?id=2dnO3LLiJ1", "authors": "Timoth\u00e9e Darcet,Maxime Oquab,Julien Mairal,Piotr Bojanowski", "tags": "ICLR 2024,Oral", "abstract": "Transformers have recently emerged as a powerful tool for learning visual representations. In this paper, we identify and characterize artifacts in feature maps of both supervised and self-supervised ViT networks. The artifacts correspond to high-norm tokens appearing during inference primarily in low-informative background areas of images, that are repurposed for internal computations. We propose a simple yet effective solution based on providing additional tokens to the input sequence of the Vision Transformer to fill that role. We show that this solution fixes that problem entirely for both supervised and self-supervised models, sets a new state of the art for self-supervised visual models on dense visual prediction tasks, enables object discovery methods with larger models, and most importantly leads to smoother feature maps and attention maps for downstream visual processing.", "pdf": "https://openreview.net/pdf/1db45cd6c97acf30b37c4ee9ac6e79d4f3ac7763.pdf"} {"title": "An Analytical Solution to Gauss-Newton Loss for Direct Image Alignment", "url": "https://openreview.net/forum?id=mE52zURNGc", "detail_url": "https://openreview.net/forum?id=mE52zURNGc", "authors": "Sergei Solonets,Daniil Sinitsyn,Lukas Von Stumberg,Nikita Araslanov,Daniel Cremers", "tags": "ICLR 2024,Oral", "abstract": "Direct image alignment is a widely used technique for relative 6DoF pose estimation between two images, but its accuracy strongly depends on pose initialization.\nTherefore, recent end-to-end frameworks increase the convergence basin of the learned feature descriptors with special training objectives, such as the Gauss-Newton loss.\nHowever, the training data may exhibit bias toward a specific type of motion and pose initialization,\nthus limiting the generalization of these methods.\nIn this work, we derive a closed-form solution to the expected optimum of the Gauss-Newton loss. \nThe solution is agnostic to the underlying feature representation and allows us to dynamically adjust the basin of convergence according to our assumptions about the uncertainty in the current estimates. These properties allow for effective control over the convergence in the alignment process.\nDespite using self-supervised feature embeddings, our solution achieves compelling accuracy w.r.t. the state-of-the-art direct image alignment methods trained end-to-end with pose supervision, and demonstrates improved robustness to pose initialization.\nOur analytical solution exposes some inherent limitations of end-to-end learning with the Gauss-Newton loss, and establishes an intriguing connection between direct image alignment and feature-matching approaches.", "pdf": "https://openreview.net/pdf/8cc6141bff9dadb82d553ab8ac1b1ff6d4f434a9.pdf"} {"title": "Learning Energy Decompositions for Partial Inference in GFlowNets", "url": "https://openreview.net/forum?id=P15CHILQlg", "detail_url": "https://openreview.net/forum?id=P15CHILQlg", "authors": "Hyosoon Jang,Minsu Kim,Sungsoo Ahn", "tags": "ICLR 2024,Oral", "abstract": "This paper studies generative flow networks (GFlowNets) to sample objects from the Boltzmann energy distribution via a sequence of actions. In particular, we focus on improving GFlowNet with partial inference: training flow functions with the evaluation of the intermediate states or transitions. To this end, the recently developed forward-looking GFlowNet reparameterizes the flow functions based on evaluating the energy of intermediate states. However, such an evaluation of intermediate energies may (i) be too expensive or impossible to evaluate and (ii) even provide misleading training signals under large energy fluctuations along the sequence of actions. To resolve this issue, we propose learning energy decompositions for GFlowNets (LED-GFN). Our main idea is to (i) decompose the energy of an object into learnable potential functions defined on state transitions and (ii) reparameterize the flow functions using the potential functions. In particular, to produce informative local credits, we propose to regularize the potential to change smoothly over the sequence of actions. It is also noteworthy that training GFlowNet with our learned potential can preserve the optimal policy. We empirically verify the superiority of LED-GFN in five problems including the generation of unstructured and maximum independent sets, molecular graphs, and RNA sequences.", "pdf": "https://openreview.net/pdf/54bfe1a393ed4a31554ead18c45d5b62548007be.pdf"} {"title": "Approximating Nash Equilibria in Normal-Form Games via Stochastic Optimization", "url": "https://openreview.net/forum?id=cc8h3I3V4E", "detail_url": "https://openreview.net/forum?id=cc8h3I3V4E", "authors": "Ian Gemp,Luke Marris,Georgios Piliouras", "tags": "ICLR 2024,Oral", "abstract": "We propose the first loss function for approximate Nash equilibria of normal-form games that is amenable to unbiased Monte Carlo estimation. This construction allows us to deploy standard non-convex stochastic optimization techniques for approximating Nash equilibria, resulting in novel algorithms with provable guarantees. We complement our theoretical analysis with experiments demonstrating that stochastic gradient descent can outperform previous state-of-the-art approaches.", "pdf": "https://openreview.net/pdf/6116af6dc392a3153d1462f038b9dac4f8305ca6.pdf"} {"title": "Multi-Source Diffusion Models for Simultaneous Music Generation and Separation", "url": "https://openreview.net/forum?id=h922Qhkmx1", "detail_url": "https://openreview.net/forum?id=h922Qhkmx1", "authors": "Giorgio Mariani,Irene Tallini,Emilian Postolache,Michele Mancusi,Luca Cosmo,Emanuele Rodol\u00e0", "tags": "ICLR 2024,Oral", "abstract": "In this work, we define a diffusion-based generative model capable of both music generation and source separation by learning the score of the joint probability density of sources sharing a context. Alongside the classic total inference tasks (i.e., generating a mixture, separating the sources), we also introduce and experiment on the partial generation task of source imputation, where we generate a subset of the sources given the others (e.g., play a piano track that goes well with the drums). Additionally, we introduce a novel inference method for the separation task based on Dirac likelihood functions. We train our model on Slakh2100, a standard dataset for musical source separation, provide qualitative results in the generation settings, and showcase competitive quantitative results in the source separation setting. Our method is the first example of a single model that can handle both generation and separation tasks, thus representing a step toward general audio models.", "pdf": "https://openreview.net/pdf/e9d4d9aabe25aa6dc764b915d5844871ff4bcd7c.pdf"} {"title": "LEGO-Prover: Neural Theorem Proving with Growing Libraries", "url": "https://openreview.net/forum?id=3f5PALef5B", "detail_url": "https://openreview.net/forum?id=3f5PALef5B", "authors": "Haiming Wang,Huajian Xin,Chuanyang Zheng,Zhengying Liu,Qingxing Cao,Yinya Huang,Jing Xiong,Han Shi,Enze Xie,Jian Yin,Zhenguo Li,Xiaodan Liang", "tags": "ICLR 2024,Oral", "abstract": "Despite the success of large language models (LLMs), the task of theorem proving still remains one of the hardest reasoning tasks that is far from being fully solved. Prior methods using language models have demonstrated promising results, but they still struggle to prove even middle school level theorems. One common limitation of these methods is that they assume a fixed theorem library during the whole theorem proving process. However, as we all know, creating new useful theorems or even new theories is not only helpful but crucial and necessary for advancing mathematics and proving harder and deeper results. In this work, we present LEGO-Prover, which employs a growing skill library containing verified lemmas as skills to augment the capability of LLMs used in theorem proving. By constructing the proof modularly, LEGO-Prover enables LLMs to utilize existing skills retrieved from the library and to create new skills during the proving process. These skills are further evolved (by prompting an LLM) to enrich the library on another scale. Modular and reusable skills are constantly added to the library to enable tackling increasingly intricate mathematical problems. Moreover, the learned library further bridges the gap between human proofs and formal proofs by making it easier to impute missing steps. LEGO-Prover advances the state-of-the-art pass rate on miniF2F-valid (48.0\\% to 57.0\\%) and miniF2F-test (45.5\\% to 50.0\\%). During the proving process, LEGO-Prover also generates over 20,000 skills (theorems/lemmas) and adds them to the growing library. Our ablation study indicates that these newly added skills are indeed helpful for proving theorems, resulting in a 4.9\\% improvement in success rate", "pdf": "https://openreview.net/pdf/3133380a86db246c6a9e18dabc0a301196b70cd6.pdf"} {"title": "ASID: Active Exploration for System Identification in Robotic Manipulation", "url": "https://openreview.net/forum?id=jNR6s6OSBT", "detail_url": "https://openreview.net/forum?id=jNR6s6OSBT", "authors": "Marius Memmel,Andrew Wagenmaker,Chuning Zhu,Dieter Fox,Abhishek Gupta", "tags": "ICLR 2024,Oral", "abstract": "Model-free control strategies such as reinforcement learning have shown the ability to learn control strategies without requiring an accurate model or simulator of the world. While this is appealing due to the lack of modeling requirements, such methods can be sample inefficient, making them impractical in many real-world domains. On the other hand, model-based control techniques leveraging accurate simulators can circumvent these challenges and use a large amount of cheap simulation data to learn controllers that can effectively transfer to the real world. The challenge with such model-based techniques is the requirement for an extremely accurate simulation, requiring both the specification of appropriate simulation assets and physical parameters. This requires considerable human effort to design for every environment being considered. In this work, we propose a learning system that can leverage a small amount of real-world data to autonomously refine a simulation model and then plan an accurate control strategy that can be deployed in the real world. Our approach critically relies on utilizing an initial (possibly inaccurate) simulator to design effective exploration policies that, when deployed in the real world, collect high-quality data. We demonstrate the efficacy of this paradigm in identifying articulation, mass, and other physical parameters in several challenging robotic manipulation tasks, and illustrate that only a small amount of real-world data can allow for effective sim-to-real transfer.", "pdf": "https://openreview.net/pdf/f456ef2115275fac2aa0977b3c7db68ed00add89.pdf"} {"title": "Towards a statistical theory of data selection under weak supervision", "url": "https://openreview.net/forum?id=HhfcNgQn6p", "detail_url": "https://openreview.net/forum?id=HhfcNgQn6p", "authors": "Germain Kolossov,Andrea Montanari,Pulkit Tandon", "tags": "ICLR 2024,Oral", "abstract": "Given a sample of size $N$, it is often useful to select a subsample of smaller size $n 0.9$) between the accuracy of rotation prediction and the performance of OCC. This suggests that a classifier that effectively distinguishes different rotations is more likely to excel in OCC, and vice versa. The root cause of this phenomenon can be attributed to the transformation bias in the dataset, where representations learned from transformations already present in the dataset tend to be less effective, making it essential to accurately estimate the transformation distribution before utilizing pretext tasks involving these transformations for reliable self-supervised representation learning. To the end, we propose a novel two-stage method to estimate the transformation distribution within the dataset. In the first stage, we learn general representations through standard contrastive pre-training. In the second stage, we select potentially semantics-preserving samples from the entire augmented dataset, which includes all rotations, by employing density matching with the provided reference distribution. By sorting samples based on semantics-preserving versus shifting transformations, we achieve improved performance on OCC benchmarks.", "pdf": "https://openreview.net/pdf/7bf21feb9bcffeac0dec9c48b1b50d05663b5a16.pdf"} {"title": "Realistic Evaluation of Semi-supervised Learning Algorithms in Open Environments", "url": "https://openreview.net/forum?id=RvUVMjfp8i", "detail_url": "https://openreview.net/forum?id=RvUVMjfp8i", "authors": "Lin-Han Jia,Lan-Zhe Guo,Zhi Zhou,Yu-Feng Li", "tags": "ICLR 2024,Spotlight", "abstract": "Semi-supervised learning (SSL) is a powerful paradigm for leveraging unlabeled data and has been proven to be successful across various tasks. Conventional SSL studies typically assume close environment scenarios where labeled and unlabeled examples are independently sampled from the same distribution. However, real-world tasks often involve open environment scenarios where the data distribution, label space, and feature space could differ between labeled and unlabeled data. This inconsistency introduces robustness challenges for SSL algorithms. In this paper, we first propose several robustness metrics for SSL based on the Robustness Analysis Curve (RAC), secondly, we establish a theoretical framework for studying the generalization performance and robustness of SSL algorithms in open environments, thirdly, we re-implement widely adopted SSL algorithms within a unified SSL toolkit and evaluate their performance on proposed open environment SSL benchmarks, including both image, text, and tabular datasets. By investigating the empirical and theoretical results, insightful discussions on enhancing the robustness of SSL algorithms in open environments are presented. The re-implementation and benchmark datasets are all publicly available. More details can be found at https://ygzwqzd.github.io/Robust-SSL-Benchmark.", "pdf": "https://openreview.net/pdf/3ab2c500841b15c2298e78836e9a91caa20ef54d.pdf"} {"title": "Efficient Inverse Multiagent Learning", "url": "https://openreview.net/forum?id=JzvIWvC9MG", "detail_url": "https://openreview.net/forum?id=JzvIWvC9MG", "authors": "Denizalp Goktas,Amy Greenwald,Sadie Zhao,Alec Koppel,Sumitra Ganesh", "tags": "ICLR 2024,Spotlight", "abstract": "In this paper, we study inverse game theory (resp. inverse multiagent learning) in\nwhich the goal is to find parameters of a game\u2019s payoff functions for which the\nexpected (resp. sampled) behavior is an equilibrium. We formulate these problems\nas generative-adversarial (i.e., min-max) optimization problems, which we develop\npolynomial-time algorithms to solve, the former of which relies on an exact first-\norder oracle, and the latter, a stochastic one. We extend our approach to solve\ninverse multiagent simulacral learning in polynomial time and number of samples.\nIn these problems, we seek a simulacrum, meaning parameters and an associated\nequilibrium that replicate the given observations in expectation. We find that our\napproach outperforms the widely-used ARIMA method in predicting prices in\nSpanish electricity markets based on time-series data.", "pdf": "https://openreview.net/pdf/a6e9bb008de3d4a2686aa0016448ee0ae913b390.pdf"} {"title": "On the Role of Discrete Tokenization in Visual Representation Learning", "url": "https://openreview.net/forum?id=WNLAkjUm19", "detail_url": "https://openreview.net/forum?id=WNLAkjUm19", "authors": "Tianqi Du,Yifei Wang,Yisen Wang", "tags": "ICLR 2024,Spotlight", "abstract": "In the realm of self-supervised learning (SSL), masked image modeling (MIM) has gained popularity alongside contrastive learning methods. MIM involves reconstructing masked regions of input images using their unmasked portions. A notable subset of MIM methodologies employs discrete tokens as the reconstruction target, but the theoretical underpinnings of this choice remain underexplored. In this paper, we explore the role of these discrete tokens, aiming to unravel their benefits and limitations. Building upon the connection between MIM and contrastive learning, we provide a comprehensive theoretical understanding on how discrete tokenization affects the model's generalization capabilities. Furthermore, we propose a novel metric named TCAS, which is specifically designed to assess the effectiveness of discrete tokens within the MIM framework. Inspired by this metric, we contribute an innovative tokenizer design and propose a corresponding MIM method named ClusterMIM. It demonstrates superior performance on a variety of benchmark datasets and ViT backbones. Code is available at \\url{https://github.com/PKU-ML/ClusterMIM}.", "pdf": "https://openreview.net/pdf/df6a2badfb1ca27a043b219c9e61e43688458fdf.pdf"} {"title": "The Consensus Game: Language Model Generation via Equilibrium Search", "url": "https://openreview.net/forum?id=n9xeGcI4Yg", "detail_url": "https://openreview.net/forum?id=n9xeGcI4Yg", "authors": "Athul Paul Jacob,Yikang Shen,Gabriele Farina,Jacob Andreas", "tags": "ICLR 2024,Spotlight", "abstract": "When applied to question answering and other text generation tasks, language models (LMs) may be queried generatively (by sampling answers from their output distribution) or discriminatively (by using them to score or rank a set of candidate answers). These procedures sometimes yield very different predictions. How do we reconcile mutually incompatible scoring procedures to obtain coherent LM predictions? We introduce a new, a training-free, game-theoretic procedure for language model decoding. Our approach casts language model decoding as a regularized imperfect-information sequential signaling game\u2014which we term the concensus game\u2014in which a generator seeks to communicate an abstract correctness parameter using natural language sentences to a discriminator. We develop computational procedures for finding approximate equilibria of this game, resulting in a decoding algorithm we call equilibrium-ranking. Applied to a large number of tasks (including reading comprehension, commonsense reasoning, mathematical problem-solving, and assistive dialog), equilibrium-ranking consistently improves performance over existing LM decoding procedures. These improvements are sometimes substantial\u2014on multiple benchmarks, we observe that applying equilibrium-ranking to LLaMA-7B outperforms the much larger LLaMA-65B and PaLM-540B models.", "pdf": "https://openreview.net/pdf/6a766aa0bf6a7a4f5d339309db677987d04377ce.pdf"} {"title": "AMAGO: Scalable In-Context Reinforcement Learning for Adaptive Agents", "url": "https://openreview.net/forum?id=M6XWoEdmwf", "detail_url": "https://openreview.net/forum?id=M6XWoEdmwf", "authors": "Jake Grigsby,Linxi Fan,Yuke Zhu", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce AMAGO, an in-context Reinforcement Learning (RL) agent that uses sequence models to tackle the challenges of generalization, long-term memory, and meta-learning. Recent works have shown that off-policy learning can make in-context RL with recurrent policies viable. Nonetheless, these approaches require extensive tuning and limit scalability by creating key bottlenecks in agents' memory capacity, planning horizon, and model size. AMAGO revisits and redesigns the off-policy in-context approach to successfully train long-sequence Transformers over entire rollouts in parallel with end-to-end RL. Our agent is scalable and applicable to a wide range of problems, and we demonstrate its strong performance empirically in meta-RL and long-term memory domains. AMAGO's focus on sparse rewards and off-policy data also allows in-context learning to extend to goal-conditioned problems with challenging exploration. When combined with a multi-goal hindsight relabeling scheme, AMAGO can solve a previously difficult category of open-world domains, where agents complete many possible instructions in procedurally generated environments.", "pdf": "https://openreview.net/pdf/6ffd1eb5dc0bd2b144d5d0309763b3ed5e114e8b.pdf"} {"title": "PILOT: An $\\mathcal{O}(1/K)$-Convergent Approach for Policy Evaluation with Nonlinear Function Approximation", "url": "https://openreview.net/forum?id=OkHHJcMroY", "detail_url": "https://openreview.net/forum?id=OkHHJcMroY", "authors": "Zhuqing Liu,Xin Zhang,Jia Liu,Zhengyuan Zhu,Songtao Lu", "tags": "ICLR 2024,Spotlight", "abstract": "Learning an accurate value function for a given policy is a critical step in solving reinforcement learning (RL) problems. So far, however, the convergence speed and sample complexity performances of most existing policy evaluation algorithms remain unsatisfactory, particularly with non-linear function approximation. This challenge motivates us to develop a new path-integrated primal-dual stochastic gradient (PILOT) method, that is able to achieve a fast convergence speed for RL policy evaluation with nonlinear function approximation. To further alleviate the periodic full gradient evaluation requirement, we further propose an enhanced method with an adaptive-batch adjustment called PILOT$^+$. The main advantages of our methods include: i) PILOT allows the use of {\\em{constant}} step sizes and achieves the $\\mathcal{O}(1/K)$ convergence rate to first-order stationary points of non-convex policy evaluation problems; ii) PILOT is a generic {\\em{single}}-timescale algorithm that is also applicable for solving a large class of non-convex strongly-concave minimax optimization problems; iii) By adaptively adjusting the batch size via historical stochastic gradient information, PILOT$^+$ is more sample-efficient empirically without loss of theoretical convergence rate. Our extensive numerical experiments verify our theoretical findings and showcase the high efficiency of the proposed PILOT and PILOT$^+$ algorithms compared with the state-of-the-art methods.", "pdf": "https://openreview.net/pdf/902e07eaba32bb9c32d1b7969eb59578e7e928ca.pdf"} {"title": "Confronting Reward Model Overoptimization with Constrained RLHF", "url": "https://openreview.net/forum?id=gkfUvn0fLU", "detail_url": "https://openreview.net/forum?id=gkfUvn0fLU", "authors": "Ted Moskovitz,Aaditya K Singh,DJ Strouse,Tuomas Sandholm,Ruslan Salakhutdinov,Anca Dragan,Stephen Marcus McAleer", "tags": "ICLR 2024,Spotlight", "abstract": "Large language models are typically aligned with human preferences by optimizing reward models (RMs) fitted to human feedback. However, human preferences are multi-faceted, and it is increasingly common to derive reward from a composition of simpler reward models which each capture a different aspect of language quality. This itself presents a challenge, as it is difficult to appropriately weight these component RMs when combining them. Compounding this difficulty, because any RM is only a proxy for human evaluation, this process is vulnerable to *overoptimization*, wherein past a certain point, accumulating higher reward is associated with worse human ratings. In this paper, we perform the first study on overoptimization in composite RMs, showing that correlation between component RMs has a significant effect on the locations of these points. We then introduce an approach to solve this issue using constrained reinforcement learning as a means of preventing the agent from exceeding each RM's threshold of usefulness. Our method addresses the problem of weighting component RMs by learning dynamic weights, naturally given by the Lagrange multipliers. As a result, each RM stays within the range at which it is an effective proxy, improving evaluation performance. Finally, we introduce an adaptive method using gradient-free optimization to identify and optimize towards these points during a single run.", "pdf": "https://openreview.net/pdf/9110e24405b3d1c469f8710d548cf6e5b7867692.pdf"} {"title": "LiDAR: Sensing Linear Probing Performance in Joint Embedding SSL Architectures", "url": "https://openreview.net/forum?id=f3g5XpL9Kb", "detail_url": "https://openreview.net/forum?id=f3g5XpL9Kb", "authors": "Vimal Thilak,Chen Huang,Omid Saremi,Laurent Dinh,Hanlin Goh,Preetum Nakkiran,Joshua M. Susskind,Etai Littwin", "tags": "ICLR 2024,Spotlight", "abstract": "Joint embedding (JE) architectures have emerged as a promising avenue for ac-\nquiring transferable data representations. A key obstacle to using JE methods,\nhowever, is the inherent challenge of evaluating learned representations without\naccess to a downstream task, and an annotated dataset. Without efficient and re-\nliable evaluation, it is difficult to iterate on architectural and training choices for\nJE methods. In this paper, we introduce LiDAR (Linear Discriminant Analysis\nRank), a metric designed to measure the quality of representations within JE archi-\ntectures. Our metric addresses several shortcomings of recent approaches based\non feature covariance rank by discriminating between informative and uninforma-\ntive features. In essence, LiDAR quantifies the rank of the Linear Discriminant\nAnalysis (LDA) matrix associated with the surrogate SSL task\u2014a measure that\nintuitively captures the information content as it pertains to solving the SSL task.\nWe empirically demonstrate that LiDAR significantly surpasses naive rank based\napproaches in its predictive power of optimal hyperparameters. Our proposed cri-\nterion presents a more robust and intuitive means of assessing the quality of rep-\nresentations within JE architectures, which we hope facilitates broader adoption\nof these powerful techniques in various domains.", "pdf": "https://openreview.net/pdf/fee7013fbdfbbccda18d5123b30300919a05a18f.pdf"} {"title": "Improved Efficiency Based on Learned Saccade and Continuous Scene Reconstruction From Foveated Visual Sampling", "url": "https://openreview.net/forum?id=lOwkOIUJtx", "detail_url": "https://openreview.net/forum?id=lOwkOIUJtx", "authors": "Jiayang Liu,Yiming Bu,Daniel Tso,Qinru Qiu", "tags": "ICLR 2024,Spotlight", "abstract": "High accuracy, low latency and high energy efficiency represent a set of contradictory goals when searching for system solutions for image classification and detection. While high-quality images naturally result in more precise detection and classification, they also result in a heavier computational workload for imaging and processing, reduce camera refresh rates, and increase the volume of data communication between the camera and processor. Taking inspiration from the foveal-peripheral sampling mechanism, saccade mechanism observed in the human visual system and the filling-in phenomena of brain, we have developed an active scene reconstruction architecture based on multiple foveal views. This model stitches together information from foveal and peripheral vision, which are sampled from multiple glances. Assisted by a reinforcement learning-based saccade mechanism, our model reduces the required input pixels by over 90\\% per frame while maintaining the same level of performance in image recognition as with the original images. We evaluated the effectiveness of our model using the GTSRB dataset and the ImageNet dataset. Using an equal number of input pixels, our study demonstrates a 5\\% higher image recognition accuracy compared to state-of-the-art foveal-peripheral vision systems. Furthermore, we demonstrate that our foveal sampling/saccadic scene reconstruction model exhibits significantly lower complexity and higher data efficiency during the training phase compared to existing approaches.", "pdf": "https://openreview.net/pdf/ad265fc85f17aaf422d2fd23b3440143ba832adc.pdf"} {"title": "Overthinking the Truth: Understanding how Language Models Process False Demonstrations", "url": "https://openreview.net/forum?id=Tigr1kMDZy", "detail_url": "https://openreview.net/forum?id=Tigr1kMDZy", "authors": "Danny Halawi,Jean-Stanislas Denain,Jacob Steinhardt", "tags": "ICLR 2024,Spotlight", "abstract": "Modern language models can imitate complex patterns through few-shot learning, enabling them to complete challenging tasks without fine-tuning. However, imitation can also lead models to reproduce inaccuracies or harmful content if present in the context. We study harmful imitation through the lens of a model\u2019s internal representations, and identify two related phenomena: overthinking and false induction heads. The first phenomenon, overthinking, appears when we decode predictions from intermediate layers, given correct vs. incorrect few-shot demonstrations. At early layers, both demonstrations induce similar model behavior, but the behavior diverges sharply at some \u201ccritical layer\u201d, after which the accuracy given incorrect demonstrations progressively decreases. The second phenomenon, false induction heads, are a possible mechanistic cause of overthinking: these are heads in late layers that attend to and copy false information from previous demonstrations, and whose ablation reduces overthinking. Beyond scientific understanding, our results suggest that studying intermediate model computations could be a promising avenue for understanding and guarding against harmful model behaviors.", "pdf": "https://openreview.net/pdf/55ea326a66e2cf61319ebe01bdea1b4ebbd8d775.pdf"} {"title": "MT-Ranker: Reference-free machine translation evaluation by inter-system ranking", "url": "https://openreview.net/forum?id=Rry1SeSOQL", "detail_url": "https://openreview.net/forum?id=Rry1SeSOQL", "authors": "Ibraheem Muhammad Moosa,Rui Zhang,Wenpeng Yin", "tags": "ICLR 2024,Spotlight", "abstract": "Traditionally, Machine Translation (MT) Evaluation has been treated as a regression problem -- producing an absolute translation-quality score. This approach has two limitations: i) the scores lack interpretability, and human annotators struggle with giving consistent scores; ii) most scoring methods are based on (reference, translation) pairs, limiting their applicability in real-world scenarios where references are absent. In practice, we often care about whether a new MT system is better or worse than some competitors. In addition, reference-free MT evaluation is increasingly practical and necessary. Unfortunately, these two practical considerations have yet to be jointly explored. In this work, we formulate the reference-free MT evaluation into a pairwise ranking problem. Given the source sentence and a pair of translations, our system predicts which translation is better. In addition to proposing this new formulation, we further show that this new paradigm can demonstrate superior correlation with human judgments by merely using indirect supervision from natural language inference and weak supervision from our synthetic data. In the context of reference-free evaluation, MT-Ranker, trained without any human annotations, achieves state-of-the-art results on the WMT Shared Metrics Task benchmarks DARR20, MQM20, and MQM21. On a more challenging benchmark, ACES, which contains fine-grained evaluation criteria such as addition, omission, and mistranslation errors, MT-Ranker marks state-of-the-art against reference-free as well as reference-based baselines.", "pdf": "https://openreview.net/pdf/fa181eb3a2cd5c3485b73e3829ad16f3dffa5faa.pdf"} {"title": "MuSR: Testing the Limits of Chain-of-thought with Multistep Soft Reasoning", "url": "https://openreview.net/forum?id=jenyYQzue1", "detail_url": "https://openreview.net/forum?id=jenyYQzue1", "authors": "Zayne Rea Sprague,Xi Ye,Kaj Bostrom,Swarat Chaudhuri,Greg Durrett", "tags": "ICLR 2024,Spotlight", "abstract": "While large language models (LLMs) equipped with techniques like chain-of-thought prompting have demonstrated impressive capabilities, they still fall short in their ability to reason robustly in complex settings. However, evaluating LLM reasoning is challenging because system capabilities continue to grow while benchmark datasets for tasks like logical deduction have remained static. We introduce MuSR, a dataset for evaluating language models on multistep soft reasoning tasks specified in a natural language narrative. This dataset has two crucial features. First, it is created through a novel neurosymbolic synthetic-to-natural generation algorithm, enabling the construction of complex reasoning instances that challenge GPT-4 (e.g., murder mysteries roughly 1000 words in length) and which can be scaled further as more capable LLMs are released. Second, our data instances are free text narratives corresponding to real-world domains of reasoning; this makes it simultaneously much more challenging than other synthetically-crafted benchmarks while remaining realistic and tractable for human annotators to solve with high accuracy. We evaluate a range of LLMs and prompting techniques on this dataset and characterize the gaps that remain for techniques like chain-of-thought to perform robust reasoning.", "pdf": "https://openreview.net/pdf/0fd545b50f3dd67f4d965d2b37b07aa5d08aba77.pdf"} {"title": "Harnessing Density Ratios for Online Reinforcement Learning", "url": "https://openreview.net/forum?id=THJEa8adBn", "detail_url": "https://openreview.net/forum?id=THJEa8adBn", "authors": "Philip Amortila,Dylan J Foster,Nan Jiang,Ayush Sekhari,Tengyang Xie", "tags": "ICLR 2024,Spotlight", "abstract": "The theories of offline and online reinforcement learning, despite having evolved in parallel, have begun to show signs of the possibility for a unification, with algorithms and analysis techniques for one setting often having natural counterparts in the other. However, the notion of *density ratio modeling*, an emerging paradigm in offline RL, has been largely absent from online RL, perhaps for good reason: the very existence and boundedness of density ratios relies on access to an exploratory dataset with good coverage, but the core challenge in online RL is to collect such a dataset without having one to start.\n\nIn this work we show---perhaps surprisingly---that density ratio-based algorithms have online counterparts. Assuming only the existence of an exploratory distribution with good coverage, a structural condition known as *coverability* (Xie et al., 2023), we give a new algorithm (GLOW) that uses density ratio realizability and value function realizability to perform sample-efficient online exploration. GLOW addresses unbounded density ratios via careful use of truncation, and combines this with optimism to guide exploration. GLOW is computationally inefficient; we complement it with a more efficient counterpart, HyGLOW, for the Hybrid RL setting (Song et al., 2023) wherein online RL is augmented with additional offline data. HyGLOW is derived as a special case of a more general meta-algorithm that provides a provable black-box reduction from hybrid RL to offline RL, which may be of independent interest.", "pdf": "https://openreview.net/pdf/2fffb35b07edd292dea91e77c2c874abdf3e831f.pdf"} {"title": "Predictive, scalable and interpretable knowledge tracing on structured domains", "url": "https://openreview.net/forum?id=NgaLU2fP5D", "detail_url": "https://openreview.net/forum?id=NgaLU2fP5D", "authors": "Hanqi Zhou,Robert Bamler,Charley M Wu,\u00c1lvaro Tejero-Cantero", "tags": "ICLR 2024,Spotlight", "abstract": "Intelligent tutoring systems optimize the selection and timing of learning materials to enhance understanding and long-term retention. This requires estimates of both the learner's progress (\"knowledge tracing\"; KT), and the prerequisite structure of the learning domain (\"knowledge mapping\"). While recent deep learning models achieve high KT accuracy, they do so at the expense of the interpretability of psychologically-inspired models. In this work, we present a solution to this trade-off. PSI-KT is a hierarchical generative approach that explicitly models how both individual cognitive traits and the prerequisite structure of knowledge influence learning dynamics, thus achieving interpretability by design. Moreover, by using scalable Bayesian inference, PSI-KT targets the real-world need for efficient personalization even with a growing body of learners and interaction data. Evaluated on three datasets from online learning platforms, PSI-KT achieves superior multi-step **p**redictive accuracy and **s**calable inference in continual-learning settings, all while providing **i**nterpretable representations of learner-specific traits and the prerequisite structure of knowledge that causally supports learning. In sum, predictive, scalable and interpretable knowledge tracing with solid knowledge mapping lays a key foundation for effective personalized learning to make education accessible to a broad, global audience.", "pdf": "https://openreview.net/pdf/52c2533f0586eecd513971405b1c38d7810ec28b.pdf"} {"title": "From Bricks to Bridges: Product of Invariances to Enhance Latent Space Communication", "url": "https://openreview.net/forum?id=vngVydDWft", "detail_url": "https://openreview.net/forum?id=vngVydDWft", "authors": "Irene Cannistraci,Luca Moschella,Marco Fumero,Valentino Maiorca,Emanuele Rodol\u00e0", "tags": "ICLR 2024,Spotlight", "abstract": "It has been observed that representations learned by distinct neural networks conceal structural similarities when the models are trained under similar inductive biases. From a geometric perspective, identifying the classes of transformations and the related invariances that connect these representations is fundamental to unlocking applications, such as merging, stitching, and reusing different neural modules. However, estimating task-specific transformations a priori can be challenging and expensive due to several factors (e.g., weights initialization, training hyperparameters, or data modality). To this end, we introduce a versatile method to directly incorporate a set of invariances into the representations, constructing a product space of invariant components on top of the latent representations without requiring prior knowledge about the optimal invariance to infuse. We validate our solution on classification and reconstruction tasks, observing consistent latent similarity and downstream performance improvements in a zero-shot stitching setting. The experimental analysis comprises three modalities (vision, text, and graphs), twelve pretrained foundational models, nine benchmarks, and several architectures trained from scratch.", "pdf": "https://openreview.net/pdf/4643421d3e88ae6516cd76daa15145a0b3f490d5.pdf"} {"title": "Proximal Policy Gradient Arborescence for Quality Diversity Reinforcement Learning", "url": "https://openreview.net/forum?id=TFKIfhvdmZ", "detail_url": "https://openreview.net/forum?id=TFKIfhvdmZ", "authors": "Sumeet Batra,Bryon Tjanaka,Matthew Christopher Fontaine,Aleksei Petrenko,Stefanos Nikolaidis,Gaurav S. Sukhatme", "tags": "ICLR 2024,Spotlight", "abstract": "Training generally capable agents that thoroughly explore their environment and\nlearn new and diverse skills is a long-term goal of robot learning. Quality Diversity\nReinforcement Learning (QD-RL) is an emerging research area that blends the\nbest aspects of both fields \u2013 Quality Diversity (QD) provides a principled form\nof exploration and produces collections of behaviorally diverse agents, while\nReinforcement Learning (RL) provides a powerful performance improvement\noperator enabling generalization across tasks and dynamic environments. Existing\nQD-RL approaches have been constrained to sample efficient, deterministic off-\npolicy RL algorithms and/or evolution strategies and struggle with highly stochastic\nenvironments. In this work, we, for the first time, adapt on-policy RL, specifically\nProximal Policy Optimization (PPO), to the Differentiable Quality Diversity (DQD)\nframework and propose several changes that enable efficient optimization and\ndiscovery of novel skills on high-dimensional, stochastic robotics tasks. Our new\nalgorithm, Proximal Policy Gradient Arborescence (PPGA), achieves state-of-\nthe-art results, including a 4x improvement in best reward over baselines on the\nchallenging humanoid domain.", "pdf": "https://openreview.net/pdf/f3442a5e81e1db8cb186b12c6fe204d3de0490dc.pdf"} {"title": "Memorization Capacity of Multi-Head Attention in Transformers", "url": "https://openreview.net/forum?id=MrR3rMxqqv", "detail_url": "https://openreview.net/forum?id=MrR3rMxqqv", "authors": "Sadegh Mahdavi,Renjie Liao,Christos Thrampoulidis", "tags": "ICLR 2024,Spotlight", "abstract": "Transformers have become the go-to architecture for language and vision tasks, yet their theoretical properties, especially memorization capacity, remain elusive. This paper investigates the memorization abilities of multi-head attention mechanisms, examining how many example sequences they can memorize, as a function of the number of heads and sequence length. Motivated by experimental findings on vision transformers, we introduce novel assumptions about the linear independence of input data, distinct from the commonly used general-position assumption. Under these assumptions, we demonstrate that an attention layer with $H$ heads, dimension $d$, and context size $n < d,$ featuring $\\Theta(Hd^2)$ parameters, can memorize $\\Omega(Hn)$ examples. Our analysis sheds light on how different attention heads handle various example sequences, aided by the softmax operator\u2019s saturation property. We validate our findings through experiments on synthetic data.", "pdf": "https://openreview.net/pdf/30b070a1d5057982a67ece4fdb61fca629dea9e6.pdf"} {"title": "Circuit Component Reuse Across Tasks in Transformer Language Models", "url": "https://openreview.net/forum?id=fpoAYV6Wsk", "detail_url": "https://openreview.net/forum?id=fpoAYV6Wsk", "authors": "Jack Merullo,Carsten Eickhoff,Ellie Pavlick", "tags": "ICLR 2024,Spotlight", "abstract": "Recent work in mechanistic interpretability has shown that behaviors in language models can be successfully reverse-engineered through circuit analysis. A common criticism, however, is that each circuit is task-specific, and thus such analysis cannot contribute to understanding the models at a higher level. In this work, we present evidence that insights (both low-level findings about specific heads and higher-level findings about general algorithms) can indeed generalize across tasks. Specifically, we study the circuit discovered in (Wang, 2022) for the Indirect Object Identification (IOI) task and 1.) show that it reproduces on a larger GPT2 model, and 2.) that it is mostly reused to solve a seemingly different task: Colored Objects (Ippolito & Callison-Burch, 2023). We provide evidence that the process underlying both tasks is functionally very similar, and contains about a 78% overlap in in-circuit attention heads. We further present a proof-of-concept intervention experiment, in which we adjust four attention heads in middle layers in order to \u2018repair\u2019 the Colored Objects circuit and make it behave like the IOI circuit. In doing so, we boost accuracy from 49.6% to 93.7% on the Colored Objects task and explain most sources of error. The intervention affects downstream attention heads in specific ways predicted by their interactions in the IOI circuit, indicating that this subcircuit behavior is invariant to the different task inputs. Overall, our results provide evidence that it may yet be possible to explain large language models' behavior in terms of a relatively small number of interpretable task-general algorithmic building blocks and computational components.", "pdf": "https://openreview.net/pdf/20152cb1b27ee48c1edca998e2aa13b4249cabaa.pdf"} {"title": "Likelihood Training of Cascaded Diffusion Models via Hierarchical Volume-preserving Maps", "url": "https://openreview.net/forum?id=sojpn00o8z", "detail_url": "https://openreview.net/forum?id=sojpn00o8z", "authors": "Henry Li,Ronen Basri,Yuval Kluger", "tags": "ICLR 2024,Spotlight", "abstract": "Cascaded models are multi-scale generative models with a marked capacity for producing perceptually impressive samples at high resolutions. In this work, we show that they can also be excellent likelihood models, so long as we overcome a fundamental difficulty with probabilistic multi-scale models: the intractability of the likelihood function. Chiefly, in cascaded models each intermediary scale introduces extraneous variables that cannot be tractably marginalized out for likelihood evaluation. This issue vanishes by modeling the diffusion process on latent spaces induced by a class of transformations we call hierarchical volume-preserving maps, which decompose spatially structured data in a hierarchical fashion without introducing local distortions in the latent space. We demonstrate that two such maps are well-known in the literature for multiscale modeling: Laplacian pyramids and wavelet transforms. Not only do such reparameterizations allow the likelihood function to be directly expressed as a joint likelihood over the scales, we show that the Laplacian pyramid and wavelet transform also produces significant improvements to the state-of-the-art on a selection of benchmarks in likelihood modeling, including density estimation, lossless compression, and out-of-distribution detection. Investigating the theoretical basis of our empirical gains we uncover deep connections to score matching under the Earth Mover's Distance (EMD), which is a well-known surrogate for perceptual similarity.", "pdf": "https://openreview.net/pdf/865dfe741f8cb5f9543b9889e222096ffcafed42.pdf"} {"title": "Empirical Analysis of Model Selection for Heterogeneous Causal Effect Estimation", "url": "https://openreview.net/forum?id=yuy6cGt3KL", "detail_url": "https://openreview.net/forum?id=yuy6cGt3KL", "authors": "Divyat Mahajan,Ioannis Mitliagkas,Brady Neal,Vasilis Syrgkanis", "tags": "ICLR 2024,Spotlight", "abstract": "We study the problem of model selection in causal inference, specifically for conditional average treatment effect (CATE) estimation. Unlike machine learning, there is no perfect analogue of cross-validation for model selection as we do not observe the counterfactual potential outcomes. Towards this, a variety of surrogate metrics have been proposed for CATE model selection that use only observed data. However, we do not have a good understanding regarding their effectiveness due to limited comparisons in prior studies. We conduct an extensive empirical analysis to benchmark the surrogate model selection metrics introduced in the literature, as well as the novel ones introduced in this work. We ensure a fair comparison by tuning the hyperparameters associated with these metrics via AutoML, and provide more detailed trends by incorporating realistic datasets via generative modeling. Our analysis suggests novel model selection strategies based on careful hyperparameter selection of CATE estimators and causal ensembling.", "pdf": "https://openreview.net/pdf/68047fc672c9a6819fb21499f7e2b6e8191790b9.pdf"} {"title": "Confidential-DPproof: Confidential Proof of Differentially Private Training", "url": "https://openreview.net/forum?id=PQY2v6VtGe", "detail_url": "https://openreview.net/forum?id=PQY2v6VtGe", "authors": "Ali Shahin Shamsabadi,Gefei Tan,Tudor Ioan Cebere,Aur\u00e9lien Bellet,Hamed Haddadi,Nicolas Papernot,Xiao Wang,Adrian Weller", "tags": "ICLR 2024,Spotlight", "abstract": "Post hoc privacy auditing techniques can be used to test the privacy guarantees of a model, but come with several limitations: (i) they can only establish lower bounds on the privacy loss, (ii) the intermediate model updates and some data must be shared with the auditor to get a better approximation of the privacy loss, and (iii) the auditor typically faces a steep computational cost to run a large number of attacks. In this paper, we propose to proactively generate a cryptographic certificate of privacy during training to forego such auditing limitations. We introduce Confidential-DPproof , a framework for Confidential Proof of Differentially Private Training, which enhances training with a certificate of the $(\\varepsilon,\\delta)$-DP guarantee achieved. To obtain this certificate without revealing information about the training data or model, we design a customized zero-knowledge proof protocol tailored to the requirements introduced by differentially private training, including random noise addition and privacy amplification by subsampling. In experiments on CIFAR-10, Confidential-DPproof trains a model achieving state-of-the-art $91$% test accuracy with a certified privacy guarantee of $(\\varepsilon=0.55,\\delta=10^{-5})$-DP in approximately 100 hours.", "pdf": "https://openreview.net/pdf/94732346a3d36701b0a68d02f2366498641c54ee.pdf"} {"title": "In-Context Pretraining: Language Modeling Beyond Document Boundaries", "url": "https://openreview.net/forum?id=LXVswInHOo", "detail_url": "https://openreview.net/forum?id=LXVswInHOo", "authors": "Weijia Shi,Sewon Min,Maria Lomeli,Chunting Zhou,Margaret Li,Xi Victoria Lin,Noah A. Smith,Luke Zettlemoyer,Wen-tau Yih,Mike Lewis", "tags": "ICLR 2024,Spotlight", "abstract": "Language models are currently trained to predict tokens given document prefixes, enabling them to zero shot long form generation and prompting-style tasks which can be reduced to document completion. We instead present IN-CONTEXT PRETRAINING, a new approach where language models are trained on a sequence of related documents, thereby explicitly encouraging them to read and reason across document boundaries. Our approach builds on the fact that current pipelines train by concatenating random sets of shorter documents to create longer context windows; this improves efficiency even though the prior documents provide no signal for predicting the next document. Given this fact, we can do IN-CONTEXT PRETRAINING by simply changing the document ordering so that each context contains related documents, and directly applying existing pretraining pipelines. However, this document sorting problem is challenging. There are billions of documents and we would like the sort to maximize contextual similarity for every document without repeating any data. To do this, we introduce approximate algorithms for finding related documents with efficient nearest neighbor search and constructing coherent batches with a graph cover algorithm. Our experiments show IN-CONTEXT PRETRAINING offers a scalable and simple approach to significantly enhance LM performance: we see notable improvements in tasks that require more complex contextual reasoning, including in-context learning (+8%), reading comprehension (+15%), faithfulness to previous contexts (+16%), long-context reasoning (+5%), and retrieval augmentation (+9%).", "pdf": "https://openreview.net/pdf/a7de9c0b47acd0a990190c7e40945c1f335d4201.pdf"} {"title": "What's In My Big Data?", "url": "https://openreview.net/forum?id=RvfPnOkPV4", "detail_url": "https://openreview.net/forum?id=RvfPnOkPV4", "authors": "Yanai Elazar,Akshita Bhagia,Ian Helgi Magnusson,Abhilasha Ravichander,Dustin Schwenk,Alane Suhr,Evan Pete Walsh,Dirk Groeneveld,Luca Soldaini,Sameer Singh,Hannaneh Hajishirzi,Noah A. Smith,Jesse Dodge", "tags": "ICLR 2024,Spotlight", "abstract": "Large text corpora are the backbone of language models.\nHowever, we have a limited understanding of the content of these corpora, including general statistics, quality, social factors, and inclusion of evaluation data (contamination).\nIn this work, we propose What's In My Big Data? (WIMBD), a platform and a set of sixteen analyses that allow us to reveal and compare the contents of large text corpora. WIMBD builds on two basic capabilities---count and search---*at scale*, which allows us to analyze more than 35 terabytes on a standard compute node. \nWe apply WIMBD to ten different corpora used to train popular language models, including *C4*, *The Pile*, and *RedPajama*.\nOur analysis uncovers several surprising and previously undocumented findings about these corpora, including the high prevalence of duplicate, synthetic, and low-quality content, personally identifiable information, toxic language, and benchmark contamination. \nFor instance, we find that about 50% of the documents in *RedPajama* and *LAION-2B-en* are duplicates. In addition, several datasets used for benchmarking models trained on such corpora are contaminated with respect to important benchmarks, including the Winograd Schema Challenge and parts of GLUE and SuperGLUE.\nWe open-source WIMBD's code and artifacts to provide a standard set of evaluations for new text-based corpora and to encourage more analyses and transparency around them.", "pdf": "https://openreview.net/pdf/8e645356fd6998459b2c368f65c8a2b3a44206af.pdf"} {"title": "On Diffusion Modeling for Anomaly Detection", "url": "https://openreview.net/forum?id=lR3rk7ysXz", "detail_url": "https://openreview.net/forum?id=lR3rk7ysXz", "authors": "Victor Livernoche,Vineet Jain,Yashar Hezaveh,Siamak Ravanbakhsh", "tags": "ICLR 2024,Spotlight", "abstract": "Known for their impressive performance in generative modeling, diffusion models are attractive candidates for density-based anomaly detection. This paper investigates different variations of diffusion modeling for unsupervised and semi-supervised anomaly detection. In particular, we find that Denoising Diffusion Probability Models (DDPM) are performant on anomaly detection benchmarks yet computationally expensive. By simplifying DDPM in application to anomaly detection, we are naturally led to an alternative approach called Diffusion Time Estimation (DTE). DTE estimates the distribution over diffusion time for a given input and uses the mode or mean of this distribution as the anomaly score. We derive an analytical form for this density and leverage a deep neural network to improve inference efficiency. Through empirical evaluations on the ADBench benchmark, we demonstrate that all diffusion-based anomaly detection methods perform competitively for both semi-supervised and unsupervised settings. Notably, DTE achieves orders of magnitude faster inference time than DDPM, while outperforming it on this benchmark. These results establish diffusion-based anomaly detection as a scalable alternative to traditional methods and recent deep-learning techniques for standard unsupervised and semi-supervised anomaly detection settings.", "pdf": "https://openreview.net/pdf/c6480e4c58a2924ca498ff399ce467bb1e61ed7b.pdf"} {"title": "Social Reward: Evaluating and Enhancing Generative AI through Million-User Feedback from an Online Creative Community", "url": "https://openreview.net/forum?id=tjn2YZSHUv", "detail_url": "https://openreview.net/forum?id=tjn2YZSHUv", "authors": "Arman Isajanyan,Artur Shatveryan,David Kocharian,Zhangyang Wang,Humphrey Shi", "tags": "ICLR 2024,Spotlight", "abstract": "Social reward as a form of community recognition provides a strong source of\nmotivation for users of online platforms to actively engage and contribute with\ncontent to accumulate peers approval. In the realm of text-conditioned image\nsynthesis, the recent surge in progress has ushered in a collaborative era where\nusers and AI systems coalesce to refine visual creations. This co-creative pro-\ncess in the landscape of online social networks empowers users to craft original\nvisual artworks seeking for community validation. Nevertheless, assessing these\nmodels in the context of collective community preference introduces distinct chal-\nlenges. Existing evaluation methods predominantly center on limited size user\nstudies guided by image quality and alignment with prompts. This work pio-\nneers a paradigm shift, unveiling Social Reward - an innovative reward modeling\nframework that leverages implicit feedback from social network users engaged\nin creative editing of generated images. We embark on an extensive journey of\ndataset curation and refinement, drawing from Picsart: an online visual creation\nand editing platform, yielding a first million-user-scale dataset of implicit human\npreferences for user-generated visual art named Picsart Image-Social. Our anal-\nysis exposes the shortcomings of current metrics in modeling community creative\npreference of text-to-image models\u2019 outputs, compelling us to introduce a novel\npredictive model explicitly tailored to address these limitations. Rigorous quan-\ntitative experiments and user study show that our Social Reward model aligns\nbetter with social popularity than existing metrics. Furthermore, we utilize So-\ncial Reward to fine-tune text-to-image models, yielding images that are more fa-\nvored by not only Social Reward, but also other established metrics. These find-\nings highlight the relevance and effectiveness of Social Reward in assessing com-\nmunity appreciation for AI-generated artworks, establishing a closer alignment\nwith users\u2019 creative goals: creating popular visual art. Codes can be accessed at\nhttps://github.com/Picsart-AI-Research/Social-Reward", "pdf": "https://openreview.net/pdf/2deafb9f8640664b5840f46d75dd6e361f54bd88.pdf"} {"title": "Role of Locality and Weight Sharing in Image-Based Tasks: A Sample Complexity Separation between CNNs, LCNs, and FCNs", "url": "https://openreview.net/forum?id=AfnsTnYphT", "detail_url": "https://openreview.net/forum?id=AfnsTnYphT", "authors": "Aakash Lahoti,Stefani Karp,Ezra Winston,Aarti Singh,Yuanzhi Li", "tags": "ICLR 2024,Spotlight", "abstract": "Vision tasks are characterized by the properties of locality and translation invariance. \n The superior performance of convolutional neural networks (CNNs) on these tasks is widely attributed to the inductive bias of locality and weight sharing baked into their architecture.\n Existing attempts to quantify the statistical benefits of these biases in CNNs over locally connected convolutional neural networks (LCNs) and fully connected neural networks (FCNs) fall into one of the following categories: either they disregard the optimizer and only provide uniform convergence upper bounds with no separating lower bounds, \n or they consider simplistic tasks that do not truly mirror the locality and translation invariance as found in real-world vision tasks.\n To address these deficiencies, we introduce the Dynamic Signal Distribution (DSD) classification task that models an image as consisting of $k$ patches, each of dimension $d$, and the label is determined by a $d$-sparse signal vector that can freely appear in any one of the $k$ patches. \n On this task, for any orthogonally equivariant algorithm like gradient descent, we prove that CNNs require $\\tilde{O}(k+d)$ samples, whereas LCNs require $\\Omega(kd)$ samples, establishing the statistical advantages of weight sharing in translation invariant tasks. \n Furthermore, LCNs need $\\tilde{O}(k(k+d))$ samples, compared to $\\Omega(k^2d)$ samples for FCNs, showcasing the benefits of locality in local tasks.\n Additionally, we develop information theoretic tools for analyzing randomized algorithms, which may be of interest for statistical research.", "pdf": "https://openreview.net/pdf/8956f1209e23a0bd6029fbbe9b6d595916060b56.pdf"} {"title": "Lion Secretly Solves a Constrained Optimization: As Lyapunov Predicts", "url": "https://openreview.net/forum?id=e4xS9ZarDr", "detail_url": "https://openreview.net/forum?id=e4xS9ZarDr", "authors": "Lizhang Chen,Bo Liu,Kaizhao Liang,qiang liu", "tags": "ICLR 2024,Spotlight", "abstract": "Lion (Evolved Sign Momentum), a new optimizer discovered through program search, has shown promising results in training large AI models. It achieves results comparable to AdamW but with greater memory efficiency. As what we can expect from the result of the random search, Lion blends a number of elements from existing algorithms, including signed momentum, decoupled weight decay, Polayk and Nesterov momentum, but doesn't fit into any existing category of theoretically grounded optimizers. Thus, even though Lion appears to perform well as a general-purpose optimizer for a wide range of tasks, its theoretical basis remains uncertain. This absence of theoretical clarity limits opportunities to further enhance and expand Lion's efficacy. This work aims to demystify Lion. Using both continuous-time and discrete-time analysis, we demonstrate that Lion is a novel and theoretically grounded approach for minimizing a general loss function $f(x)$ while enforcing a bound constraint $||x||_\\infty \\leq 1/\\lambda$. Lion achieves this through the incorporation of decoupled weight decay, where $\\lambda$ represents the weight decay coefficient. Our analysis is facilitated by the development of a new Lyapunov function for the Lion updates. It applies to a wide range of Lion-$\\phi$ algorithms, where the $sign(\\cdot)$ operator in Lion is replaced by the subgradient of a convex function $\\phi$, leading to the solution of the general composite optimization problem $\\min_x f(x) + \\phi^*(x)$. Our findings provide valuable insights into the dynamics of Lion and pave the way for further enhancements and extensions of Lion-related algorithms.", "pdf": "https://openreview.net/pdf/415e0c57e06c454b4abc460e3311a5bc3e5c9b6b.pdf"} {"title": "Distributionally Robust Optimization with Bias and Variance Reduction", "url": "https://openreview.net/forum?id=TTrzgEZt9s", "detail_url": "https://openreview.net/forum?id=TTrzgEZt9s", "authors": "Ronak Mehta,Vincent Roulet,Krishna Pillutla,Zaid Harchaoui", "tags": "ICLR 2024,Spotlight", "abstract": "We consider the distributionally robust optimization (DRO) problem, wherein a learner optimizes the worst-case empirical risk achievable by reweighing the observed training examples. We present Prospect, a stochastic gradient-based algorithm that only requires tuning a single learning rate hyperparameter, and prove that it enjoys linear convergence for smooth regularized losses. This contrasts with previous algorithms that either require tuning multiple hyperparameters or potentially fail to converge due to biased gradient estimates or inadequate regularization. Empirically, we show that Prospect can converge 2-3x faster than baselines such as SGD and stochastic saddle-point methods on distribution shift and fairness benchmarks spanning tabular, vision, and language domains.", "pdf": "https://openreview.net/pdf/6c3d461c90f544421c04e52861860e354c20c157.pdf"} {"title": "A Benchmark for Learning to Translate a New Language from One Grammar Book", "url": "https://openreview.net/forum?id=tbVWug9f2h", "detail_url": "https://openreview.net/forum?id=tbVWug9f2h", "authors": "Garrett Tanzer,Mirac Suzgun,Eline Visser,Dan Jurafsky,Luke Melas-Kyriazi", "tags": "ICLR 2024,Spotlight", "abstract": "Large language models (LLMs) can perform impressive feats with in-context learning or lightweight finetuning. It is natural to wonder how well these models adapt to genuinely new tasks, but how does one find tasks that are unseen in internet-scale training sets? We turn to a field that is explicitly motivated and bottlenecked by a scarcity of web data: low-resource languages. In this paper, we introduce MTOB (Machine Translation from One Book), a benchmark for learning to translate between English and Kalamang\u2014a language with less than 200 speakers and therefore virtually no presence on the web\u2014using several hundred pages of field linguistics reference materials. This task framing is novel in that it asks a model to learn a language from a single human-readable book of grammar explanations, rather than a large mined corpus of in-domain data, more akin to L2 language learning than L1 language acquisition. We demonstrate that baselines using current LLMs are promising but fall short of human performance, achieving 44.7 chrF on Kalamang to English translation and 45.8 chrF on English to Kalamang translation, compared to 51.6 and 57.0 chrF by a human who learned Kalamang from the same reference materials. We hope that MTOB will help measure LLM capabilities along a new dimension, and that the methods developed to solve it could help expand access to language technology for underserved communities by leveraging qualitatively different kinds of data than traditional machine translation.", "pdf": "https://openreview.net/pdf/410fe170bf3313698302de77b1641b5a9ea9eaa3.pdf"} {"title": "Improving Offline RL by Blending Heuristics", "url": "https://openreview.net/forum?id=MCl0TLboP1", "detail_url": "https://openreview.net/forum?id=MCl0TLboP1", "authors": "Sinong Geng,Aldo Pacchiano,Andrey Kolobov,Ching-An Cheng", "tags": "ICLR 2024,Spotlight", "abstract": "We propose **H**e**u**ristic **Bl**ending (HUBL), a simple performance-improving technique for a broad class of offline RL algorithms based on value bootstrapping. HUBL modifies the Bellman operators used in these algorithms, partially replacing the bootstrapped values with heuristic ones that are estimated with Monte-Carlo returns. For trajectories with higher returns, HUBL relies more on the heuristic values and less on bootstrapping; otherwise, it leans more heavily on bootstrapping. HUBL is very easy to combine with many existing offline RL implementations by relabeling the offline datasets with adjusted rewards and discount factors. We derive a theory that explains HUBL's effect on offline RL as reducing offline RL's complexity and thus increasing its finite-sample performance. Furthermore, we empirically demonstrate that HUBL consistently improves the policy quality of four state-of-the-art bootstrapping-based offline RL algorithms (ATAC, CQL, TD3+BC, and IQL), by 9% on average over 27 datasets of the D4RL and Meta-World benchmarks.", "pdf": "https://openreview.net/pdf/16cac64e07aa3a54bc305bc6025b1b1e3326c1f7.pdf"} {"title": "Decision ConvFormer: Local Filtering in MetaFormer is Sufficient for Decision Making", "url": "https://openreview.net/forum?id=af2c8EaKl8", "detail_url": "https://openreview.net/forum?id=af2c8EaKl8", "authors": "Jeonghye Kim,Suyoung Lee,Woojun Kim,Youngchul Sung", "tags": "ICLR 2024,Spotlight", "abstract": "The recent success of Transformer in natural language processing has sparked its use in various domains. In offline reinforcement learning (RL), Decision Transformer (DT) is emerging as a promising model based on Transformer. However, we discovered that the attention module of DT is not appropriate to capture the inherent local dependence pattern in trajectories of RL modeled as a Markov decision process. To overcome the limitations of DT, we propose a novel action sequence predictor, named Decision ConvFormer (DC), based on the architecture of MetaFormer, which is a general structure to process multiple entities in parallel and understand the interrelationship among the multiple entities. DC employs local convolution filtering as the token mixer and can effectively capture the inherent local associations of the RL dataset. In extensive experiments, DC achieved state-of-the-art performance across various standard RL benchmarks while requiring fewer resources. Furthermore, we show that DC better understands the underlying meaning in data and exhibits enhanced generalization capability.", "pdf": "https://openreview.net/pdf/d34443d3c25936596abdb44faed04a15cdf3e290.pdf"} {"title": "How Many Pretraining Tasks Are Needed for In-Context Learning of Linear Regression?", "url": "https://openreview.net/forum?id=vSh5ePa0ph", "detail_url": "https://openreview.net/forum?id=vSh5ePa0ph", "authors": "Jingfeng Wu,Difan Zou,Zixiang Chen,Vladimir Braverman,Quanquan Gu,Peter Bartlett", "tags": "ICLR 2024,Spotlight", "abstract": "Transformers pretrained on diverse tasks exhibit remarkable in-context learning (ICL) capabilities, enabling them to solve unseen tasks solely based on input contexts without adjusting model parameters. In this paper, we study ICL in one of its simplest setups: pretraining a single-layer linear attention model for linear regression with a Gaussian prior. We establish a statistical task complexity bound for the attention model pretraining, showing that effective pretraining only requires a small number of independent tasks. Furthermore, we prove that the pretrained model closely matches the Bayes optimal algorithm, i.e., optimally tuned ridge regression, by achieving nearly Bayes optimal risk on unseen tasks under a fixed context length. These theoretical findings complement prior experimental research and shed light on the statistical foundations of ICL.", "pdf": "https://openreview.net/pdf/2afad356cf9372d0067c51eea7b8c4169effbe2c.pdf"} {"title": "Tool-Augmented Reward Modeling", "url": "https://openreview.net/forum?id=d94x0gWTUX", "detail_url": "https://openreview.net/forum?id=d94x0gWTUX", "authors": "Lei Li,Yekun Chai,Shuohuan Wang,Yu Sun,Hao Tian,Ningyu Zhang,Hua Wu", "tags": "ICLR 2024,Spotlight", "abstract": "Reward modeling (*a.k.a.*, preference modeling) is instrumental for aligning large language models with human preferences, particularly within the context of reinforcement learning from human feedback (RLHF). While conventional reward models (RMs) have exhibited remarkable scalability, they oft struggle with fundamental functionality such as arithmetic computation, code execution, and factual lookup. In this paper, we propose a tool-augmented preference modeling approach, named Themis, to address these limitations by empowering RMs with access to external environments, including calculators and search engines. This approach not only fosters synergy between tool utilization and reward grading but also enhances interpretive capacity and scoring reliability. Our study delves into the integration of external tools into RMs, enabling them to interact with diverse external sources and construct task-specific tool engagement and reasoning traces in an autoregressive manner. We validate our approach across a wide range of domains, incorporating seven distinct external tools. Our experimental results demonstrate a noteworthy overall improvement of 17.7% across eight tasks in preference ranking. Furthermore, our approach outperforms Gopher 280B by 7.3% on TruthfulQA task in zero-shot evaluation. In human evaluations, RLHF trained with Themis attains an average win rate of 32% when compared to baselines across four distinct tasks. Additionally, we provide a comprehensive collection of tool-related RM datasets, incorporating data from seven distinct tool APIs, totaling 15,000 instances. We have made the code, data, and model checkpoints publicly available to facilitate and inspire further research advancements (https://github.com/ernie-research/Tool-Augmented-Reward-Model).", "pdf": "https://openreview.net/pdf/65b055ecf7ec43b68562fc8ca3ce916f8c400085.pdf"} {"title": "Reward-Consistent Dynamics Models are Strongly Generalizable for Offline Reinforcement Learning", "url": "https://openreview.net/forum?id=GSBHKiw19c", "detail_url": "https://openreview.net/forum?id=GSBHKiw19c", "authors": "Fan-Ming Luo,Tian Xu,Xingchen Cao,Yang Yu", "tags": "ICLR 2024,Spotlight", "abstract": "Learning a precise dynamics model can be crucial for offline reinforcement learning, which, unfortunately, has been found to be quite challenging. Dynamics models that are learned by fitting historical transitions often struggle to generalize to unseen transitions. In this study, we identify a hidden but pivotal factor termed dynamics reward that remains consistent across transitions, offering a pathway to better generalization. Therefore, we propose the idea of reward-consistent dynamics models: any trajectory generated by the dynamics model should maximize the dynamics reward derived from the data. We implement this idea as the MOREC (Model-based Offline reinforcement learning with Reward Consistency) method, which can be seamlessly integrated into previous offline model-based reinforcement learning (MBRL) methods. MOREC learns a generalizable dynamics reward function from offline data, which is subsequently employed as a transition filter in any offline MBRL method: when generating transitions, the dynamics model generates a batch of transitions and selects the one with the highest dynamics reward value. On a synthetic task, we visualize that MOREC has a strong generalization ability and can surprisingly recover some distant unseen transitions. On 21 offline tasks in D4RL and NeoRL benchmarks, MOREC improves the previous state-of-the-art performance by a significant margin, i.e., 4.6\\% on D4RL tasks and 25.9\\% on NeoRL tasks. Notably, MOREC is the first method that can achieve above 95\\% online RL performance in 6 out of 12 D4RL tasks and 3 out of 9 NeoRL tasks. Code is available at https://github.com/polixir/morec.", "pdf": "https://openreview.net/pdf/b755072e1d772c902d9b57e9ad4a0ff78b0df063.pdf"} {"title": "Towards Optimal Regret in Adversarial Linear MDPs with Bandit Feedback", "url": "https://openreview.net/forum?id=6yv8UHVJn4", "detail_url": "https://openreview.net/forum?id=6yv8UHVJn4", "authors": "Haolin Liu,Chen-Yu Wei,Julian Zimmert", "tags": "ICLR 2024,Spotlight", "abstract": "We study online reinforcement learning in linear Markov decision processes with adversarial losses and bandit feedback. We introduce two algorithms that achieve improved regret performance compared to existing approaches. The first algorithm, although computationally inefficient, achieves a regret of $\\widetilde{O}(\\sqrt{K})$ without relying on simulators, where $K$ is the number of episodes. This is the first rate-optimal result in the considered setting. The second algorithm is computationally efficient and achieves a regret of $\\widetilde{O}(K^{\\frac{3}{4}})$ . These results significantly improve over the prior state-of-the-art: a computationally inefficient algorithm by Kong et al. (2023) with $\\widetilde{O}(K^{\\frac{4}{5}}+1/\\lambda_{\\min})$ regret, and a computationally efficient algorithm by Sherman et al. (2023b) with $\\widetilde{O}(K^{\\frac{6}{7}})$ regret.", "pdf": "https://openreview.net/pdf/364346cf68aee0e5e5761a8aad4c6a42391b9e05.pdf"} {"title": "Dual RL: Unification and New Methods for Reinforcement and Imitation Learning", "url": "https://openreview.net/forum?id=xt9Bu66rqv", "detail_url": "https://openreview.net/forum?id=xt9Bu66rqv", "authors": "Harshit Sikchi,Qinqing Zheng,Amy Zhang,Scott Niekum", "tags": "ICLR 2024,Spotlight", "abstract": "The goal of reinforcement learning (RL) is to find a policy that maximizes the expected cumulative return. It has been shown that this objective can be represented as an optimization problem of state-action visitation distribution under linear constraints. The dual problem of this formulation, which we refer to as *dual RL*, is unconstrained and easier to optimize. In this work, we first cast several state-of-the-art offline RL and offline imitation learning (IL) algorithms as instances of dual RL approaches with shared structures. Such unification allows us to identify the root cause of the shortcomings of prior methods. For offline IL, our analysis shows that prior methods are based on a restrictive coverage assumption that greatly limits their performance in practice. To fix this limitation, we propose a new discriminator-free method ReCOIL that learns to imitate from arbitrary off-policy data to obtain near-expert performance. For offline RL, our analysis frames a recent offline RL method XQL in the dual framework, and we further propose a new method $f$-DVL that provides alternative choices to the Gumbel regression loss that fixes the known training instability issue of XQL. The performance improvements by both of our proposed methods, ReCOIL and $f$-DVL, in IL and RL are validated on an extensive suite of simulated robot locomotion and manipulation tasks.", "pdf": "https://openreview.net/pdf/981af9a85be76008d6063c6ddd1477450c3bf463.pdf"} {"title": "Out-Of-Domain Unlabeled Data Improves Generalization", "url": "https://openreview.net/forum?id=Bo6GpQ3B9a", "detail_url": "https://openreview.net/forum?id=Bo6GpQ3B9a", "authors": "seyed amir hossein saberi,Amir Najafi,Alireza Heidari,Mohammad Hosein Movasaghinia,Abolfazl Motahari,Babak Khalaj", "tags": "ICLR 2024,Spotlight", "abstract": "We propose a novel framework for incorporating unlabeled data into semi-supervised classification problems, where scenarios involving the minimization of either i) adversarially robust or ii) non-robust loss functions have been considered. Notably, we allow the unlabeled samples to deviate slightly (in total variation sense) from the in-domain distribution. The core idea behind our framework is to combine Distributionally Robust Optimization (DRO) with self-supervised training. As a result, we also leverage efficient polynomial-time algorithms for the training stage. From a theoretical standpoint, we apply our framework on the classification problem of a mixture of two Gaussians in $\\mathbb{R}^d$, where in addition to the $m$ independent and labeled samples from the true distribution, a set of $n$ (usually with $n\\gg m$) out of domain and unlabeled samples are gievn as well. Using only the labeled data, it is known that the generalization error can be bounded by $\\propto\\left(d/m\\right)^{1/2}$. However, using our method on both isotropic and non-isotropic Gaussian mixture models, one can derive a new set of analytically explicit and non-asymptotic bounds which show substantial improvement on the generalization error compared ERM. Our results underscore two significant insights: 1) out-of-domain samples, even when unlabeled, can be harnessed to narrow the generalization gap, provided that the true data distribution adheres to a form of the \"cluster assumption\", and 2) the semi-supervised learning paradigm can be regarded as a special case of our framework when there are no distributional shifts. We validate our claims through experiments conducted on a variety of synthetic and real-world datasets.", "pdf": "https://openreview.net/pdf/db772b639656e2c5123c187ceacfb10805558c17.pdf"} {"title": "Catastrophic Jailbreak of Open-source LLMs via Exploiting Generation", "url": "https://openreview.net/forum?id=r42tSSCHPh", "detail_url": "https://openreview.net/forum?id=r42tSSCHPh", "authors": "Yangsibo Huang,Samyak Gupta,Mengzhou Xia,Kai Li,Danqi Chen", "tags": "ICLR 2024,Spotlight", "abstract": "The rapid progress in open-source large language models (LLMs) is significantly advancing AI development. Extensive efforts have been made before model release to align their behavior with human values, with the primary goal of ensuring their helpfulness and harmlessness. However, even carefully aligned models can be manipulated maliciously, leading to unintended behaviors, known as ``jailbreaks\". These jailbreaks are typically triggered by specific text inputs, often referred to as adversarial prompts. In this work, we propose the generation exploitation attack, an extremely simple approach that disrupts model alignment by only manipulating variations of decoding methods. By exploiting different generation strategies, including varying decoding hyper-parameters and sampling methods, we increase the attack success rate from $0\\%$ to more than $95\\%$ across 11 language models including LLaMA2, Vicuna, Falcon, and MPT families, outperforming state-of-the-art attacks with $30\\times$ lower computational cost. Finally, we propose an effective alignment method that explores diverse generation strategies, which can reasonably reduce the attack success rate under our attack. Altogether, our study underscores a major failure in current safety evaluation and alignment procedures for open-source LLMs, strongly advocating for more comprehensive red teaming and better alignment before releasing such models.", "pdf": "https://openreview.net/pdf/721e3e663ee57023772f0b3b63424ccc43e71e44.pdf"} {"title": "PolyGCL: GRAPH CONTRASTIVE LEARNING via Learnable Spectral Polynomial Filters", "url": "https://openreview.net/forum?id=y21ZO6M86t", "detail_url": "https://openreview.net/forum?id=y21ZO6M86t", "authors": "Jingyu Chen,Runlin Lei,Zhewei Wei", "tags": "ICLR 2024,Spotlight", "abstract": "Recently, Graph Contrastive Learning (GCL) has achieved significantly superior performance in self-supervised graph representation learning. \nHowever, the existing GCL technique has inherent smooth characteristics because of its low-pass GNN encoder and objective based on homophily assumption, which poses a challenge when applying it to heterophilic graphs.\nIn supervised learning tasks, spectral GNNs with polynomial approximation excel in both homophilic and heterophilic settings by adaptively fitting graph filters of arbitrary shapes. \nYet, their applications in unsupervised learning are rarely explored.\nBased on the above analysis, a natural question arises: Can we incorporate the excellent properties of spectral polynomial filters into graph contrastive learning?\nIn this paper, we address the question by studying the necessity of introducing high-pass information for heterophily from a spectral perspective.\nWe propose PolyGCL, a GCL pipeline that utilizes polynomial filters to achieve contrastive learning between the low-pass and high-pass views.\nSpecifically, PolyGCL utilizes polynomials with learnable filter functions to generate different spectral views and an objective that incorporates high-pass information through a linear combination. \nWe theoretically prove that PolyGCL outperforms previous GCL paradigms when applied to graphs with varying levels of homophily.\nWe conduct extensive experiments on both synthetic and real-world datasets, which demonstrate the promising performance of PolyGCL on homophilic and heterophilic graphs.", "pdf": "https://openreview.net/pdf/e0bdb5536d418b614a12c003721153c1e6fbaf4b.pdf"} {"title": "Solving Homogeneous and Heterogeneous Cooperative Tasks with Greedy Sequential Execution", "url": "https://openreview.net/forum?id=hB2hXtxIPH", "detail_url": "https://openreview.net/forum?id=hB2hXtxIPH", "authors": "Shanqi Liu,Dong Xing,Pengjie Gu,Xinrun Wang,Bo An,Yong Liu", "tags": "ICLR 2024,Spotlight", "abstract": "Cooperative multi-agent reinforcement learning (MARL) is extensively used for solving complex cooperative tasks, and value decomposition methods are a prevalent approach for this domain. However, these methods have not been successful in addressing both homogeneous and heterogeneous tasks simultaneously which is a crucial aspect for the practical application of cooperative agents. \nOn one hand, value decomposition methods demonstrate superior performance in homogeneous tasks. Nevertheless, they tend to produce agents with similar policies, which is unsuitable for heterogeneous tasks. On the other hand, solutions based on personalized observation or assigned roles are well-suited for heterogeneous tasks. However, they often lead to a trade-off situation where the agent's performance in homogeneous scenarios is negatively affected due to the aggregation of distinct policies. An alternative approach is to adopt sequential execution policies, which offer a flexible form for learning both types of tasks. However, learning sequential execution policies poses challenges in terms of credit assignment, and the limited information about subsequently executed agents can lead to sub-optimal solutions, which is known as the relative over-generalization problem. To tackle these issues, this paper proposes Greedy Sequential Execution (GSE) as a solution to learn the optimal policy that covers both scenarios. In the proposed GSE framework, we introduce an individual utility function into the framework of value decomposition to consider the complex interactions between agents. \nThis function is capable of representing both the homogeneous and heterogeneous optimal policies. Furthermore, we utilize greedy marginal contribution calculated by the utility function as the credit value of the sequential execution policy to address the credit assignment and relative over-generalization problem. We evaluated GSE in both homogeneous and heterogeneous scenarios. The results demonstrate that GSE achieves significant improvement in performance across multiple domains, especially in scenarios involving both homogeneous and heterogeneous tasks.", "pdf": "https://openreview.net/pdf/e91e10a20d5e17ec8a9a469f872e7d0ec680cb37.pdf"} {"title": "Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from Offline Data", "url": "https://openreview.net/forum?id=Xkf2EBj4w3", "detail_url": "https://openreview.net/forum?id=Xkf2EBj4w3", "authors": "Chongyi Zheng,Benjamin Eysenbach,Homer Rich Walke,Patrick Yin,Kuan Fang,Ruslan Salakhutdinov,Sergey Levine", "tags": "ICLR 2024,Spotlight", "abstract": "Robotic systems that rely primarily on self-supervised learning have the potential to decrease the amount of human annotation and engineering effort required to learn control strategies. In the same way that prior robotic systems have leveraged self-supervised techniques from computer vision (CV) and natural language processing (NLP), our work builds on prior work showing that the reinforcement learning (RL) itself can be cast as a self-supervised problem: learning to reach any goal without human-specified rewards or labels. Despite the seeming appeal, little (if any) prior work has demonstrated how self-supervised RL methods can be practically deployed on robotic systems. By first studying a challenging simulated version of this task, we discover design decisions about architectures and hyperparameters that increase the success rate by $2 \\times$. These findings lay the groundwork for our main result: we demonstrate that a self-supervised RL algorithm based on contrastive learning can solve real-world, image-based robotic manipulation tasks, with tasks being specified by a single goal image provided after training.", "pdf": "https://openreview.net/pdf/d61ac5544c95c3239516e7c46f315bce6b00ce8b.pdf"} {"title": "Multi-View Causal Representation Learning with Partial Observability", "url": "https://openreview.net/forum?id=OGtnhKQJms", "detail_url": "https://openreview.net/forum?id=OGtnhKQJms", "authors": "Dingling Yao,Danru Xu,Sebastien Lachapelle,Sara Magliacane,Perouz Taslakian,Georg Martius,Julius von K\u00fcgelgen,Francesco Locatello", "tags": "ICLR 2024,Spotlight", "abstract": "We present a unified framework for studying the identifiability of representations learned from simultaneously observed views, such as different data modalities. We allow a partially observed setting in which each view constitutes a nonlinear mixture of a subset of underlying latent variables, which can be causally related.\nWe prove that the information shared across all subsets of any number of views can be learned up to a smooth bijection using contrastive learning and a single encoder per view. \nWe also provide graphical criteria indicating which latent variables can be identified through a simple set of rules, which we refer to as identifiability algebra. Our general framework and theoretical results unify and extend several previous work on multi-view nonlinear ICA, disentanglement, and causal representation learning. We experimentally validate our claims on numerical, image, and multi-modal data sets. Further, we demonstrate that the performance of prior methods is recovered in different special cases of our setup. \nOverall, we find that access to multiple partial views offers unique opportunities for identifiable representation learning, enabling the discovery of latent structures from purely observational data.", "pdf": "https://openreview.net/pdf/c4477da8d3ff31861069faeb5e0c7ebdb054e07f.pdf"} {"title": "CABINET: Content Relevance-based Noise Reduction for Table Question Answering", "url": "https://openreview.net/forum?id=SQrHpTllXa", "detail_url": "https://openreview.net/forum?id=SQrHpTllXa", "authors": "Sohan Patnaik,Heril Changwal,Milan Aggarwal,Sumit Bhatia,Yaman Kumar,Balaji Krishnamurthy", "tags": "ICLR 2024,Spotlight", "abstract": "Table understanding capability of Large Language Models (LLMs) has been extensively studied through the task of question-answering (QA) over tables. Typically, only a small part of the whole table is relevant to derive the answer for a given question. The irrelevant parts act as noise and are distracting information, resulting in sub-optimal performance due to the vulnerability of LLMs to noise. To mitigate this, we propose CABINET (Content RelevAnce-Based NoIse ReductioN for TablE QuesTion-Answering) \u2013 a framework to enable LLMs to focus on relevant tabular data by suppressing extraneous information. CABINET comprises an Unsupervised Relevance Scorer (URS), trained differentially with the QA LLM, that weighs the table content based on its relevance to the input question before feeding it to the question answering LLM (QA LLM). To further aid the relevance scorer, CABINET employs a weakly supervised module that generates a parsing statement describing the criteria of rows and columns relevant to the question and highlights the content of corresponding table cells. CABINET significantly outperforms various tabular LLM baselines, as well as GPT3-based in-context learning methods, is more robust to noise, maintains outperformance on tables of varying sizes, and establishes new SoTA performance on WikiTQ, FeTaQA, and WikiSQL datasets. We release our code and datasets here.", "pdf": "https://openreview.net/pdf/0a15c1a222a5d423ce19524261f01484f4e7b695.pdf"} {"title": "Safe RLHF: Safe Reinforcement Learning from Human Feedback", "url": "https://openreview.net/forum?id=TyFrPOKYXw", "detail_url": "https://openreview.net/forum?id=TyFrPOKYXw", "authors": "Josef Dai,Xuehai Pan,Ruiyang Sun,Jiaming Ji,Xinbo Xu,Mickel Liu,Yizhou Wang,Yaodong Yang", "tags": "ICLR 2024,Spotlight", "abstract": "With the development of large language models (LLMs), striking a balance between the performance and safety of AI systems has never been more critical. However, the inherent tension between the objectives of helpfulness and harmlessness presents a significant challenge during LLM training. To address this issue, we propose Safe Reinforcement Learning from Human Feedback (Safe RLHF), a novel algorithm for human value alignment. Safe RLHF explicitly decouples human preferences regarding helpfulness and harmlessness, effectively avoiding the crowd workers' confusion about the tension and allowing us to train separate reward and cost models. We formalize the safety concern of LLMs as an optimization task of maximizing the reward function while satisfying specified cost constraints. Leveraging the Lagrangian method to solve this constrained problem, Safe RLHF dynamically adjusts the balance between the two objectives during fine-tuning. Through a three-round fine-tuning using Safe RLHF, we demonstrate a superior ability to mitigate harmful responses while enhancing model performance compared to existing value-aligned algorithms. Experimentally, we fine-tuned the Alpaca-7B using Safe RLHF and aligned it with collected human preferences, significantly improving its helpfulness and harmlessness according to human evaluations.\n\nCode is available at https://github.com/PKU-Alignment/safe-rlhf.\n\n\nWarning: This paper contains example data that may be offensive or harmful.", "pdf": "https://openreview.net/pdf/4db509db9de557fb05dd265958739cb86ea87827.pdf"} {"title": "Benchmarking Algorithms for Federated Domain Generalization", "url": "https://openreview.net/forum?id=wprSv7ichW", "detail_url": "https://openreview.net/forum?id=wprSv7ichW", "authors": "Ruqi Bai,Saurabh Bagchi,David I. Inouye", "tags": "ICLR 2024,Spotlight", "abstract": "While prior federated learning (FL) methods mainly consider client heterogeneity, we focus on the *Federated Domain Generalization (DG)* task, which introduces train-test heterogeneity in the FL context. Existing evaluations in this field are limited in terms of the scale of the clients and dataset diversity. Thus, we propose a Federated DG benchmark that aim to test the limits of current methods with high client heterogeneity, large numbers of clients, and diverse datasets. Towards this objective, we introduce a novel data partition method that allows us to distribute any domain dataset among few or many clients while controlling client heterogeneity. We then introduce and apply our methodology to evaluate 14 DG methods, which include centralized DG methods adapted to the FL context, FL methods that handle client heterogeneity, and methods designed specifically for Federated DG on 7 datasets. Our results suggest that, despite some progress, significant performance gaps remain in Federated DG, especially when evaluating with a large number of clients, high client heterogeneity, or more realistic datasets. Furthermore, our extendable benchmark code will be publicly released to aid in benchmarking future Federated DG approaches.", "pdf": "https://openreview.net/pdf/216358074a4ed3ebfdc3beb60b624a2b6647445d.pdf"} {"title": "CrossQ: Batch Normalization in Deep Reinforcement Learning for Greater Sample Efficiency and Simplicity", "url": "https://openreview.net/forum?id=PczQtTsTIX", "detail_url": "https://openreview.net/forum?id=PczQtTsTIX", "authors": "Aditya Bhatt,Daniel Palenicek,Boris Belousov,Max Argus,Artemij Amiranashvili,Thomas Brox,Jan Peters", "tags": "ICLR 2024,Spotlight", "abstract": "Sample efficiency is a crucial problem in deep reinforcement learning. Recent algorithms, such as REDQ and DroQ, found a way to improve the sample efficiency by increasing the update-to-data (UTD) ratio to 20 gradient update steps on the critic per environment sample.\nHowever, this comes at the expense of a greatly increased computational cost. To reduce this computational burden, we introduce CrossQ:\nA lightweight algorithm for continuous control tasks that makes careful use of Batch Normalization and removes target networks to surpass the current state-of-the-art in sample efficiency while maintaining a low UTD ratio of 1. Notably, CrossQ does not rely on advanced bias-reduction schemes used in current methods. CrossQ's contributions are threefold: (1) it matches or surpasses current state-of-the-art methods in terms of sample efficiency, (2) it substantially reduces the computational cost compared to REDQ and DroQ, (3) it is easy to implement, requiring just a few lines of code on top of SAC.", "pdf": "https://openreview.net/pdf/750ae12418a1dc0f2dd3d9ff5ef5013234515fe6.pdf"} {"title": "Blending Imitation and Reinforcement Learning for Robust Policy Improvement", "url": "https://openreview.net/forum?id=eJ0dzPJq1F", "detail_url": "https://openreview.net/forum?id=eJ0dzPJq1F", "authors": "Xuefeng Liu,Takuma Yoneda,Rick Stevens,Matthew Walter,Yuxin Chen", "tags": "ICLR 2024,Spotlight", "abstract": "While reinforcement learning (RL) has shown promising performance, its sample complexity continues to be a substantial hurdle, restricting its broader application across a variety of domains. Imitation learning (IL) utilizes oracles to improve sample efficiency, yet it is often constrained by the quality of the oracles deployed. To address the demand for robust policy improvement in real-world scenarios, we introduce a novel algorithm, Robust Policy Improvement (RPI), which actively interleaves between IL and RL based on an online estimate of their performance. RPI draws on the strengths of IL, using oracle queries to facilitate exploration\u2014an aspect that is notably challenging in sparse-reward RL\u2014particularly during the early stages of learning. As learning unfolds, RPI gradually transitions to RL, effectively treating the learned policy as an improved oracle. This algorithm is capable of learning from and improving upon a diverse set of black-box oracles. Integral to RPI are Robust Active Policy Selection (RAPS) and Robust Policy Gradient (RPG), both of which reason over whether to perform state-wise imitation from the oracles or learn from its own value function when the learner\u2019s performance surpasses that of the oracles in a specific state. Empirical evaluations and theoretical analysis validate that RPI excels in comparison to existing state-of-the-art methodologies, demonstrating superior performance across various benchmark domains.", "pdf": "https://openreview.net/pdf/bdad46427f76d0c9b2c72d8012d5b33aeddc4e8e.pdf"} {"title": "H-GAP: Humanoid Control with a Generalist Planner", "url": "https://openreview.net/forum?id=LYG6tBlEX0", "detail_url": "https://openreview.net/forum?id=LYG6tBlEX0", "authors": "zhengyao jiang,Yingchen Xu,Nolan Wagener,Yicheng Luo,Michael Janner,Edward Grefenstette,Tim Rockt\u00e4schel,Yuandong Tian", "tags": "ICLR 2024,Spotlight", "abstract": "Humanoid control is an important research challenge offering avenues for integration into human-centric infrastructures and enabling physics-driven humanoid animations.\nThe daunting challenges in this field stem from the difficulty of optimizing in high-dimensional action spaces and the instability introduced by the bipedal morphology of humanoids. \nHowever, the extensive collection of human motion-captured data and the derived datasets of humanoid trajectories, such as MoCapAct, paves the way to tackle these challenges. In this context, we present Humanoid Generalist Autoencoding Planner (H-GAP), a state-action trajectory generative model trained on humanoid trajectories derived from human motion-captured data, capable of adeptly handling downstream control tasks with Model Predictive Control (MPC).\nFor 56 degrees of freedom humanoid, we empirically demonstrate that H-GAP learns to represent and generate a wide range of motor behaviors. Further, without any learning from online interactions, it can also flexibly transfer these behaviours to solve novel downstream control tasks via planning. Notably, H-GAP excels established MPC baselines with access to the ground truth model, and is superior or comparable to offline RL methods trained for individual tasks.\nFinally, we do a series of empirical studies on the scaling properties of H-GAP, showing the potential for performance gains via additional data but not computing.", "pdf": "https://openreview.net/pdf/5572333360c59f829af61902b8f2157f4a2e4109.pdf"} {"title": "Unlocking the Power of Representations in Long-term Novelty-based Exploration", "url": "https://openreview.net/forum?id=OwtMhMSybu", "detail_url": "https://openreview.net/forum?id=OwtMhMSybu", "authors": "Alaa Saade,Steven Kapturowski,Daniele Calandriello,Charles Blundell,Pablo Sprechmann,Leopoldo Sarra,Oliver Groth,Michal Valko,Bilal Piot", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce Robust Exploration via Clustering-based Online Density Estimation (RECODE), a non-parametric method for novelty-based exploration that estimates visitation counts for clusters of states based on their similarity in a chosen embedding space. By adapting classical clustering to the nonstationary setting of Deep RL, RECODE can efficiently track state visitation counts over thousands of episodes. We further propose a novel generalization of the inverse dynamics loss, which leverages masked transformer architectures for multi-step prediction; which in conjunction with \\DETOCS achieves a new state-of-the-art in a suite of challenging 3D-exploration tasks in DM-Hard-8. RECODE also sets new state-of-the-art in hard exploration Atari games, and is the first agent to reach the end screen in \"Pitfall!\"", "pdf": "https://openreview.net/pdf/97bd16e3685d8761abc10cf118ffd94e8ae31b77.pdf"} {"title": "Accelerating Data Generation for Neural Operators via Krylov Subspace Recycling", "url": "https://openreview.net/forum?id=UpgRVWexaD", "detail_url": "https://openreview.net/forum?id=UpgRVWexaD", "authors": "Hong Wang,Zhongkai Hao,Jie Wang,Zijie Geng,Zhen Wang,Bin Li,Feng Wu", "tags": "ICLR 2024,Spotlight", "abstract": "Learning neural operators for solving partial differential equations (PDEs) has attracted great attention due to its high inference efficiency.\nHowever, training such operators requires generating a substantial amount of labeled data, i.e., PDE problems together with their solutions.\nThe data generation process is exceptionally time-consuming, as it involves solving numerous systems of linear equations to obtain numerical solutions to the PDEs.\nMany existing methods solve these systems independently without considering their inherent similarities, resulting in extremely redundant computations.\nTo tackle this problem, we propose a novel method, namely **S**orting **K**rylov **R**ecycling (**SKR**), to boost the efficiency of solving these systems, thus significantly accelerating data generation for neural operators training.\nTo the best of our knowledge, SKR is the first attempt to address the time-consuming nature of data generation for learning neural operators.\nThe working horse of SKR is Krylov subspace recycling, a powerful technique for solving a series of interrelated systems by leveraging their inherent similarities.\nSpecifically, SKR employs a sorting algorithm to arrange these systems in a sequence, where adjacent systems exhibit high similarities.\nThen it equips a solver with Krylov subspace recycling to solve the systems sequentially instead of independently, thus effectively enhancing the solving efficiency.\nBoth theoretical analysis and extensive experiments demonstrate that SKR can significantly accelerate neural operator data generation, achieving a remarkable speedup of up to 13.9 times.", "pdf": "https://openreview.net/pdf/ff0efe03064ef3409c6562ca6ccdf6ff02fb9cba.pdf"} {"title": "Deep Orthogonal Hypersphere Compression for Anomaly Detection", "url": "https://openreview.net/forum?id=cJs4oE4m9Q", "detail_url": "https://openreview.net/forum?id=cJs4oE4m9Q", "authors": "Yunhe Zhang,Yan Sun,Jinyu Cai,Jicong Fan", "tags": "ICLR 2024,Spotlight", "abstract": "Many well-known and effective anomaly detection methods assume that a reasonable decision boundary has a hypersphere shape, which however is difficult to obtain in practice and is not sufficiently compact, especially when the data are in high-dimensional spaces. In this paper, we first propose a novel deep anomaly detection model that improves the original hypersphere learning through an orthogonal projection layer, which ensures that the training data distribution is consistent with the hypersphere hypothesis, thereby increasing the true positive rate and decreasing the false negative rate. Moreover, we propose a bi-hypersphere compression method to obtain a hyperspherical shell that yields a more compact decision region than a hyperball, which is demonstrated theoretically and numerically. The proposed methods are not confined to common datasets such as image and tabular data, but are also extended to a more challenging but promising scenario, graph-level anomaly detection, which learns graph representation with maximum mutual information between the substructure and global structure features while exploring orthogonal single- or bi-hypersphere anomaly decision boundaries. The numerical and visualization results on benchmark datasets demonstrate the superiority of our methods in comparison to many baselines and state-of-the-art methods.", "pdf": "https://openreview.net/pdf/b8052c3c7a3cfeccc963ae2d0d24045831b4d84e.pdf"} {"title": "On the Role of General Function Approximation in Offline Reinforcement Learning", "url": "https://openreview.net/forum?id=JSS9rKHySk", "detail_url": "https://openreview.net/forum?id=JSS9rKHySk", "authors": "Chenjie Mao,Qiaosheng Zhang,Zhen Wang,Xuelong Li", "tags": "ICLR 2024,Spotlight", "abstract": "We study offline reinforcement learning (RL) with general function approximation. General function approximation is a powerful tool for algorithm design and analysis, but its adaptation to offline RL encounters several challenges due to varying approximation targets and assumptions that blur the real meanings of function assumptions. In this paper, we try to formulate and clarify the treatment of general function approximation in offline RL in two aspects: (1) analyzing different types of assumptions and their practical usage, and (2) understanding its role as a restriction on underlying MDPs from information-theoretic perspectives. Additionally, we introduce a new insight for lower bound establishing: one can exploit model-realizability to establish general-purpose lower bounds that can be generalized into other functions. Building upon this insight, we propose two generic lower bounds that contribute to a better understanding of offline RL with general function approximation.", "pdf": "https://openreview.net/pdf/8621c59f51343401b65a5d8c4ba33ef5a631dd93.pdf"} {"title": "Analyzing Feed-Forward Blocks in Transformers through the Lens of Attention Maps", "url": "https://openreview.net/forum?id=mYWsyTuiRp", "detail_url": "https://openreview.net/forum?id=mYWsyTuiRp", "authors": "Goro Kobayashi,Tatsuki Kuribayashi,Sho Yokoi,Kentaro Inui", "tags": "ICLR 2024,Spotlight", "abstract": "Transformers are ubiquitous in wide tasks.\nInterpreting their internals is a pivotal goal. \nNevertheless, their particular components, feed-forward (FF) blocks, have typically been less analyzed despite their substantial parameter amounts.\nWe analyze the input contextualization effects of FF blocks by rendering them in the attention maps as a human-friendly visualization scheme.\nOur experiments with both masked- and causal-language models reveal that FF networks modify the input contextualization to emphasize specific types of linguistic compositions. \nIn addition, FF and its surrounding components tend to cancel out each other's effects, suggesting potential redundancy in the processing of the Transformer layer.", "pdf": "https://openreview.net/pdf/bf95c5609d1e4a6df3810208c09e97d13c14a953.pdf"} {"title": "Asymptotically Free Sketched Ridge Ensembles: Risks, Cross-Validation, and Tuning", "url": "https://openreview.net/forum?id=i9Vs5NGDpk", "detail_url": "https://openreview.net/forum?id=i9Vs5NGDpk", "authors": "Pratik Patil,Daniel LeJeune", "tags": "ICLR 2024,Spotlight", "abstract": "We employ random matrix theory to establish consistency of generalized cross validation (GCV) for estimating prediction risks of sketched ridge regression ensembles, enabling efficient and consistent tuning of regularization and sketching parameters. Our results hold for a broad class of asymptotically free sketches under very mild data assumptions. For squared prediction risk, we provide a decomposition into an unsketched equivalent implicit ridge bias and a sketching-based variance, and prove that the risk can be globally optimized by only tuning sketch size in infinite ensembles. For general subquadratic prediction risk functionals, we extend GCV to construct consistent risk estimators, and thereby obtain distributional convergence of the GCV-corrected predictions in Wasserstein-2 metric. This in particular allows construction of prediction intervals with asymptotically correct coverage conditional on the training data. We also propose an \"ensemble trick\" whereby the risk for unsketched ridge regression can be efficiently estimated via GCV using small sketched ridge ensembles. We empirically validate our theoretical results using both synthetic and real large-scale datasets with practical sketches including CountSketch and subsampled randomized discrete cosine transforms.", "pdf": "https://openreview.net/pdf/28ddc17b08b8df3bdcb23d2393cdb2d882b39eef.pdf"} {"title": "Nemesis: Normalizing the Soft-prompt Vectors of Vision-Language Models", "url": "https://openreview.net/forum?id=zmJDzPh1Dm", "detail_url": "https://openreview.net/forum?id=zmJDzPh1Dm", "authors": "Shuai Fu,Xiequn Wang,Qiushi Huang,Yu Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "With the prevalence of large-scale pretrained vision-language models (VLMs), such as CLIP, soft-prompt tuning has become a popular method for adapting these models to various downstream tasks. However, few works delve into the inherent properties of learnable soft-prompt vectors, specifically the impact of their norms to the performance of VLMs. This motivates us to pose an unexplored research question: ``Do we need to normalize the soft prompts in VLMs?'' To fill this research gap, we first uncover a phenomenon, called the $\\textbf{Low-Norm Effect}$ by performing extensive corruption experiments, suggesting that reducing the norms of certain learned prompts occasionally enhances the performance of VLMs, while increasing them often degrades it. To harness this effect, we propose a novel method named $\\textbf{N}$ormalizing th$\\textbf{e}$ soft-pro$\\textbf{m}$pt v$\\textbf{e}$ctors of vi$\\textbf{si}$on-language model$\\textbf{s}$ ($\\textbf{Nemesis}$) to normalize soft-prompt vectors in VLMs. To the best of our knowledge, our work is the first to systematically investigate the role of norms of soft-prompt vector in VLMs, offering valuable insights for future research in soft-prompt tuning.", "pdf": "https://openreview.net/pdf/13b62a20df81ff095be76d919e6602f6f86780bf.pdf"} {"title": "Towards Understanding Factual Knowledge of Large Language Models", "url": "https://openreview.net/forum?id=9OevMUdods", "detail_url": "https://openreview.net/forum?id=9OevMUdods", "authors": "Xuming Hu,Junzhe Chen,Xiaochuan Li,Yufei Guo,Lijie Wen,Philip S. Yu,Zhijiang Guo", "tags": "ICLR 2024,Spotlight", "abstract": "Large language models (LLMs) have recently driven striking performance improvements across a range of natural language processing tasks. The factual knowledge acquired during pretraining and instruction tuning can be useful in various downstream tasks, such as question answering, and language generation. Unlike conventional Knowledge Bases (KBs) that explicitly store factual knowledge, LLMs implicitly store facts in their parameters. Content generated by the LLMs can often exhibit inaccuracies or deviations from the truth, due to facts that can be incorrectly induced or become obsolete over time. To this end, we aim to explore the extent and scope of factual knowledge within LLMs by designing the benchmark Pinocchio. Pinocchio contains 20K diverse factual questions that span different sources, timelines, domains, regions, and languages. Furthermore, we investigate whether LLMs can compose multiple facts, update factual knowledge temporally, reason over multiple pieces of facts, identify subtle factual differences, and resist adversarial examples. Extensive experiments on different sizes and types of LLMs show that existing LLMs still lack factual knowledge and suffer from various spurious correlations. We believe this is a critical bottleneck for realizing trustworthy artificial intelligence. The dataset Pinocchio and our codes are publicly available at: https://github.com/THU-BPM/Pinocchio.", "pdf": "https://openreview.net/pdf/025c1d2beaec72d3724b68ca610dd61083362fdc.pdf"} {"title": "CAS: A Probability-Based Approach for Universal Condition Alignment Score", "url": "https://openreview.net/forum?id=E78OaH2s3f", "detail_url": "https://openreview.net/forum?id=E78OaH2s3f", "authors": "Chunsan Hong,ByungHee Cha,Tae-Hyun Oh", "tags": "ICLR 2024,Spotlight", "abstract": "Recent conditional diffusion models have shown remarkable advancements and have been widely applied in fascinating real-world applications. However, samples generated by these models often do not strictly comply with user-provided conditions. Due to this, there have been few attempts to evaluate this alignment via pre-trained scoring models to select well-generated samples. Nonetheless, current studies are confined to the text-to-image domain and require large training datasets. This suggests that crafting alignment scores for various conditions will demand considerable resources in the future. In this context, we introduce a universal condition alignment score that leverages the conditional probability measurable through the diffusion process. Our technique operates across all conditions and requires no additional models beyond the diffusion model used for generation, effectively enabling self-rejection. Our experiments validate that our met- ric effectively applies in diverse conditional generations, such as text-to-image, {instruction, image}-to-image, edge-/scribble-to-image, and text-to-audio.", "pdf": "https://openreview.net/pdf/f985d33f22b4298cf2016099ca2640c09738d070.pdf"} {"title": "Demystifying CLIP Data", "url": "https://openreview.net/forum?id=5BCFlnfE1g", "detail_url": "https://openreview.net/forum?id=5BCFlnfE1g", "authors": "Hu Xu,Saining Xie,Xiaoqing Tan,Po-Yao Huang,Russell Howes,Vasu Sharma,Shang-Wen Li,Gargi Ghosh,Luke Zettlemoyer,Christoph Feichtenhofer", "tags": "ICLR 2024,Spotlight", "abstract": "Contrastive Language-Image Pre-training (CLIP) is an approach that has advanced research and applications in computer vision, fueling modern recognition systems and generative models. We believe that the main ingredient to the success of CLIP is its \\textit{data} and \\textit{not} the \\textit{model} architecture or pre-training {objective}. However, CLIP only provides very limited information about its data and how it has been collected, leading to works that aim to reproduce CLIP's data by filtering with its model parameters. In this work, we intend to reveal CLIP's data curation approach and in our pursuit of making it open to the community introduce Metadata-Curated Language-Image Pre-training (MetaCLIP). MetaCLIP takes a raw data pool and metadata (derived from CLIP's concepts) and yields a balanced subset over the metadata distribution. Our experimental study rigorously isolates the model and training settings, concentrating solely on data. MetaCLIP applied to CommonCrawl with 400M image-text data pairs outperforms CLIP's data on multiple standard benchmarks. In zero-shot ImageNet classification, MetaCLIP achieves 70.8\\% accuracy, surpassing CLIP's 68.3\\% on \\mbox{ViT-B} models. Scaling to 1B data, while maintaining the same training budget, attains \\textbf{72.4\\%}. Our observations hold across various model sizes, exemplified by ViT-H achieving \\textbf{80.5\\%}, without any bells-and-whistles. Curation code and training data distribution over metadata will be made available.", "pdf": "https://openreview.net/pdf/9c2e91f6f4a65a812be7ea55d1d89e0fd4844e26.pdf"} {"title": "Adversarial AutoMixup", "url": "https://openreview.net/forum?id=o8tjamaJ80", "detail_url": "https://openreview.net/forum?id=o8tjamaJ80", "authors": "Huafeng Qin,Xin Jin,Yun Jiang,Moun\u00eem El-Yacoubi,Xinbo Gao", "tags": "ICLR 2024,Spotlight", "abstract": "Data mixing augmentation has been widely applied to improve the generalization ability of deep neural networks. Recently, offline data mixing augmentation, e.g. handcrafted and saliency information-based mixup, has been gradually replaced by automatic mixing approaches. Through minimizing two sub-tasks, namely, mixed sample generation and mixup classification in an end-to-end way, AutoMix significantly improves accuracy on image classification tasks. However, as the optimization objective is consistent for the two sub-tasks, this approach is prone to generating consistent instead of diverse mixed samples, which results in overfitting for target task training. In this paper, we propose AdAutomixup, an adversarial automatic mixup augmentation approach that generates challenging samples to train a robust classifier for image classification, by alternatively optimizing the classifier and the mixup sample generator. AdAutomixup comprises two modules, a mixed example generator, and a target classifier. The mixed sample generator aims to produce hard mixed examples to challenge the target classifier, while the target classifier's aim is to learn robust features from hard mixed examples to improve generalization. To prevent the collapse of the inherent meanings of images, we further introduce an exponential moving average (EMA) teacher and cosine similarity to train AdAutomixup in an end-to-end way. Extensive experiments on seven image benchmarks consistently prove that our approach outperforms the state of the art in various classification scenarios. The source code is available at \nhttps://github.com/JinXins/Adversarial-AutoMixup.", "pdf": "https://openreview.net/pdf/47a0f838e2a5b9d0158f56d2adb6432f9f878803.pdf"} {"title": "Spatially-Aware Transformers for Embodied Agents", "url": "https://openreview.net/forum?id=Ts95eXsPBc", "detail_url": "https://openreview.net/forum?id=Ts95eXsPBc", "authors": "Junmo Cho,Jaesik Yoon,Sungjin Ahn", "tags": "ICLR 2024,Spotlight", "abstract": "Episodic memory plays a crucial role in various cognitive processes, such as the ability to mentally recall past events. While cognitive science emphasizes the significance of spatial context in the formation and retrieval of episodic memory, the current primary approach to implementing episodic memory in AI systems is through transformers that store temporally ordered experiences, which overlooks the spatial dimension. As a result, it is unclear how the underlying structure could be extended to incorporate the spatial axis beyond temporal order alone and thereby what benefits can be obtained. To address this, this paper explores the use of Spatially-Aware Transformer models that incorporate spatial information. These models enable the creation of place-centric episodic memory that considers both temporal and spatial dimensions. Adopting this approach, we demonstrate that memory utilization efficiency can be improved, leading to enhanced accuracy in various place-centric downstream tasks. Additionally, we propose the Adaptive Memory Allocator, a memory management method based on reinforcement learning that aims to optimize efficiency of memory utilization. Our experiments demonstrate the advantages of our proposed model in various environments and across multiple downstream tasks, including prediction, generation, reasoning, and reinforcement learning. The source code for our models and experiments will be available at \\href{https://github.com/spatially_aware_transformer}{https://github.com/spatially_aware_transformer}.", "pdf": "https://openreview.net/pdf/1dc0ff45874e872749597c38bcfc2df0fa7ed0d6.pdf"} {"title": "Grounding Language Plans in Demonstrations Through Counterfactual Perturbations", "url": "https://openreview.net/forum?id=qoHeuRAcSl", "detail_url": "https://openreview.net/forum?id=qoHeuRAcSl", "authors": "Yanwei Wang,Tsun-Hsuan Wang,Jiayuan Mao,Michael Hagenow,Julie Shah", "tags": "ICLR 2024,Spotlight", "abstract": "Grounding the common-sense reasoning of Large Language Models in physical domains remains a pivotal yet unsolved problem for embodied AI. Whereas prior works have focused on leveraging LLMs directly for planning in symbolic spaces, this work uses LLMs to guide the search of task structures and constraints implicit in multi-step demonstrations. Specifically, we borrow from manipulation planning literature the concept of mode families, which group robot configurations by specific motion constraints, to serve as an abstraction layer between the high-level language representations of an LLM and the low-level physical trajectories of a robot. By replaying a few human demonstrations with synthetic perturbations, we generate coverage over the demonstrations' state space with additional successful executions as well as counterfactuals that fail the task. Our explanation-based learning framework trains an end-to-end differentiable neural network to predict successful trajectories from failures and as a by-product learns classifiers that ground low-level states and images in mode families without dense labeling. The learned grounding classifiers can further be used to translate language plans into reactive policies in the physical domain in an interpretable manner. We show our approach improves the interpretability and reactivity of imitation learning through 2D navigation and simulated and real robot manipulation tasks. Website: https://yanweiw.github.io/glide/", "pdf": "https://openreview.net/pdf/095205c11ca0cd7a485da10923a605bbfd899160.pdf"} {"title": "Beyond Worst-case Attacks: Robust RL with Adaptive Defense via Non-dominated Policies", "url": "https://openreview.net/forum?id=DFTHW0MyiW", "detail_url": "https://openreview.net/forum?id=DFTHW0MyiW", "authors": "Xiangyu Liu,Chenghao Deng,Yanchao Sun,Yongyuan Liang,Furong Huang", "tags": "ICLR 2024,Spotlight", "abstract": "In light of the burgeoning success of reinforcement learning (RL) in diverse real-world applications, considerable focus has been directed towards ensuring RL policies are robust to adversarial attacks during test time. Current approaches largely revolve around solving a minimax problem to prepare for potential worst-case scenarios. While effective against strong attacks, these methods often compromise performance in the absence of attacks or the presence of only weak attacks. To address this, we study policy robustness under the well-accepted state-adversarial attack model, extending our focus beyond merely worst-case attacks. We first formalize this task at test time as a regret minimization problem and establish its intrinsic difficulty in achieving sublinear regret when the baseline policy is from a general continuous policy class, $\\Pi$. This finding prompts us to \\textit{refine} the baseline policy class $\\Pi$ prior to test time, aiming for efficient adaptation within a compact, finite policy class $\\tilde{\\Pi}$, which can resort to an adversarial bandit subroutine. In light of the importance of a finite and compact $\\tilde{\\Pi}$, we propose a novel training-time algorithm to iteratively discover \\textit{non-dominated policies}, forming a near-optimal and minimal $\\tilde{\\Pi}$, thereby ensuring both robustness and test-time efficiency. Empirical validation on the Mujoco corroborates the superiority of our approach in terms of natural and robust performance, as well as adaptability to various attack scenarios.", "pdf": "https://openreview.net/pdf/a7e93789afc93e70c42ccbaecde9d7a6bb824d01.pdf"} {"title": "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy", "url": "https://openreview.net/forum?id=eFWG9Cy3WK", "detail_url": "https://openreview.net/forum?id=eFWG9Cy3WK", "authors": "Pingzhi Li,Zhenyu Zhang,Prateek Yadav,Yi-Lin Sung,Yu Cheng,Mohit Bansal,Tianlong Chen", "tags": "ICLR 2024,Spotlight", "abstract": "Sparsely activated Mixture-of-Experts (SMoE) has shown promise to scale up the learning capacity of neural networks, however, they have issues like: ($a$) $\\textit{High Memory Usage,}$ due to duplication of the network layers into multiple copies as experts; and ($b$) $\\textit{Redundancy in Experts,}$ as common learning-based routing policies suffer from representational collapse. Therefore, vanilla SMoE models are memory inefficient and non-scalable, especially for resource-constrained downstream scenarios. In this paper, we ask: Can we craft a compact SMoE model by consolidating expert information? What is the best recipe to merge multiple experts into fewer but more knowledgeable experts? Our pilot investigation reveals that conventional model merging methods fail to be effective in such expert merging for SMoE. The potential reasons are: ($1$) redundant information overshadows critical experts; ($2$) appropriate neuron permutation for each expert is missing to bring all of them in alignment. To address these challenges, we propose a novel merging algorithm for SMoE, $\\textit{i.e.}$, $\\texttt{M-SMoE}$, which leverages routing statistics to guide expert merging. Specifically, it starts with neuron permutation alignment for experts; then, dominant experts and their \"group members\" are formed based on routing policies; lastly, every expert group is merged into a single expert by utilizing each expert's activation frequency as their weight for merging, thus diminishing the impact of insignificant experts. Moreover, we draw an interesting observation that our proposed merging promotes a low dimensionality in the merged expert's weight space, naturally paving the way for additional compression. Hence, our final method, $\\texttt{MC-SMoE}$ ($\\textit{i.e.}$, Merge, then Compress SMoE), further decomposes the merged experts into low-rank and structural sparse alternatives. Extensive experiments across $8$ benchmarks validate the effectiveness of our proposals. For instance, our $\\texttt{MC-SMoE}$ achieves up to $80\\%$ memory and a $20\\%$ FLOPs reduction, with virtually no loss in performance. Our code is provided as supplementary material.", "pdf": "https://openreview.net/pdf/93b33eef04948ce274bfc922884b3a34062628c7.pdf"} {"title": "On Bias-Variance Alignment in Deep Models", "url": "https://openreview.net/forum?id=i2Phucne30", "detail_url": "https://openreview.net/forum?id=i2Phucne30", "authors": "Lin Chen,Michal Lukasik,Wittawat Jitkrittum,Chong You,Sanjiv Kumar", "tags": "ICLR 2024,Spotlight", "abstract": "Classical wisdom in machine learning holds that the generalization error can be decomposed into bias and variance, and these two terms exhibit a \\emph{trade-off}. However, in this paper, we show that for an ensemble of deep learning based classification models, bias and variance are \\emph{aligned} at a sample level, where squared bias is approximately \\emph{equal} to variance for correctly classified sample points. We present empirical evidence confirming this phenomenon in a variety of deep learning models and datasets. Moreover, we study this phenomenon from two theoretical perspectives: calibration and neural collapse. We first show theoretically that under the assumption that the models are well calibrated, we can observe the bias-variance alignment. Second, starting from the picture provided by the neural collapse theory, we show an approximate correlation between bias and variance.", "pdf": "https://openreview.net/pdf/178a29fe5352a6013ac367fcb6cc4e69d501cb81.pdf"} {"title": "SEGNO: Generalizing Equivariant Graph Neural Networks with Physical Inductive Biases", "url": "https://openreview.net/forum?id=3oTPsORaDH", "detail_url": "https://openreview.net/forum?id=3oTPsORaDH", "authors": "Yang Liu,Jiashun Cheng,Haihong Zhao,Tingyang Xu,Peilin Zhao,Fugee Tsung,Jia Li,Yu Rong", "tags": "ICLR 2024,Spotlight", "abstract": "Graph Neural Networks (GNNs) with equivariant properties have emerged as powerful tools for modeling complex dynamics of multi-object physical systems. However, their generalization ability is limited by the inadequate consideration of physical inductive biases: (1) Existing studies overlook the continuity of transitions among system states, opting to employ several discrete transformation layers to learn the direct mapping between two adjacent states; (2) Most models only account for first-order velocity information, despite the fact that many physical systems are governed by second-order motion laws. To incorporate these inductive biases, we propose the Second-order Equivariant Graph Neural Ordinary Differential Equation (SEGNO). Specifically, we show how the second-order continuity can be incorporated into GNNs while maintaining the equivariant property. Furthermore, we offer theoretical insights into SEGNO, highlighting that it can learn a unique trajectory between adjacent states, which is crucial for model generalization. Additionally, we prove that the discrepancy between this learned trajectory of SEGNO and the true trajectory is bounded. Extensive experiments on complex dynamical systems including molecular dynamics and motion capture demonstrate that our model yields a significant improvement over the state-of-the-art baselines.", "pdf": "https://openreview.net/pdf/4acee75d595246c171ea11abc7187435b10862ea.pdf"} {"title": "Spectrally Transformed Kernel Regression", "url": "https://openreview.net/forum?id=OeQE9zsztS", "detail_url": "https://openreview.net/forum?id=OeQE9zsztS", "authors": "Runtian Zhai,Rattana Pukdee,Roger Jin,Maria Florina Balcan,Pradeep Kumar Ravikumar", "tags": "ICLR 2024,Spotlight", "abstract": "Unlabeled data is a key component of modern machine learning. In general, the role\nof unlabeled data is to impose a form of smoothness, usually from the similarity\ninformation encoded in a base kernel, such as the \u03f5-neighbor kernel or the adjacency\nmatrix of a graph. This work revisits the classical idea of spectrally transformed\nkernel regression (STKR), and provides a new class of general and scalable STKR\nestimators able to leverage unlabeled data. Intuitively, via spectral transformation,\nSTKR exploits the data distribution for which unlabeled data can provide additional\ninformation. First, we show that STKR is a principled and general approach,\nby characterizing a universal type of \u201ctarget smoothness\u201d, and proving that any\nsufficiently smooth function can be learned by STKR. Second, we provide scalable\nSTKR implementations for the inductive setting and a general transformation\nfunction, while prior work is mostly limited to the transductive setting. Third, we\nderive statistical guarantees for two scenarios: STKR with a known polynomial\ntransformation, and STKR with kernel PCA when the transformation is unknown.\nOverall, we believe that this work helps deepen our understanding of how to work\nwith unlabeled data, and its generality makes it easier to inspire new methods.", "pdf": "https://openreview.net/pdf/cb1d12d196c77a2a26e09caff4a57bf370c27871.pdf"} {"title": "Online GNN Evaluation Under Test-time Graph Distribution Shifts", "url": "https://openreview.net/forum?id=KbetDM33YG", "detail_url": "https://openreview.net/forum?id=KbetDM33YG", "authors": "Xin Zheng,Dongjin Song,Qingsong Wen,Bo Du,Shirui Pan", "tags": "ICLR 2024,Spotlight", "abstract": "Evaluating the performance of a well-trained GNN model on real-world graphs is a pivotal step for reliable GNN online deployment and serving. \nDue to a lack of test node labels and unknown potential training-test graph data distribution shifts, conventional model evaluation encounters limitations in calculating performance metrics (e.g., test error) and measuring graph data-level discrepancies, particularly when the training graph used for developing GNNs remains unobserved during test time.\nIn this paper, we study a new research problem, online GNN evaluation, which aims to provide valuable insights into the well-trained GNNs's ability to effectively generalize to real-world unlabeled graphs under the test-time graph distribution shifts.\nConcretely, we develop an effective learning behavior discrepancy score, dubbed LeBeD, to estimate the test-time generalization errors of well-trained GNN models. \nThrough a novel GNN re-training strategy with a parameter-free optimality criterion, the proposed LeBeD comprehensively integrates learning behavior discrepancies from both node prediction and structure reconstruction perspectives.\nThis enables the effective evaluation of the well-trained GNNs' ability to capture test node semantics and structural representations, making it an expressive metric for estimating the generalization error in online GNN evaluation.\nExtensive experiments on real-world test graphs under diverse graph distribution shifts could verify the effectiveness of the proposed method, revealing its strong correlation with ground-truth test errors on various well-trained GNN models.", "pdf": "https://openreview.net/pdf/3b5dbcc5108726d3892b0ca05c69f9fa4eca4529.pdf"} {"title": "Understanding and Mitigating the Label Noise in Pre-training on Downstream Tasks", "url": "https://openreview.net/forum?id=TjhUtloBZU", "detail_url": "https://openreview.net/forum?id=TjhUtloBZU", "authors": "Hao Chen,Jindong Wang,Ankit Shah,Ran Tao,Hongxin Wei,Xing Xie,Masashi Sugiyama,Bhiksha Raj", "tags": "ICLR 2024,Spotlight", "abstract": "Pre-training on large-scale datasets and then fine-tuning on downstream tasks have become a standard practice in deep learning. However, pre-training data often contain label noise that may adversely affect the generalization of the model. This paper aims to understand the nature of noise in pre-training datasets and to mitigate its impact on downstream tasks. More specifically, through extensive experiments of supervised pre-training models on synthetic noisy ImageNet-1K and YFCC15M datasets, we demonstrate that while slight noise in pre-training can benefit in-domain (ID) transfer performance, where the training and testing data share the same distribution, it always deteriorates out-of-domain (OOD) performance, where training and testing data distribution are different. We empirically verify that the reason behind is noise in pre-training shapes the feature space differently. We then propose a light-weight black-box tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization on both ID and OOD tasks, considering one may not be able to fully fine-tune or even access the pre-trained models. We conduct practical experiments on popular vision and language models that are pre-trained on noisy data for evaluation of our approach. Our analysis and results show the importance of this interesting and novel research direction, which we term Noisy Model Learning.", "pdf": "https://openreview.net/pdf/ae1d6d7106b28a7cb9cacfd2f2b5dd34c10f0ce0.pdf"} {"title": "WildChat: 1M ChatGPT Interaction Logs in the Wild", "url": "https://openreview.net/forum?id=Bl8u7ZRlbM", "detail_url": "https://openreview.net/forum?id=Bl8u7ZRlbM", "authors": "Wenting Zhao,Xiang Ren,Jack Hessel,Claire Cardie,Yejin Choi,Yuntian Deng", "tags": "ICLR 2024,Spotlight", "abstract": "Chatbots such as GPT-4 and ChatGPT are now serving millions of users. Despite their widespread use, there remains a lack of public datasets showcasing how these tools are used by a population of users in practice. To bridge this gap, we offered free access to ChatGPT for online users in exchange for their affirmative, consensual opt-in to anonymously collect their chat transcripts and request headers. From this, we compiled WildChat, a corpus of 1 million user-ChatGPT conversations, which consists of over 2.5 million interaction turns. We compare WildChat with other popular user-chatbot interaction datasets, and find that our dataset offers the most diverse user prompts, contains the largest number of languages, and presents the richest variety of potentially toxic use-cases for researchers to study. In addition to timestamped chat transcripts, we enrich the dataset with demographic data, including state, country, and hashed IP addresses, alongside request headers. This augmentation allows for more detailed analysis of user behaviors across different geographical regions and temporal dimensions. Finally, because it captures a broad range of use cases, we demonstrate the dataset\u2019s potential utility in fine-tuning instruction-following models. WildChat is released at https://wildchat.allen.ai under AI2 ImpACT Licenses.", "pdf": "https://openreview.net/pdf/40986fdeaa994d9dc8bbbe1a2c320eff8eff10a7.pdf"} {"title": "Learning Hierarchical Image Segmentation For Recognition and By Recognition", "url": "https://openreview.net/forum?id=IRcv4yFX6z", "detail_url": "https://openreview.net/forum?id=IRcv4yFX6z", "authors": "Tsung-Wei Ke,Sangwoo Mo,Stella X. Yu", "tags": "ICLR 2024,Spotlight", "abstract": "Large vision and language models learned directly through image-text associations often lack detailed visual substantiation, whereas image segmentation tasks are treated separately from recognition, supervisedly learned without interconnections.\n\nOur key observation is that, while an image can be recognized in multiple ways, each has a consistent part-and-whole visual organization. Segmentation thus should be treated not as an end task to be mastered through supervised learning, but as an internal process that evolves with and supports the ultimate goal of recognition. \n\nWe propose to integrate a hierarchical segmenter into the recognition process, \n{\\it train} and {\\it adapt} the entire model solely on image-level recognition objectives. We learn hierarchical segmentation {\\it for free} alongside recognition, automatically uncovering part-to-whole relationships that not only underpin but also enhance recognition. \n\nEnhancing the Vision Transformer (ViT) with adaptive segment tokens and graph pooling, our model surpasses ViT in unsupervised part-whole discovery, semantic segmentation, image classification, and efficiency. Notably, our model (trained on {\\it unlabeled} 1M ImageNet images) outperforms SAM (trained on 11M images and 1 billion masks) by absolute 8\\% in mIoU on PartImageNet object segmentation.", "pdf": "https://openreview.net/pdf/9442c745c4c7ad1fb08e97f6a98ae46f4593aea4.pdf"} {"title": "Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models", "url": "https://openreview.net/forum?id=plmBsXHxgR", "detail_url": "https://openreview.net/forum?id=plmBsXHxgR", "authors": "Erfan Shayegani,Yue Dong,Nael Abu-Ghazaleh", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce new jailbreak attacks on vision language models (VLMs), which use aligned LLMs and are resilient to text-only jailbreak attacks. Specifically, we develop cross-modality attacks on alignment where we pair adversarial images going through the vision encoder with textual prompts to break the alignment of the language model. Our attacks employ a novel compositional strategy that combines an image, adversarially targeted towards toxic embeddings, with generic prompts to accomplish the jailbreak. Thus, the LLM draws the context to answer the generic prompt from the adversarial image. The generation of benign-appearing adversarial images leverages a novel embedding-space-based methodology, operating with no access to the LLM model. Instead, the attacks require access only to the vision encoder and utilize one of our four embedding space targeting strategies. By not requiring access to the LLM, the attacks lower the entry barrier for attackers, particularly when vision encoders such as CLIP are embedded in closed-source LLMs. The attacks achieve a high success rate across different VLMs, highlighting the risk of cross-modality alignment vulnerabilities, and the need for new alignment approaches for multi-modal models.", "pdf": "https://openreview.net/pdf/73245653c0cb13379877051e65dbc93ef4aa85cd.pdf"} {"title": "DreamFlow: High-quality text-to-3D generation by Approximating Probability Flow", "url": "https://openreview.net/forum?id=GURqUuTebY", "detail_url": "https://openreview.net/forum?id=GURqUuTebY", "authors": "Kyungmin Lee,Kihyuk Sohn,Jinwoo Shin", "tags": "ICLR 2024,Spotlight", "abstract": "Recent progress in text-to-3D generation has been achieved through the utilization of score distillation methods: they make use of the pre-trained text-to-image (T2I) diffusion models by distilling via the diffusion model training objective. However, such an approach inevitably results in the use of random timesteps at each update, which increases the variance of the gradient and ultimately prolongs the optimization process. In this paper, we propose to enhance the text-to-3D optimization by leveraging the T2I diffusion prior in the generative sampling process with a predetermined timestep schedule. To this end, we interpret text-to-3D optimization as a multi-view image-to-image translation problem, and propose a solution by approximating the probability flow. By leveraging the proposed novel optimization algorithm, we design DreamFlow, a practical three-stage coarse-to-fine text-to-3D optimization framework that enables fast generation of high-quality and high-resolution (i.e., 1024\u00d71024) 3D contents. For example, we demonstrate that DreamFlow is 5 times faster than the existing state-of-the-art text-to-3D method, while producing more photorealistic 3D contents.", "pdf": "https://openreview.net/pdf/65b059da26adcac6aded2561eda017af0181ec6d.pdf"} {"title": "Stack Attention: Improving the Ability of Transformers to Model Hierarchical Patterns", "url": "https://openreview.net/forum?id=XVhm3X8Fum", "detail_url": "https://openreview.net/forum?id=XVhm3X8Fum", "authors": "Brian DuSell,David Chiang", "tags": "ICLR 2024,Spotlight", "abstract": "Attention, specifically scaled dot-product attention, has proven effective for natural language, but it does not have a mechanism for handling hierarchical patterns of arbitrary nesting depth, which limits its ability to recognize certain syntactic structures. To address this shortcoming, we propose stack attention: an attention operator that incorporates stacks, inspired by their theoretical connections to context-free languages (CFLs). We show that stack attention is analogous to standard attention, but with a latent model of syntax that requires no syntactic supervision. We propose two variants: one related to deterministic pushdown automata (PDAs) and one based on nondeterministic PDAs, which allows transformers to recognize arbitrary CFLs. We show that transformers with stack attention are very effective at learning CFLs that standard transformers struggle on, achieving strong results on a CFL with theoretically maximal parsing difficulty. We also show that stack attention is more effective at natural language modeling under a constrained parameter budget, and we include results on machine translation.", "pdf": "https://openreview.net/pdf/1725d2a5bfd546b5acdabab5eb5d281caf92d1e4.pdf"} {"title": "SOTOPIA: Interactive Evaluation for Social Intelligence in Language Agents", "url": "https://openreview.net/forum?id=mM7VurbA4r", "detail_url": "https://openreview.net/forum?id=mM7VurbA4r", "authors": "Xuhui Zhou,Hao Zhu,Leena Mathur,Ruohong Zhang,Haofei Yu,Zhengyang Qi,Louis-Philippe Morency,Yonatan Bisk,Daniel Fried,Graham Neubig,Maarten Sap", "tags": "ICLR 2024,Spotlight", "abstract": "*Humans are social beings*; we pursue social goals in our daily interactions, which is a crucial aspect of social intelligence. Yet, AI systems' abilities in this realm remain elusive. We present SOTOPIA, an open-ended environment to simulate complex social interactions between artificial agents and evaluate their social intelligence. In our environment, agents role-play and *interact* under a wide variety of scenarios; they coordinate, collaborate, exchange, and compete with each other to achieve complex social goals. We simulate the role-play interaction between LLM-based agents and humans within this task space and evaluate their performance with a holistic evaluation framework called SOTOPIA-Eval. With SOTOPIA, we find significant differences between these models in terms of their social intelligence, and we identify a subset of SOTOPIA scenarios, SOTOPIA-hard, that is generally challenging for all models. We find that on this subset, GPT-4 achieves a significantly lower goal completion rate than humans and struggles to exhibit social commonsense reasoning and strategic communication skills. These findings demonstrate SOTOPIA's promise as a general platform for research on evaluating and improving social intelligence in artificial agents.", "pdf": "https://openreview.net/pdf/6aece1f9088fc415196df5830f1ca62e6dbf37d3.pdf"} {"title": "Privileged Sensing Scaffolds Reinforcement Learning", "url": "https://openreview.net/forum?id=EpVe8jAjdx", "detail_url": "https://openreview.net/forum?id=EpVe8jAjdx", "authors": "Edward S. Hu,James Springer,Oleh Rybkin,Dinesh Jayaraman", "tags": "ICLR 2024,Spotlight", "abstract": "We need to look at our shoelaces as we first learn to tie them but having mastered this skill, can do it from touch alone. We call this phenomenon \u201csensory scaffolding\u201d: observation streams that are not needed by a master might yet aid a novice learner. We consider such sensory scaffolding setups for training artificial agents. For example, a robot arm may need to be deployed with just a low-cost, robust, general-purpose camera; yet its performance may improve by having privileged training-time-only access to informative albeit expensive and unwieldy motion capture rigs or fragile tactile sensors. For these settings, we propose \u201cScaffolder\u201d, a reinforcement learning approach which effectively exploits privileged sensing in critics, world models, reward estimators, and other such auxiliary components that are only used at training time, to improve the target policy. For evaluating sensory scaffolding agents, we design a new \u201cS3\u201d suite of ten diverse simulated robotic tasks that explore a wide range of practical sensor setups. Agents must use privileged camera sensing to train blind hurdlers, privileged active visual perception to help robot arms overcome visual occlusions, privileged touch sensors to train robot hands, and more. Scaffolder easily outperforms relevant prior baselines and frequently performs comparably even to policies that have test-time access to the privileged sensors. Website: https://penn-pal-lab.github.io/scaffolder/", "pdf": "https://openreview.net/pdf/8ab7ec56b6e56ac45ce2b68e95b277283dbb8377.pdf"} {"title": "Learning to Act without Actions", "url": "https://openreview.net/forum?id=rvUq3cxpDF", "detail_url": "https://openreview.net/forum?id=rvUq3cxpDF", "authors": "Dominik Schmidt,Minqi Jiang", "tags": "ICLR 2024,Spotlight", "abstract": "Pre-training large models on vast amounts of web data has proven to be an effective approach for obtaining powerful, general models in domains such as language and vision. However, this paradigm has not yet taken hold in reinforcement learning. This is because videos, the most abundant form of embodied behavioral data on the web, lack the action labels required by existing methods for imitating behavior from demonstrations. We introduce **Latent Action Policies** (LAPO), a method for recovering latent action information\u2014and thereby latent-action policies, world models, and inverse dynamics models\u2014purely from videos. LAPO is the first method able to recover the structure of the true action space just from observed dynamics, even in challenging procedurally-generated environments. LAPO enables training latent-action policies that can be rapidly fine-tuned into expert-level policies, either offline using a small action-labeled dataset, or online with rewards. LAPO takes a first step towards pre-training powerful, generalist policies and world models on the vast amounts of videos readily available on the web. Our code is available here: \n[https://github.com/schmidtdominik/LAPO](https://github.com/schmidtdominik/LAPO).", "pdf": "https://openreview.net/pdf/3828f51b4c06dfddf96ff09f99344482461b30d4.pdf"} {"title": "Error Norm Truncation: Robust Training in the Presence of Data Noise for Text Generation Models", "url": "https://openreview.net/forum?id=zMvMwNvs4R", "detail_url": "https://openreview.net/forum?id=zMvMwNvs4R", "authors": "Tianjian Li,Haoran Xu,Philipp Koehn,Daniel Khashabi,Kenton Murray", "tags": "ICLR 2024,Spotlight", "abstract": "Text generation models are notoriously vulnerable to errors in the training data. With the wide-spread availability of massive amounts of web-crawled data becoming more commonplace, how can we enhance the robustness of models trained on a massive amount of noisy web-crawled text? In our work, we propose Error Norm Truncation (ENT), a robust enhancement method to the standard training objective that truncates noisy data. Compared to methods that only uses the negative log-likelihood loss to estimate data quality, our method provides a more accurate estimation by considering the distribution of non-target tokens, which is often overlooked by previous work. Through comprehensive experiments across language modeling, machine translation, and text summarization, we show that equipping text generation models with ENT improves generation quality over standard training and previous soft and hard truncation methods. Furthermore, we show that our method improves the robustness of models against two of the most detrimental types of noise in machine translation, resulting in an increase of more than 2 BLEU points over the MLE baseline when up to 50\\% of noise is added to the data.", "pdf": "https://openreview.net/pdf/1f26c396130ee811d490f099f908c0b8c99a3382.pdf"} {"title": "Massively Scalable Inverse Reinforcement Learning in Google Maps", "url": "https://openreview.net/forum?id=z3L59iGALM", "detail_url": "https://openreview.net/forum?id=z3L59iGALM", "authors": "Matt Barnes,Matthew Abueg,Oliver F. Lange,Matt Deeds,Jason Trader,Denali Molitor,Markus Wulfmeier,Shawn O'Banion", "tags": "ICLR 2024,Spotlight", "abstract": "Inverse reinforcement learning (IRL) offers a powerful and general framework for learning humans' latent preferences in route recommendation, yet no approach has successfully addressed planetary-scale problems with hundreds of millions of states and demonstration trajectories. In this paper, we introduce scaling techniques based on graph compression, spatial parallelization, and improved initialization conditions inspired by a connection to eigenvector algorithms. We revisit classic IRL methods in the routing context, and make the key observation that there exists a trade-off between the use of cheap, deterministic planners and expensive yet robust stochastic policies. This insight is leveraged in Receding Horizon Inverse Planning (RHIP), a new generalization of classic IRL algorithms that provides fine-grained control over performance trade-offs via its planning horizon. Our contributions culminate in a policy that achieves a 16-24% improvement in route quality at a global scale, and to the best of our knowledge, represents the largest published study of IRL algorithms in a real-world setting to date. We conclude by conducting an ablation study of key components, presenting negative results from alternative eigenvalue solvers, and identifying opportunities to further improve scalability via IRL-specific batching strategies.", "pdf": "https://openreview.net/pdf/4d44d3413d097944d3e6328c02e01436c56a3fac.pdf"} {"title": "Thin-Shell Object Manipulations With Differentiable Physics Simulations", "url": "https://openreview.net/forum?id=KsUh8MMFKQ", "detail_url": "https://openreview.net/forum?id=KsUh8MMFKQ", "authors": "Yian Wang,Juntian Zheng,Zhehuan Chen,Zhou Xian,Gu Zhang,Chao Liu,Chuang Gan", "tags": "ICLR 2024,Spotlight", "abstract": "In this work, we aim to teach robots to manipulate various thin-shell materials. \nPrior works studying thin-shell object manipulation mostly rely on heuristic policies or learn policies from real-world video demonstrations, and only focus on limited material types and tasks (e.g., cloth unfolding). However, these approaches face significant challenges when extended to a wider variety of thin-shell materials and a diverse range of tasks.\nOn the other hand, while virtual simulations are shown to be effective in diverse robot skill learning and evaluation, prior thin-shell simulation environments only support a subset of thin-shell materials, which also limits their supported range of tasks. \nTo fill in this gap, we introduce ThinShellLab - a fully differentiable simulation platform tailored for robotic interactions with diverse thin-shell materials possessing varying material properties, enabling flexible thin-shell manipulation skill learning and evaluation. Building on top of our developed simulation engine, we design a diverse set of manipulation tasks centered around different thin-shell objects. Our experiments suggest that manipulating thin-shell objects presents several unique challenges: 1) thin-shell manipulation relies heavily on frictional forces due to the objects' co-dimensional nature, 2) the materials being manipulated are highly sensitive to minimal variations in interaction actions, and 3) the constant and frequent alteration in contact pairs makes trajectory optimization methods susceptible to local optima, and neither standard reinforcement learning algorithms nor trajectory optimization methods (either gradient-based or gradient-free) are able to solve the tasks alone. To overcome these challenges, we present an optimization scheme that couples sampling-based trajectory optimization and gradient-based optimization, boosting both learning efficiency and converged performance across various proposed tasks. In addition, the differentiable nature of our platform facilitates a smooth sim-to-real transition. By tuning simulation parameters with a minimal set of real-world data, we demonstrate successful deployment of the learned skills to real-robot settings. ThinShellLab will be publicly available. Video demonstration and more information can be found on the project website https://vis-www.cs.umass.edu/ThinShellLab/.", "pdf": "https://openreview.net/pdf/edbd3be2c1ca4369cdf41d2d892af284a0c21cc3.pdf"} {"title": "Can Sensitive Information Be Deleted From LLMs? Objectives for Defending Against Extraction Attacks", "url": "https://openreview.net/forum?id=7erlRDoaV8", "detail_url": "https://openreview.net/forum?id=7erlRDoaV8", "authors": "Vaidehi Patil,Peter Hase,Mohit Bansal", "tags": "ICLR 2024,Spotlight", "abstract": "Pretrained language models sometimes possess knowledge that we do not wish them to, including memorized personal information and knowledge that could be used to harm people. They can also output toxic or harmful text. To mitigate these safety and informational issues, we propose an attack-and-defense framework for studying the task of deleting sensitive information directly from model weights. We study direct edits to model weights because (1) this approach should guarantee that particular deleted information is never extracted by future prompt attacks, and (2) it should protect against whitebox attacks, which is necessary for making claims about safety/privacy in a setting where publicly available model weights could be used to elicit sensitive information. Our threat model assumes that an attack succeeds if the answer to a sensitive question is located among a set of B generated candidates, based on scenarios where the information would be insecure if the answer is among B candidates. Experimentally, we show that even state-of-the-art model editing methods such as ROME struggle to truly delete factual information from models like GPT-J, as our whitebox and blackbox attacks can recover \u201cdeleted\u201d information from an edited model 38% of the time. These attacks leverage two key observations: (1) that traces of deleted information can be found in intermediate model hidden states, and (2) that applying an editing method for one question may not delete information across rephrased versions of the question. Finally, we provide new defense methods that protect against some extraction attacks, but we do not find a single universally effective defense method. Our results suggest that truly deleting sensitive information is a tractable but difficult problem, since even relatively low attack success rates have potentially severe implications for the deployment of language models in a world where individuals enjoy ownership of their personal data, a right to privacy, and safety from harmful model outputs.", "pdf": "https://openreview.net/pdf/58461c89ea8f999b82924f883aa34b627914dc42.pdf"} {"title": "Learning to Reject Meets Long-tail Learning", "url": "https://openreview.net/forum?id=ta26LtNq2r", "detail_url": "https://openreview.net/forum?id=ta26LtNq2r", "authors": "Harikrishna Narasimhan,Aditya Krishna Menon,Wittawat Jitkrittum,Neha Gupta,Sanjiv Kumar", "tags": "ICLR 2024,Spotlight", "abstract": "Learning to reject (L2R) is a classical problem where one seeks a classifier capable of abstaining on low-confidence samples. Most prior work on L2R has focused on minimizing the standard misclassification error. However, in many real-world applications, the label distribution is highly imbalanced, necessitating alternate evaluation metrics such as the balanced error or the worst-group error that enforce equitable performance across both the head and tail classes. In this paper, we establish that traditional L2R methods can be grossly sub-optimal for such metrics, and show that this is due to an intricate dependence in the objective between the label costs and the rejector. We then derive the form of the Bayes-optimal classifier and rejector for the balanced error, propose a novel plug-in approach to mimic this solution, and extend our results to general evaluation metrics. Through experiments on benchmark image classification tasks, we show that our approach yields better trade-offs in both the balanced and worst-group error compared to L2R baselines.", "pdf": "https://openreview.net/pdf/2d19da352871db6a1e7954f4651e9b44594126b8.pdf"} {"title": "On the Foundations of Shortcut Learning", "url": "https://openreview.net/forum?id=Tj3xLVuE9f", "detail_url": "https://openreview.net/forum?id=Tj3xLVuE9f", "authors": "Katherine Hermann,Hossein Mobahi,Thomas FEL,Michael Curtis Mozer", "tags": "ICLR 2024,Spotlight", "abstract": "Deep-learning models can extract a rich assortment of features from data. Which features a model uses depends not only on *predictivity*---how reliably a feature indicates training-set labels---but also on *availability*---how easily the feature can be extracted from inputs. The literature on shortcut learning has noted examples in which models privilege one feature over another, for example texture over shape and image backgrounds over foreground objects. Here, we test hypotheses about which input properties are more available to a model, and systematically study how predictivity and availability interact to shape models' feature use. We construct a minimal, explicit generative framework for synthesizing classification datasets with two latent features that vary in predictivity and in factors we hypothesize to relate to availability, and we quantify a model's shortcut bias---its over-reliance on the shortcut (more available, less predictive) feature at the expense of the core (less available, more predictive) feature. We find that linear models are relatively unbiased, but introducing a single hidden layer with ReLU or Tanh units yields a bias. Our empirical findings are consistent with a theoretical account based on Neural Tangent Kernels. Finally, we study how models used in practice trade off predictivity and availability in naturalistic datasets, discovering availability manipulations which increase models' degree of shortcut bias. Taken together, these findings suggest that the propensity to learn shortcut features is a fundamental characteristic of deep nonlinear architectures warranting systematic study given its role in shaping how models solve tasks.", "pdf": "https://openreview.net/pdf/3f47b29f0e35691e7047d9fbfa0e4c47ea966e49.pdf"} {"title": "Synaptic Weight Distributions Depend on the Geometry of Plasticity", "url": "https://openreview.net/forum?id=x5txICnnjC", "detail_url": "https://openreview.net/forum?id=x5txICnnjC", "authors": "Roman Pogodin,Jonathan Cornford,Arna Ghosh,Gauthier Gidel,Guillaume Lajoie,Blake Aaron Richards", "tags": "ICLR 2024,Spotlight", "abstract": "A growing literature in computational neuroscience leverages gradient descent and learning algorithms that approximate it to study synaptic plasticity in the brain. However, the vast majority of this work ignores a critical underlying assumption: the choice of distance for synaptic changes - i.e. the geometry of synaptic plasticity. Gradient descent assumes that the distance is Euclidean, but many other distances are possible, and there is no reason that biology necessarily uses Euclidean geometry. Here, using the theoretical tools provided by mirror descent, we show that the distribution of synaptic weights will depend on the geometry of synaptic plasticity. We use these results to show that experimentally-observed log-normal weight distributions found in several brain areas are not consistent with standard gradient descent (i.e. a Euclidean geometry), but rather with non-Euclidean distances. Finally, we show that it should be possible to experimentally test for different synaptic geometries by comparing synaptic weight distributions before and after learning. Overall, our work shows that the current paradigm in theoretical work on synaptic plasticity that assumes Euclidean synaptic geometry may be misguided and that it should be possible to experimentally determine the true geometry of synaptic plasticity in the brain.", "pdf": "https://openreview.net/pdf/bfba522bee31428aed87f14e95fb1c48baff7080.pdf"} {"title": "Graph Metanetworks for Processing Diverse Neural Architectures", "url": "https://openreview.net/forum?id=ijK5hyxs0n", "detail_url": "https://openreview.net/forum?id=ijK5hyxs0n", "authors": "Derek Lim,Haggai Maron,Marc T. Law,Jonathan Lorraine,James Lucas", "tags": "ICLR 2024,Spotlight", "abstract": "Neural networks efficiently encode learned information within their parameters. Consequently, many tasks can be unified by treating neural networks themselves as input data. When doing so, recent studies demonstrated the importance of accounting for the symmetries and geometry of parameter spaces. However, those works developed architectures tailored to specific networks such as MLPs and CNNs without normalization layers, and generalizing such architectures to other types of networks can be challenging. In this work, we overcome these challenges by building new metanetworks --- neural networks that take weights from other neural networks as input. Put simply, we carefully build graphs representing the input neural networks and process the graphs using graph neural networks. Our approach, Graph Metanetworks (GMNs), generalizes to neural architectures where competing methods struggle, such as multi-head attention layers, normalization layers, convolutional layers, ResNet blocks, and group-equivariant linear layers. We prove that GMNs are expressive and equivariant to parameter permutation symmetries that leave the input neural network functions unchanged. We validate the effectiveness of our method on several metanetwork tasks over diverse neural network architectures.", "pdf": "https://openreview.net/pdf/461000be3cc4ed15947307020c46aafa982c40ae.pdf"} {"title": "Dropout Enhanced Bilevel Training", "url": "https://openreview.net/forum?id=06lrITXVAx", "detail_url": "https://openreview.net/forum?id=06lrITXVAx", "authors": "Peiran Yu,Junyi Li,Heng Huang", "tags": "ICLR 2024,Spotlight", "abstract": "Bilevel optimization problems appear in many widely used machine learning tasks. Bilevel optimization models are sensitive to small changes, and bilevel training tasks typically involve limited datasets. Therefore, overfitting is a common challenge in bilevel training tasks. This paper considers the use of dropout to address this problem. We propose a bilevel optimization model that depends on the distribution of dropout masks. We investigate how the dropout rate affects the hypergradient of this model. We propose a dropout bilevel method to solve the dropout bilevel optimization model. Subsequently, we analyze the resulting dropout bilevel method from an optimization perspective. Analyzing the optimization properties of methods with dropout is essential because it provides convergence guarantees for methods using dropout. However, there has been limited investigation in this research direction. We provide the complexity of the resulting dropout bilevel method in terms of reaching an $\\epsilon$ stationary point of the proposed stochastic bilevel model. Empirically, we demonstrate that overfitting occurs in data cleaning problems, and the method proposed in this work mitigates this issue.", "pdf": "https://openreview.net/pdf/09304d5bf3e31448450004ee461830870db26085.pdf"} {"title": "Privacy Amplification for Matrix Mechanisms", "url": "https://openreview.net/forum?id=xUzWmFdglP", "detail_url": "https://openreview.net/forum?id=xUzWmFdglP", "authors": "Christopher A. Choquette-Choo,Arun Ganesh,Thomas Steinke,Abhradeep Guha Thakurta", "tags": "ICLR 2024,Spotlight", "abstract": "Privacy amplification exploits randomness in data selection to provide tighter differential privacy (DP) guarantees. This analysis is key to DP-SGD's success in machine learning (ML), but, is not readily applicable to the newer state-of-the-art (SOTA) algorithms. This is because these algorithms, known as DP-FTRL, use the matrix mechanism to add correlated noise instead of independent noise as in DP-SGD.\n\nIn this paper, we propose \"MMCC'' (matrix mechanism conditional composition), the first algorithm to analyze privacy amplification via sampling for any generic matrix mechanism. MMCC is nearly tight in that it approaches a lower bound as $\\epsilon\\to0$. \nTo analyze correlated outputs in MMCC, we prove that they can be analyzed as if they were independent, by conditioning them on prior outputs. Our \"conditional composition theorem'' has broad utility: we use it to show that the noise added to binary-tree-DP-FTRL can asymptotically match the noise added to DP-SGD with amplification. Our algorithm also has practical empirical utility. We show that amplification leads to significant improvement in the privacy/utility trade-offs for DP-FTRL style algorithms for standard benchmark tasks.", "pdf": "https://openreview.net/pdf/1f4107c113add2f5ec1249dac4d4dfa313bfb5c5.pdf"} {"title": "Bandits Meet Mechanism Design to Combat Clickbait in Online Recommendation", "url": "https://openreview.net/forum?id=lsxeNvYqCj", "detail_url": "https://openreview.net/forum?id=lsxeNvYqCj", "authors": "Thomas Kleine Buening,Aadirupa Saha,Christos Dimitrakakis,Haifeng Xu", "tags": "ICLR 2024,Spotlight", "abstract": "We study a strategic variant of the multi-armed bandit problem, which we coin the strategic click-bandit. This model is motivated by applications in online recommendation where the choice of recommended items depends on both the click-through rates and the post-click rewards. Like in classical bandits, rewards follow a fixed unknown distribution. However, we assume that the click-rate of each arm is chosen strategically by the arm (e.g., a host on Airbnb) in order to maximize the number of times it gets clicked. The algorithm designer does not know the post-click rewards nor the arms' actions (i.e., strategically chosen click-rates) in advance, and must learn both values over time. To solve this problem, we design an incentive-aware learning algorithm, UCB-S, which achieves two goals simultaneously: (a) incentivizing desirable arm behavior under uncertainty; (b) minimizing regret by learning unknown parameters. We approximately characterize all Nash equilibria of the arms under UCB-S and show a $\\tilde{\\mathcal{O}} (\\sqrt{KT})$ regret bound uniformly in every equilibrium. We also show that incentive-unaware algorithms generally fail to achieve low regret in the strategic click-bandit. Finally, we support our theoretical results by simulations of strategic arm behavior which confirm the effectiveness and robustness of our proposed incentive design.", "pdf": "https://openreview.net/pdf/8dcfb207122a45da36a445b1322405bc5aafef25.pdf"} {"title": "Towards Principled Representation Learning from Videos for Reinforcement Learning", "url": "https://openreview.net/forum?id=3mnWvUZIXt", "detail_url": "https://openreview.net/forum?id=3mnWvUZIXt", "authors": "Dipendra Misra,Akanksha Saran,Tengyang Xie,Alex Lamb,John Langford", "tags": "ICLR 2024,Spotlight", "abstract": "We study pre-training representations for decision-making using video data, which is abundantly available for tasks such as game agents and software testing. Even though significant empirical advances have been made on this problem, a theoretical understanding remains absent. We initiate the theoretical investigation into principled approaches for representation learning and focus on learning the latent state representations of the underlying MDP using video data. We study two types of settings: one where there is iid noise in the observation, and a more challenging setting where there is also the presence of exogenous noise, which is non-iid noise that is temporally correlated, such as the motion of people or cars in the background. We study three commonly used approaches: autoencoding, temporal contrastive learning, and forward modeling. We prove upper bounds for temporal contrastive learning and forward modeling in the presence of only iid noise. We show that these approaches can learn the latent state and use it to do efficient downstream RL with polynomial sample complexity. When exogenous noise is also present, we establish a lower bound result showing that the sample complexity of learning from video data can be exponentially worse than learning from action-labeled trajectory data. This partially explains why reinforcement learning with video pre-training is hard. We evaluate these representational learning methods in two visual domains, yielding results that are consistent with our theoretical findings.", "pdf": "https://openreview.net/pdf/030d9ee1692e81df35ac143b32eb5beb1c384730.pdf"} {"title": "Optimal Sample Complexity of Contrastive Learning", "url": "https://openreview.net/forum?id=NU9AYHJvYe", "detail_url": "https://openreview.net/forum?id=NU9AYHJvYe", "authors": "Noga Alon,Dmitrii Avdiukhin,Dor Elboim,Orr Fischer,Grigory Yaroslavtsev", "tags": "ICLR 2024,Spotlight", "abstract": "Contrastive learning is a highly successful technique for learning representations of data from labeled tuples, specifying the distance relations within the tuple. We study the sample complexity of contrastive learning, i.e. the minimum number of labeled tuples sufficient for getting high generalization accuracy. We give tight bounds on the sample complexity in a variety of settings, focusing on arbitrary distance functions, $\\ell_p$-distances, and tree metrics. Our main result is an (almost) optimal bound on the sample complexity of learning $\\ell_p$-distances for integer $p$. For any $p \\ge 1$, we show that $\\tilde \\Theta(nd)$ labeled tuples are necessary and sufficient for learning $d$-dimensional representations of $n$-point datasets. Our results hold for an arbitrary distribution of the input samples and are based on giving the corresponding bounds on the Vapnik-Chervonenkis/Natarajan dimension of the associated problems. We further show that the theoretical bounds on sample complexity obtained via VC/Natarajan dimension can have strong predictive power for experimental results, in contrast with the folklore belief about a substantial gap between the statistical learning theory and the practice of deep learning.", "pdf": "https://openreview.net/pdf/00a6d5623a59e634ef01b7ebdd71bd1b30b22500.pdf"} {"title": "Post-hoc bias scoring is optimal for fair classification", "url": "https://openreview.net/forum?id=FM5xfcaR2Y", "detail_url": "https://openreview.net/forum?id=FM5xfcaR2Y", "authors": "Wenlong Chen,Yegor Klochkov,Yang Liu", "tags": "ICLR 2024,Spotlight", "abstract": "We consider a binary classification problem under group fairness constraints, which can be one of Demographic Parity (DP), Equalized Opportunity (EOp), or Equalized Odds (EO). We propose an explicit characterization of Bayes optimal classifier under the fairness constraints, which turns out to be a simple modification rule of the unconstrained classifier. Namely, we introduce a novel instance-level measure of bias, which we call bias score, and the modification rule is a simple linear rule on top of the finite amount of bias scores. Based on this characterization, we develop a post-hoc approach that allows us to adapt to fairness constraints while maintaining high accuracy. In the case of DP and EOp constraints, the modification rule is thresholding a single bias score, while in the case of EO constraints we are required to fit a linear modification rule with 2 parameters. The method can also be applied for composite group-fairness criteria, such as ones involving several sensitive attributes. We achieve competitive or better performance compared to both in-processing and post-processing methods across three datasets: Adult, COMPAS, and CelebA. Unlike most post-processing methods, we do not require access to sensitive attributes during the inference time.", "pdf": "https://openreview.net/pdf/76bd92c39da18e17fb34a06d39a4c5bb7e7573ba.pdf"} {"title": "Sharpness-Aware Data Poisoning Attack", "url": "https://openreview.net/forum?id=bxITGFPVWh", "detail_url": "https://openreview.net/forum?id=bxITGFPVWh", "authors": "Pengfei He,Han Xu,Jie Ren,Yingqian Cui,Shenglai Zeng,Hui Liu,Charu C. Aggarwal,Jiliang Tang", "tags": "ICLR 2024,Spotlight", "abstract": "Recent research has highlighted the vulnerability of Deep Neural Networks (DNNs) against data poisoning attacks. These attacks aim to inject poisoning samples into the models' training dataset such that the trained models have inference failures. While previous studies have executed different types of attacks, one major challenge that greatly limits their effectiveness is the \nuncertainty of the re-training process after the injection of poisoning samples. It includes the uncertainty of training initialization, algorithm and model architecture. To address this challenge, we propose a new strategy called **Sharpness-Aware Data Poisoning Attack (SAPA)**. In particular, it leverages the concept of DNNs' loss landscape sharpness to optimize the poisoning effect on the (approximately) worst re-trained model. Extensive experiments demonstrate that SAPA offers a general and principled strategy that significantly enhances various types of poisoning attacks against various types of re-training uncertainty.", "pdf": "https://openreview.net/pdf/c5b9990eda79f25ade2e1cff4e9d53d63490c7fc.pdf"} {"title": "Pre-training with Random Orthogonal Projection Image Modeling", "url": "https://openreview.net/forum?id=z4Hcegjzph", "detail_url": "https://openreview.net/forum?id=z4Hcegjzph", "authors": "Maryam Haghighat,Peyman Moghadam,Shaheer Mohamed,Piotr Koniusz", "tags": "ICLR 2024,Spotlight", "abstract": "Masked Image Modeling (MIM) is a powerful self-supervised strategy for visual pre-training without the use of labels. MIM applies random crops to input images, processes them with an encoder, and then recovers the masked inputs with a decoder, which encourages the network to capture and learn structural information about objects and scenes. The intermediate feature representations obtained from MIM are suitable for fine-tuning on downstream tasks. In this paper, we propose an Image Modeling framework based on random orthogonal projection instead of binary masking as in MIM. Our proposed Random Orthogonal Projection Image Modeling (ROPIM) reduces spatially-wise token information under guaranteed bound on the noise variance and can be considered as masking entire spatial image area under locally varying masking degrees. Since ROPIM uses a random subspace for the projection that realizes the masking step, the readily available complement of the subspace can be used during unmasking to promote recovery of removed information. In this paper, we show that using random orthogonal projection leads to superior performance compared to crop-based masking. We demonstrate state-of-the-art results on several popular benchmarks.", "pdf": "https://openreview.net/pdf/10fbc610db66b05362ea4434c03c92835ba7f338.pdf"} {"title": "Lagrangian Flow Networks for Conservation Laws", "url": "https://openreview.net/forum?id=Nshk5YpdWE", "detail_url": "https://openreview.net/forum?id=Nshk5YpdWE", "authors": "Fabricio Arend Torres,Marcello Massimo Negri,Marco Inversi,Jonathan Aellen,Volker Roth", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce Lagrangian Flow Networks (LFlows) for modeling fluid densities and velocities continuously in space and time.\nBy construction, the proposed LFlows satisfy the continuity equation,\na PDE describing mass conservation in its differential form. \nOur model is based on the insight that solutions to the continuity equation can be expressed as\ntime-dependent density transformations via differentiable and invertible maps.\nThis follows from classical theory of the existence and uniqueness of Lagrangian flows for smooth vector fields.\nHence, we model fluid densities by transforming a base density with parameterized diffeomorphisms conditioned on time.\nThe key benefit compared to methods relying on numerical ODE solvers or PINNs is that the analytic expression of the velocity is always consistent with changes in density.\nFurthermore, we require neither expensive numerical solvers, nor additional penalties to enforce the PDE.\nLFlows show higher predictive accuracy in density modeling tasks compared to competing models in 2D and 3D,\nwhile being computationally efficient.\nAs a real-world application, we model bird migration based on sparse weather radar measurements.", "pdf": "https://openreview.net/pdf/53607a671ffb6900fcf4870e6bd3c5866146e9b4.pdf"} {"title": "Linearity of Relation Decoding in Transformer Language Models", "url": "https://openreview.net/forum?id=w7LU2s14kE", "detail_url": "https://openreview.net/forum?id=w7LU2s14kE", "authors": "Evan Hernandez,Arnab Sen Sharma,Tal Haklay,Kevin Meng,Martin Wattenberg,Jacob Andreas,Yonatan Belinkov,David Bau", "tags": "ICLR 2024,Spotlight", "abstract": "Much of the knowledge encoded in transformer language models (LMs) may be expressed in terms of relations: relations between words and their synonyms, entities and their attributes, etc. We show that, for a subset of relations, this computation is well-approximated by a single linear transformation on the subject representation. Linear relation representations may be obtained by constructing a first-order approximation to the LM from a single prompt, and they exist for a variety of factual, commonsense, and linguistic relations. However, we also identify many cases in which LM predictions capture relational knowledge accurately, but this knowledge is not linearly encoded in their representations. Our results thus reveal a simple, interpretable, but heterogeneously deployed knowledge representation strategy in transformer LMs.", "pdf": "https://openreview.net/pdf/548eac40ba455c0509185e199cc8f49f2f96523c.pdf"} {"title": "Subtractive Mixture Models via Squaring: Representation and Learning", "url": "https://openreview.net/forum?id=xIHi5nxu9P", "detail_url": "https://openreview.net/forum?id=xIHi5nxu9P", "authors": "Lorenzo Loconte,Aleksanteri Mikulus Sladek,Stefan Mengel,Martin Trapp,Arno Solin,Nicolas Gillis,Antonio Vergari", "tags": "ICLR 2024,Spotlight", "abstract": "Mixture models are traditionally represented and learned by adding several distributions as components. Allowing mixtures to subtract probability mass or density can drastically reduce the number of components needed to model complex distributions. However, learning such subtractive mixtures while ensuring they still encode a non-negative function is challenging. We investigate how to learn and perform inference on deep subtractive mixtures by squaring them. We do this in the framework of probabilistic circuits, which enable us to represent tensorized mixtures and generalize several other subtractive models. We theoretically prove that the class of squared circuits allowing subtractions can be exponentially more expressive than traditional additive mixtures; and, we empirically show this increased expressiveness on a series of real-world distribution estimation tasks.", "pdf": "https://openreview.net/pdf/3c3d0732ef99262e1fa4d9fc10ce89f35d07da28.pdf"} {"title": "On the Provable Advantage of Unsupervised Pretraining", "url": "https://openreview.net/forum?id=rmXXKxQpOR", "detail_url": "https://openreview.net/forum?id=rmXXKxQpOR", "authors": "Jiawei Ge,Shange Tang,Jianqing Fan,Chi Jin", "tags": "ICLR 2024,Spotlight", "abstract": "Unsupervised pretraining, which learns a useful representation using a large amount of unlabeled data to facilitate the learning of downstream tasks, is a critical component of modern large-scale machine learning systems. Despite its tremendous empirical success, the rigorous theoretical understanding of why unsupervised pretraining generally helps remains rather limited---most existing results are restricted to particular methods or approaches for unsupervised pretraining with specialized structural assumptions. This paper studies a generic framework,\nwhere the unsupervised representation learning task is specified by an abstract class of latent variable models $\\Phi$ and the downstream task is specified by a class of prediction functions $\\Psi$. We consider a natural approach of using Maximum Likelihood Estimation (MLE) for unsupervised pretraining and Empirical Risk Minimization (ERM) for learning downstream tasks. We prove that, under a mild ``informative'' condition, our algorithm achieves an excess risk of $\\\\tilde{\\\\mathcal{O}}(\\sqrt{\\mathcal{C}\\_\\Phi/m} + \\sqrt{\\mathcal{C}\\_\\Psi/n})$ for downstream tasks, where $\\mathcal{C}\\_\\Phi, \\mathcal{C}\\_\\Psi$ are complexity measures of function classes $\\Phi, \\Psi$, and $m, n$ are the number of unlabeled and labeled data respectively. Comparing to the baseline of $\\tilde{\\mathcal{O}}(\\sqrt{\\mathcal{C}\\_{\\Phi \\circ \\Psi}/n})$ achieved by performing supervised learning using only the labeled data, our result rigorously shows the benefit of unsupervised pretraining when $m \\gg n$ and $\\mathcal{C}\\_{\\Phi\\circ \\Psi} > \\mathcal{C}\\_\\Psi$. This paper further shows that our generic framework covers a wide range of approaches for unsupervised pretraining, including factor models, Gaussian mixture models, and contrastive learning.", "pdf": "https://openreview.net/pdf/0acbd62d42fa6cae47b05447238cec98212499f9.pdf"} {"title": "TorchRL: A data-driven decision-making library for PyTorch", "url": "https://openreview.net/forum?id=QxItoEAVMb", "detail_url": "https://openreview.net/forum?id=QxItoEAVMb", "authors": "Albert Bou,Matteo Bettini,Sebastian Dittert,Vikash Kumar,Shagun Sodhani,Xiaomeng Yang,Gianni De Fabritiis,Vincent Moens", "tags": "ICLR 2024,Spotlight", "abstract": "PyTorch has ascended as a premier machine learning framework, yet it lacks a native and comprehensive library for decision and control tasks suitable for large development teams dealing with complex real-world data and environments. To address this issue, we propose TorchRL, a generalistic control library for PyTorch that provides well-integrated, yet standalone components. We introduce a new and flexible PyTorch primitive, the TensorDict, which facilitates streamlined algorithm development across the many branches of Reinforcement Learning (RL) and control. We provide a detailed description of the building blocks and an extensive overview of the library across domains and tasks. Finally, we experimentally demonstrate its reliability and flexibility, and show comparative benchmarks to demonstrate its computational efficiency. TorchRL fosters long-term support and is publicly available on GitHub for greater reproducibility and collaboration within the research community. The code is open-sourced on GitHub.", "pdf": "https://openreview.net/pdf/74ac7ade66fbb7c4ade0e5391457625d3599377c.pdf"} {"title": "Towards Robust Offline Reinforcement Learning under Diverse Data Corruption", "url": "https://openreview.net/forum?id=5hAMmCU0bK", "detail_url": "https://openreview.net/forum?id=5hAMmCU0bK", "authors": "Rui Yang,Han Zhong,Jiawei Xu,Amy Zhang,Chongjie Zhang,Lei Han,Tong Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "Offline reinforcement learning (RL) presents a promising approach for learning reinforced policies from offline datasets without the need for costly or unsafe interactions with the environment. However, datasets collected by humans in real-world environments are often noisy and may even be maliciously corrupted, which can significantly degrade the performance of offline RL. In this work, we first investigate the performance of current offline RL algorithms under comprehensive data corruption, including states, actions, rewards, and dynamics. Our extensive experiments reveal that implicit Q-learning (IQL) demonstrates remarkable resilience to data corruption among various offline RL algorithms. Furthermore, we conduct both empirical and theoretical analyses to understand IQL's robust performance, identifying its supervised policy learning scheme as the key factor. Despite its relative robustness, IQL still suffers from heavy-tail targets of Q functions under dynamics corruption. To tackle this challenge, we draw inspiration from robust statistics to employ the Huber loss to handle the heavy-tailedness and utilize quantile estimators to balance penalization for corrupted data and learning stability. By incorporating these simple yet effective modifications into IQL, we propose a more robust offline RL approach named Robust IQL (RIQL). Extensive experiments demonstrate that RIQL exhibits highly robust performance when subjected to diverse data corruption scenarios.", "pdf": "https://openreview.net/pdf/60415c1f68bdc23484417721aa0069e22923b3b7.pdf"} {"title": "Variational Bayesian Last Layers", "url": "https://openreview.net/forum?id=Sx7BIiPzys", "detail_url": "https://openreview.net/forum?id=Sx7BIiPzys", "authors": "James Harrison,John Willes,Jasper Snoek", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce a deterministic variational formulation for training Bayesian last layer neural networks. This yields a sampling-free, single-pass model and loss that effectively improves uncertainty estimation. Our variational Bayesian last layer (VBLL) can be trained and evaluated with only quadratic complexity in last layer width, and is thus (nearly) computationally free to add to standard architectures. We experimentally investigate VBLLs, and show that they improve predictive accuracy, calibration, and out of distribution detection over baselines across both regression and classification. Finally, we investigate combining VBLL layers with variational Bayesian feature learning, yielding a lower variance collapsed variational inference method for Bayesian neural networks.", "pdf": "https://openreview.net/pdf/60aaa131dccd0ce0a263698149f1661e9ffe3a5e.pdf"} {"title": "EQA-MX: Embodied Question Answering using Multimodal Expression", "url": "https://openreview.net/forum?id=7gUrYE50Rb", "detail_url": "https://openreview.net/forum?id=7gUrYE50Rb", "authors": "Md Mofijul Islam,Alexi Gladstone,Riashat Islam,Tariq Iqbal", "tags": "ICLR 2024,Spotlight", "abstract": "Humans predominantly use verbal utterances and nonverbal gestures (e.g., eye gaze and pointing gestures) in their natural interactions. For instance, pointing gestures and verbal information is often required to comprehend questions such as \"what object is that?\" Thus, this question-answering (QA) task involves complex reasoning of multimodal expressions (verbal utterances and nonverbal gestures). However, prior works have explored QA tasks in non-embodied settings, where questions solely contain verbal utterances from a single verbal and visual perspective. In this paper, we have introduced 8 novel embodied question answering (EQA) tasks to develop learning models to comprehend embodied questions with multimodal expressions. We have developed a novel large-scale dataset, EQA-MX, with over 8 million diverse embodied QA data samples involving multimodal expressions from multiple visual and verbal perspectives. To learn salient multimodal representations from discrete verbal embeddings and continuous wrapping of multiview visual representations, we propose a vector-quantization (VQ) based multimodal representation learning model, VQ-Fusion, for the EQA tasks. Our extensive experimental results suggest that VQ-Fusion can improve the performance of existing state-of-the-art visual-language models up to 13% across EQA tasks.", "pdf": "https://openreview.net/pdf/f2829b2f4bb32e56e54abc38d0fe382ae50c7361.pdf"} {"title": "Retrieval-based Disentangled Representation Learning with Natural Language Supervision", "url": "https://openreview.net/forum?id=ZlQRiFmq7Y", "detail_url": "https://openreview.net/forum?id=ZlQRiFmq7Y", "authors": "Jiawei Zhou,Xiaoguang Li,Lifeng Shang,Xin Jiang,Qun Liu,Lei Chen", "tags": "ICLR 2024,Spotlight", "abstract": "Disentangled representation learning remains challenging as the underlying factors of variation in the data do not naturally exist. The inherent complexity of real-world data makes it unfeasible to exhaustively enumerate and encapsulate all its variations within a finite set of factors. However, it is worth noting that most real-world data have linguistic equivalents, typically in the form of textual descriptions. These linguistic counterparts can represent the data and effortlessly decomposed into distinct tokens. In light of this, we present Vocabulary Disentangled Retrieval (VDR), a retrieval-based framework that harnesses natural language as proxies of the underlying data variation to drive disentangled representation learning. Our approach employ a bi-encoder model to represent both data and natural language in a vocabulary space, enabling the model to distinguish dimensions that capture intrinsic characteristics within data through its natural language counterpart, thus facilitating disentanglement. We extensively assess the performance of VDR across 15 retrieval benchmark datasets, covering text-to-text and cross-modal retrieval scenarios, as well as human evaluation. Our experimental results compellingly demonstrate the superiority of VDR over previous bi-encoder retrievers with comparable model size and training costs, achieving an impressive 8.7% improvement in NDCG@10 on the BEIR benchmark, a 5.3\\% increase on MS COCO, and a 6.0% increase on Flickr30k in terms of mean recall in the zero-shot setting. Moreover, The results from human evaluation indicate that interpretability of our method is on par with SOTA captioning models.", "pdf": "https://openreview.net/pdf/806a04ba3fc6094730d982164ed4de6b3cf4f351.pdf"} {"title": "On the Markov Property of Neural Algorithmic Reasoning: Analyses and Methods", "url": "https://openreview.net/forum?id=Kn7tWhuetn", "detail_url": "https://openreview.net/forum?id=Kn7tWhuetn", "authors": "Montgomery Bohde,Meng Liu,Alexandra Saxton,Shuiwang Ji", "tags": "ICLR 2024,Spotlight", "abstract": "Neural algorithmic reasoning is an emerging research direction that endows neural networks with the ability to mimic algorithmic executions step-by-step. A common paradigm in existing designs involves the use of historical embeddings in predicting the results of future execution steps. Our observation in this work is that such historical dependence intrinsically contradicts the Markov nature of algorithmic reasoning tasks. Based on this motivation, we present our ForgetNet, which does not use historical embeddings and thus is consistent with the Markov nature of the tasks. To address challenges in training ForgetNet at early stages, we further introduce G-ForgetNet, which uses a gating mechanism to allow for the selective integration of historical embeddings. Such an enhanced capability provides valuable computational pathways during the model's early training phase. Our extensive experiments, based on the CLRS-30 algorithmic reasoning benchmark, demonstrate that both ForgetNet and G-ForgetNet achieve better generalization capability than existing methods. Furthermore, we investigate the behavior of the gating mechanism, highlighting its degree of alignment with our intuitions and its effectiveness for robust performance. Our code is publicly available at https://github.com/divelab/ForgetNet.", "pdf": "https://openreview.net/pdf/46ea9907175ecd6c88621bba3b5478fb9390eea8.pdf"} {"title": "TRAM: Bridging Trust Regions and Sharpness Aware Minimization", "url": "https://openreview.net/forum?id=kxebDHZ7b7", "detail_url": "https://openreview.net/forum?id=kxebDHZ7b7", "authors": "Tom Sherborne,Naomi Saphra,Pradeep Dasigi,Hao Peng", "tags": "ICLR 2024,Spotlight", "abstract": "Sharpness-aware minimization (SAM) reports improving domain generalization by\nreducing the loss surface curvature in the parameter space. However,\ngeneralization during _fine-tuning_ is often more dependent on the\ntransferability of _representations_ in the function space. Trust-region\nmethods (TR) target this goal by regularizing representation curvature to reduce\ncatastrophic forgetting of pre-trained task-agnostic information while adopting\ntask-specific skills. We consider unifying these strategies for low curvature in\nboth parameter space and function space to improve out-of-domain (OOD)\ngeneralization. We propose **Trust Region Aware Minimization** (TRAM), a\nSAM algorithm fine-tuning for low parameter sharpness and smooth, informative\nrepresentations preserving pre-trained structure. TRAM uses a trust region bound\nto inform the SAM adversarial neighborhood, introducing an awareness of function\ncurvature within optimization for flatter minima. We empirically validate TRAM\nin vision (cross-dataset adaptation) and text (OOD language modeling, zero-shot\ncross-lingual transfer) tasks where robust domain transfer and representation\ngenerality are critical. TRAM outperforms SAM- and TR-based optimization across\nall tasks, notably surpassing competing methods for hard transfer between\n_anticorrelated_ domains. TRAM establishes a novel standard in\nfine-tuning for domain-generalizable models with minimal additional computation\nover previous sharpness-aware methods.", "pdf": "https://openreview.net/pdf/15fa46e9fb64654d30da84732fc37543dd3a94ca.pdf"} {"title": "CAMIL: Context-Aware Multiple Instance Learning for Cancer Detection and Subtyping in Whole Slide Images", "url": "https://openreview.net/forum?id=rzBskAEmoc", "detail_url": "https://openreview.net/forum?id=rzBskAEmoc", "authors": "Olga Fourkioti,Matt De Vries,Chris Bakal", "tags": "ICLR 2024,Spotlight", "abstract": "The visual examination of tissue biopsy sections is fundamental for cancer diagnosis, with pathologists analyzing sections at multiple magnifications to discern tumor cells and their subtypes. However, existing attention-based multiple instance learning (MIL) models used for analyzing Whole Slide Images (WSIs) in cancer diagnostics often overlook the contextual information of tumor and neighboring tiles, leading to misclassifications. To address this, we propose the Context-Aware Multiple Instance Learning (CAMIL) architecture. CAMIL incorporates neighbor-constrained attention to consider dependencies among tiles within a WSI and integrates contextual constraints as prior knowledge into the MIL model. We evaluated CAMIL on subtyping non-small cell lung cancer (TCGA-NSCLC) and detecting lymph node (CAMELYON16 and CAMELYON17) metastasis, achieving test AUCs of 97.5\\%, 95.9\\%, and 88.1\\%, respectively, outperforming other state-of-the-art methods. Additionally, CAMIL enhances model interpretability by identifying regions of high diagnostic value. Our code is available at https://github.com/olgarithmics/ICLR_CAMIL.", "pdf": "https://openreview.net/pdf/b4d6251d3b1639d170a910826e7643be5d050285.pdf"} {"title": "DyST: Towards Dynamic Neural Scene Representations on Real-World Videos", "url": "https://openreview.net/forum?id=MnMWa94t12", "detail_url": "https://openreview.net/forum?id=MnMWa94t12", "authors": "Maximilian Seitzer,Sjoerd van Steenkiste,Thomas Kipf,Klaus Greff,Mehdi S. M. Sajjadi", "tags": "ICLR 2024,Spotlight", "abstract": "Visual understanding of the world goes beyond the semantics and flat structure of individual images. In this work, we aim to capture both the 3D structure and dynamics of real-world scenes from monocular real-world videos. Our Dynamic Scene Transformer (DyST) model leverages recent work in neural scene representation to learn a latent decomposition of monocular real-world videos into scene content, per-view scene dynamics, and camera pose. This separation is achieved through a novel co-training scheme on monocular videos and our new synthetic dataset DySO. DyST learns tangible latent representations for dynamic scenes that enable view generation with separate control over the camera and the content of the scene.", "pdf": "https://openreview.net/pdf/cb1c5f7dc44ea3c18ca42146caaee182fe578c30.pdf"} {"title": "Bilevel Optimization under Unbounded Smoothness: A New Algorithm and Convergence Analysis", "url": "https://openreview.net/forum?id=LqRGsGWOTX", "detail_url": "https://openreview.net/forum?id=LqRGsGWOTX", "authors": "Jie Hao,Xiaochuan Gong,Mingrui Liu", "tags": "ICLR 2024,Spotlight", "abstract": "Bilevel optimization is an important formulation for many machine learning problems, such as meta-learning and hyperparameter optimization. Current bilevel optimization algorithms assume that the gradient of the upper-level function is Lipschitz (i.e., the upper-level function has a bounded smoothness parameter). However, recent studies reveal that certain neural networks such as recurrent neural networks (RNNs) and long-short-term memory networks (LSTMs) exhibit potential unbounded smoothness, rendering conventional bilevel optimization algorithms unsuitable for these neural networks. In this paper, we design a new bilevel optimization algorithm, namely BO-REP, to address this challenge. This algorithm updates the upper-level variable using normalized momentum and incorporates two novel techniques for updating the lower-level variable: \\textit{initialization refinement} and \\textit{periodic updates}. Specifically, once the upper-level variable is initialized, a subroutine is invoked to obtain a refined estimate of the corresponding optimal lower-level variable, and the lower-level variable is updated only after every specific period instead of each iteration. When the upper-level problem is nonconvex and unbounded smooth, and the lower-level problem is strongly convex, we prove that our algorithm requires $\\widetilde{O}(1/\\epsilon^4)$ \\footnote{Here $\\widetilde{O}(\\cdot)$ compresses logarithmic factors of $1/\\epsilon$ and $1/\\delta$, where $\\delta\\in(0,1)$ denotes the failure probability.} iterations to find an $\\epsilon$-stationary point in the stochastic setting, where each iteration involves calling a stochastic gradient or Hessian-vector product oracle. Notably, this result matches the state-of-the-art complexity results under the bounded smoothness setting and without mean-squared smoothness of the stochastic gradient, up to logarithmic factors. Our proof relies on novel technical lemmas for the periodically updated lower-level variable, which are of independent interest. Our experiments on hyper-representation learning, hyperparameter optimization, and data hyper-cleaning for text classification tasks demonstrate the effectiveness of our proposed algorithm. The code is available at [https://github.com/MingruiLiu-ML-Lab/Bilevel-Optimization-under-Unbounded-Smoothness](https://github.com/MingruiLiu-ML-Lab/Bilevel-Optimization-under-Unbounded-Smoothness).", "pdf": "https://openreview.net/pdf/1a34c4fa191cbbf4c1a8a8ca78bf84ce2094b701.pdf"} {"title": "Bounds on Representation-Induced Confounding Bias for Treatment Effect Estimation", "url": "https://openreview.net/forum?id=d3xKPQVjSc", "detail_url": "https://openreview.net/forum?id=d3xKPQVjSc", "authors": "Valentyn Melnychuk,Dennis Frauen,Stefan Feuerriegel", "tags": "ICLR 2024,Spotlight", "abstract": "State-of-the-art methods for conditional average treatment effect (CATE) estimation make widespread use of representation learning. Here, the idea is to reduce the variance of the low-sample CATE estimation by a (potentially constrained) low-dimensional representation. However, low-dimensional representations can lose information about the observed confounders and thus lead to bias, because of which the validity of representation learning for CATE estimation is typically violated. In this paper, we propose a new, representation-agnostic refutation framework for estimating bounds on the representation-induced confounding bias that comes from dimensionality reduction (or other constraints on the representations) in CATE estimation. First, we establish theoretically under which conditions CATE is non-identifiable given low-dimensional (constrained) representations. Second, as our remedy, we propose a neural refutation framework which performs partial identification of CATE or, equivalently, aims at estimating lower and upper bounds of the representation-induced confounding bias. We demonstrate the effectiveness of our bounds in a series of experiments. In sum, our refutation framework is of direct relevance in practice where the validity of CATE estimation is of importance.", "pdf": "https://openreview.net/pdf/d06dd3ea5318958c6924d08f905235b1512fde33.pdf"} {"title": "DSPy: Compiling Declarative Language Model Calls into State-of-the-Art Pipelines", "url": "https://openreview.net/forum?id=sY5N0zY5Od", "detail_url": "https://openreview.net/forum?id=sY5N0zY5Od", "authors": "Omar Khattab,Arnav Singhvi,Paridhi Maheshwari,Zhiyuan Zhang,Keshav Santhanam,Sri Vardhamanan A,Saiful Haq,Ashutosh Sharma,Thomas T. Joshi,Hanna Moazam,Heather Miller,Matei Zaharia,Christopher Potts", "tags": "ICLR 2024,Spotlight", "abstract": "The ML community is rapidly exploring techniques for prompting language models (LMs) and for stacking them into pipelines that solve complex tasks. Unfortunately, existing LM pipelines are typically implemented using hard-coded \u201cprompt templates\u201d, i.e. lengthy strings discovered via trial and error. Toward a more systematic approach for developing and optimizing LM pipelines, we introduce DSPy, a programming model that abstracts LM pipelines as text transformation graphs, or imperative computational graphs where LMs are invoked through declarative modules. DSPy modules are parameterized, meaning they can learn how to apply compositions of prompting, finetuning, augmentation, and reasoning techniques. We design a compiler that will optimize any DSPy pipeline to maximize a given metric, by creating and collecting demonstrations. We conduct two case studies, showing that succinct DSPy programs can express and optimize pipelines that reason about math word problems, tackle multi-hop retrieval, answer complex questions, and control agent loops. Within minutes of compiling, DSPy can automatically produce pipelines that outperform out-of-the-box few-shot prompting as well as expert-created demonstrations for GPT-3.5 and Llama2-13b-chat. On top of that, DSPy programs compiled for relatively small LMs like 770M parameter T5 and Llama2-13b-chat are competitive with many approaches that rely on large and proprietary LMs like GPT-3.5 and on expert-written prompt chains. DSPy is available at https://github.com/stanfordnlp/dspy", "pdf": "https://openreview.net/pdf/41028bc2988c119c4fb5c213ab3919ceae696846.pdf"} {"title": "Impact of Computation in Integral Reinforcement Learning for Continuous-Time Control", "url": "https://openreview.net/forum?id=xJEd8PkdNz", "detail_url": "https://openreview.net/forum?id=xJEd8PkdNz", "authors": "Wenhan Cao,Wei Pan", "tags": "ICLR 2024,Spotlight", "abstract": "Integral reinforcement learning (IntRL) demands the precise computation of the utility function's integral at its policy evaluation (PEV) stage. This is achieved through quadrature rules, which are weighted sums of utility functions evaluated from state samples obtained in discrete time. Our research reveals a critical yet underexplored phenomenon: the choice of the computational method -- in this case, the quadrature rule -- can significantly impact control performance. This impact is traced back to the fact that computational errors introduced in the PEV stage can affect the policy iteration's convergence behavior, which in turn affects the learned controller. To elucidate how computation impacts control, we draw a parallel between IntRL's policy iteration and Newton's method applied to the Hamilton-Jacobi-Bellman equation. In this light, computational error in PEV manifests as an extra error term in each iteration of Newton's method, with its upper bound proportional to the computational error. Further, we demonstrate that when the utility function resides in a reproducing kernel Hilbert space (RKHS), the optimal quadrature is achievable by employing Bayesian quadrature with the RKHS-inducing kernel function. We prove that the local convergence rates for IntRL using the trapezoidal rule and Bayesian quadrature with a Mat\u00e9rn kernel to be $O(N^{-2})$ and $O(N^{-b})$, where $N$ is the number of evenly-spaced samples and $b$ is the Mat\u00e9rn kernel's smoothness parameter. These theoretical findings are finally validated by two canonical control tasks.", "pdf": "https://openreview.net/pdf/9eca44e1414a070f87a6a21de74fc149bd37de96.pdf"} {"title": "Masks, Signs, And Learning Rate Rewinding", "url": "https://openreview.net/forum?id=qODvxQ8TXW", "detail_url": "https://openreview.net/forum?id=qODvxQ8TXW", "authors": "Advait Harshal Gadhikar,Rebekka Burkholz", "tags": "ICLR 2024,Spotlight", "abstract": "Learning Rate Rewinding (LRR) has been established as a strong variant of Iterative Magnitude Pruning (IMP) to find lottery tickets in deep overparameterized neural networks. While both iterative pruning schemes couple structure and parameter learning, understanding how LRR excels in both aspects can bring us closer to the design of more flexible deep learning algorithms that can optimize diverse sets of sparse architectures. To this end, we conduct experiments that disentangle the effect of mask learning and parameter optimization and how both benefit from overparameterization. The ability of LRR to flip parameter signs early and stay robust to sign perturbations seems to make it not only more effective in mask identification but also in optimizing diverse sets of masks, including random ones. In support of this hypothesis, we prove in a simplified single hidden neuron setting that LRR succeeds in more cases than IMP, as it can escape initially problematic sign configurations.", "pdf": "https://openreview.net/pdf/8049c38689012fa79be944abb2bec1446b8ed012.pdf"} {"title": "Gradual Domain Adaptation via Gradient Flow", "url": "https://openreview.net/forum?id=iTTZFKrlGV", "detail_url": "https://openreview.net/forum?id=iTTZFKrlGV", "authors": "Zhan Zhuang,Yu Zhang,Ying Wei", "tags": "ICLR 2024,Spotlight", "abstract": "Domain shift degrades classification models on new data distributions. Conventional unsupervised domain adaptation (UDA) aims to learn features that bridge labeled source and unlabeled target domains. In contrast to feature learning, gradual domain adaptation (GDA) leverages extra continuous intermediate domains with pseudo-labels to boost the source classifier. However, real intermediate domains are sometimes unavailable or ineffective. In this paper, we propose $\\textbf{G}$radual Domain Adaptation via $\\textbf{G}$radient $\\textbf{F}$low (GGF) to generate intermediate domains with preserving labels, thereby enabling us a fine-tuning method for GDA. We employ the Wasserstein gradient flow in Kullback\u2013Leibler divergence to transport samples from the source to the target domain. To simulate the dynamics, we utilize the Langevin algorithm. Since the Langevin algorithm disregards label information and introduces diffusion noise, we introduce classifier-based and sample-based potentials to avoid label switching and dramatic deviations in the sampling process. For the proposed GGF model, we analyze its generalization bound. Experiments on several benchmark datasets demonstrate the superiority of the proposed GGF method compared to state-of-the-art baselines.", "pdf": "https://openreview.net/pdf/ff915349976b783c6976376bdd9392b8a18f7773.pdf"} {"title": "Maximum Entropy Heterogeneous-Agent Reinforcement Learning", "url": "https://openreview.net/forum?id=tmqOhBC4a5", "detail_url": "https://openreview.net/forum?id=tmqOhBC4a5", "authors": "Jiarong Liu,Yifan Zhong,Siyi Hu,Haobo Fu,QIANG FU,Xiaojun Chang,Yaodong Yang", "tags": "ICLR 2024,Spotlight", "abstract": "*Multi-agent reinforcement learning* (MARL) has been shown effective for cooperative games in recent years. However, existing state-of-the-art methods face challenges related to sample complexity, training instability, and the risk of converging to a suboptimal Nash Equilibrium. In this paper, we propose a unified framework for learning \\emph{stochastic} policies to resolve these issues. We embed cooperative MARL problems into probabilistic graphical models, from which we derive the maximum entropy (MaxEnt) objective for MARL. Based on the MaxEnt framework, we propose *Heterogeneous-Agent Soft Actor-Critic* (HASAC) algorithm. Theoretically, we prove the monotonic improvement and convergence to *quantal response equilibrium* (QRE) properties of HASAC. Furthermore, we generalize a unified template for MaxEnt algorithmic design named *Maximum Entropy Heterogeneous-Agent Mirror Learning* (MEHAML), which provides any induced method with the same guarantees as HASAC. We evaluate HASAC on six benchmarks: Bi-DexHands, Multi-Agent MuJoCo, StarCraft Multi-Agent Challenge, Google Research Football, Multi-Agent Particle Environment, and Light Aircraft Game. Results show that HASAC consistently outperforms strong baselines, exhibiting better sample efficiency, robustness, and sufficient exploration.", "pdf": "https://openreview.net/pdf/82bacc9b0a9551bf4922e43270f4c315044f70af.pdf"} {"title": "Hybrid Directional Graph Neural Network for Molecules", "url": "https://openreview.net/forum?id=BBD6KXIGJL", "detail_url": "https://openreview.net/forum?id=BBD6KXIGJL", "authors": "Junyi An,Chao Qu,Zhipeng Zhou,Fenglei Cao,Xu Yinghui,Yuan Qi,Furao Shen", "tags": "ICLR 2024,Spotlight", "abstract": "Equivariant message passing neural networks have emerged as the prevailing approach for predicting chemical properties of molecules due to their ability to leverage translation and rotation symmetries, resulting in a strong inductive bias. However, the equivariant operations in each layer can impose excessive constraints on the function form and network flexibility. To address these challenges, we introduce a novel network called the Hybrid Directional Graph Neural Network (HDGNN), which effectively combines strictly equivariant operations with learnable modules. We evaluate the performance of HDGNN on the QM9 dataset and the IS2RE dataset of OC20, demonstrating its state-of-the-art performance on several tasks and competitive performance on others. Our code is anonymously released on https://github.com/ajy112/HDGNN.", "pdf": "https://openreview.net/pdf/fdaba18af51e693376f79fad547fdec1e1913044.pdf"} {"title": "Unbiased Watermark for Large Language Models", "url": "https://openreview.net/forum?id=uWVC5FVidc", "detail_url": "https://openreview.net/forum?id=uWVC5FVidc", "authors": "Zhengmian Hu,Lichang Chen,Xidong Wu,Yihan Wu,Hongyang Zhang,Heng Huang", "tags": "ICLR 2024,Spotlight", "abstract": "The recent advancements in large language models (LLMs) have sparked a growing apprehension regarding the potential misuse. One approach to mitigating this risk is to incorporate watermarking techniques into LLMs, allowing for the tracking and attribution of model outputs. This study examines a crucial aspect of watermarking: how significantly watermarks impact the quality of model-generated outputs. Previous studies have suggested a trade-off between watermark strength and output quality. However, our research demonstrates that it is possible to integrate watermarks without affecting the output probability distribution with appropriate implementation. We refer to this type of watermark as an unbiased watermark. This has significant implications for the use of LLMs, as it becomes impossible for users to discern whether a service provider has incorporated watermarks or not. Furthermore, the presence of watermarks does not compromise the performance of the model in downstream tasks, ensuring that the overall utility of the language model is preserved. Our findings contribute to the ongoing discussion around responsible AI development, suggesting that unbiased watermarks can serve as an effective means of tracking and attributing model outputs without sacrificing output quality.", "pdf": "https://openreview.net/pdf/fdb6b7b2517ce71ee9ed99a12175e4a0273d2b3f.pdf"} {"title": "Leveraging Low-Rank and Sparse Recurrent Connectivity for Robust Closed-Loop Control", "url": "https://openreview.net/forum?id=EriR6Ec69a", "detail_url": "https://openreview.net/forum?id=EriR6Ec69a", "authors": "Neehal Tumma,Mathias Lechner,Noel Loo,Ramin Hasani,Daniela Rus", "tags": "ICLR 2024,Spotlight", "abstract": "Developing autonomous agents that can interact with changing environments is an open challenge in machine learning. Robustness is particularly important in these settings as agents are often fit offline on expert demonstrations but deployed online where they must generalize to the closed feedback loop within the environment. In this work, we explore the application of recurrent neural networks to tasks of this nature and understand how a parameterization of their recurrent connectivity influences robustness in closed-loop settings. Specifically, we represent the recurrent connectivity as a function of rank and sparsity and show both theoretically and empirically that modulating these two variables has desirable effects on network dynamics. The proposed low-rank, sparse connectivity induces an interpretable prior on the network that proves to be most amenable for a class of models known as closed-form continuous-time neural networks (CfCs). We find that CfCs with fewer parameters can outperform their full-rank, fully-connected counterparts in the online setting under distribution shift. This yields memory-efficient and robust agents while opening a new perspective on how we can modulate network dynamics through connectivity.", "pdf": "https://openreview.net/pdf/480e9c477c5c570d2bb4494763d1237fdf11f122.pdf"} {"title": "CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction", "url": "https://openreview.net/forum?id=DjzvJCRsVf", "detail_url": "https://openreview.net/forum?id=DjzvJCRsVf", "authors": "Size Wu,Wenwei Zhang,Lumin Xu,Sheng Jin,Xiangtai Li,Wentao Liu,Chen Change Loy", "tags": "ICLR 2024,Spotlight", "abstract": "Open-vocabulary dense prediction tasks including object detection and image segmentation have been advanced by the success of Contrastive Language-Image Pre-training (CLIP). CLIP models, particularly those incorporating vision transformers (ViTs), have exhibited remarkable generalization ability in zero-shot image classification. However, when transferring the vision-language alignment of CLIP from global image representation to local region representation for the open-vocabulary dense prediction tasks, CLIP ViTs suffer from the domain shift from full images to local image regions. In this paper, we embark on an in-depth analysis of the region-language alignment in CLIP models, which is essential for downstream open-vocabulary dense prediction tasks. Subsequently, we propose an approach named CLIPSelf, which adapts the image-level recognition ability of CLIP ViT to local image regions without needing any region-text pairs. CLIPSelf empowers ViTs to distill itself by aligning a region representation extracted from its dense feature map with the image-level representation of the corresponding image crop. With the enhanced CLIP ViTs, we achieve new state-of-the-art performance on open-vocabulary object detection, semantic segmentation, and panoptic segmentation across various benchmarks. Models and code are released at https://github.com/wusize/CLIPSelf.", "pdf": "https://openreview.net/pdf/126c5bcbf7072558944cfd391f4b42a43cdd40b1.pdf"} {"title": "Large Brain Model for Learning Generic Representations with Tremendous EEG Data in BCI", "url": "https://openreview.net/forum?id=QzTpTRVtrP", "detail_url": "https://openreview.net/forum?id=QzTpTRVtrP", "authors": "Weibang Jiang,Liming Zhao,Bao-liang Lu", "tags": "ICLR 2024,Spotlight", "abstract": "The current electroencephalogram (EEG) based deep learning models are typically designed for specific datasets and applications in brain-computer interaction (BCI), limiting the scale of the models and thus diminishing their perceptual capabilities and generalizability. Recently, Large Language Models (LLMs) have achieved unprecedented success in text processing, prompting us to explore the capabilities of Large EEG Models (LEMs). We hope that LEMs can break through the limitations of different task types of EEG datasets, and obtain universal perceptual capabilities of EEG signals through unsupervised pre-training. Then the models can be fine-tuned for different downstream tasks. However, compared to text data, the volume of EEG datasets is generally small and the format varies widely. For example, there can be mismatched numbers of electrodes, unequal length data samples, varied task designs, and low signal-to-noise ratio. To overcome these challenges, we propose a unified foundation model for EEG called Large Brain Model (LaBraM). LaBraM enables cross-dataset learning by segmenting the EEG signals into EEG channel patches. Vector-quantized neural spectrum prediction is used to train a semantically rich neural tokenizer that encodes continuous raw EEG channel patches into compact neural codes. We then pre-train neural Transformers by predicting the original neural codes for the masked EEG channel patches. The LaBraMs were pre-trained on about 2,500 hours of various types of EEG signals from around 20 datasets and validated on multiple different types of downstream tasks. Experiments on abnormal detection, event type classification, emotion recognition, and gait prediction show that our LaBraM outperforms all compared SOTA methods in their respective fields. Our code is available at https://github.com/935963004/LaBraM.", "pdf": "https://openreview.net/pdf/ce4dc6959056452394dc1ad7b8f64005d65d2165.pdf"} {"title": "Towards LLM4QPE: Unsupervised Pretraining of Quantum Property Estimation and A Benchmark", "url": "https://openreview.net/forum?id=vrBVFXwAmi", "detail_url": "https://openreview.net/forum?id=vrBVFXwAmi", "authors": "Yehui Tang,Hao Xiong,Nianzu Yang,Tailong Xiao,Junchi Yan", "tags": "ICLR 2024,Spotlight", "abstract": "Estimating the properties of quantum systems such as quantum phase has been critical in addressing the essential quantum many-body problems in physics and chemistry. Deep learning models have been recently introduced to property estimation, surpassing conventional statistical approaches. However, these methods are tailored to the specific task and quantum data at hand. It remains an open and attractive question for devising a more universal task-agnostic pretraining model for quantum property estimation. In this paper, we propose LLM4QPE, a large language model style quantum task-agnostic pretraining and finetuning paradigm that 1) performs unsupervised pretraining on diverse quantum systems with different physical conditions; 2) uses the pretrained model for supervised finetuning and delivers high performance with limited training data, on downstream tasks. It mitigates the cost for quantum data collection and speeds up convergence. Extensive experiments show the promising efficacy of LLM4QPE in various tasks including classifying quantum phases of matter on Rydberg atom model and predicting two-body correlation function on anisotropic Heisenberg model.", "pdf": "https://openreview.net/pdf/511ed0e0d3b143e5589b96afaba84da894f71df7.pdf"} {"title": "GTMGC: Using Graph Transformer to Predict Molecule\u2019s Ground-State Conformation", "url": "https://openreview.net/forum?id=F7QnIKlC1N", "detail_url": "https://openreview.net/forum?id=F7QnIKlC1N", "authors": "Guikun Xu,Yongquan Jiang,PengChuan Lei,Yan Yang,Jim Chen", "tags": "ICLR 2024,Spotlight", "abstract": "The ground-state conformation of a molecule is often decisive for its properties. However, experimental or computational methods, such as density functional theory (DFT), are time-consuming and labor-intensive for obtaining this conformation. Deep learning (DL) based molecular representation learning (MRL) has made significant advancements in molecular modeling and has achieved remarkable results in various tasks. Consequently, it has emerged as a promising approach for directly predicting the ground-state conformation of molecules. In this regard, we introduce GTMGC, a novel network based on Graph-Transformer (GT) that seamlessly predicts the spatial configuration of molecules in a 3D space from their 2D topological architecture in an end-to-end manner. Moreover, we propose a novel self-attention mechanism called Molecule Structural Residual Self-Attention (MSRSA) for molecular structure modeling. This mechanism not only guarantees high model performance and easy implementation but also lends itself well to other molecular modeling tasks. Our method has been evaluated on the Molecule3D benchmark dataset and the QM9 dataset. Experimental results demonstrate that our approach achieves remarkable performance and outperforms current state-of-the-art methods as well as the widely used open-source software RDkit.", "pdf": "https://openreview.net/pdf/c141834e1d331e6055ab503795d49d4e4b8548fb.pdf"} {"title": "Generalization of Scaled Deep ResNets in the Mean-Field Regime", "url": "https://openreview.net/forum?id=tMzPZTvz2H", "detail_url": "https://openreview.net/forum?id=tMzPZTvz2H", "authors": "Yihang Chen,Fanghui Liu,Yiping Lu,Grigorios Chrysos,Volkan Cevher", "tags": "ICLR 2024,Spotlight", "abstract": "Despite the widespread empirical success of ResNet, the generalization properties of deep ResNet are rarely explored beyond the lazy training regime. In this work, we investigate scaled ResNet in the limit of infinitely deep and wide neural networks, of which the gradient flow is described by a partial differential equation in the large-neural network limit, i.e., the mean-field regime. To derive the generalization bounds under this setting, our analysis necessitates a shift from the conventional time-invariant Gram matrix employed in the lazy training regime to a time-variant, distribution-dependent version. To this end, we provide a global lower bound on the minimum eigenvalue of the Gram matrix under the mean-field regime. Besides, for the traceability of the dynamic of Kullback-Leibler (KL) divergence, we establish the linear convergence of the empirical error and estimate the upper bound of the KL divergence over parameters distribution. Finally, we build the uniform convergence for generalization bound via Rademacher complexity. Our results offer new insights into the generalization ability of deep ResNet beyond the lazy training regime and contribute to advancing the understanding of the fundamental properties of deep neural networks.", "pdf": "https://openreview.net/pdf/72b4830ed0321f0098f96447794bfcc965134752.pdf"} {"title": "ODE Discovery for Longitudinal Heterogeneous Treatment Effects Inference", "url": "https://openreview.net/forum?id=pxI5IPeWgW", "detail_url": "https://openreview.net/forum?id=pxI5IPeWgW", "authors": "Krzysztof Kacprzyk,Samuel Holt,Jeroen Berrevoets,Zhaozhi Qian,Mihaela van der Schaar", "tags": "ICLR 2024,Spotlight", "abstract": "Inferring unbiased treatment effects has received widespread attention in the machine learning community. In recent years, our community has proposed numerous solutions in standard settings, high-dimensional treatment settings, and even longitudinal settings. While very diverse, the solution has mostly relied on neural networks for inference and simultaneous correction of assignment bias. New approaches typically build on top of previous approaches by proposing new (or refined) architectures and learning algorithms. However, the end result\u2014a neural-network-based inference machine\u2014remains unchallenged. In this paper, we introduce a different type of solution in the longitudinal setting: a closed-form ordinary differential equation (ODE). While we still rely on continuous optimization to learn an ODE, the resulting inference machine is no longer a neural network. Doing so yields several advantages such as interpretability, irregular sampling, and a different set of identification assumptions. Above all, we consider the introduction of a completely new type of solution to be our most important contribution as it may spark entirely new innovations in treatment effects in general. We facilitate this by formulating our contribution as a framework that can transform any ODE discovery method into a treatment effects method.", "pdf": "https://openreview.net/pdf/2e710a9328ce1b12daf4fde40da8165ca071d5db.pdf"} {"title": "Learning Hierarchical World Models with Adaptive Temporal Abstractions from Discrete Latent Dynamics", "url": "https://openreview.net/forum?id=TjCDNssXKU", "detail_url": "https://openreview.net/forum?id=TjCDNssXKU", "authors": "Christian Gumbsch,Noor Sajid,Georg Martius,Martin V. Butz", "tags": "ICLR 2024,Spotlight", "abstract": "Hierarchical world models can significantly improve model-based reinforcement learning (MBRL) and planning by enabling reasoning across multiple time scales. Nonetheless, the majority of state-of-the-art MBRL methods employ flat, non-hierarchical models. We propose Temporal Hierarchies from Invariant Context Kernels (THICK), an algorithm that learns a world model hierarchy via discrete latent dynamics. The lower level of THICK updates parts of its latent state sparsely in time, forming invariant contexts. The higher level exclusively predicts situations involving context changes. Our experiments demonstrate that THICK learns categorical, interpretable, temporal abstractions on the high level, while maintaining precise low-level predictions. Furthermore, we show that the emergent hierarchical predictive model seamlessly enhances the abilities of MBRL or planning methods. We believe that THICK contributes to the further development of hierarchical agents capable of more sophisticated planning and reasoning abilities.", "pdf": "https://openreview.net/pdf/3e5df2ed6659f21032c8784d5836ef8147d1413a.pdf"} {"title": "Prediction without Preclusion: Recourse Verification with Reachable Sets", "url": "https://openreview.net/forum?id=SCQfYpdoGE", "detail_url": "https://openreview.net/forum?id=SCQfYpdoGE", "authors": "Avni Kothari,Bogdan Kulynych,Tsui-Wei Weng,Berk Ustun", "tags": "ICLR 2024,Spotlight", "abstract": "Machine learning models are often used to decide who receives a loan, a job interview, or a public benefit. Models in such settings use features without considering their *actionability*. As a result, they can assign predictions that are \\emph{fixed} -- meaning that individuals who are denied loans and interviews are, in fact, *precluded from access* to credit and employment. In this work, we introduce a procedure called *recourse verification* to test if a model assigns fixed predictions to its decision subjects. We propose a model-agnostic approach for verification with *reachable sets* -- i.e., the set of all points that a person can reach through their actions in feature space. We develop methods to construct reachable sets for discrete feature spaces, which can certify the responsiveness of *any model* by simply querying its predictions. We conduct a comprehensive empirical study on the infeasibility of recourse on datasets from consumer finance. Our results highlight how models can inadvertently preclude access by assigning fixed predictions and underscore the need to account for actionability in model development.", "pdf": "https://openreview.net/pdf/08180baa9640c55bee2805e14b90ecc715509ee3.pdf"} {"title": "ODICE: Revealing the Mystery of Distribution Correction Estimation via Orthogonal-gradient Update", "url": "https://openreview.net/forum?id=L8UNn7Llt4", "detail_url": "https://openreview.net/forum?id=L8UNn7Llt4", "authors": "Liyuan Mao,Haoran Xu,Weinan Zhang,Xianyuan Zhan", "tags": "ICLR 2024,Spotlight", "abstract": "In this study, we investigate the DIstribution Correction Estimation (DICE) methods, an important line of work in offline reinforcement learning (RL) and imitation learning (IL). DICE-based methods impose state-action-level behavior constraint, which is an ideal choice for offline learning. However, they typically perform much worse than current state-of-the-art (SOTA) methods that solely use action-level behavior constraint. After revisiting DICE-based methods, we find there exist two gradient terms when learning the value function using true-gradient update: forward gradient (taken on the current state) and backward gradient (taken on the next state). Using forward gradient bears a large similarity to many offline RL methods, and thus can be regarded as applying action-level constraint. However, directly adding the backward gradient may degenerate or cancel out its effect if these two gradients have conflicting directions. To resolve this issue, we propose a simple yet effective modification that projects the backward gradient onto the normal plane of the forward gradient, resulting in an orthogonal-gradient update, a new learning rule for DICE-based methods. We conduct thorough theoretical analyses and find that the projected backward gradient brings state-level behavior regularization, which reveals the mystery of DICE-based methods: the value learning objective does try to impose state-action-level constraint, but needs to be used in a corrected way. Through toy examples and extensive experiments on complex offline RL and IL tasks, we demonstrate that DICE-based methods using orthogonal-gradient updates achieve SOTA performance and great robustness.", "pdf": "https://openreview.net/pdf/833ece7fade579c01692e5603d476db35ce59989.pdf"} {"title": "Improving Non-Transferable Representation Learning by Harnessing Content and Style", "url": "https://openreview.net/forum?id=FYKVPOHCpE", "detail_url": "https://openreview.net/forum?id=FYKVPOHCpE", "authors": "Ziming Hong,Zhenyi Wang,Li Shen,Yu Yao,Zhuo Huang,Shiming Chen,Chuanwu Yang,Mingming Gong,Tongliang Liu", "tags": "ICLR 2024,Spotlight", "abstract": "Non-transferable learning (NTL) aims to restrict the generalization of models toward the target domain(s). To this end, existing works learn non-transferable representations by reducing statistical dependence between the source and target domain. However, such statistical methods essentially neglect to distinguish between *styles* and *contents*, leading them to inadvertently fit (i) spurious correlation between *styles* and *labels*, and (ii) fake independence between *contents* and *labels*. Consequently, their performance will be limited when natural distribution shifts occur or malicious intervention is imposed. In this paper, we propose a novel method (dubbed as H-NTL) to understand and advance the NTL problem by introducing a causal model to separately model *content* and *style* as two latent factors, based on which we disentangle and harness them as guidances for learning non-transferable representations with intrinsically causal relationships. Speci\ufb01cally, to avoid fitting spurious correlation and fake independence, we propose a variational inference framework to disentangle the naturally mixed *content factors* and *style factors* under our causal model. Subsequently, based on dual-path knowledge distillation, we harness the disentangled two *factors* as guidances for non-transferable representation learning: (i) we constraint the source domain representations to fit *content factors* (which are the intrinsic cause of *labels*), and (ii) we enforce that the target domain representations fit *style factors* which barely can predict labels. As a result, the learned feature representations follow optimal untransferability toward the target domain and minimal negative influence on the source domain, thus enabling better NTL performance. Empirically, the proposed H-NTL signi\ufb01cantly outperforms competing methods by a large margin.", "pdf": "https://openreview.net/pdf/4d359626d33d8cb2e10f6d1cf6728b805a2b5316.pdf"} {"title": "ModernTCN: A Modern Pure Convolution Structure for General Time Series Analysis", "url": "https://openreview.net/forum?id=vpJMJerXHU", "detail_url": "https://openreview.net/forum?id=vpJMJerXHU", "authors": "Luo donghao,wang xue", "tags": "ICLR 2024,Spotlight", "abstract": "Recently, Transformer-based and MLP-based models have emerged rapidly and\nwon dominance in time series analysis. In contrast, convolution is losing steam\nin time series tasks nowadays for inferior performance. This paper studies the\nopen question of how to better use convolution in time series analysis and makes\nefforts to bring convolution back to the arena of time series analysis. To this end,\nwe modernize the traditional TCN and conduct time series related modifications\nto make it more suitable for time series tasks. As the outcome, we propose\nModernTCN and successfully solve this open question through a seldom-explored\nway in time series community. As a pure convolution structure, ModernTCN still\nachieves the consistent state-of-the-art performance on five mainstream time series\nanalysis tasks while maintaining the efficiency advantage of convolution-based\nmodels, therefore providing a better balance of efficiency and performance than\nstate-of-the-art Transformer-based and MLP-based models. Our study further\nreveals that, compared with previous convolution-based models, our ModernTCN\nhas much larger effective receptive fields (ERFs), therefore can better unleash the\npotential of convolution in time series analysis. Code is available at this repository:\nhttps://github.com/luodhhh/ModernTCN.", "pdf": "https://openreview.net/pdf/c0de77eed380b4b2736dfe855ed3cf0d62f7d8c1.pdf"} {"title": "Towards Robust Out-of-Distribution Generalization Bounds via Sharpness", "url": "https://openreview.net/forum?id=tPEwSYPtAC", "detail_url": "https://openreview.net/forum?id=tPEwSYPtAC", "authors": "Yingtian Zou,Kenji Kawaguchi,Yingnan Liu,Jiashuo Liu,Mong-Li Lee,Wynne Hsu", "tags": "ICLR 2024,Spotlight", "abstract": "Generalizing to out-of-distribution (OOD) data or unseen domain, termed OOD generalization, still lacks appropriate theoretical guarantees. Canonical OOD bounds focus on different distance measurements between source and target domains but fail to consider the optimization property of the learned model. As empirically shown in recent work, sharpness of learned minimum influences OOD generalization. To bridge this gap between optimization and OOD generalization, we study the effect of sharpness on how a model tolerates data change in domain shift which is usually captured by \"robustness\" in generalization. In this paper, we give a rigorous connection between sharpness and robustness, which gives better OOD guarantees for robust algorithms. It also provides a theoretical backing for \"flat minima leads to better OOD generalization\". Overall, we propose a sharpness-based OOD generalization bound by taking robustness into consideration, resulting in a tighter bound than non-robust guarantees. Our findings are supported by the experiments on a ridge regression model, as well as the experiments on deep learning classification tasks.", "pdf": "https://openreview.net/pdf/36904eee9458f8a5da9944cdcd92446a053dfa88.pdf"} {"title": "MAPE-PPI: Towards Effective and Efficient Protein-Protein Interaction Prediction via Microenvironment-Aware Protein Embedding", "url": "https://openreview.net/forum?id=itGkF993gz", "detail_url": "https://openreview.net/forum?id=itGkF993gz", "authors": "Lirong Wu,Yijun Tian,Yufei Huang,Siyuan Li,Haitao Lin,Nitesh V Chawla,Stan Z. Li", "tags": "ICLR 2024,Spotlight", "abstract": "Protein-Protein Interactions (PPIs) are fundamental in various biological processes and play a key role in life activities. The growing demand and cost of experimental PPI assays require computational methods for efficient PPI prediction. While existing methods rely heavily on protein sequence for PPI prediction, it is the protein structure that is the key to determine the interactions. To take both protein modalities into account, we define the microenvironment of an amino acid residue by its sequence and structural contexts, which describe the surrounding chemical properties and geometric features. In addition, microenvironments defined in previous work are largely based on experimentally assayed physicochemical properties, for which the \"vocabulary\" is usually extremely small. This makes it difficult to cover the diversity and complexity of microenvironments. In this paper, we propose Microenvironment-Aware Protein Embedding for PPI prediction (MPAE-PPI), which encodes microenvironments into chemically meaningful discrete codes via a sufficiently large microenvironment \"vocabulary\" (i.e., codebook). Moreover, we propose a novel pre-training strategy, namely Masked Codebook Modeling (MCM), to capture the dependencies between different microenvironments by randomly masking the codebook and reconstructing the input. With the learned microenvironment codebook, we can reuse it as an off-the-shelf tool to efficiently and effectively encode proteins of different sizes and functions for large-scale PPI prediction. Extensive experiments show that MAPE-PPI can scale to PPI prediction with millions of PPIs with superior trade-offs between effectiveness and computational efficiency than the state-of-the-art competitors.", "pdf": "https://openreview.net/pdf/72464b5e34ef4de8f928bfdd6309981dbe271cf6.pdf"} {"title": "Negative Label Guided OOD Detection with Pretrained Vision-Language Models", "url": "https://openreview.net/forum?id=xUO1HXz4an", "detail_url": "https://openreview.net/forum?id=xUO1HXz4an", "authors": "Xue Jiang,Feng Liu,Zhen Fang,Hong Chen,Tongliang Liu,Feng Zheng,Bo Han", "tags": "ICLR 2024,Spotlight", "abstract": "Out-of-distribution (OOD) detection aims at identifying samples from unknown classes, playing a crucial role in trustworthy models against errors on unexpected inputs. \nExtensive research has been dedicated to exploring OOD detection in the vision modality. \n{Vision-language models (VLMs) can leverage both textual and visual information for various multi-modal applications, whereas few OOD detection methods take into account information from the text modality. \nIn this paper, we propose a novel post hoc OOD detection method, called NegLabel, which takes a vast number of negative labels from extensive corpus databases. We design a novel scheme for the OOD score collaborated with negative labels.\nTheoretical analysis helps to understand the mechanism of negative labels. Extensive experiments demonstrate that our method NegLabel achieves state-of-the-art performance on various OOD detection benchmarks and generalizes well on multiple VLM architectures. Furthermore, our method NegLabel exhibits remarkable robustness against diverse domain shifts. The codes are available at https://github.com/tmlr-group/NegLabel.", "pdf": "https://openreview.net/pdf/b9ad30ff96f366ad87a0053257956ba3b2a4ece6.pdf"} {"title": "OPTIMAL ROBUST MEMORIZATION WITH RELU NEURAL NETWORKS", "url": "https://openreview.net/forum?id=47hDbAMLbc", "detail_url": "https://openreview.net/forum?id=47hDbAMLbc", "authors": "Lijia Yu,Xiao-Shan Gao,Lijun Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "Memorization with neural networks is to study the expressive power of neural networks to interpolate a finite classification data set, which is closely related to the generalizability of deep learning. However, the important problem of robust memorization has not been thoroughly studied. In this paper, several basic problems about robust memorization are solved. First, we prove that it is NP-hard to compute neural networks with certain simple structures, which are robust memorization. A network hypothesis space is called optimal robust memorization for a data set if it can achieve robust memorization for any budget less than half the separation bound of the data set. Second, we explicitly construct neural networks with O(N n) parameters for optimal robust memorization of any data set with dimension n and size N . We also give a lower bound for the width of networks to achieve optimal robust memorization. Finally, we explicitly construct neural networks with\nO(N n log n) parameters for optimal robust memorization of any binary classification data set by controlling the Lipschitz constant of the network.", "pdf": "https://openreview.net/pdf/c5be8a576cab367723fcf91c4b950557846a3e1a.pdf"} {"title": "Neural Contractive Dynamical Systems", "url": "https://openreview.net/forum?id=iAYIRHOYy8", "detail_url": "https://openreview.net/forum?id=iAYIRHOYy8", "authors": "Hadi Beik Mohammadi,S\u00f8ren Hauberg,Georgios Arvanitidis,Nadia Figueroa,Gerhard Neumann,Leonel Rozo", "tags": "ICLR 2024,Spotlight", "abstract": "Stability guarantees are crucial when ensuring that a fully autonomous robot does not take undesirable or potentially harmful actions. Unfortunately, global stability guarantees are hard to provide in dynamical systems learned from data, especially when the learned dynamics are governed by neural networks. We propose a novel methodology to learn \\emph{neural contractive dynamical systems}, where our neural architecture ensures contraction, and hence, global stability. To efficiently scale the method to high-dimensional dynamical systems, we develop a variant of the variational autoencoder that learns dynamics in a low-dimensional latent representation space while retaining contractive stability after decoding. We further extend our approach to learning contractive systems on the Lie group of rotations to account for full-pose end-effector dynamic motions. The result is the first highly flexible learning architecture that provides contractive stability guarantees with capability to perform obstacle avoidance. Empirically, we demonstrate that our approach encodes the desired dynamics more accurately than the current state-of-the-art, which provides less strong stability guarantees.", "pdf": "https://openreview.net/pdf/a89591eec311a0efbd01f7135555a21d2d682c1c.pdf"} {"title": "Scaling Laws for Associative Memories", "url": "https://openreview.net/forum?id=Tzh6xAJSll", "detail_url": "https://openreview.net/forum?id=Tzh6xAJSll", "authors": "Vivien Cabannes,Elvis Dohmatob,Alberto Bietti", "tags": "ICLR 2024,Spotlight", "abstract": "Learning arguably involves the discovery and memorization of abstract rules. The aim of this paper is to study associative memory mechanisms. Our model is based on high-dimensional matrices consisting of outer products of embeddings, which relates to the inner layers of transformer language models. We derive precise scaling laws with respect to sample size and parameter size, and discuss the statistical efficiency of different estimators, including optimization-based algorithms. We provide extensive numerical experiments to validate and interpret theoretical results, including fine-grained visualizations of the stored memory associations.", "pdf": "https://openreview.net/pdf/ba075a88abc0ad2b7f00577253a950d3264c5f2f.pdf"} {"title": "Text2Reward: Reward Shaping with Language Models for Reinforcement Learning", "url": "https://openreview.net/forum?id=tUM39YTRxH", "detail_url": "https://openreview.net/forum?id=tUM39YTRxH", "authors": "Tianbao Xie,Siheng Zhao,Chen Henry Wu,Yitao Liu,Qian Luo,Victor Zhong,Yanchao Yang,Tao Yu", "tags": "ICLR 2024,Spotlight", "abstract": "Designing reward functions is a longstanding challenge in reinforcement learning (RL); it requires specialized knowledge or domain data, leading to high costs for development. To address this, we introduce Text2Reward, a data-free framework that automates the generation and shaping of dense reward functions based on large language models (LLMs). Given a goal described in natural language, Text2Reward generates shaped dense reward functions as an executable program grounded in a compact representation of the environment. Unlike inverse RL and recent work that uses LLMs to write sparse reward codes or unshaped dense rewards with a constant function across timesteps, Text2Reward produces interpretable, free-form dense reward codes that cover a wide range of tasks, utilize existing packages, and allow iterative refinement with human feedback. We evaluate Text2Reward on two robotic manipulation benchmarks (ManiSkill2, MetaWorld) and two locomotion environments of MuJoCo. On 13 of the 17 manipulation tasks, policies trained with generated reward codes achieve similar or better task success rates and convergence speed than expert-written reward codes. For locomotion tasks, our method learns six novel locomotion behaviors with a success rate exceeding 94%. Furthermore, we show that the policies trained in the simulator with our method can be deployed in the real world. Finally, Text2Reward further improves the policies by refining their reward functions with human feedback. Video results are available at https://text-to-reward.github.io/", "pdf": "https://openreview.net/pdf/a52e7202163a42116fae8ada42123e37f2aef287.pdf"} {"title": "Towards Meta-Pruning via Optimal Transport", "url": "https://openreview.net/forum?id=sMoifbuxjB", "detail_url": "https://openreview.net/forum?id=sMoifbuxjB", "authors": "Alexander Theus,Olin Geimer,Friedrich Wicke,Thomas Hofmann,Sotiris Anagnostidis,Sidak Pal Singh", "tags": "ICLR 2024,Spotlight", "abstract": "Structural pruning of neural networks conventionally relies on identifying and discarding less important neurons, a practice often resulting in significant accuracy loss that necessitates subsequent fine-tuning efforts. This paper introduces a novel approach named Intra-Fusion, challenging this prevailing pruning paradigm.\nUnlike existing methods that focus on designing meaningful neuron importance metrics, Intra-Fusion redefines the overlying pruning procedure.\nThrough utilizing the concepts of model fusion and Optimal Transport, we leverage an agnostically given importance metric to arrive at a more effective sparse model representation.\nNotably, our approach achieves substantial accuracy recovery without the need for resource-intensive fine-tuning, making it an efficient and promising tool for neural network compression.\nAdditionally, we explore how fusion can be added to the pruning process to significantly decrease the training time while maintaining competitive performance. We benchmark our results for various networks on commonly used datasets such as CIFAR-10, CIFAR-100, and ImageNet. More broadly, we hope that the proposed Intra-Fusion approach invigorates exploration into a fresh alternative to the predominant compression approaches.\nOur code is available [here](https://github.com/alexandertheus/Intra-Fusion).", "pdf": "https://openreview.net/pdf/07560e42af2e42df14ac71025723b0b97a0924dd.pdf"} {"title": "InternVid: A Large-scale Video-Text Dataset for Multimodal Understanding and Generation", "url": "https://openreview.net/forum?id=MLBdiWu4Fw", "detail_url": "https://openreview.net/forum?id=MLBdiWu4Fw", "authors": "Yi Wang,Yinan He,Yizhuo Li,Kunchang Li,Jiashuo Yu,Xin Ma,Xinhao Li,Guo Chen,Xinyuan Chen,Yaohui Wang,Ping Luo,Ziwei Liu,Yali Wang,Limin Wang,Yu Qiao", "tags": "ICLR 2024,Spotlight", "abstract": "This paper introduces InternVid, a large-scale video-centric multimodal dataset that enables learning powerful and transferable video-text representations for multimodal understanding and generation. InternVid contains over 7 million videos lasting nearly 760K hours, yielding 234M video clips accompanied by detailed descriptions of total 4.1B words. Our core contribution is to develop a scalable approach to autonomously build a high-quality video-text dataset with large language models (LLM), thereby showcasing its efficacy in learning video-language representation at scale. Specifically, we utilize a multi-scale approach to generate video-related descriptions. Furthermore, we introduce ViCLIP, a video-text representation learning model based on ViT-L. Learned on InternVid via contrastive learning, this model demonstrates leading zero-shot action recognition and competitive video retrieval performance. Beyond basic video understanding tasks like recognition and retrieval, our dataset and model have broad applications. They are particularly beneficial for generating interleaved video-text data for learning a video-centric dialogue system, advancing video-to-text and text-to-video generation research. These proposed resources provide a tool for researchers and practitioners interested in multimodal video understanding and generation.", "pdf": "https://openreview.net/pdf/5355ce2fec3ff26dca65a969b767fd7b1102bb05.pdf"} {"title": "Dictionary Contrastive Learning for Efficient Local Supervision without Auxiliary Networks", "url": "https://openreview.net/forum?id=Gg7cXo3S8l", "detail_url": "https://openreview.net/forum?id=Gg7cXo3S8l", "authors": "Suhwan Choi,Myeongho Jeon,Yeonjung Hwang,Jeonglyul Oh,Sungjun Lim,Joonseok Lee,Myungjoo Kang", "tags": "ICLR 2024,Spotlight", "abstract": "While backpropagation (BP) has achieved widespread success in deep learning, it\nfaces two prominent challenges: computational inefficiency and biological implausibility.\nIn response to these challenges, local supervision, encompassing Local\nLearning (LL) and Forward Learning (FL), has emerged as a promising research\ndirection. LL employs module-wise BP to achieve competitive results yet relies on\nmodule-wise auxiliary networks, which increase memory and parameter demands.\nConversely, FL updates layer weights without BP and auxiliary networks but falls\nshort of BP\u2019s performance. This paper proposes a simple yet effective objective\nwithin a contrastive learning framework for local supervision without auxiliary\nnetworks. Given the insight that the existing contrastive learning framework for\nlocal supervision is susceptible to task-irrelevant information without auxiliary\nnetworks, we present DICTIONARY CONTRASTIVE LEARNING (DCL) that optimizes\nthe similarity between local features and label embeddings. Our method\nusing static label embeddings yields substantial performance improvements in the\nFL scenario, outperforming state-of-the-art FL approaches. Moreover, our method\nusing adaptive label embeddings closely approaches the performance achieved by\nLL while achieving superior memory and parameter efficiency.", "pdf": "https://openreview.net/pdf/f9734ebbb92e7bdafcdb35c2da50c63e5e5ad16d.pdf"} {"title": "Bounding Box Stability against Feature Dropout Reflects Detector Generalization across Environments", "url": "https://openreview.net/forum?id=lmM4Ecm4HJ", "detail_url": "https://openreview.net/forum?id=lmM4Ecm4HJ", "authors": "Yang Yang,Wenhai Wang,Zhe Chen,Jifeng Dai,Liang Zheng", "tags": "ICLR 2024,Spotlight", "abstract": "Bounding boxes uniquely characterize object detection, where a good detector gives accurate bounding boxes of categories of interest. However, in the real-world where test ground truths are not provided, it is non-trivial to find out whether bounding boxes are accurate, thus preventing us from assessing the detector generalization ability. In this work, we find under feature map dropout, good detectors tend to output bounding boxes whose locations do not change much, while bounding boxes of poor detectors will undergo noticeable position changes. We compute the box stability score (BS score) to reflect this stability. Specifically, given an image, we compute a normal set of bounding boxes and a second set after feature map dropout. To obtain BS score, we use bipartite matching to find the corresponding boxes between the two sets and compute the average Intersection over Union (IoU) across the entire test set. We contribute to finding that BS score has a strong, positive correlation with detection accuracy measured by mean average precision (mAP) under various test environments. This relationship allows us to predict the accuracy of detectors on various real-world test sets without accessing test ground truths, verified on canonical detection tasks such as vehicle detection and pedestrian detection.", "pdf": "https://openreview.net/pdf/5510c4a1e453a12979e2d2a9f12b836fdc0436c8.pdf"} {"title": "Deep Geodesic Canonical Correlation Analysis for Covariance-Based Neuroimaging Data", "url": "https://openreview.net/forum?id=PnR1MNen7u", "detail_url": "https://openreview.net/forum?id=PnR1MNen7u", "authors": "Ce Ju,Reinmar J Kobler,Liyao Tang,Cuntai Guan,Motoaki Kawanabe", "tags": "ICLR 2024,Spotlight", "abstract": "In human neuroimaging, multi-modal imaging techniques are frequently combined to enhance our comprehension of whole-brain dynamics and improve diagnosis in clinical practice. Modalities like electroencephalography and functional magnetic resonance imaging provide distinct views to the brain dynamics due to diametral spatiotemporal sensitivities and underlying neurophysiological coupling mechanisms. These distinct views pose a considerable challenge to learning a shared representation space, especially when dealing with covariance-based data characterized by their geometric structure. To capitalize on the geometric structure, we introduce a measure called geodesic correlation which expands traditional correlation consistency to covariance-based data on the symmetric positive definite (SPD) manifold. This measure is derived from classical canonical correlation analysis and serves to evaluate the consistency of latent representations obtained from paired views. For multi-view, self-supervised learning where one or both latent views are SPD we propose an innovative geometric deep learning framework termed DeepGeoCCA. Its primary objective is to enhance the geodesic correlation of unlabeled, paired data, thereby generating novel representations while retaining the geometric structures. In simulations and experiments with multi-view and multi-modal human neuroimaging data, we find that DeepGeoCCA learns latent representations with high geodesic correlation for unseen data while retaining relevant information for downstream tasks.", "pdf": "https://openreview.net/pdf/4ccf9cac26244e14e3fd2742852e226018c0e4b8.pdf"} {"title": "SWAP-NAS: Sample-Wise Activation Patterns for Ultra-fast NAS", "url": "https://openreview.net/forum?id=tveiUXU2aa", "detail_url": "https://openreview.net/forum?id=tveiUXU2aa", "authors": "Yameng Peng,Andy Song,Haytham M. Fayek,Vic Ciesielski,Xiaojun Chang", "tags": "ICLR 2024,Spotlight", "abstract": "Training-free metrics (a.k.a. zero-cost proxies) are widely used to avoid resource-intensive neural network training, especially in Neural Architecture Search (NAS). Recent studies show that existing training-free metrics have several limitations, such as limited correlation and poor generalisation across different search spaces and tasks. Hence, we propose Sample-Wise Activation Patterns and its derivative, SWAP-Score, a novel high-performance training-free metric. It measures the expressivity of networks over a batch of input samples. The SWAP-Score is strongly correlated with ground-truth performance across various search spaces and tasks, outperforming 15 existing training-free metrics on NAS-Bench-101/201/301 and TransNAS-Bench-101. The SWAP-Score can be further enhanced by regularisation, which leads to even higher correlations in cell-based search space and enables model size control during the search. For example, Spearman\u2019s rank correlation coefficient between regularised SWAP-Score and CIFAR-100 validation accuracies on NAS-Bench-201 networks is 0.90, significantly higher than 0.80 from the second-best metric, NWOT. When integrated with an evolutionary algorithm for NAS, our SWAP-NAS achieves competitive performance on CIFAR-10 and ImageNet in approximately 6 minutes and 9 minutes of GPU time respectively.", "pdf": "https://openreview.net/pdf/37b8588fc5e1d4701d0dd7f69b3af45b36b148e9.pdf"} {"title": "RT-Trajectory: Robotic Task Generalization via Hindsight Trajectory Sketches", "url": "https://openreview.net/forum?id=F1TKzG8LJO", "detail_url": "https://openreview.net/forum?id=F1TKzG8LJO", "authors": "Jiayuan Gu,Sean Kirmani,Paul Wohlhart,Yao Lu,Montserrat Gonzalez Arenas,Kanishka Rao,Wenhao Yu,Chuyuan Fu,Keerthana Gopalakrishnan,Zhuo Xu,Priya Sundaresan,Peng Xu,Hao Su,Karol Hausman,Chelsea Finn,Quan Vuong,Ted Xiao", "tags": "ICLR 2024,Spotlight", "abstract": "Generalization remains one of the most important desiderata for robust robot learning systems. While recently proposed approaches show promise in generalization to novel objects, semantic concepts, or visual distribution shifts, generalization to new tasks remains challenging. For example, a language-conditioned policy trained on pick-and-place tasks will not be able to generalize to a folding task, even if the arm trajectory of folding is similar to pick-and-place. Our key insight is that this kind of generalization becomes feasible if we represent the task through rough trajectory sketches. We propose a policy conditioning method using such rough trajectory sketches, which we call RT-Trajectory, that is practical, easy to specify, and allows the policy to effectively perform new tasks that would otherwise be challenging to perform. We find that trajectory sketches strike a balance between being detailed enough to express low-level motion-centric guidance while being coarse enough to allow the learned policy to interpret the trajectory sketch in the context of situational visual observations. In addition, we show how trajectory sketches can provide a useful interface to communicate with robotic policies -- they can be specified through simple human inputs like drawings or videos, or through automated methods such as modern image-generating or waypoint-generating methods. We evaluate RT-Trajectory at scale on a variety of real-world robotic tasks, and find that RT-Trajectory is able to perform a wider range of tasks compared to language-conditioned and goal-conditioned policies, when provided the same training data.", "pdf": "https://openreview.net/pdf/99c49fe414f0c5349b9a1f94d32198a847626df5.pdf"} {"title": "NaturalSpeech 2: Latent Diffusion Models are Natural and Zero-Shot Speech and Singing Synthesizers", "url": "https://openreview.net/forum?id=Rc7dAwVL3v", "detail_url": "https://openreview.net/forum?id=Rc7dAwVL3v", "authors": "Kai Shen,Zeqian Ju,Xu Tan,Eric Liu,Yichong Leng,Lei He,Tao Qin,sheng zhao,Jiang Bian", "tags": "ICLR 2024,Spotlight", "abstract": "Scaling text-to-speech (TTS) to large-scale, multi-speaker, and in-the-wild datasets is important to capture the diversity in human speech such as speaker identities, prosodies, and styles (e.g., singing). Current large TTS systems usually quantize speech into discrete tokens and use language models to generate these tokens one by one, which suffer from unstable prosody, word skipping/repeating issue, and poor voice quality. In this paper, we develop NaturalSpeech 2, a TTS system that leverages a neural audio codec with residual vector quantizers to get the quantized latent vectors and uses a diffusion model to generate these latent vectors conditioned on text input. To enhance the zero-shot capability that is important to achieve diverse speech synthesis, we design a speech prompting mechanism to facilitate in-context learning in the diffusion model and the duration/pitch predictor. We scale NaturalSpeech 2 to large-scale datasets with 44K hours of speech and singing data and evaluate its voice quality on unseen speakers. NaturalSpeech 2 outperforms previous TTS systems by a large margin in terms of prosody/timbre similarity, robustness, and voice quality in a zero-shot setting, and performs novel zero-shot singing synthesis with only a speech prompt. Audio samples are available at https://naturalspeech2.github.io/.", "pdf": "https://openreview.net/pdf/509cda476b8eb36072d873b3fb7a5b5868bb7ce7.pdf"} {"title": "Submodular Reinforcement Learning", "url": "https://openreview.net/forum?id=loYSzjSaAK", "detail_url": "https://openreview.net/forum?id=loYSzjSaAK", "authors": "Manish Prajapat,Mojmir Mutny,Melanie Zeilinger,Andreas Krause", "tags": "ICLR 2024,Spotlight", "abstract": "In reinforcement learning (RL), rewards of states are typically considered additive, and following the Markov assumption, they are independent of states visited previously. In many important applications, such as coverage control, experiment design and informative path planning, rewards naturally have diminishing returns, i.e., their value decreases in light of similar states visited previously. To tackle this, we propose Submodular RL (subRL), a paradigm which seeks to optimize more general, non-additive (and history-dependent) rewards modelled via submodular set functions, which capture diminishing returns. Unfortunately, in general, even in tabular settings, we show that the resulting optimization problem is hard to approximate. On the other hand, motivated by the success of greedy algorithms in classical submodular optimization, we propose subPO, a simple policy gradient-based algorithm for subRL that handles non-additive rewards by greedily maximizing marginal gains. Indeed, under some assumptions on the underlying Markov Decision Process (MDP), subPO recovers optimal constant factor approximations of submodular bandits. Moreover, we derive a natural policy gradient approach for locally optimizing subRL instances even in large state- and action- spaces. We showcase the versatility of our approach by applying subPO to several applications, such as biodiversity monitoring, Bayesian experiment design, informative path planning, and coverage maximization. Our results demonstrate sample efficiency, as well as scalability to high-dimensional state-action spaces.", "pdf": "https://openreview.net/pdf/8fc77d8529744661d87719d0416984370812942f.pdf"} {"title": "Making Pre-trained Language Models Great on Tabular Prediction", "url": "https://openreview.net/forum?id=anzIzGZuLi", "detail_url": "https://openreview.net/forum?id=anzIzGZuLi", "authors": "Jiahuan Yan,Bo Zheng,Hongxia Xu,Yiheng Zhu,Danny Chen,Jimeng Sun,Jian Wu,Jintai Chen", "tags": "ICLR 2024,Spotlight", "abstract": "The transferability of deep neural networks (DNNs) has made significant progress in image and language processing. However, due to the heterogeneity among tables, such DNN bonus is still far from being well exploited on tabular data prediction (e.g., regression or classification tasks). Condensing knowledge from diverse domains, language models (LMs) possess the capability to comprehend feature names from various tables, potentially serving as versatile learners in transferring knowledge across distinct tables and diverse prediction tasks, but their discrete text representation space is inherently incompatible with numerical feature values in tables. In this paper, we present TP-BERTa, a specifically pre-trained LM for tabular data prediction. Concretely, a novel relative magnitude tokenization converts scalar numerical feature values to finely discrete, high-dimensional tokens, and an intra-feature attention approach integrates feature values with the corresponding feature names. Comprehensive experiments demonstrate that our pre-trained TP-BERTa leads the performance among tabular DNNs and is competitive with Gradient Boosted Decision Tree models in typical tabular data regime.", "pdf": "https://openreview.net/pdf/c4a9c6bae09d696686e4f491b7316a399127722b.pdf"} {"title": "Solving Inverse Problems with Latent Diffusion Models via Hard Data Consistency", "url": "https://openreview.net/forum?id=j8hdRqOUhN", "detail_url": "https://openreview.net/forum?id=j8hdRqOUhN", "authors": "Bowen Song,Soo Min Kwon,Zecheng Zhang,Xinyu Hu,Qing Qu,Liyue Shen", "tags": "ICLR 2024,Spotlight", "abstract": "Latent diffusion models have been demonstrated to generate high-quality images, while offering efficiency in model training compared to diffusion models operating in the pixel space. However, incorporating latent diffusion models to solve inverse problems remains a challenging problem due to the nonlinearity of the encoder and decoder. To address these issues, we propose ReSample, an algorithm that can solve general inverse problems with pre-trained latent diffusion models. Our algorithm incorporates data consistency by solving an optimization problem during the reverse sampling process, a concept that we term as hard data consistency. Upon solving this optimization problem, we propose a novel resampling scheme to map the measurement-consistent sample back onto the noisy data manifold and theoretically demonstrate its benefits. Lastly, we apply our algorithm to solve a wide range of linear and nonlinear inverse problems in both natural and medical images, demonstrating that our approach outperforms existing state-of-the-art approaches, including those based on pixel-space diffusion models.", "pdf": "https://openreview.net/pdf/da11a915f62958de563c258cf1a15b945a4040f0.pdf"} {"title": "The False Promise of Imitating Proprietary Language Models", "url": "https://openreview.net/forum?id=Kz3yckpCN5", "detail_url": "https://openreview.net/forum?id=Kz3yckpCN5", "authors": "Arnav Gudibande,Eric Wallace,Charlie Victor Snell,Xinyang Geng,Hao Liu,Pieter Abbeel,Sergey Levine,Dawn Song", "tags": "ICLR 2024,Spotlight", "abstract": "An emerging method to cheaply improve a weaker language model is to finetune it on outputs from a stronger model, such as a proprietary system like ChatGPT (e.g., Alpaca, Self-Instruct, and others). In this work, we critically analyze this approach of imitating language models. We first finetune a series of LMs that imitate ChatGPT using varying base model sizes (1.5B--13B), data sources, and imitation data amounts (0.3M--150M tokens). We then evaluate the models using crowd raters and canonical NLP benchmarks. Initially, we were surprised by the output quality of our imitation models---they appear far better at following instructions, and crowd workers rate their outputs as competitive with ChatGPT. However, when conducting more targeted automatic evaluations, we find that imitation models close little to none of the gap from the base LM to ChatGPT on tasks that are not heavily supported in the imitation data. We show that these performance discrepancies may slip past human raters because imitation models are adept at mimicking ChatGPT\u2019s style but not its factuality. Overall, we conclude that while model imitation can be useful for training models to follow instructions and avoid toxic outputs, it falls short its full promise in many ways. In particular, there exists a substantial capabilities gap between open and closed LMs that we find cannot be bridged merely by adding more imitation data. Instead, we find that fine-tuning more capable base LMs has a significantly more substantial effect on closing this gap. In turn, we argue that the higher leverage action for improving open-source models is to tackle the difficult challenge of developing better base LMs, rather than taking the shortcut of imitating proprietary systems.", "pdf": "https://openreview.net/pdf/b4739f190faa45d202f9d7847e19ebde7844eb25.pdf"} {"title": "Sample-Efficient Linear Representation Learning from Non-IID Non-Isotropic Data", "url": "https://openreview.net/forum?id=Tr3fZocrI6", "detail_url": "https://openreview.net/forum?id=Tr3fZocrI6", "authors": "Thomas TCK Zhang,Leonardo Felipe Toso,James Anderson,Nikolai Matni", "tags": "ICLR 2024,Spotlight", "abstract": "A powerful concept behind much of the recent progress in machine learning is the extraction of common features across data from heterogeneous sources or tasks. Intuitively, using all of one's data to learn a common representation function benefits both computational effort and statistical generalization by leaving a smaller number of parameters to fine-tune on a given task. Toward theoretically grounding these merits, we propose a general setting of recovering linear operators $M$\nfrom noisy vector measurements $y = Mx + w$, where the covariates $x$ may be both non-i.i.d. and non-isotropic. We demonstrate that existing isotropy-agnostic meta-learning approaches incur biases on the representation update, which causes the scaling of the noise terms to lose favorable dependence on the number of source tasks. This in turn can cause the sample complexity of representation learning to be bottlenecked by the single-task data size. We introduce an adaptation, $\\texttt{De-bias}$ & $\\texttt{Feature-Whiten}$ ($\\texttt{DFW}$), of the popular alternating minimization-descent (AMD) scheme proposed in Collins et al., (2021), and establish linear convergence to the optimal representation with noise level scaling down with the $\\textit{total}$ source data size. This leads to generalization bounds on the same order as an oracle empirical risk minimizer. We verify the vital importance of $\\texttt{DFW}$ on various numerical simulations. In particular, we show that vanilla alternating-minimization descent fails catastrophically even for iid, but mildly non-isotropic data.\nOur analysis unifies and generalizes prior work, and provides a flexible framework for a wider range of applications, such as in controls and dynamical systems.", "pdf": "https://openreview.net/pdf/e7dccb9a39e6905c3e57dd5f906c31fbd3cab350.pdf"} {"title": "Information Retention via Learning Supplemental Features", "url": "https://openreview.net/forum?id=o83eu4H9Mb", "detail_url": "https://openreview.net/forum?id=o83eu4H9Mb", "authors": "Zhipeng Xie,Yahe Li", "tags": "ICLR 2024,Spotlight", "abstract": "The information bottleneck principle provides an information-theoretic method for learning a good representation as a trade-off between conciseness and predictive ability, which can reduce information redundancy, eliminate irrelevant and superfluous features, and thus enhance the in-domain generalizability. However, in low-resource or out-of-domain scenarios where the assumption of i.i.d does not necessarily hold true, superfluous (or redundant) relevant features may be supplemental to the mainline features of the model, and be beneficial in making prediction for test dataset with distribution shift. Therefore, instead of squeezing the input information by information bottleneck, we propose to keep as much relevant information as possible in use for making predictions. A three-stage supervised learning framework is designed and implemented to jointly learn the mainline and supplemental features, relieving supplemental features from the suppression of mainline features. Extensive experiments have shown that the learned representations of our method have good in-domain and out-of-domain generalization abilities, especially in low-resource cases.", "pdf": "https://openreview.net/pdf/7f425cda0816de3fc282c27ce87697f5b5c44077.pdf"} {"title": "Mayfly: a Neural Data Structure for Graph Stream Summarization", "url": "https://openreview.net/forum?id=n7Sr8SW4bn", "detail_url": "https://openreview.net/forum?id=n7Sr8SW4bn", "authors": "Yuan Feng,Yukun Cao,Wang Hairu,Xike Xie,S Kevin Zhou", "tags": "ICLR 2024,Spotlight", "abstract": "A graph is a structure made up of vertices and edges used to represent complex relationships between entities, while a graph stream is a continuous flow of graph updates that convey evolving relationships between entities. The massive volume and high dynamism of graph streams promote research on data structures of graph summarization, which provides a concise and approximate view of graph streams with sub-linear space and linear construction time, enabling real-time graph analytics in various domains, such as social networking, financing, and cybersecurity.\nIn this work, we propose the Mayfly, the first neural data structure for summarizing graph streams. The Mayfly replaces handcrafted data structures with better accuracy and adaptivity.\nTo cater to practical applications, Mayfly incorporates two offline training phases.\nDuring the larval phase, the Mayfly learns basic summarization abilities from automatically and synthetically constituted meta-tasks, and in the metamorphosis phase, it rapidly adapts to real graph streams via meta-tasks.\nWith specific configurations of information pathways, the Mayfly enables flexible support for miscellaneous graph queries, including edge, node, and connectivity queries.\nExtensive empirical studies show that the Mayfly significantly outperforms its handcrafted competitors.", "pdf": "https://openreview.net/pdf/e7d88ca4c807b194ba332ff83811bbd8c79934bc.pdf"} {"title": "Exploring the Common Appearance-Boundary Adaptation for Nighttime Optical Flow", "url": "https://openreview.net/forum?id=776lhoaulC", "detail_url": "https://openreview.net/forum?id=776lhoaulC", "authors": "Hanyu Zhou,Yi Chang,Haoyue Liu,YAN WENDING,Yuxing Duan,Zhiwei Shi,Luxin Yan", "tags": "ICLR 2024,Spotlight", "abstract": "We investigate a challenging task of nighttime optical flow, which suffers from weakened texture and amplified noise. These degradations weaken discriminative visual features, thus causing invalid motion feature matching. Typically, existing methods employ domain adaptation to transfer knowledge from auxiliary domain to nighttime domain in either input visual space or output motion space. However, this direct adaptation is ineffective, since there exists a large domain gap due to the intrinsic heterogeneous nature of the feature representations between auxiliary and nighttime domains. To overcome this issue, we explore a common-latent space as the intermediate bridge to reinforce the feature alignment between auxiliary and nighttime domains. In this work, we exploit two auxiliary daytime and event domains, and propose a novel common appearance-boundary adaptation framework for nighttime optical flow. In appearance adaptation, we employ the intrinsic image decomposition to embed the auxiliary daytime image and the nighttime image into a reflectance-aligned common space. We discover that motion distributions of the two reflectance maps are very similar, benefiting us to consistently transfer motion appearance knowledge from daytime to nighttime domain. In boundary adaptation, we theoretically derive the motion correlation formula between nighttime image and accumulated events within a spatiotemporal gradient-aligned common space. We figure out that the correlation of the two spatiotemporal gradient maps shares significant discrepancy, benefitting us to contrastively transfer boundary knowledge from event to nighttime domain. Moreover, appearance adaptation and boundary adaptation are complementary to each other, since they could jointly transfer global motion and local boundary knowledge to the nighttime domain. Extensive experiments have been performed to verify the superiority of the proposed method.", "pdf": "https://openreview.net/pdf/e3328346222dfff5580d6899ec8ffbac04ef6de9.pdf"} {"title": "Graphical Multioutput Gaussian Process with Attention", "url": "https://openreview.net/forum?id=6N8TW504aa", "detail_url": "https://openreview.net/forum?id=6N8TW504aa", "authors": "Yijue Dai,Wenzhong Yan,Feng Yin", "tags": "ICLR 2024,Spotlight", "abstract": "Integrating information while recognizing dependence from multiple data sources and enhancing the predictive performance of the multi-output regression are challenging tasks. Multioutput Gaussian Process (MOGP) methods offer outstanding solutions with tractable predictions and uncertainty quantification. However, their practical applications are hindered by high computational complexity and storage demand. Additionally, there exist model mismatches in existing MOGP models when dealing with non-Gaussian data. To improve the model representation ability in terms of flexibility, optimality, and scalability, this paper introduces a novel multi-output regression framework, termed Graphical MOGP (GMOGP), which is empowered by: (i) Generating flexible Gaussian process priors consolidated from dentified parents, (ii) providing dependent processes with attention-based graphical representations, and (iii) achieving Pareto optimal solutions of kernel hyperparameters via a distributed learning framework. Numerical results confirm that the proposed GMOGP significantly outperforms state-of-the-art MOGP alternatives in predictive performance, as well as in time and memory efficiency, across various synthetic and real datasets.", "pdf": "https://openreview.net/pdf/0b6ac06d4a4184388fc33af01e76741a7603c341.pdf"} {"title": "Soft Contrastive Learning for Time Series", "url": "https://openreview.net/forum?id=pAsQSWlDUf", "detail_url": "https://openreview.net/forum?id=pAsQSWlDUf", "authors": "Seunghan Lee,Taeyoung Park,Kibok Lee", "tags": "ICLR 2024,Spotlight", "abstract": "Contrastive learning has shown to be effective to learn representations from time series in a self-supervised way.\nHowever, contrasting similar time series instances or values from adjacent timestamps within a time series leads to ignore their inherent correlations, which results in deteriorating the quality of learned representations.\nTo address this issue, we propose \\textit{SoftCLT}, a simple yet effective soft contrastive learning strategy for time series.\nThis is achieved by introducing instance-wise and temporal contrastive loss with soft assignments ranging from zero to one.\nSpecifically, we define soft assignments for 1) instance-wise contrastive loss by distance between time series on the data space, warping and 2) temporal contrastive loss by the difference of timestamps.\nSoftCLT is a plug-and-play method for time series contrastive learning that improves the quality of learned representations without bells and whistles.\nIn experiments, we demonstrate that SoftCLT consistently improves the performance in various downstream tasks including classification, semi-supervised learning, transfer learning, and anomaly detection, showing state-of-the-art performance.\nCode is available at this repository: https://github.com/seunghan96/softclt.", "pdf": "https://openreview.net/pdf/310a449b3f99f247f4a3f30cd2a2f8806296770d.pdf"} {"title": "Enhancing Group Fairness in Online Settings Using Oblique Decision Forests", "url": "https://openreview.net/forum?id=E1NxN5QMOE", "detail_url": "https://openreview.net/forum?id=E1NxN5QMOE", "authors": "Somnath Basu Roy Chowdhury,Nicholas Monath,Ahmad Beirami,Rahul Kidambi,Kumar Avinava Dubey,Amr Ahmed,Snigdha Chaturvedi", "tags": "ICLR 2024,Spotlight", "abstract": "Fairness, especially group fairness, is an important consideration in the context of machine learning systems. The most commonly adopted group fairness-enhancing techniques are in-processing methods that rely on a mixture of a fairness objective (e.g., demographic parity) and a task-specific objective (e.g., cross-entropy) during the training process. However, when data arrives in an online fashion \u2013 one instance at a time \u2013 optimizing such fairness objectives poses several challenges. In particular, group fairness objectives are defined using expectations of predictions across different demographic groups. In the online setting, where the algorithm has access to a single instance at a time, estimating the group fairness objective requires additional storage and significantly more computation (e.g., forward/backward passes) than the task-specific objective at every time step. In this paper, we propose Aranyani, an ensemble of oblique decision trees, to make fair decisions in online settings. The hierarchical tree structure of Aranyani enables parameter isolation and allows us to efficiently compute the fairness gradients using aggregate statistics of previous decisions, eliminating the need for additional storage and forward/backward passes. We also present an efficient framework to train Aranyani and theoretically analyze several of its properties. We conduct empirical evaluations on 5 publicly available benchmarks (including vision and language datasets) to show that Aranyani achieves a better accuracy-fairness trade-off compared to baseline approaches.", "pdf": "https://openreview.net/pdf/a8b785960a7be6f38289cdad3923ad1ba27c3a26.pdf"} {"title": "Generative Learning for Financial Time Series with Irregular and Scale-Invariant Patterns", "url": "https://openreview.net/forum?id=CdjnzWsQax", "detail_url": "https://openreview.net/forum?id=CdjnzWsQax", "authors": "Hongbin Huang,Minghua Chen,Xiao Qiao", "tags": "ICLR 2024,Spotlight", "abstract": "Limited data availability poses a major obstacle in training deep learning models for financial applications. Synthesizing financial time series to augment real-world data is challenging due to the irregular and scale-invariant patterns uniquely associated with financial time series - temporal dynamics that repeat with varying duration and magnitude. Such dynamics cannot be captured by existing approaches, which often assume regularity and uniformity in the underlying data. We develop a novel generative framework called FTS-Diffusion to model irregular and scale-invariant patterns that consists of three modules. First, we develop a scale-invariant pattern recognition algorithm to extract recurring patterns that vary in duration and magnitude. Second, we construct a diffusion-based generative network to synthesize segments of patterns. Third, we model the temporal transition of patterns in order to aggregate the generated segments. Extensive experiments show that FTS-Diffusion generates synthetic financial time series highly resembling observed data, outperforming state-of-the-art alternatives. Two downstream experiments demonstrate that augmenting real-world data with synthetic data generated by FTS-Diffusion reduces the error of stock market prediction by up to 17.9%. To the best of our knowledge, this is the first work on generating intricate time series with irregular and scale-invariant patterns, addressing data limitation issues in finance.", "pdf": "https://openreview.net/pdf/afa4bb323e04cbec65604b1a8df0f2eebc2962f3.pdf"} {"title": "Multiscale Positive-Unlabeled Detection of AI-Generated Texts", "url": "https://openreview.net/forum?id=5Lp6qU9hzV", "detail_url": "https://openreview.net/forum?id=5Lp6qU9hzV", "authors": "Yuchuan Tian,Hanting Chen,Xutao Wang,Zheyuan Bai,QINGHUA ZHANG,Ruifeng Li,Chao Xu,Yunhe Wang", "tags": "ICLR 2024,Spotlight", "abstract": "Recent releases of Large Language Models (LLMs), e.g. ChatGPT, are astonishing at generating human-like texts, but they may impact the authenticity of texts. Previous works proposed methods to detect these AI-generated texts, including simple ML classifiers, pretrained-model-based zero-shot methods, and finetuned language classification models. However, mainstream detectors always fail on short texts, like SMSes, Tweets, and reviews. In this paper, a Multiscale Positive-Unlabeled (MPU) training framework is proposed to address the difficulty of short-text detection without sacrificing long-texts. Firstly, we acknowledge the human-resemblance property of short machine texts, and rephrase AI text detection as a partial Positive-Unlabeled (PU) problem by regarding these short machine texts as partially \"unlabeled\". Then in this PU context, we propose the length-sensitive Multiscale PU Loss, where a recurrent model in abstraction is used to estimate positive priors of scale-variant corpora. Additionally, we introduce a Text Multiscaling module to enrich training corpora. Experiments show that our MPU method augments detection performance on long AI-generated texts, and significantly improves short-text detection of language model detectors. Language Models trained with MPU could outcompete existing detectors on various short-text and long-text detection benchmarks. The codes are available at https://github.com/mindspore-lab/mindone/tree/master/examples/detect_chatgpt and https://github.com/YuchuanTian/AIGC_text_detector.", "pdf": "https://openreview.net/pdf/bd6826c79f81e0e0ac6f4c84f2b46d80eb3d130b.pdf"} {"title": "A Lightweight Method for Tackling Unknown Participation Statistics in Federated Averaging", "url": "https://openreview.net/forum?id=ZKEuFKfCKA", "detail_url": "https://openreview.net/forum?id=ZKEuFKfCKA", "authors": "Shiqiang Wang,Mingyue Ji", "tags": "ICLR 2024,Spotlight", "abstract": "In federated learning (FL), clients usually have diverse participation statistics that are unknown a priori, which can significantly harm the performance of FL if not handled properly. Existing works aiming at addressing this problem are usually based on global variance reduction, which requires a substantial amount of additional memory in a multiplicative factor equal to the total number of clients. An important open problem is to find a lightweight method for FL in the presence of clients with unknown participation rates. In this paper, we address this problem by adapting the aggregation weights in federated averaging (FedAvg) based on the participation history of each client. We first show that, with heterogeneous participation statistics, FedAvg with non-optimal aggregation weights can diverge from the optimal solution of the original FL objective, indicating the need of finding optimal aggregation weights. However, it is difficult to compute the optimal weights when the participation statistics are unknown. To address this problem, we present a new algorithm called FedAU, which improves FedAvg by adaptively weighting the client updates based on online estimates of the optimal weights without knowing the statistics of client participation. We provide a theoretical convergence analysis of FedAU using a novel methodology to connect the estimation error and convergence. Our theoretical results reveal important and interesting insights, while showing that FedAU converges to an optimal solution of the original objective and has desirable properties such as linear speedup. Our experimental results also verify the advantage of FedAU over baseline methods with various participation patterns.", "pdf": "https://openreview.net/pdf/deb3da6004c1c25ab01ed64fde43ebe424d7a09c.pdf"} {"title": "Identifying the Risks of LM Agents with an LM-Emulated Sandbox", "url": "https://openreview.net/forum?id=GEcwtMk1uA", "detail_url": "https://openreview.net/forum?id=GEcwtMk1uA", "authors": "Yangjun Ruan,Honghua Dong,Andrew Wang,Silviu Pitis,Yongchao Zhou,Jimmy Ba,Yann Dubois,Chris J. Maddison,Tatsunori Hashimoto", "tags": "ICLR 2024,Spotlight", "abstract": "Recent advances in Language Model (LM) agents and tool use, exemplified by applications like ChatGPT Plugins, enable a rich set of capabilities but also amplify potential risks\u2014such as leaking private data or causing financial losses. Identifying these risks is labor-intensive, necessitating implementing the tools, setting up the environment for each test scenario manually, and finding risky cases. As tools and agents become more complex, the high cost of testing these agents will make it increasingly difficult to find high-stakes, long-tail risks. To address these challenges, we introduce ToolEmu: a framework that uses an LM to emulate tool execution and enables scalable testing of LM agents against a diverse range of tools and scenarios. Alongside the emulator, we develop an LM-based automatic safety evaluator that examines agent failures and quantifies associated risks. We test both the tool emulator and evaluator through human evaluation and find that 68.8% of failures identified with ToolEmu would be valid real-world agent failures. Using our curated initial benchmark consisting of 36 high-stakes toolkits and 144 test cases, we provide a quantitative risk analysis of current LM agents and identify numerous failures with potentially severe outcomes. Notably, even the safest LM agent exhibits such failures 23.9% of the time according to our evaluator, underscoring the need to develop safer LM agents for real-world deployment.", "pdf": "https://openreview.net/pdf/d1601f78407737fc216de9e6ec0085038f8c885f.pdf"} {"title": "Coeditor: Leveraging Repo-level Diffs for Code Auto-editing", "url": "https://openreview.net/forum?id=ALVwQjZRS8", "detail_url": "https://openreview.net/forum?id=ALVwQjZRS8", "authors": "Jiayi Wei,Greg Durrett,Isil Dillig", "tags": "ICLR 2024,Spotlight", "abstract": "Developers often dedicate significant time to maintaining and refactoring existing code. However, most prior work on generative models for code focuses solely on creating new code, overlooking the distinctive needs of editing existing code. In this work, we explore a multi-round code auto-editing setting, aiming to predict edits to a code region based on recent changes within the same codebase. Our model, Coeditor, is a fine-tuned language model specifically designed for code editing tasks. We represent code changes using a line diff format and employ static analysis to form large customized model contexts, ensuring the availability of appropriate information for prediction. We collect a code editing dataset from the commit histories of 1650 open-source Python projects for training and evaluation. In a simplified single-round, single-edit task, Coeditor significantly outperforms GPT-3.5 and SOTA open-source code completion models (bringing exact-match accuracy from 34.7 up to 60.4), demonstrating the benefits of incorporating editing history for code completion. In a multi-round, multi-edit setting, we observe substantial gains by iteratively conditioning on additional user edits. We have open-sourced our code, data, and model weights to encourage future research and have released a VSCode extension powered by our model for interactive IDE usage.", "pdf": "https://openreview.net/pdf/a68ee5b156d07bd4d39e7718b01a1ecdc5b5c3cb.pdf"} {"title": "FITS: Modeling Time Series with $10k$ Parameters", "url": "https://openreview.net/forum?id=bWcnvZ3qMb", "detail_url": "https://openreview.net/forum?id=bWcnvZ3qMb", "authors": "Zhijian Xu,Ailing Zeng,Qiang Xu", "tags": "ICLR 2024,Spotlight", "abstract": "In this paper, we introduce FITS, a lightweight yet powerful model for time series analysis. Unlike existing models that directly process raw time-domain data, FITS operates on the principle that time series can be manipulated through interpolation in the complex frequency domain, achieving performance comparable to state-of-the-art models for time series forecasting and anomaly detection tasks. Notably, FITS accomplishes this with a svelte profile of just about $10k$ parameters, making it ideally suited for edge devices and paving the way for a wide range of applications. The code is available for review at: \\url{https://anonymous.4open.science/r/FITS}.", "pdf": "https://openreview.net/pdf/b24cfba5a0bb5ddb925050c72614c266f677f9a0.pdf"} {"title": "MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models", "url": "https://openreview.net/forum?id=N8N0hgNDRt", "detail_url": "https://openreview.net/forum?id=N8N0hgNDRt", "authors": "Longhui Yu,Weisen Jiang,Han Shi,Jincheng YU,Zhengying Liu,Yu Zhang,James Kwok,Zhenguo Li,Adrian Weller,Weiyang Liu", "tags": "ICLR 2024,Spotlight", "abstract": "Large language models (LLMs) have pushed the limits of natural language understanding and exhibited excellent problem-solving ability. Despite the great success, most existing open-source LLMs (\\eg, LLaMA-2) are still far away from satisfactory for solving mathematical problems due to the complex reasoning procedures. To bridge this gap, we propose \\emph{MetaMath}, a finetuned language model that specializes in mathematical reasoning. Specifically, we start by bootstrapping mathematical questions by rewriting the question from multiple perspectives, which results in a new dataset called MetaMathQA. Then we finetune the LLaMA-2 models on MetaMathQA. Experimental results on two popular benchmarks (\\ie, GSM8K and MATH) for mathematical reasoning demonstrate that MetaMath outperforms a suite of open-source LLMs by a significant margin. Our MetaMath-7B model achieves $66.5\\%$ on GSM8K and $19.8\\%$ on MATH, exceeding the state-of-the-art models of the same size by $11.5\\%$ and $8.7\\%$. Particularly, MetaMath-70B achieves an accuracy of $82.3\\%$ on GSM8K, slightly better than GPT-3.5-Turbo. We release the MetaMathQA dataset, the MetaMath models with different model sizes and the training code for public use.", "pdf": "https://openreview.net/pdf/f6e244230affa5173ef87947c86c25bd2891100d.pdf"} {"title": "Query-Policy Misalignment in Preference-Based Reinforcement Learning", "url": "https://openreview.net/forum?id=UoBymIwPJR", "detail_url": "https://openreview.net/forum?id=UoBymIwPJR", "authors": "Xiao Hu,Jianxiong Li,Xianyuan Zhan,Qing-Shan Jia,Ya-Qin Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "Preference-based reinforcement learning (PbRL) provides a natural way to align RL agents\u2019 behavior with human desired outcomes, but is often restrained by costly human feedback. To improve feedback efficiency, most existing PbRL methods focus on selecting queries to maximally improve the overall quality of the reward model, but counter-intuitively, we find that this may not necessarily lead to improved performance. To unravel this mystery, we identify a long-neglected issue in the query selection schemes of existing PbRL studies: Query-Policy Misalignment. We show that the seemingly informative queries selected to improve the overall quality of reward model actually may not align with RL agents\u2019 interests, thus offering little help on policy learning and eventually resulting in poor feedback efficiency. We show that this issue can be effectively addressed via policy-aligned query and a specially designed hybrid experience replay, which together enforce the bidirectional query-policy alignment. Simple yet elegant, our method can be easily incorporated into existing approaches by changing only a few lines of code. We showcase in comprehensive experiments that our method achieves substantial gains in both human feedback and RL sample efficiency, demonstrating the importance of addressing query-policy misalignment in PbRL tasks.", "pdf": "https://openreview.net/pdf/7731e84686a5dad0613dd42e1b05f91259ca9066.pdf"} {"title": "Feature-aligned N-BEATS with Sinkhorn divergence", "url": "https://openreview.net/forum?id=TS8HoIWAPQ", "detail_url": "https://openreview.net/forum?id=TS8HoIWAPQ", "authors": "Joonhun Lee,Myeongho Jeon,Myungjoo Kang,Kyunghyun Park", "tags": "ICLR 2024,Spotlight", "abstract": "We propose Feature-aligned N-BEATS as a domain-generalized time series forecasting model. It is a nontrivial extension of N-BEATS with doubly residual stacking principle (Oreshkin et al. [45]) into a representation learning framework. In particular, it revolves around marginal feature probability measures induced by the intricate composition of residual and feature extracting operators of N-BEATS in each stack and aligns them stack-wise via an approximate of an optimal transport distance referred to as the Sinkhorn divergence. The training loss consists of an empirical risk minimization from multiple source domains, i.e., forecasting loss, and an alignment loss calculated with the Sinkhorn divergence, which allows the model to learn invariant features stack-wise across multiple source data sequences while retaining N-BEATS\u2019s interpretable design and forecasting power. Comprehensive experimental evaluations with ablation studies are provided and the corresponding results demonstrate the proposed model\u2019s forecasting and generalization capabilities.", "pdf": "https://openreview.net/pdf/47a2e8bbe6cc10d3bff9d06eb5871437eede86e2.pdf"} {"title": "Instructive Decoding: Instruction-Tuned Large Language Models are Self-Refiner from Noisy Instructions", "url": "https://openreview.net/forum?id=LebzzClHYw", "detail_url": "https://openreview.net/forum?id=LebzzClHYw", "authors": "Taehyeon Kim,Joonkee Kim,Gihun Lee,Se-Young Yun", "tags": "ICLR 2024,Spotlight", "abstract": "While instruction-tuned language models have demonstrated impressive zero-shot generalization, these models often struggle to generate accurate responses when faced with instructions that fall outside their training set. This paper presents Instructive Decoding (ID), a simple yet effective approach that augments the efficacy of instruction-tuned models. Specifically, ID adjusts the logits for next-token prediction in a contrastive manner, utilizing predictions generated from a manipulated version of the original instruction, referred to as a noisy instruction. This noisy instruction aims to elicit responses that could diverge from the intended instruction yet remain plausible. We conduct experiments across a spectrum of such noisy instructions, ranging from those that insert semantic noise via random words to others like 'opposite' that elicit the deviated responses. Our approach achieves considerable performance gains across various instruction-tuned models and tasks without necessitating any additional parameter updates. Notably, utilizing 'opposite' as the noisy instruction in ID, which shows the maximum divergence from the original instruction, consistently produces the most significant performance gains across multiple models and tasks.", "pdf": "https://openreview.net/pdf/41130f3ca565e158b2e1217fa3f5da2ba15efd6e.pdf"} {"title": "Consistent Multi-Class Classification from Multiple Unlabeled Datasets", "url": "https://openreview.net/forum?id=fW7DOHDQvF", "detail_url": "https://openreview.net/forum?id=fW7DOHDQvF", "authors": "Zixi Wei,Senlin Shu,Yuzhou Cao,Hongxin Wei,Bo An,Lei Feng", "tags": "ICLR 2024,Spotlight", "abstract": "Weakly supervised learning aims to construct effective predictive models from imperfectly labeled data. The recent trend of weakly supervised learning has focused on how to learn an accurate classifier from completely unlabeled data, given little supervised information such as class priors. In this paper, we consider a newly proposed weakly supervised learning problem called multi-class classification from multiple unlabeled datasets, where only multiple sets of unlabeled data and their class priors (i.e., the proportions of each class) are provided for training the classifier. To solve this problem, we first propose a classifier-consistent method (CCM) based on a probability transition matrix. However, CCM cannot guarantee risk consistency and lacks of purified supervision information during training. Therefore, we further propose a risk-consistent method (RCM) that progressively purifies supervision information during training by importance weighting. We provide comprehensive theoretical analyses for our methods to demonstrate the statistical consistency. Experimental results on multiple benchmark datasets and various prior matrices demonstrate the superiority of our proposed methods.", "pdf": "https://openreview.net/pdf/c4cacd1a99a9f6cd491de8f23cdb492b4906cbce.pdf"} {"title": "SpikePoint: An Efficient Point-based Spiking Neural Network for Event Cameras Action Recognition", "url": "https://openreview.net/forum?id=7etoNfU9uF", "detail_url": "https://openreview.net/forum?id=7etoNfU9uF", "authors": "Hongwei Ren,Yue Zhou,Xiaopeng LIN,Yulong Huang,Haotian FU,Jie Song,Bojun Cheng", "tags": "ICLR 2024,Spotlight", "abstract": "Event cameras are bio-inspired sensors that respond to local changes in light intensity and feature low latency, high energy efficiency, and high dynamic range. Meanwhile, Spiking Neural Networks (SNNs) have gained significant attention due to their remarkable efficiency and fault tolerance. By synergistically harnessing the energy efficiency inherent in event cameras and the spike-based processing capabilities of SNNs, their integration could enable ultra-low-power application scenarios, such as action recognition tasks. However, existing approaches often entail converting asynchronous events into conventional frames, leading to additional data mapping efforts and a loss of sparsity, contradicting the design concept of SNNs and event cameras. To address this challenge, we propose SpikePoint, a novel end-to-end point-based SNN architecture. SpikePoint excels at processing sparse event cloud data, effectively extracting both global and local features through a singular-stage structure. Leveraging the surrogate training method, SpikePoint achieves high accuracy with few parameters and maintains low power consumption, specifically employing the identity mapping feature extractor on diverse datasets. SpikePoint achieves state-of-the-art (SOTA) performance on four event-based action recognition datasets using only 16 timesteps, surpassing other SNN methods. Moreover, it also achieves SOTA performance across all methods on three datasets, utilizing approximately 0.3 % of the parameters and 0.5 % of power consumption employed by artificial neural networks (ANNs). These results emphasize the significance of Point Cloud and pave the way for many ultra-low-power event-based data processing applications.", "pdf": "https://openreview.net/pdf/1d9c8889139a212409d5faf9dc557045e96dcc89.pdf"} {"title": "Inverse Approximation Theory for Nonlinear Recurrent Neural Networks", "url": "https://openreview.net/forum?id=yC2waD70Vj", "detail_url": "https://openreview.net/forum?id=yC2waD70Vj", "authors": "Shida Wang,Zhong Li,Qianxiao Li", "tags": "ICLR 2024,Spotlight", "abstract": "We prove an inverse approximation theorem for the approximation of nonlinear sequence-to-sequence relationships using recurrent neural networks (RNNs). This is a so-called Bernstein-type result in approximation theory, which deduces properties of a target function under the assumption that it can be effectively approximated by a hypothesis space. In particular, we show that nonlinear sequence relationships that can be stably approximated by nonlinear RNNs must have an exponential decaying memory structure - a notion that can be made precise. This extends the previously identified curse of memory in linear RNNs into the general nonlinear setting, and quantifies the essential limitations of the RNN architecture for learning sequential relationships with long-term memory. Based on the analysis, we propose a principled reparameterization method to overcome the limitations. Our theoretical results are confirmed by numerical experiments.", "pdf": "https://openreview.net/pdf/a89df38f3e96bab890df4328af64ca3eb34b8df0.pdf"} {"title": "Kernel Metric Learning for In-Sample Off-Policy Evaluation of Deterministic RL Policies", "url": "https://openreview.net/forum?id=plebgsdiiV", "detail_url": "https://openreview.net/forum?id=plebgsdiiV", "authors": "Haanvid Lee,Tri Wahyu Guntara,Jongmin Lee,Yung-Kyun Noh,Kee-Eung Kim", "tags": "ICLR 2024,Spotlight", "abstract": "We consider off-policy evaluation (OPE) of deterministic target policies for reinforcement learning (RL) in environments with continuous action spaces. While it is common to use importance sampling for OPE, it suffers from high variance when the behavior policy deviates significantly from the target policy. In order to address this issue, some recent works on OPE proposed in-sample learning with importance resampling. Yet, these approaches are not applicable to deterministic target policies for continuous action spaces. To address this limitation, we propose to relax the deterministic target policy using a kernel and learn the kernel metrics that minimize the overall mean squared error of the estimated temporal difference update vector of an action value function, where the action value function is used for policy evaluation. We derive the bias and variance of the estimation error due to this relaxation and provide analytic solutions for the optimal kernel metric. In empirical studies using various test domains, we show that the OPE with in-sample learning using the kernel with optimized metric achieves significantly improved accuracy than other baselines.", "pdf": "https://openreview.net/pdf/24aac816499705b0d6a509f4908ba2e27ed10775.pdf"} {"title": "Large Language Models are Efficient Learners of Noise-Robust Speech Recognition", "url": "https://openreview.net/forum?id=ceATjGPTUD", "detail_url": "https://openreview.net/forum?id=ceATjGPTUD", "authors": "Yuchen Hu,CHEN CHEN,Chao-Han Huck Yang,Ruizhe Li,Chao Zhang,Pin-Yu Chen,EngSiong Chng", "tags": "ICLR 2024,Spotlight", "abstract": "Recent advances in large language models (LLMs) have promoted generative error correction (GER) for automatic speech recognition (ASR), which leverages the rich linguistic knowledge and powerful reasoning ability of LLMs to improve recognition results. The latest work proposes a GER benchmark with \"HyPoradise\" dataset to learn the mapping from ASR N-best hypotheses to ground-truth transcription by efficient LLM finetuning, which shows great effectiveness but lacks specificity on noise-robust ASR. In this work, we extend the benchmark to noisy conditions and investigate if we can teach LLMs to perform denoising for GER just like what robust ASR do, where one solution is introducing noise information as a conditioner into LLM. However, directly incorporating noise embeddings from audio encoder could harm the LLM tuning due to cross-modality gap. To this end, we propose to extract a language-space noise embedding from the N-best list to represent the noise conditions of source speech, which can promote the denoising process in GER. Furthermore, in order to enhance its representation ability of audio noise, we design a knowledge distillation (KD) approach via mutual information estimation to distill the real noise information in audio embeddings to our language embedding. Experiments on various latest LLMs demonstrate our approach achieves a new breakthrough with up to 53.9% correction improvement in terms of word error rate while with limited training data. Analysis shows that our language-space noise embedding can well represent the noise conditions of source speech, under which off-the-shelf LLMs show strong ability of language-space denoising.", "pdf": "https://openreview.net/pdf/2403a00daa1fa0949fa21d4c5bb972bd398f4dea.pdf"} {"title": "H2O-SDF: Two-phase Learning for 3D Indoor Reconstruction using Object Surface Fields", "url": "https://openreview.net/forum?id=P1ANzoGg3W", "detail_url": "https://openreview.net/forum?id=P1ANzoGg3W", "authors": "Minyoung Park,Mirae Do,Yeon Jae Shin,Jaeseok Yoo,Jongkwang Hong,Joongrock Kim,Chul Lee", "tags": "ICLR 2024,Spotlight", "abstract": "Advanced techniques using Neural Radiance Fields (NeRF), Signed Distance Fields (SDF), and Occupancy Fields have recently emerged as solutions for 3D indoor scene reconstruction. We introduce a novel two-phase learning approach, H2O-SDF, that discriminates between object and non-object regions within indoor environments. This method achieves a nuanced balance, carefully preserving the geometric integrity of room layouts while also capturing intricate surface details of specific objects. A cornerstone of our two-phase learning framework is the introduction of the Object Surface Field (OSF), a novel concept designed to mitigate the persistent vanishing gradient problem that has previously hindered the capture of high-frequency details in other methods. Our proposed approach is validated through several experiments that include ablation studies.", "pdf": "https://openreview.net/pdf/efccbd7a6e50a44711e740d5009616c9e19fb6e9.pdf"} {"title": "Sample-Efficient Quality-Diversity by Cooperative Coevolution", "url": "https://openreview.net/forum?id=JDud6zbpFv", "detail_url": "https://openreview.net/forum?id=JDud6zbpFv", "authors": "Ke Xue,Ren-Jian Wang,Pengyi Li,Dong Li,Jianye HAO,Chao Qian", "tags": "ICLR 2024,Spotlight", "abstract": "Quality-Diversity (QD) algorithms, as a subset of evolutionary algorithms, have emerged as a powerful optimization paradigm with the aim of generating a set of high-quality and diverse solutions. Although QD has demonstrated competitive performance in reinforcement learning, its low sample efficiency remains a significant impediment for real-world applications. Recent research has primarily focused on augmenting sample efficiency by refining selection and variation operators of QD. However, one of the less considered yet crucial factors is the inherently large-scale issue of the QD optimization problem. In this paper, we propose a novel Cooperative Coevolution QD (CCQD) framework, which decomposes a policy network naturally into two types of layers, corresponding to representation and decision respectively, and thus simplifies the problem significantly. The resulting two (representation and decision) subpopulations are coevolved cooperatively. CCQD can be implemented with different selection and variation operators. Experiments on several popular tasks within the QDAX suite demonstrate that an instantiation of CCQD achieves approximately a 200% improvement in sample efficiency.", "pdf": "https://openreview.net/pdf/fcc91cb60f0dd347bc02c8beadb05d7d55b9f04f.pdf"} {"title": "SILO Language Models: Isolating Legal Risk In a Nonparametric Datastore", "url": "https://openreview.net/forum?id=ruk0nyQPec", "detail_url": "https://openreview.net/forum?id=ruk0nyQPec", "authors": "Sewon Min,Suchin Gururangan,Eric Wallace,Weijia Shi,Hannaneh Hajishirzi,Noah A. Smith,Luke Zettlemoyer", "tags": "ICLR 2024,Spotlight", "abstract": "The legality of training language models (LMs) on copyrighted or otherwise restricted data is under intense debate. However, as we show, model performance significantly degrades if trained only on low-risk text (e.g., out-of-copyright books or government documents), due to its limited size and domain coverage. We present SILO, a new language model that manages this risk-performance tradeoff during inference. SILO is built by (1) training a parametric LM on the Open License Corpus (OLC), a new corpus we curate with 228B tokens of public domain and permissively licensed text and (2) augmenting it with a more general and easily modifiable nonparametric datastore (e.g., containing copyrighted books or news) that is only queried during inference. The datastore allows use of high-risk data without training on it, supports sentence-level data attribution, and enables data producers to opt out from the model by removing content from the store. These capabilities can foster compliance with data-use regulations such as the fair use doctrine in the United States and the GDPR in the European Union. Our experiments show that the parametric LM struggles on its own with domains not covered by OLC. However, access to the datastore greatly improves out of domain performance, closing 90% of the performance gap with an LM trained on the Pile, a more diverse corpus with mostly high-risk text. We also analyze which nonparametric approach works best, where the remaining errors lie, and how performance scales with datastore size. Our results suggest that it is possible to build high quality language models while mitigating legal risk.", "pdf": "https://openreview.net/pdf/34c8fb489d21452c21be4b0037700e7157e99c21.pdf"} {"title": "Dynamic Discounted Counterfactual Regret Minimization", "url": "https://openreview.net/forum?id=6PbvbLyqT6", "detail_url": "https://openreview.net/forum?id=6PbvbLyqT6", "authors": "Hang Xu,Kai Li,Haobo Fu,QIANG FU,Junliang Xing,Jian Cheng", "tags": "ICLR 2024,Spotlight", "abstract": "Counterfactual regret minimization (CFR) is a family of iterative algorithms showing promising results in solving imperfect-information games. Recent novel CFR variants (e.g., CFR+, DCFR) have significantly improved the convergence rate of the vanilla CFR. The key to these CFR variants\u2019 performance is weighting each iteration non-uniformly, i.e., discounting earlier iterations. However, these algorithms use a fixed, manually-specified scheme to weight each iteration, which enormously limits their potential. In this work, we propose Dynamic Discounted CFR (DDCFR), the first equilibrium-finding framework that discounts prior iterations using a dynamic, automatically-learned scheme. We formalize CFR\u2019s iteration process as a carefully designed Markov decision process and transform the discounting scheme learning problem into a policy optimization problem within it. The learned discounting scheme dynamically weights each iteration on the fly using information available at runtime. Experimental results across multiple games demonstrate that DDCFR\u2019s dynamic discounting scheme has a strong generalization ability and leads to faster convergence with improved performance. The code is available at https://github.com/rpSebastian/DDCFR.", "pdf": "https://openreview.net/pdf/336422d3878e37b0144f3b3da58f90bde675aa6a.pdf"} {"title": "GIO: Gradient Information Optimization for Training Dataset Selection", "url": "https://openreview.net/forum?id=3NnfJnbJT2", "detail_url": "https://openreview.net/forum?id=3NnfJnbJT2", "authors": "Dante Everaert,Christopher Potts", "tags": "ICLR 2024,Spotlight", "abstract": "It is often advantageous to train models on a subset of the available train examples, because the examples are of variable quality or because one would like to train with fewer examples, without sacrificing performance. We present Gradient Information Optimization (GIO), a scalable, task-agnostic approach to this data selection problem that requires only a small set of (unlabeled) examples representing a target distribution. GIO begins from a natural, information-theoretic objective that is intractable in practice. Our contribution is in showing that it can be made highly scalable through a simple relaxation of the objective and a highly efficient implementation. In experiments with machine translation, spelling correction, and image recognition, we show that GIO delivers outstanding results with very small train sets. These findings are robust to different representation models and hyperparameters for GIO itself. GIO is task- and domain-agnostic and can be applied out-of-the-box to new datasets and domains. We open source a pip-installable implementation of the algorithm as \"pip install grad-info-opt\".", "pdf": "https://openreview.net/pdf/5ca46b1a1d6645c98d731e33e243896ae32be3d3.pdf"} {"title": "SNIP: Bridging Mathematical Symbolic and Numeric Realms with Unified Pre-training", "url": "https://openreview.net/forum?id=KZSEgJGPxu", "detail_url": "https://openreview.net/forum?id=KZSEgJGPxu", "authors": "Kazem Meidani,Parshin Shojaee,Chandan K. Reddy,Amir Barati Farimani", "tags": "ICLR 2024,Spotlight", "abstract": "In an era where symbolic mathematical equations are indispensable for modeling complex natural phenomena, scientific inquiry often involves collecting observations and translating them into mathematical expressions. Recently, deep learning has emerged as a powerful tool for extracting insights from data. However, existing models typically specialize in either numeric or symbolic domains, and are usually trained in a supervised manner tailored to specific tasks. This approach neglects the substantial benefits that could arise from a task-agnostic multi-modal understanding between symbolic equations and their numeric counterparts. To bridge the gap, we introduce SNIP, a Symbolic-Numeric Integrated Pre-training model, which employs contrastive learning between symbolic and numeric domains, enhancing their mutual similarities in the embeddings. By performing latent space analysis, we observe that SNIP provides cross-domain insights into the representations, revealing that symbolic supervision enhances the embeddings of numeric data and vice versa. We evaluate SNIP across diverse tasks, including symbolic-to-numeric mathematical property prediction and numeric-to-symbolic equation discovery, commonly known as symbolic regression. Results show that SNIP effectively transfers to various tasks, consistently outperforming fully supervised baselines and competing strongly with established task-specific methods, especially in the low data regime scenarios where available data is limited.", "pdf": "https://openreview.net/pdf/1ba8f83f76d43ebf7625a6e87d0060d6361310f6.pdf"} {"title": "Fantastic Gains and Where to Find Them: On the Existence and Prospect of General Knowledge Transfer between Any Pretrained Model", "url": "https://openreview.net/forum?id=m50eKHCttz", "detail_url": "https://openreview.net/forum?id=m50eKHCttz", "authors": "Karsten Roth,Lukas Thede,A. Sophia Koepke,Oriol Vinyals,Olivier J Henaff,Zeynep Akata", "tags": "ICLR 2024,Spotlight", "abstract": "Training deep networks requires various design decisions regarding for instance their architecture, data augmentation, or optimization. In this work, we find these training variations to result in networks learning unique feature sets from the data. Using public model libraries comprising thousands of models trained on canonical datasets like ImageNet, we observe that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other \u2013 independent of overall performance. Given any arbitrary pairing of pretrained models and no external rankings (such as separate test sets, e.g. due to data privacy), we investigate if it is possible to transfer such \"complementary\" knowledge from one model to another without performance degradation \u2013 a task made particularly difficult as additional knowledge can be contained in stronger, equiperformant or weaker models. Yet facilitating robust transfer in scenarios agnostic to pretrained model pairings would unlock auxiliary gains and knowledge fusion from any model repository without restrictions on model and problem specifics - including from weaker, lower-performance models. This work therefore provides an initial, in-depth exploration on the viability of such general-purpose knowledge transfer. Across large-scale experiments, we first reveal the shortcomings of standard knowledge distillation techniques, and then propose a much more general extension through data partitioning for successful transfer between nearly all pretrained models, which we show can also be done unsupervised. Finally, we assess both the scalability and impact of fundamental model properties on successful model-agnostic knowledge transfer.", "pdf": "https://openreview.net/pdf/122f5389127b21435f80c82696204c736a116976.pdf"} {"title": "Robustifying State-space Models for Long Sequences via Approximate Diagonalization", "url": "https://openreview.net/forum?id=DjeQ39QoLQ", "detail_url": "https://openreview.net/forum?id=DjeQ39QoLQ", "authors": "Annan Yu,Arnur Nigmetov,Dmitriy Morozov,Michael W. Mahoney,N. Benjamin Erichson", "tags": "ICLR 2024,Spotlight", "abstract": "State-space models (SSMs) have recently emerged as a framework for learning long-range sequence tasks. An example is the structured state-space sequence (S4) layer, which uses the diagonal-plus-low-rank structure of the HiPPO initialization framework. However, the complicated structure of the S4 layer poses challenges; and, in an effort to address these challenges, models such as S4D and S5 have considered a purely diagonal structure. This choice simplifies the implementation, improves computational efficiency, and allows channel communication. However, diagonalizing the HiPPO framework is itself an ill-posed problem. In this paper, we propose a general solution for this and related ill-posed diagonalization problems in machine learning. We introduce a generic, backward-stable ``perturb-then-diagonalize'' (PTD) methodology, which is based on the pseudospectral theory of non-normal operators, and which may be interpreted as the approximate diagonalization of the non-normal matrices defining SSMs. Based on this, we introduce the S4-PTD and S5-PTD models. Through theoretical analysis of the transfer functions of different initialization schemes, we demonstrate that the S4-PTD/S5-PTD initialization strongly converges to the HiPPO framework, while the S4D/S5 initialization only achieves weak convergences. As a result, our new models show resilience to Fourier-mode noise-perturbed inputs, a crucial property not achieved by the S4D/S5 models. In addition to improved robustness, our S5-PTD model averages 87.6% accuracy on the Long-Range Arena benchmark, demonstrating that the PTD methodology helps to improve the accuracy of deep learning models.", "pdf": "https://openreview.net/pdf/204207dab9f475c4c40ddb4a399f19c5fac72105.pdf"} {"title": "Provable Offline Preference-Based Reinforcement Learning", "url": "https://openreview.net/forum?id=tVMPfEGT2w", "detail_url": "https://openreview.net/forum?id=tVMPfEGT2w", "authors": "Wenhao Zhan,Masatoshi Uehara,Nathan Kallus,Jason D. Lee,Wen Sun", "tags": "ICLR 2024,Spotlight", "abstract": "In this paper, we investigate the problem of offline Preference-based Reinforcement Learning (PbRL) with human feedback where feedback is available in the form of preference between trajectory pairs rather than explicit rewards. Our proposed algorithm consists of two main steps: (1) estimate the implicit reward using Maximum Likelihood Estimation (MLE) with general function approximation from offline data and (2) solve a distributionally robust planning problem over a confidence set around the MLE. We consider the general reward setting where the reward can be defined over the whole trajectory and provide a novel guarantee that allows us to learn any target policy with a polynomial number of samples, as long as the target policy is covered by the offline data. This guarantee is the first of its kind with general function approximation. To measure the coverage of the target policy, we introduce a new single-policy concentrability coefficient, which can be upper bounded by the per-trajectory concentrability coefficient. We also establish lower bounds that highlight the necessity of such concentrability and the difference from standard RL, where state-action-wise rewards are directly observed. We further extend and analyze our algorithm when the feedback is given over action pairs.", "pdf": "https://openreview.net/pdf/ef2a33a9b6e9fd7ea7de7dbba6688f49e0e58206.pdf"} {"title": "Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory", "url": "https://openreview.net/forum?id=gmg7t8b4s0", "detail_url": "https://openreview.net/forum?id=gmg7t8b4s0", "authors": "Niloofar Mireshghallah,Hyunwoo Kim,Xuhui Zhou,Yulia Tsvetkov,Maarten Sap,Reza Shokri,Yejin Choi", "tags": "ICLR 2024,Spotlight", "abstract": "Existing efforts on quantifying privacy implications for large language models (LLMs) solely focus on measuring leakage of training data. In this work, we shed light on the often-overlooked interactive settings where an LLM receives information from multiple sources and generates an output to be shared with other entities, creating the potential of exposing sensitive input data in inappropriate contexts. In these scenarios, humans nat- urally uphold privacy by choosing whether or not to disclose information depending on the context. We ask the question \u201cCan LLMs demonstrate an equivalent discernment and reasoning capability when considering privacy in context?\u201d We propose CONFAIDE, a benchmark grounded in the theory of contextual integrity and designed to identify critical weaknesses in the privacy reasoning capabilities of instruction-tuned LLMs. CONFAIDE consists of four tiers, gradually increasing in complexity, with the final tier evaluating contextual privacy reasoning and theory of mind capabilities. Our experiments show that even commercial models such as GPT-4 and ChatGPT reveal private information in contexts that humans would not, 39% and 57% of the time, respectively, highlighting the urgent need for a new direction of privacy-preserving approaches as we demonstrate a larger underlying problem stemmed in the models\u2019 lack of reasoning capabilities.", "pdf": "https://openreview.net/pdf/915e98b16264c3e1d6d3db0a8d69afc76b90ae14.pdf"} {"title": "Provable Reward-Agnostic Preference-Based Reinforcement Learning", "url": "https://openreview.net/forum?id=yTBXeXdbMf", "detail_url": "https://openreview.net/forum?id=yTBXeXdbMf", "authors": "Wenhao Zhan,Masatoshi Uehara,Wen Sun,Jason D. Lee", "tags": "ICLR 2024,Spotlight", "abstract": "Preference-based Reinforcement Learning (PbRL) is a paradigm in which an RL agent learns to optimize a task using pair-wise preference-based feedback over trajectories, rather than explicit reward signals. While PbRL has demonstrated practical success in fine-tuning language models, existing theoretical work focuses on regret minimization and fails to capture most of the practical frameworks. In this study, we fill in such a gap between theoretical PbRL and practical algorithms by proposing a theoretical reward-agnostic PbRL framework where exploratory trajectories that enable accurate learning of hidden reward functions are acquired before collecting any human feedback. Theoretical analysis demonstrates that our algorithm requires less human feedback for learning the optimal policy under preference-based models with linear parameterization and unknown transitions, compared to the existing theoretical literature. Specifically, our framework can incorporate linear and low-rank MDPs with efficient sample complexity. Additionally, we investigate reward-agnostic RL with action-based comparison feedback and introduce an efficient querying algorithm tailored to this scenario.", "pdf": "https://openreview.net/pdf/cc1b1fb83857ac10b10a13b0a2c7d061594bfdd8.pdf"} {"title": "Unleashing the Potential of Fractional Calculus in Graph Neural Networks with FROND", "url": "https://openreview.net/forum?id=wcka3bd7P4", "detail_url": "https://openreview.net/forum?id=wcka3bd7P4", "authors": "Qiyu Kang,Kai Zhao,Qinxu Ding,Feng Ji,Xuhao Li,Wenfei Liang,Yang Song,Wee Peng Tay", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce the FRactional-Order graph Neural Dynamical network (FROND), a new continuous graph neural network (GNN) framework. Unlike traditional continuous GNNs that rely on integer-order differential equations, FROND employs the Caputo fractional derivative to leverage the non-local properties of fractional calculus. This approach enables the capture of long-term dependencies in feature updates, moving beyond the Markovian update mechanisms in conventional integer-order models and offering enhanced capabilities in graph representation learning. \nWe offer an interpretation of the node feature updating process in FROND from a non-Markovian random walk perspective when the feature updating is particularly governed by a diffusion process.\nWe demonstrate analytically that oversmoothing can be mitigated in this setting.\nExperimentally, we validate the FROND framework by comparing the fractional adaptations of various established integer-order continuous GNNs, demonstrating their consistently improved performance and underscoring the framework's potential as an effective extension to enhance traditional continuous GNNs.\nThe code is available at \\url{https://github.com/zknus/ICLR2024-FROND}.", "pdf": "https://openreview.net/pdf/62b51824b9a914534dd00158380ffd4aa835c48a.pdf"} {"title": "MetaPhysiCa: Improving OOD Robustness in Physics-informed Machine Learning", "url": "https://openreview.net/forum?id=KrWuDiW4Qm", "detail_url": "https://openreview.net/forum?id=KrWuDiW4Qm", "authors": "S Chandra Mouli,Muhammad Alam,Bruno Ribeiro", "tags": "ICLR 2024,Spotlight", "abstract": "A fundamental challenge in physics-informed machine learning (PIML) is the design of robust PIML methods for out-of-distribution (OOD) forecasting tasks. These OOD tasks require learning-to-learn from observations of the same (ODE) dynamical system with different unknown ODE parameters, and demand accurate forecasts even under out-of-support initial conditions and out-of-support ODE parameters. In this work we propose to improve the OOD robustness of PIML via a meta-learning procedure for causal structure discovery. Using three different OOD tasks, we empirically observe that the proposed approach significantly outperforms existing state-of-the-art PIML and deep learning methods (with $2\\times$ to $28\\times$ lower OOD errors).", "pdf": "https://openreview.net/pdf/e8efd660b312112ef0fd22c7e460d8a72eb51253.pdf"} {"title": "Views Can Be Deceiving: Improved SSL Through Feature Space Augmentation", "url": "https://openreview.net/forum?id=mutJBk3ILg", "detail_url": "https://openreview.net/forum?id=mutJBk3ILg", "authors": "Kimia Hamidieh,Haoran Zhang,Swami Sankaranarayanan,Marzyeh Ghassemi", "tags": "ICLR 2024,Spotlight", "abstract": "Supervised learning methods have been found to exhibit inductive biases favoring simpler features. When such features are spuriously correlated with the label, this can result in suboptimal performance on minority subgroups. Despite the growing popularity of methods which learn from unlabeled data, the extent to which these representations rely on spurious features for prediction is unclear. In this work, we explore the impact of spurious features on Self-Supervised Learning (SSL) for visual representation learning. We first empirically show that commonly used augmentations in SSL can cause undesired invariances in the image space, and illustrate this with a simple example. We further show that classical approaches in combating spurious correlations, such as dataset re-sampling during SSL, do not consistently lead to invariant representations. Motivated by these findings, we propose LateTVG to remove spurious information from these representations during pre-training, by regularizing later layers of the encoder via pruning. We find that our method produces representations which outperform the baselines on several benchmarks, without the need for group or label information during SSL.", "pdf": "https://openreview.net/pdf/b3e9f812dd9a2de2308f2211b33b7d419ab89fc1.pdf"} {"title": "Project and Probe: Sample-Efficient Adaptation by Interpolating Orthogonal Features", "url": "https://openreview.net/forum?id=f6CBQYxXvr", "detail_url": "https://openreview.net/forum?id=f6CBQYxXvr", "authors": "Annie S Chen,Yoonho Lee,Amrith Setlur,Sergey Levine,Chelsea Finn", "tags": "ICLR 2024,Spotlight", "abstract": "Transfer learning with a small amount of target data is an effective and common approach to adapting a pre-trained model to distribution shifts. In some situations, target data labels may be expensive to obtain, so we may only have access to a limited number of target data points. To make the most of a very small target dataset, we propose a lightweight, sample-efficient approach that learns a diverse set of features and adapts to a target distribution by interpolating these features. Our approach, Project and Probe (Pro$^2$), first learns a linear projection that maps a pre-trained embedding onto orthogonal directions while being predictive of labels in the source dataset. The goal of this step is to learn a variety of predictive features, so that at least some of them remain useful after distribution shift. Pro$^2$ then learns a linear classifier on top of these projected features using a small target dataset. Theoretically, we find that Pro$^2$ results in more sample-efficient generalization by inducing a favorable bias-variance tradeoff. Our experiments on four datasets, with multiple distribution shift settings for each, show that Pro$^2$ improves performance by 5-15% when given limited target data compared to prior methods such as standard linear probing.", "pdf": "https://openreview.net/pdf/c7ebd7fa822b912a9fa27ca0702572a707ec85e6.pdf"} {"title": "Implicit bias of SGD in $L_2$-regularized linear DNNs: One-way jumps from high to low rank", "url": "https://openreview.net/forum?id=P1aobHnjjj", "detail_url": "https://openreview.net/forum?id=P1aobHnjjj", "authors": "Zihan Wang,Arthur Jacot", "tags": "ICLR 2024,Spotlight", "abstract": "The $L_{2}$-regularized loss of Deep Linear Networks (DLNs) with\nmore than one hidden layers has multiple local minima, corresponding\nto matrices with different ranks. In tasks such as matrix completion,\nthe goal is to converge to the local minimum with the smallest rank\nthat still fits the training data. While rank-underestimating minima\ncan be avoided since they do not fit the data, GD might get\nstuck at rank-overestimating minima. We show that with SGD, there is always a probability to jump\nfrom a higher rank minimum to a lower rank one, but the probability\nof jumping back is zero. More precisely, we define a sequence of sets\n$B_{1}\\subset B_{2}\\subset\\cdots\\subset B_{R}$ so that $B_{r}$\ncontains all minima of rank $r$ or less (and not more) that are absorbing\nfor small enough ridge parameters $\\lambda$ and learning rates $\\eta$:\nSGD has prob. 0 of leaving $B_{r}$, and from any starting point there\nis a non-zero prob. for SGD to go in $B_{r}$.", "pdf": "https://openreview.net/pdf/c809967fdabfa761319f0239253e78d629fa1684.pdf"} {"title": "Whole-Song Hierarchical Generation of Symbolic Music Using Cascaded Diffusion Models", "url": "https://openreview.net/forum?id=sn7CYWyavh", "detail_url": "https://openreview.net/forum?id=sn7CYWyavh", "authors": "Ziyu Wang,Lejun Min,Gus Xia", "tags": "ICLR 2024,Spotlight", "abstract": "Recent deep music generation studies have put much emphasis on long-term generation with structures. However, we are yet to see high-quality, well-structured **whole-song** generation. In this paper, we make the first attempt to model a full music piece under the realization of *compositional hierarchy*. With a focus on symbolic representations of pop songs, we define a hierarchical language, in which each level of hierarchy focuses on the semantics and context dependency at a certain music scope. The high-level languages reveal whole-song form, phrase, and cadence, whereas the low-level languages focus on notes, chords, and their local patterns. A cascaded diffusion model is trained to model the hierarchical language, where each level is conditioned on its upper levels. Experiments and analysis show that our model is capable of generating full-piece music with recognizable global verse-chorus structure and cadences, and the music quality is higher than the baselines. Additionally, we show that the proposed model is *controllable* in a flexible way. By sampling from the interpretable hierarchical languages or adjusting pre-trained external representations, users can control the music flow via various features such as phrase harmonic structures, rhythmic patterns, and accompaniment texture.", "pdf": "https://openreview.net/pdf/36e2505cb773c92384616d6a2cc198d112c0cfab.pdf"} {"title": "Evaluating the Zero-shot Robustness of Instruction-tuned Language Models", "url": "https://openreview.net/forum?id=g9diuvxN6D", "detail_url": "https://openreview.net/forum?id=g9diuvxN6D", "authors": "Jiuding Sun,Chantal Shaib,Byron C Wallace", "tags": "ICLR 2024,Spotlight", "abstract": "Instruction fine-tuning has recently emerged as a promising approach for improving the zero-shot capabilities of Large Language Models (LLMs) on new tasks. This technique has shown particular strength in improving the performance of modestly sized LLMs, sometimes inducing performance competitive with much larger model variants. In this paper, we ask two questions: (1) How sensitive are instruction-tuned models to the particular phrasings of instructions, and, (2) How can we make them more robust to such natural language variation? To answer the former, we collect a set of 319 instructions manually written by NLP practitioners for over 80 unique tasks included in widely used benchmarks, and we evaluate the variance and average performance of these instructions as compared to instruction phrasings observed during instruction fine-tuning. We find that using novel (unobserved) but appropriate instruction phrasings consistently degrades model performance, sometimes substantially so. Further, such natural instructions yield a wide variance in downstream performance, despite their semantic equivalence. Put another way, instruction-tuned models are not especially robust to instruction re-phrasings. \nWe propose a simple method to mitigate this issue by introducing ``soft prompt'' embedding parameters and optimizing these to maximize the similarity between representations of semantically equivalent instructions. We show that this method consistently improves the robustness of instruction-tuned models.", "pdf": "https://openreview.net/pdf/96fd9e94bdfa38510258729a25b2ba3b5aa4064d.pdf"} {"title": "Critical Learning Periods Emerge Even in Deep Linear Networks", "url": "https://openreview.net/forum?id=Aq35gl2c1k", "detail_url": "https://openreview.net/forum?id=Aq35gl2c1k", "authors": "Michael Kleinman,Alessandro Achille,Stefano Soatto", "tags": "ICLR 2024,Spotlight", "abstract": "Critical learning periods are periods early in development where temporary sensory deficits can have a permanent effect on behavior and learned representations. \nDespite the radical differences between biological and artificial networks, critical learning periods have been empirically observed in both systems. This suggests that critical periods may be fundamental to learning and not an accident of biology.\nYet, why exactly critical periods emerge in deep networks is still an open question, and in particular it is unclear whether the critical periods observed in both systems depend on particular architectural or optimization details. To isolate the key underlying factors, we focus on deep linear network models, and show that, surprisingly, such networks also display much of the behavior seen in biology and artificial networks, while being amenable to analytical treatment. We show that critical periods depend on the depth of the model and structure of the data distribution. We also show analytically and in simulations that the learning of features is tied to competition between sources. Finally, we extend our analysis to multi-task learning to show that pre-training on certain tasks can damage the transfer performance on new tasks, and show how this depends on the relationship between tasks and the duration of the pre-training stage. To the best of our knowledge, our work provides the first analytically tractable model that sheds light into why critical learning periods emerge in biological and artificial networks.", "pdf": "https://openreview.net/pdf/62e86f3312ebf0b894a2af8cffc4f37094ff6695.pdf"} {"title": "MOTOR: A Time-to-Event Foundation Model For Structured Medical Records", "url": "https://openreview.net/forum?id=NialiwI2V6", "detail_url": "https://openreview.net/forum?id=NialiwI2V6", "authors": "Ethan Steinberg,Jason Alan Fries,Yizhe Xu,Nigam Shah", "tags": "ICLR 2024,Spotlight", "abstract": "We present a self-supervised, time-to-event (TTE) foundation model called MOTOR (Many Outcome Time Oriented Representations) which is pretrained on timestamped sequences of events in electronic health records (EHR) and health insurance claims. TTE models are used for estimating the probability distribution of the time until a specific event occurs, which is an important task in medical settings. TTE models provide many advantages over classification using fixed time horizons, including naturally handling censored observations, but are challenging to train with limited labeled data. MOTOR addresses this challenge by pretraining on up to 55M patient records (9B clinical events). We evaluate MOTOR's transfer learning performance on 19 tasks, across 3 patient databases (a private EHR system, MIMIC-IV, and Merative claims data). Task-specific models adapted from MOTOR improve time-dependent C statistics by 4.6\\% over state-of-the-art, improve label efficiency by up to 95\\%, and are more robust to temporal distributional shifts. We further evaluate cross-site portability by adapting our MOTOR foundation model for six prediction tasks on the MIMIC-IV dataset, where it outperforms all baselines. MOTOR is the first foundation model for medical TTE predictions and we release a 143M parameter pretrained model for research use at https://huggingface.co/StanfordShahLab/motor-t-base.", "pdf": "https://openreview.net/pdf/4183f6aeee58dddcc690f9265adb21cd0dac6757.pdf"} {"title": "GenSim: Generating Robotic Simulation Tasks via Large Language Models", "url": "https://openreview.net/forum?id=OI3RoHoWAN", "detail_url": "https://openreview.net/forum?id=OI3RoHoWAN", "authors": "Lirui Wang,Yiyang Ling,Zhecheng Yuan,Mohit Shridhar,Chen Bao,Yuzhe Qin,Bailin Wang,Huazhe Xu,Xiaolong Wang", "tags": "ICLR 2024,Spotlight", "abstract": "Collecting large amounts of real-world interaction data to train general robotic policies is often prohibitively expensive, thus motivating the use of simulation data. However, existing methods for data generation have generally focused on scene-level diversity (e.g., object instances and poses) rather than task-level diversity, due to the human effort required to come up with and verify novel tasks. This has made it challenging for policies trained on simulation data to demonstrate significant task-level generalization. In this paper, we propose to automatically generate rich simulation environments and expert demonstrations by exploiting a large language models' (LLM) grounding and coding ability. Our approach, dubbed GenSim, has two modes: goal-directed generation, wherein a target task is given to the LLM and the LLM proposes a task curriculum to solve the target task, and exploratory generation, wherein the LLM bootstraps from previous tasks and iteratively proposes novel tasks that would be helpful in solving more complex tasks. We use GPT4 to expand the existing benchmark by ten times to over 100 tasks, on which we conduct supervised finetuning and evaluate several LLMs including finetuned GPTs and Code Llama on code generation for robotic simulation tasks. Furthermore, we observe that LLMs-generated simulation programs can enhance task-level generalization significantly when used for multitask policy training. We further find that with minimal sim-to-real adaptation, the multitask policies pretrained on GPT4-generated simulation tasks exhibit stronger transfer to unseen long-horizon tasks in the real world and outperform baselines by 25%. See our project website (https://gen-sim.github.io) and demo (https://huggingface.co/spaces/Gen-Sim/Gen-Sim) for visualizations and open-source models and datasets.", "pdf": "https://openreview.net/pdf/d84b32393144549665a7888268a368b1eb84b7c3.pdf"} {"title": "Understanding Augmentation-based Self-Supervised Representation Learning via RKHS Approximation and Regression", "url": "https://openreview.net/forum?id=Ax2yRhCQr1", "detail_url": "https://openreview.net/forum?id=Ax2yRhCQr1", "authors": "Runtian Zhai,Bingbin Liu,Andrej Risteski,J Zico Kolter,Pradeep Kumar Ravikumar", "tags": "ICLR 2024,Spotlight", "abstract": "Data augmentation is critical to the empirical success of modern self-supervised representation learning, such as contrastive learning and masked language modeling.\nHowever, a theoretical understanding of the exact role of the augmentation remains limited.\nRecent work has built the connection between self-supervised learning and the approximation of the top eigenspace of a graph Laplacian operator, suggesting that learning a linear probe atop such representation can be connected to RKHS regression.\nBuilding on this insight, this work delves into a statistical analysis of augmentation-based pretraining.\nStarting from the isometry property, a geometric characterization of the target function given by the augmentation, we disentangle the effects of the model and the augmentation,\nand prove two generalization bounds that are free of model complexity.\nOur first bound works for an arbitrary encoder, and it is the sum of an estimation error bound incurred by fitting a linear probe, and an approximation error bound by RKHS approximation.\nOur second bound specifically addresses the case\nwhere the encoder extracts the top-d eigenspace of a finite-sample-based approximation of the underlying RKHS.\nA key ingredient in our analysis is the *augmentation complexity*,\nwhich we use to quantitatively compare different augmentations and analyze their impact on downstream performance.", "pdf": "https://openreview.net/pdf/2e845de474870e7f97f44a9beeff24e04dd224b6.pdf"} {"title": "Sudden Drops in the Loss: Syntax Acquisition, Phase Transitions, and Simplicity Bias in MLMs", "url": "https://openreview.net/forum?id=MO5PiKHELW", "detail_url": "https://openreview.net/forum?id=MO5PiKHELW", "authors": "Angelica Chen,Ravid Shwartz-Ziv,Kyunghyun Cho,Matthew L Leavitt,Naomi Saphra", "tags": "ICLR 2024,Spotlight", "abstract": "Most interpretability research in NLP focuses on understanding the behavior and features of a fully trained model. However, certain insights into model behavior may only be accessible by observing the trajectory of the training process. We present a case study of syntax acquisition in masked language models (MLMs) that demonstrates how analyzing the evolution of interpretable artifacts throughout training deepens our understanding of emergent behavior. In particular, we study Syntactic Attention Structure (SAS), a naturally emerging property of MLMs wherein specific Transformer heads tend to focus on specific syntactic relations. We identify a brief window in pretraining when models abruptly acquire SAS, concurrent with a steep drop in loss. This breakthrough precipitates the subsequent acquisition of linguistic capabilities. We then examine the causal role of SAS by manipulating SAS during training, and demonstrate that SAS is necessary for the development of grammatical capabilities. We further find that SAS competes with other beneficial traits during training, and that briefly suppressing SAS improves model quality. These findings offer an interpretation of a real-world example of both simplicity bias and breakthrough training dynamics.", "pdf": "https://openreview.net/pdf/8d2d2b9084da09d4b41f5ad2da660350019c5412.pdf"} {"title": "SE(3)-Stochastic Flow Matching for Protein Backbone Generation", "url": "https://openreview.net/forum?id=kJFIH23hXb", "detail_url": "https://openreview.net/forum?id=kJFIH23hXb", "authors": "Joey Bose,Tara Akhound-Sadegh,Guillaume Huguet,Kilian FATRAS,Jarrid Rector-Brooks,Cheng-Hao Liu,Andrei Cristian Nica,Maksym Korablyov,Michael M. Bronstein,Alexander Tong", "tags": "ICLR 2024,Spotlight", "abstract": "The computational design of novel protein structures has the potential to impact numerous scientific disciplines greatly. Toward this goal, we introduce \\foldflow, a series of novel generative models of increasing modeling power based on the flow-matching paradigm over $3\\mathrm{D}$ rigid motions---i.e. the group $\\mathrm{SE(3)}$---enabling accurate modeling of protein backbones. We first introduce $\\text{FoldFlow-Base}$, a simulation-free approach to learning deterministic continuous-time dynamics and matching invariant target distributions on $\\mathrm{SE(3)}$. We next accelerate training by incorporating Riemannian optimal transport to create $\\text{FoldFlow-OT}$, leading to the construction of both more simple and stable flows. Finally, we design \\foldflowsfm, coupling both Riemannian OT and simulation-free training to learn stochastic continuous-time dynamics over $\\mathrm{SE(3)}$. Our family of $\\text{FoldFlow}$, generative models offers several key advantages over previous approaches to the generative modeling of proteins: they are more stable and faster to train than diffusion-based approaches, and our models enjoy the ability to map any invariant source distribution to any invariant target distribution over $\\mathrm{SE(3)}$. Empirically, we validate $\\text{FoldFlow}$, on protein backbone generation of up to $300$ amino acids leading to high-quality designable, diverse, and novel samples.", "pdf": "https://openreview.net/pdf/2ecf8626dae97e88a2770c4d2e119db485d03748.pdf"} {"title": "DP-OPT: Make Large Language Model Your Privacy-Preserving Prompt Engineer", "url": "https://openreview.net/forum?id=Ifz3IgsEPX", "detail_url": "https://openreview.net/forum?id=Ifz3IgsEPX", "authors": "Junyuan Hong,Jiachen T. Wang,Chenhui Zhang,Zhangheng LI,Bo Li,Zhangyang Wang", "tags": "ICLR 2024,Spotlight", "abstract": "Large Language Models (LLMs) have emerged as dominant tools for various tasks, particularly when tailored for a specific target by prompt tuning. Nevertheless, concerns surrounding data privacy present obstacles due to the tuned prompts' dependency on sensitive private information. A practical solution is to host a local LLM and optimize a soft prompt privately using data. Yet, hosting a local model becomes problematic when model ownership is protected. Alternative methods, like sending data to the model's provider for training, intensify these privacy issues facing an untrusted provider. In this paper, we present a novel solution called Differentially-Private Offsite Prompt Tuning (DP-OPT) to address this challenge. Our approach involves tuning a discrete prompt on the client side and then applying it to the desired cloud models. We demonstrate that prompts suggested by LLMs themselves can be transferred without compromising performance significantly. To ensure that the prompts do not leak private information, we introduce the first private prompt generation mechanism, by a differentially-private (DP) ensemble of in-context learning with private demonstrations. With DP-OPT, generating privacy-preserving prompts by Vicuna-7b can yield competitive performance compared to non-private in-context learning on GPT3.5 or local private prompt tuning.\nCodes are available at https://github.com/VITA-Group/DP-OPT.", "pdf": "https://openreview.net/pdf/6dfeb74c7c420594a6132f6bfe094a53dbf73317.pdf"} {"title": "Geographic Location Encoding with Spherical Harmonics and Sinusoidal Representation Networks", "url": "https://openreview.net/forum?id=PudduufFLa", "detail_url": "https://openreview.net/forum?id=PudduufFLa", "authors": "Marc Ru\u00dfwurm,Konstantin Klemmer,Esther Rolf,Robin Zbinden,Devis Tuia", "tags": "ICLR 2024,Spotlight", "abstract": "Learning representations of geographical space is vital for any machine learning model that integrates geolocated data, spanning application domains such as remote sensing, ecology, or epidemiology. Recent work embeds coordinates using sine and cosine projections based on Double Fourier Sphere (DFS) features. These embeddings assume a rectangular data domain even on global data, which can lead to artifacts, especially at the poles. At the same time, little attention has been paid to the exact design of the neural network architectures with which these functional embeddings are combined. This work proposes a novel location encoder for globally distributed geographic data that combines spherical harmonic basis functions, natively defined on spherical surfaces, with sinusoidal representation networks (SirenNets) that can be interpreted as learned Double Fourier Sphere embedding. We systematically evaluate positional embeddings and neural network architectures across various benchmarks and synthetic evaluation datasets. In contrast to previous approaches that require the combination of both positional encoding and neural networks to learn meaningful representations, we show that both spherical harmonics and sinusoidal representation networks are competitive on their own but set state-of-the-art performances across tasks when combined. The model code and experiments are available at https://github.com/marccoru/locationencoder.", "pdf": "https://openreview.net/pdf/11eead9eb25de1cd772e111da1b931604a8fe49a.pdf"} {"title": "A General Framework for User-Guided Bayesian Optimization", "url": "https://openreview.net/forum?id=NjU0jtXcYn", "detail_url": "https://openreview.net/forum?id=NjU0jtXcYn", "authors": "Carl Hvarfner,Frank Hutter,Luigi Nardi", "tags": "ICLR 2024,Spotlight", "abstract": "The optimization of expensive-to-evaluate black-box functions is prevalent in various scientific disciplines. Bayesian optimization is an automatic, general and sample-efficient method to solve these problems with minimal knowledge of the the underlying function dynamics. However, the ability of Bayesian optimization to incorporate prior knowledge or beliefs about the function at hand in order to accelerate the optimization is limited, which reduces its appeal for knowledgeable practitioners with tight budgets. To allow domain experts to customize the optimization routine, we propose ColaBO, the first Bayesian-principled framework for incorporating prior beliefs beyond the typical kernel structure, such as the likely location of the optimizer or the optimal value. The generality of ColaBO makes it applicable across different Monte Carlo acquisition functions and types of user beliefs. We empirically demonstrate ColaBO's ability to substantially accelerate optimization when the prior information is accurate, and to retain approximately default performance when it is misleading.", "pdf": "https://openreview.net/pdf/9fcb921b69956d08aba85b088ea1c5d6c9b8c037.pdf"} {"title": "Lemur: Harmonizing Natural Language and Code for Language Agents", "url": "https://openreview.net/forum?id=hNhwSmtXRh", "detail_url": "https://openreview.net/forum?id=hNhwSmtXRh", "authors": "Yiheng Xu,Hongjin SU,Chen Xing,Boyu Mi,Qian Liu,Weijia Shi,Binyuan Hui,Fan Zhou,Yitao Liu,Tianbao Xie,Zhoujun Cheng,Siheng Zhao,Lingpeng Kong,Bailin Wang,Caiming Xiong,Tao Yu", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce Lemur and Lemur-Chat, openly accessible language models optimized\nfor both natural language and coding capabilities to serve as the backbone\nof versatile language agents. The evolution from language chat models to\nfunctional language agents demands that models not only master human interaction,\nreasoning, and planning but also ensure grounding in the relevant environments.\nThis calls for a harmonious blend of language and coding capabilities\nin the models. Lemur and Lemur-Chat are proposed to address this necessity,\ndemonstrating balanced proficiencies in both domains, unlike existing\nopen-source models that tend to specialize in either. Through meticulous pretraining\nusing a code-intensive corpus and instruction fine-tuning on text and code\ndata, our models achieve state-of-the-art averaged performance across diverse\ntext and coding benchmarks. Comprehensive experiments demonstrate Lemur\u2019s\nsuperiority over existing open-source models and its proficiency across various\nagent tasks involving human communication, tool usage, and interaction under\nfully- and partially- observable environments. The harmonization between natural\nand programming languages enables Lemur-Chat to significantly narrow the\ngap with proprietary models on agent abilities, providing key insights into developing\nadvanced open-source agents adept at reasoning, planning, and operating\nseamlessly across environments. Our model and code have been open-sourced at\nhttps://github.com/OpenLemur/Lemur.", "pdf": "https://openreview.net/pdf/8ef62990871ebf2cac77dc6ea498085f167f070a.pdf"} {"title": "A path-norm toolkit for modern networks: consequences, promises and challenges", "url": "https://openreview.net/forum?id=hiHZVUIYik", "detail_url": "https://openreview.net/forum?id=hiHZVUIYik", "authors": "Antoine Gonon,Nicolas Brisebarre,Elisa Riccietti,R\u00e9mi Gribonval", "tags": "ICLR 2024,Spotlight", "abstract": "This work introduces the first toolkit around path-norms that fully encompasses general DAG ReLU networks with biases, skip connections and any operation based on the extraction of order statistics: max pooling, GroupSort etc.\nThis toolkit notably allows us to establish generalization bounds for modern neural networks that are not only the most widely applicable path-norm based ones, but also recover or beat the sharpest known bounds of this type. \nThese extended path-norms further enjoy the usual benefits of path-norms: ease of computation, invariance under the symmetries of the network, and improved sharpness on layered fully-connected networks compared to the product of operator norms, another complexity measure most commonly used.\n\nThe versatility of the toolkit and its ease of implementation allow us to challenge the concrete promises of path-norm-based generalization bounds, by numerically evaluating the sharpest known bounds for ResNets on ImageNet.", "pdf": "https://openreview.net/pdf/6dba7f474e1381840f1c444d21ab27a1c1a22129.pdf"} {"title": "Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages", "url": "https://openreview.net/forum?id=Kuh5qgCGCp", "detail_url": "https://openreview.net/forum?id=Kuh5qgCGCp", "authors": "Jinyi Hu,Yuan Yao,Chongyi Wang,SHAN WANG,Yinxu Pan,Qianyu Chen,Tianyu Yu,Hanghao Wu,Yue Zhao,Haoye Zhang,Xu Han,Yankai Lin,Jiao Xue,dahai li,Zhiyuan Liu,Maosong Sun", "tags": "ICLR 2024,Spotlight", "abstract": "Recently there has been a significant surge in multimodal learning in terms of both image-to-text and text-to-image generation. However, the success is typically limited to English, leaving other languages largely behind. Building a competitive counterpart in other languages is highly challenging due to the low-resource nature of non-English multimodal data (i.e., lack of large-scale, high-quality image-text data). In this work, we propose MPM, an effective training paradigm for training large multimodal models in low-resource languages. MPM demonstrates that Multilingual language models can Pivot zero-shot Multimodal learning across languages. Specifically, based on a strong multilingual large language model, multimodal models pretrained on English-only image-text data can well generalize to other languages in a (quasi)-zero-shot manner, even surpassing models trained on image-text data in native languages. Taking Chinese as a practice of MPM, we build large multimodal models VisCPM in image-to-text and text-to-image generation, which achieve state-of-the-art (open-source) performance in Chinese. To facilitate future research, we open-source codes and model weights at https://github.com/OpenBMB/VisCPM.", "pdf": "https://openreview.net/pdf/07df702c6aa71499ac1bb0cc1988bd883407f9de.pdf"} {"title": "From Sparse to Soft Mixtures of Experts", "url": "https://openreview.net/forum?id=jxpsAj7ltE", "detail_url": "https://openreview.net/forum?id=jxpsAj7ltE", "authors": "Joan Puigcerver,Carlos Riquelme Ruiz,Basil Mustafa,Neil Houlsby", "tags": "ICLR 2024,Spotlight", "abstract": "Sparse mixture of expert architectures (MoEs) scale model capacity without significant increases in training or inference costs.\nDespite their success, MoEs suffer from a number of issues: training instability, token dropping, inability to scale the number of experts, or ineffective finetuning.\nIn this work, we propose Soft MoE, a fully-differentiable sparse Transformer that addresses these challenges, while maintaining the benefits of MoEs.\nSoft MoE performs an implicit soft assignment by passing different weighted combinations of all input tokens to each expert.\nAs in other MoEs, experts in Soft MoE only process a subset of the (combined) tokens, enabling larger model capacity (and performance) at lower inference cost.\nIn the context of visual recognition, Soft MoE greatly outperforms dense Transformers (ViTs) and popular MoEs (Tokens Choice and Experts Choice).\nSoft MoE scales well: Soft MoE Huge/14 with 128 experts in 16 MoE layers has over 40x more parameters than ViT Huge/14, with only 2% increased inference time, and substantially better quality.", "pdf": "https://openreview.net/pdf/fd68ff38ff599fb1021a7e6add08b00e8fec95b9.pdf"} {"title": "Selective Mixup Fine-Tuning for Optimizing Non-Decomposable Objectives", "url": "https://openreview.net/forum?id=rxVBKhyfSo", "detail_url": "https://openreview.net/forum?id=rxVBKhyfSo", "authors": "Shrinivas Ramasubramanian,Harsh Rangwani,Sho Takemori,Kunal Samanta,Yuhei Umeda,Venkatesh Babu Radhakrishnan", "tags": "ICLR 2024,Spotlight", "abstract": "The rise in internet usage has led to the generation of massive amounts of data, resulting in the adoption of various supervised and semi-supervised machine learning algorithms, which can effectively utilize the colossal amount of data to train models. However, before deploying these models in the real world, these must be strictly evaluated on performance measures like worst-case recall and satisfy constraints such as fairness. We find that current state-of-the-art empirical techniques offer sub-optimal performance on these practical, non-decomposable performance objectives. On the other hand, the theoretical techniques necessitate training a new model from scratch for each performance objective. To bridge the gap, we propose SelMix, a selective mixup-based inexpensive fine-tuning technique for pre-trained models, to optimize for the desired objective. The core idea of our framework is to determine a sampling distribution to perform a mixup of features between samples from particular classes such that it optimizes the given objective. We comprehensively evaluate our technique against the existing empirical and theoretically principled methods on standard benchmark datasets for imbalanced classification. We find that proposed SelMix fine-tuning significantly improves the performance for various practical non-decomposable objectives across benchmarks.", "pdf": "https://openreview.net/pdf/42154f6a78eb07727368d3e4f20969606728ec4b.pdf"} {"title": "NoiseDiffusion: Correcting Noise for Image Interpolation with Diffusion Models beyond Spherical Linear Interpolation", "url": "https://openreview.net/forum?id=6O3Q6AFUTu", "detail_url": "https://openreview.net/forum?id=6O3Q6AFUTu", "authors": "PengFei Zheng,Yonggang Zhang,Zhen Fang,Tongliang Liu,Defu Lian,Bo Han", "tags": "ICLR 2024,Spotlight", "abstract": "Image interpolation based on diffusion models is promising in creating fresh and interesting images. \nAdvanced interpolation methods mainly focus on spherical linear interpolation, where images are encoded into the noise space and then interpolated for denoising to images. \nHowever, existing methods face challenges in effectively interpolating natural images (not generated by diffusion models), thereby restricting their practical applicability. \nOur experimental investigations reveal that these challenges stem from the invalidity of the encoding noise, which may no longer obey the expected noise distribution, e.g., a normal distribution. \nTo address these challenges, we propose a novel approach to correct noise for image interpolation, NoiseDiffusion. Specifically, NoiseDiffusion approaches the invalid noise to the expected distribution by introducing subtle Gaussian noise and introduces a constraint to suppress noise with extreme values. In this context, promoting noise validity contributes to mitigating image artifacts, but the constraint and introduced exogenous noise typically lead to a reduction in signal-to-noise ratio, i.e., loss of original image information. Hence, NoiseDiffusion performs interpolation within the noisy image space and injects raw images into these noisy counterparts to address the challenge of information loss. Consequently, NoiseDiffusion enables us to interpolate natural images without causing artifacts or information loss, thus achieving the best interpolation results.", "pdf": "https://openreview.net/pdf/9bcefae56342d9cecd1e962c0a0c0cab8b325854.pdf"} {"title": "SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis", "url": "https://openreview.net/forum?id=di52zR8xgf", "detail_url": "https://openreview.net/forum?id=di52zR8xgf", "authors": "Dustin Podell,Zion English,Kyle Lacey,Andreas Blattmann,Tim Dockhorn,Jonas M\u00fcller,Joe Penna,Robin Rombach", "tags": "ICLR 2024,Spotlight", "abstract": "We present Stable Diffusion XL (SDXL), a latent diffusion model for text-to-image synthesis. Compared to previous versions of Stable Diffusion, SDXL leverages a three times larger UNet backbone, achieved by significantly increasing the number of attention blocks and including a second text encoder. Further, we design multiple novel conditioning schemes and train SDXL on multiple aspect ratios. To ensure highest quality results, we also introduce a refinement model which is used to improve the visual fidelity of samples generated by SDXL using a post-hoc image-to-image technique. We demonstrate that SDXL improves dramatically over previous versions of Stable Diffusion and achieves results competitive with those of black-box state-of-the-art image generators such as Midjourney.", "pdf": "https://openreview.net/pdf/13e7093b32d96bdf98f6af87ef27a4f89585f067.pdf"} {"title": "Entity-Centric Reinforcement Learning for Object Manipulation from Pixels", "url": "https://openreview.net/forum?id=uDxeSZ1wdI", "detail_url": "https://openreview.net/forum?id=uDxeSZ1wdI", "authors": "Dan Haramati,Tal Daniel,Aviv Tamar", "tags": "ICLR 2024,Spotlight", "abstract": "Manipulating objects is a hallmark of human intelligence, and an important task in domains such as robotics. In principle, Reinforcement Learning (RL) offers a general approach to learn object manipulation. In practice, however, domains with more than a few objects are difficult for RL agents due to the curse of dimensionality, especially when learning from raw image observations. In this work we propose a structured approach for visual RL that is suitable for representing multiple objects and their interaction, and use it to learn goal-conditioned manipulation of several objects. Key to our method is the ability to handle goals with dependencies between the objects (e.g., moving objects in a certain order). We further relate our architecture to the generalization capability of the trained agent, based on a theoretical result for compositional generalization, and demonstrate agents that learn with 3 objects but generalize to similar tasks with over 10 objects. Videos and code are available on the project website: https://sites.google.com/view/entity-centric-rl", "pdf": "https://openreview.net/pdf/2627c85192603f1981d419b5964256fa755a98c8.pdf"} {"title": "Constrained Bi-Level Optimization: Proximal Lagrangian Value Function Approach and Hessian-free Algorithm", "url": "https://openreview.net/forum?id=xJ5N8qrEPl", "detail_url": "https://openreview.net/forum?id=xJ5N8qrEPl", "authors": "Wei Yao,Chengming Yu,Shangzhi Zeng,Jin Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "This paper presents a new approach and algorithm for solving a class of constrained Bi-Level Optimization (BLO) problems in which the lower-level problem involves constraints coupling both upper-level and lower-level variables. Such problems have recently gained significant attention due to their broad applicability in machine learning. However, conventional gradient-based methods unavoidably rely on computationally intensive calculations related to the Hessian matrix. To address this challenge, we devise a smooth proximal Lagrangian value function to handle the constrained lower-level problem. Utilizing this construct, we introduce a single-level reformulation for constrained BLOs that transforms the original BLO problem into an equivalent optimization problem with smooth constraints. Enabled by this reformulation, we develop a Hessian-free gradient-based algorithm\u2014termed proximal Lagrangian Value function-based Hessian-free Bi-level Algorithm (LV-HBA)\u2014that is straightforward to implement in a single loop manner. Consequently, LV-HBA is especially well-suited for machine learning applications. Furthermore, we offer non-asymptotic convergence analysis for LV-HBA, eliminating the need for traditional strong convexity assumptions for the lower-level problem while also being capable of accommodating non-singleton scenarios. Empirical results substantiate the algorithm's superior practical performance.", "pdf": "https://openreview.net/pdf/83a8490d92a68039f334e84d5eef6e970f4315e5.pdf"} {"title": "Inherently Interpretable Time Series Classification via Multiple Instance Learning", "url": "https://openreview.net/forum?id=xriGRsoAza", "detail_url": "https://openreview.net/forum?id=xriGRsoAza", "authors": "Joseph Early,Gavin Cheung,Kurt Cutajar,Hanting Xie,Jas Kandola,Niall Twomey", "tags": "ICLR 2024,Spotlight", "abstract": "Conventional Time Series Classification (TSC) methods are often black boxes that obscure inherent interpretation of their decision-making processes. In this work, we leverage Multiple Instance Learning (MIL) to overcome this issue, and propose a new framework called MILLET: Multiple Instance Learning for Locally Explainable Time series classification. We apply MILLET to existing deep learning TSC models and show how they become inherently interpretable without compromising (and in some cases, even improving) predictive performance. We evaluate MILLET on 85 UCR TSC datasets and also present a novel synthetic dataset that is specially designed to facilitate interpretability evaluation. On these datasets, we show MILLET produces sparse explanations quickly that are of higher quality than other well-known interpretability methods. To the best of our knowledge, our work with MILLET is the first to develop general MIL methods for TSC and apply them to an extensive variety of domains.", "pdf": "https://openreview.net/pdf/b63fbc2c94da98976eead6f787efe4df88f0e8a7.pdf"} {"title": "A Mutual Information Perspective on Federated Contrastive Learning", "url": "https://openreview.net/forum?id=JrmPG9ufKg", "detail_url": "https://openreview.net/forum?id=JrmPG9ufKg", "authors": "Christos Louizos,Matthias Reisser,Denis Korzhenkov", "tags": "ICLR 2024,Spotlight", "abstract": "We investigate contrastive learning in the federated setting through the lens of Sim- CLR and multi-view mutual information maximization. In doing so, we uncover a connection between contrastive representation learning and user verification; by adding a user verification loss to each client\u2019s local SimCLR loss we recover a lower bound to the global multi-view mutual information. To accommodate for the case of when some labelled data are available at the clients, we extend our SimCLR variant to the federated semi-supervised setting. We see that a supervised SimCLR objective can be obtained with two changes: a) the contrastive loss is computed between datapoints that share the same label and b) we require an additional auxiliary head that predicts the correct labels from either of the two views. Along with the proposed SimCLR extensions, we also study how different sources of non-i.i.d.-ness can impact the performance of federated unsupervised learning through global mutual information maximization; we find that a global objective is beneficial for some sources of non-i.i.d.-ness but can be detrimental for others. We empirically evaluate our proposed extensions in various tasks to validate our claims and furthermore demonstrate that our proposed modifications generalize to other pretraining methods.", "pdf": "https://openreview.net/pdf/5dc1a3c5ed8b71611436f05451243c4bea50fd66.pdf"} {"title": "MMD Graph Kernel: Effective Metric Learning for Graphs via Maximum Mean Discrepancy", "url": "https://openreview.net/forum?id=GZ6AcZwA8r", "detail_url": "https://openreview.net/forum?id=GZ6AcZwA8r", "authors": "Yan Sun,Jicong Fan", "tags": "ICLR 2024,Spotlight", "abstract": "This paper focuses on graph metric learning. First, we present a class of maximum mean discrepancy (MMD) based graph kernels, called MMD-GK. These kernels are computed by applying MMD to the node representations of two graphs with message-passing propagation. \nSecondly, we provide a class of deep MMD-GKs that are able to learn graph kernels and implicit graph features adaptively in an unsupervised manner. Thirdly, we propose a class of supervised deep MMD-GKs that are able to utilize label information of graphs and hence yield more discriminative metrics. Besides the algorithms, we provide theoretical analysis for the proposed methods. The proposed methods are evaluated in comparison to many baselines such as graph kernels and graph neural networks in the tasks of graph clustering and graph classification. The numerical results demonstrate the effectiveness and superiority of our methods.", "pdf": "https://openreview.net/pdf/618ed39d01f843e5ce4cab3fd3086ea9061d926b.pdf"} {"title": "SGD Finds then Tunes Features in Two-Layer Neural Networks with near-Optimal Sample Complexity: A Case Study in the XOR problem", "url": "https://openreview.net/forum?id=HgOJlxzB16", "detail_url": "https://openreview.net/forum?id=HgOJlxzB16", "authors": "Margalit Glasgow", "tags": "ICLR 2024,Spotlight", "abstract": "In this work, we consider the optimization process of minibatch stochastic gradient descent (SGD) on a 2-layer neural network with data separated by a quadratic ground truth function. We prove that with data drawn from the Boolean hypercube labeled by the quadratic ``XOR'' function $y = -x_ix_j$ , it is possible to train to a population error $o(1)$\n with $\\Theta(d\\text{polylog}(d))$ samples. Our result considers simultaneously training both layers of the two-layer-neural network with ReLU activations via standard minibatch SGD on the logistic loss. To our knowledge, this work is the first to give a sample complexity of \n for efficiently learning the XOR function on isotropic data on a standard neural network with standard training. Our main technique is showing that the network evolves in two phases: a \\em signal-finding \\em phase where the network is small and many of the neurons evolve independently to find features, and a \\em signal-heavy \\em phase, where SGD maintains and balances the features. We leverage the simultaneous training of the layers to show that it is sufficient for only a small fraction of the neurons to learn features, since those neurons will be amplified by the simultaneous growth of their second layer weights.", "pdf": "https://openreview.net/pdf/d00625c7d5210386b43eec8c623aa8b9a3158712.pdf"} {"title": "DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks", "url": "https://openreview.net/forum?id=gjfOL9z5Xr", "detail_url": "https://openreview.net/forum?id=gjfOL9z5Xr", "authors": "Kaijie Zhu,Jiaao Chen,Jindong Wang,Neil Zhenqiang Gong,Diyi Yang,Xing Xie", "tags": "ICLR 2024,Spotlight", "abstract": "Large language models (LLMs) have achieved remarkable performance in various evaluation benchmarks. However, concerns are raised about potential data contamination in their considerable volume of training corpus. Moreover, the static nature and fixed complexity of current benchmarks may inadequately gauge the advancing capabilities of LLMs. \nIn this paper, we introduce DyVal, a general and flexible protocol for dynamic evaluation of LLMs. Based on our framework, we build graph-informed DyVal by leveraging the structural advantage of directed acyclic graphs to dynamically generate evaluation samples with controllable complexities. DyVal generates challenging evaluation sets on reasoning tasks including mathematics, logical reasoning, and algorithm problems. We evaluate various LLMs ranging from Flan-T5-large to GPT-3.5-Turbo and GPT-4. Experiments show that LLMs perform worse in DyVal-generated evaluation samples with different complexities, highlighting the significance of dynamic evaluation.\nWe also analyze the failure cases and results of different prompting methods.\nMoreover, DyVal-generated samples are not only evaluation sets, but also helpful data for fine-tuning to improve the performance of LLMs on existing benchmarks.\nWe hope that DyVal can shed light on future evaluation research of LLMs. Code is available at: https://github.com/microsoft/promptbench.", "pdf": "https://openreview.net/pdf/997160014cb22fbfafa276c6b83c6cb0ebc92e9f.pdf"} {"title": "Illusory Attacks: Information-theoretic detectability matters in adversarial attacks", "url": "https://openreview.net/forum?id=F5dhGCdyYh", "detail_url": "https://openreview.net/forum?id=F5dhGCdyYh", "authors": "Tim Franzmeyer,Stephen Marcus McAleer,Joao F. Henriques,Jakob Nicolaus Foerster,Philip Torr,Adel Bibi,Christian Schroeder de Witt", "tags": "ICLR 2024,Spotlight", "abstract": "Autonomous agents deployed in the real world need to be robust against adversarial attacks on sensory inputs. \nRobustifying agent policies requires anticipating the strongest attacks possible.\nWe demonstrate that existing observation-space attacks on reinforcement learning agents have a common weakness: while effective, their lack of information-theoretic detectability constraints makes them \\textit{detectable} using automated means or human inspection. \nDetectability is undesirable to adversaries as it may trigger security escalations.\nWe introduce \\textit{\\eattacks{}}, a novel form of adversarial attack on sequential decision-makers that is both effective and of $\\epsilon-$bounded statistical detectability. \nWe propose a novel dual ascent algorithm to learn such attacks end-to-end.\nCompared to existing attacks, we empirically find \\eattacks{} to be significantly harder to detect with automated methods, and a small study with human participants\\footnote{IRB approval under reference R84123/RE001} suggests they are similarly harder to detect for humans. \nOur findings suggest the need for better anomaly detectors, as well as effective hardware- and system-level defenses. The project website can be found at https://tinyurl.com/illusory-attacks.", "pdf": "https://openreview.net/pdf/d66e1f533261c55834c022529b10774843d7e452.pdf"} {"title": "Addressing Signal Delay in Deep Reinforcement Learning", "url": "https://openreview.net/forum?id=Z8UfDs4J46", "detail_url": "https://openreview.net/forum?id=Z8UfDs4J46", "authors": "Wei Wang,Dongqi Han,Xufang Luo,Dongsheng Li", "tags": "ICLR 2024,Spotlight", "abstract": "Despite the notable advancements in deep reinforcement learning (DRL) in recent years, a prevalent issue that is often overlooked is the impact of signal delay. Signal delay occurs when there is a lag between an agent's perception of the environment and its corresponding actions. In this paper, we first formalize delayed-observation Markov decision processes (DOMDP) by extending the standard MDP framework to incorporate signal delays. Next, we elucidate the challenges posed by the presence of signal delay in DRL, showing that trivial DRL algorithms and generic methods for partially observable tasks suffer greatly from delays. Lastly, we propose effective strategies to overcome these challenges. Our methods achieve remarkable performance in continuous robotic control tasks with large delays, yielding results comparable to those in non-delayed cases. Overall, our work contributes to a deeper understanding of DRL in the presence of signal delays and introduces novel approaches to address the associated challenges.", "pdf": "https://openreview.net/pdf/56d8f487fd9b7b18c6c5bbea5d5d311be7d0eb5e.pdf"} {"title": "Relay Diffusion: Unifying diffusion process across resolutions for image synthesis", "url": "https://openreview.net/forum?id=qTlcbLSm4p", "detail_url": "https://openreview.net/forum?id=qTlcbLSm4p", "authors": "Jiayan Teng,Wendi Zheng,Ming Ding,Wenyi Hong,Jianqiao Wangni,Zhuoyi Yang,Jie Tang", "tags": "ICLR 2024,Spotlight", "abstract": "Diffusion models achieved great success in image synthesis, but still face challenges in high-resolution generation. Through the lens of discrete cosine transformation, we find the main reason is that *the same noise level on a higher resolution results in a higher Signal-to-Noise Ratio in the frequency domain*. In this work, we present Relay Diffusion Model (RDM), which transfers a low-resolution image or noise into an equivalent high-resolution one for diffusion model via blurring diffusion and block noise. Therefore, the diffusion process can continue seamlessly in any new resolution or model without restarting from pure noise or low-resolution conditioning. RDM achieves state-of-the-art FID on CelebA-HQ and sFID on ImageNet 256$\\times$256, surpassing previous works such as ADM, LDM and DiT by a large margin. All the codes and checkpoints are open-sourced at \\url{https://github.com/THUDM/RelayDiffusion}.", "pdf": "https://openreview.net/pdf/6c6bde8db08956621d943921475d8db656889420.pdf"} {"title": "ScaleCrafter: Tuning-free Higher-Resolution Visual Generation with Diffusion Models", "url": "https://openreview.net/forum?id=u48tHG5f66", "detail_url": "https://openreview.net/forum?id=u48tHG5f66", "authors": "Yingqing He,Shaoshu Yang,Haoxin Chen,Xiaodong Cun,Menghan Xia,Yong Zhang,Xintao Wang,Ran He,Qifeng Chen,Ying Shan", "tags": "ICLR 2024,Spotlight", "abstract": "In this work, we investigate the capability of generating images from pre-trained diffusion models at much higher resolutions than the training image sizes. In addition, the generated images should have arbitrary image aspect ratios. When generating images directly at a higher resolution, 1024 x 1024, with the pre-trained Stable Diffusion using training images of resolution 512 x 512, we observe persistent problems of object repetition and unreasonable object structures. Existing works for higher-resolution generation, such as attention-based and joint-diffusion approaches, cannot well address these issues. As a new perspective, we examine the structural components of the U-Net in diffusion models and identify the crucial cause as the limited perception field of convolutional kernels. Based on this key observation, we propose a simple yet effective re-dilation that can dynamically adjust the convolutional perception field during inference. We further propose the dispersed convolution and noise-damped classifier-free guidance, which can enable ultra-high-resolution image generation (e.g., 4096 x 4096). Notably, our approach does not require any training or optimization. Extensive experiments demonstrate that our approach can address the repetition issue well and achieve state-of-the-art performance on higher-resolution image synthesis, especially in texture details. Our work also suggests that a pre-trained diffusion model trained on low-resolution images can be directly used for high-resolution visual generation without further tuning, which may provide insights for future research on ultra-high-resolution image and video synthesis. More results are available at the anonymous website: https://scalecrafter.github.io/ScaleCrafter/", "pdf": "https://openreview.net/pdf/b3127432d984fcf8fdd23acb6f94df6a0f0d8752.pdf"} {"title": "DrM: Mastering Visual Reinforcement Learning through Dormant Ratio Minimization", "url": "https://openreview.net/forum?id=MSe8YFbhUE", "detail_url": "https://openreview.net/forum?id=MSe8YFbhUE", "authors": "Guowei Xu,Ruijie Zheng,Yongyuan Liang,Xiyao Wang,Zhecheng Yuan,Tianying Ji,Yu Luo,Xiaoyu Liu,Jiaxin Yuan,Pu Hua,Shuzhen Li,Yanjie Ze,Hal Daum\u00e9 III,Furong Huang,Huazhe Xu", "tags": "ICLR 2024,Spotlight", "abstract": "Visual reinforcement learning (RL) has shown promise in continuous control tasks.\nDespite its progress, current algorithms are still unsatisfactory in virtually every aspect of the performance such as sample efficiency, asymptotic performance, and their robustness to the choice of random seeds.\nIn this paper, we identify a major shortcoming in existing visual RL methods that is the agents often exhibit sustained inactivity during early training, thereby limiting their ability to explore effectively. \nExpanding upon this crucial observation, we additionally unveil a significant correlation between the agents' inclination towards motorically inactive exploration and the absence of neuronal activity within their policy networks.\nTo quantify this inactivity, we adopt dormant ratio as a metric to measure inactivity in the RL agent's network.\nEmpirically, we also recognize that the dormant ratio can act as a standalone indicator of an agent's activity level, regardless of the received reward signals.\nLeveraging the aforementioned insights, we introduce DrM, a method that uses three core mechanisms to guide agents' exploration-exploitation trade-offs by actively minimizing the dormant ratio. \nExperiments demonstrate that DrM achieves significant improvements in sample efficiency and asymptotic performance with no broken seeds (76 seeds in total) across three continuous control benchmark environments, including DeepMind Control Suite, MetaWorld, and Adroit.\nMost importantly, DrM is the first model-free algorithm that consistently solves tasks in both the Dog and Manipulator domains from the DeepMind Control Suite as well as three dexterous hand manipulation tasks without demonstrations in Adroit, all based on pixel observations.", "pdf": "https://openreview.net/pdf/1c921e84904e3fe06d5ed433201fa0fb2d619046.pdf"} {"title": "How Over-Parameterization Slows Down Gradient Descent in Matrix Sensing: The Curses of Symmetry and Initialization", "url": "https://openreview.net/forum?id=xGvPKAiOhq", "detail_url": "https://openreview.net/forum?id=xGvPKAiOhq", "authors": "Nuoya Xiong,Lijun Ding,Simon Shaolei Du", "tags": "ICLR 2024,Spotlight", "abstract": "This paper rigorously shows how over-parameterization dramatically changes the convergence behaviors of gradient descent (GD) for the matrix sensing problem, where the goal is to recover an unknown low-rank ground-truth matrix from near-isotropic linear measurements.\nFirst, we consider the symmetric setting with the symmetric parameterization where $M^* \\in \\mathbb{R}^{n \\times n}$ is a positive semi-definite unknown matrix of rank $r \\ll n$, and one uses a symmetric parameterization $XX^\\top$ to learn $M^*$. Here $X \\in \\mathbb{R}^{n \\times k}$ with $k > r$ is the factor matrix. We give a novel $\\Omega\\left(1/T^2\\right)$ lower bound of randomly initialized GD for the over-parameterized case ($k >r$) where $T$ is the number of iterations. This is in stark contrast to the exact-parameterization scenario ($k=r$) where the convergence rate is $\\exp\\left(-\\Omega\\left(T\\right)\\right)$. Next, we study asymmetric setting where $M^* \\in \\mathbb{R}^{n_1 \\times n_2}$ is the unknown matrix of rank $r \\ll \\min\\{n_1,n_2\\}$, and one uses an asymmetric parameterization $FG^\\top$ to learn $M^*$ where $F \\in \\mathbb{R}^{n_1 \\times k}$ and $G \\in \\mathbb{R}^{n_2 \\times k}$. We give the first global exact convergence result of randomly initialized GD for the exact-parameterization case ($k=r$) with an $\\exp\\left(-\\Omega\\left(T\\right)\\right)$ rate. Furthermore, we give the first global exact convergence result for the over-parameterization case ($k>r$) with an $\\exp\\left(-\\Omega\\left(\\alpha^2 T\\right)\\right)$ rate where $\\alpha$ is the initialization scale. This linear convergence result in the over-parameterization case is especially significant because one can apply the asymmetric parameterization to the symmetric setting to speed up from $\\Omega\\left(1/T^2\\right)$ to linear convergence. Therefore, we identify a surprising phenomenon: asymmetric parameterization can exponentially speed up convergence. Equally surprising is our analysis that highlights the importance of imbalance between $F$ and $G$. This is in sharp contrast to prior works which emphasize balance. We further give an example showing the dependency on $\\alpha$ in the convergence rate is unavoidable in the worst case. On the other hand, we propose a novel method that only modifies one step of GD and obtains a convergence rate independent of $\\alpha$, recovering the rate in the exact-parameterization case. We provide empirical studies to verify our theoretical findings.", "pdf": "https://openreview.net/pdf/85f424e5914d2f97c9b19f6dc6f35512876e93ac.pdf"} {"title": "AnyText: Multilingual Visual Text Generation and Editing", "url": "https://openreview.net/forum?id=ezBH9WE9s2", "detail_url": "https://openreview.net/forum?id=ezBH9WE9s2", "authors": "Yuxiang Tuo,Wangmeng Xiang,Jun-Yan He,Yifeng Geng,Xuansong Xie", "tags": "ICLR 2024,Spotlight", "abstract": "Diffusion model based Text-to-Image has achieved impressive achievements recently. Although current technology for synthesizing images is highly advanced and capable of generating images with high fidelity, it is still possible to give the show away when focusing on the text area in the generated image, as synthesized text often contains blurred, unreadable, or incorrect characters, making visual text generation one of the most challenging issues in this field. To address this issue, we introduce AnyText, a diffusion-based multilingual visual text generation and editing model, that focuses on rendering accurate and coherent text in the image. AnyText comprises a diffusion pipeline with two primary elements: an auxiliary latent module and a text embedding module. The former uses inputs like text glyph, position, and masked image to generate latent features for text generation or editing. The latter employs an OCR model for encoding stroke data as embeddings, which blend with image caption embeddings from the tokenizer to generate texts that seamlessly integrate with the background. We employed text-control diffusion loss and text perceptual loss for training to further enhance writing accuracy. AnyText can write characters in multiple languages, to the best of our knowledge, this is the first work to address multilingual visual text generation. It is worth mentioning that AnyText can be plugged into existing diffusion models from the community for rendering or editing text accurately. After conducting extensive evaluation experiments, our method has outperformed all other approaches by a significant margin. Additionally, we contribute the first large-scale multilingual text images dataset, AnyWord-3M, containing 3 million image-text pairs with OCR annotations in multiple languages. Based on AnyWord-3M dataset, we propose AnyText-benchmark for the evaluation of visual text generation accuracy and quality. Our project will be open-sourced soon to improve and promote the development of text generation technology.", "pdf": "https://openreview.net/pdf/edddf44179f4afd7e80deb61d69c562453ed37e7.pdf"} {"title": "At Which Training Stage Does Code Data Help LLMs Reasoning?", "url": "https://openreview.net/forum?id=KIPJKST4gw", "detail_url": "https://openreview.net/forum?id=KIPJKST4gw", "authors": "YINGWEI MA,Yue Liu,Yue Yu,Yuanliang Zhang,Yu Jiang,Changjian Wang,Shanshan Li", "tags": "ICLR 2024,Spotlight", "abstract": "Large Language models (LLMs) have exhibited remarkable reasoning capabilities and become the foundation of language technologies. Inspired by the great success of code data in training LLMs, we naturally wonder at which training stage introducing code data can really help LLMs reasoning. To this end, this paper systematically explores the impact of code data on LLMs at different stages. Concretely, we introduce the code data at the pre-training stage, instruction-tuning stage, and both of them, respectively. Then, the reasoning capability of LLMs is comprehensively and fairly evaluated via six reasoning tasks. We critically analyze the experimental results and provide conclusions with insights. First, pre-training LLMs with the mixture of code and text can significantly enhance LLMs' general reasoning capability almost without negative transfer on other tasks. Besides, at the instruction-tuning stage, code data endows LLMs the task-specific reasoning capability. Moreover, the dynamic mixing strategy of code and text data assists LLMs to learn reasoning capability step-by-step during training. These insights deepen the understanding of LLMs regarding reasoning ability for their application, such as scientific question answering, legal support, etc.", "pdf": "https://openreview.net/pdf/ad3cc5f97b65866e7a809f7145c65de76936c8d1.pdf"} {"title": "Coordinate-Aware Modulation for Neural Fields", "url": "https://openreview.net/forum?id=4UiLqimGm5", "detail_url": "https://openreview.net/forum?id=4UiLqimGm5", "authors": "Joo Chan Lee,Daniel Rho,Seungtae Nam,Jong Hwan Ko,Eunbyung Park", "tags": "ICLR 2024,Spotlight", "abstract": "Neural fields, mapping low-dimensional input coordinates to corresponding signals, have shown promising results in representing various signals. Numerous methodologies have been proposed, and techniques employing MLPs and grid representations have achieved substantial success. MLPs allow compact and high expressibility, yet often suffer from spectral bias and slow convergence speed. On the other hand, methods using grids are free from spectral bias and achieve fast training speed, however, at the expense of high spatial complexity. In this work, we propose a novel way for exploiting both MLPs and grid representations in neural fields. Unlike the prevalent methods that combine them sequentially (extract features from the grids first and feed them to the MLP), we inject spectral bias-free grid representations into the intermediate features in the MLP. More specifically, we suggest a Coordinate-Aware Modulation (CAM), which modulates the intermediate features using scale and shift parameters extracted from the grid representations. This can maintain the strengths of MLPs while mitigating any remaining potential biases, facilitating the rapid learning of high-frequency components. In addition, we empirically found that the feature normalizations, which have not been successful in neural filed literature, proved to be effective when applied in conjunction with the proposed CAM. Experimental results demonstrate that CAM enhances the performance of neural representation and improves learning stability across a range of signals. Especially in the novel view synthesis task, we achieved state-of-the-art performance with the least number of parameters and fast training speed for dynamic scenes and the best performance under 1MB memory for static scenes. CAM also outperforms the best-performing video compression methods using neural fields by a large margin. Our project page is available at https://maincold2.github.io/cam/.", "pdf": "https://openreview.net/pdf/0375ec72b4d2bd03dd66aa47ec6117ed73d38054.pdf"} {"title": "Efficient ConvBN Blocks for Transfer Learning and Beyond", "url": "https://openreview.net/forum?id=lHZm9vNm5H", "detail_url": "https://openreview.net/forum?id=lHZm9vNm5H", "authors": "Kaichao You,Guo Qin,Anchang Bao,Meng Cao,Ping Huang,Jiulong Shan,Mingsheng Long", "tags": "ICLR 2024,Spotlight", "abstract": "Convolution-BatchNorm (ConvBN) blocks are integral components in various computer vision tasks and other domains. A ConvBN block can operate in three modes: Train, Eval, and Deploy. While the Train mode is indispensable for training models from scratch, the Eval mode is suitable for transfer learning and beyond, and the Deploy mode is designed for the deployment of models. This paper focuses on the trade-off between stability and efficiency in ConvBN blocks: Deploy mode is efficient but suffers from training instability; Eval mode is widely used in transfer learning but lacks efficiency. To solve the dilemma, we theoretically reveal the reason behind the diminished training stability observed in the Deploy mode. Subsequently, we propose a novel Tune mode to bridge the gap between Eval mode and Deploy mode. The proposed Tune mode is as stable as Eval mode for transfer learning, and its computational efficiency closely matches that of the Deploy mode. Through extensive experiments in object detection, classification, and adversarial example generation across $5$ datasets and $12$ model architectures, we demonstrate that the proposed Tune mode retains the performance while significantly reducing GPU memory footprint and training time, thereby contributing efficient ConvBN blocks for transfer learning and beyond. Our method has been integrated into both PyTorch (general machine learning framework) and MMCV/MMEngine (computer vision framework). Practitioners just need one line of code to enjoy our efficient ConvBN blocks thanks to PyTorch's builtin machine learning compilers.", "pdf": "https://openreview.net/pdf/da8b125b27c2e46c80329f438b05e284d75b61a8.pdf"} {"title": "Large Content And Behavior Models To Understand, Simulate, And Optimize Content And Behavior", "url": "https://openreview.net/forum?id=TrKq4Wlwcz", "detail_url": "https://openreview.net/forum?id=TrKq4Wlwcz", "authors": "Ashmit Khandelwal,Aditya Agrawal,Aanisha Bhattacharyya,Yaman Kumar,Somesh Singh,Uttaran Bhattacharya,Ishita Dasgupta,Stefano Petrangeli,Rajiv Ratn Shah,Changyou Chen,Balaji Krishnamurthy", "tags": "ICLR 2024,Spotlight", "abstract": "Shannon and Weaver's seminal information theory divides communication into three levels: technical, semantic, and effectiveness. While the technical level deals with the accurate reconstruction of transmitted symbols, the semantic and effectiveness levels deal with the inferred meaning and its effect on the receiver. Large Language Models (LLMs), with their wide generalizability, make some progress towards the second level. However, LLMs and other communication models are not conventionally designed for predicting and optimizing communication for desired receiver behaviors and intents. As a result, the effectiveness level remains largely untouched by modern communication systems. In this paper, we introduce the receivers' \"behavior tokens,\" such as shares, likes, clicks, purchases, and retweets, in the LLM's training corpora to optimize content for the receivers and predict their behaviors. Other than showing similar performance to LLMs on content understanding tasks, our trained models show generalization capabilities on the behavior dimension for behavior simulation, content simulation, behavior understanding, and behavior domain adaptation. We show results on all these capabilities using a wide range of tasks on three corpora. We call these models Large Content and Behavior Models (LCBMs). Further, to spur more research on LCBMs, we release our new Content Behavior Corpus (CBC), a repository containing communicator, message, and corresponding receiver behavior (https://behavior-in-the-wild.github.io/LCBM).", "pdf": "https://openreview.net/pdf/b2522b24d83cf73195d3d49d25c22b11f9633228.pdf"} {"title": "Beyond Reverse KL: Generalizing Direct Preference Optimization with Diverse Divergence Constraints", "url": "https://openreview.net/forum?id=2cRzmWXK9N", "detail_url": "https://openreview.net/forum?id=2cRzmWXK9N", "authors": "Chaoqi Wang,Yibo Jiang,Chenghao Yang,Han Liu,Yuxin Chen", "tags": "ICLR 2024,Spotlight", "abstract": "The increasing capabilities of large language models (LLMs) raise opportunities for artificial general intelligence but concurrently amplify safety concerns, such as potential misuse of AI systems, necessitating effective AI alignment. Reinforcement Learning from Human Feedback (RLHF) has emerged as a promising pathway towards AI alignment but brings forth challenges due to its complexity and dependence on a separate reward model. Direct Preference Optimization (DPO) has been proposed as an alternative; and it remains equivalent to RLHF under the reverse KL regularization constraint. This paper presents $f$-DPO, a generalized approach to DPO by incorporating diverse divergence constraints. We show that under certain $f$-divergences, including Jensen-Shannon divergence, forward KL divergences and $\\alpha$-divergences, the complex relationship between the reward and optimal policy can also be simplified by addressing the Karush\u2013Kuhn\u2013Tucker conditions. This eliminates the need for estimating the normalizing constant in the Bradley-Terry model and enables a tractable mapping between the reward function and the optimal policy. Our approach optimizes LLMs to align with human preferences in a more efficient and supervised manner under a broad set of divergence constraints. Empirically, adopting these divergences ensures a balance between alignment performance and generation diversity. Importantly, our $f$-DPO outperforms PPO-based methods in divergence efficiency, and divergence constraints directly influence expected calibration error (ECE).", "pdf": "https://openreview.net/pdf/161cf92d489db9ea1ae0f43eab7797245e967c2c.pdf"} {"title": "FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets", "url": "https://openreview.net/forum?id=CYmF38ysDa", "detail_url": "https://openreview.net/forum?id=CYmF38ysDa", "authors": "Seonghyeon Ye,Doyoung Kim,Sungdong Kim,Hyeonbin Hwang,Seungone Kim,Yongrae Jo,James Thorne,Juho Kim,Minjoon Seo", "tags": "ICLR 2024,Spotlight", "abstract": "Evaluation of Large Language Models (LLMs) is challenging because instruction-following necessitates alignment with human values and the required set of skills varies depending on the instruction. However, previous studies have mainly focused on coarse-grained evaluation (i.e. overall preference-based evaluation), which limits interpretability since it does not consider the nature of user instructions that require instance-wise skill composition. In this paper, we introduce FLASK (Fine-grained Language Model Evaluation based on Alignment Skill Sets), a fine-grained evaluation protocol for both human-based and model-based evaluation which decomposes coarse-level scoring to a skill set-level scoring for each instruction. We experimentally observe that the fine-graininess of evaluation is crucial for attaining a holistic view of model performance and increasing the reliability of the evaluation. Using FLASK, we compare multiple open-source and proprietary LLMs and observe a high correlation between model-based and human-based evaluations.", "pdf": "https://openreview.net/pdf/bf031191f0c9813f740339b104d14e7a4d3f03e4.pdf"} {"title": "LMSYS-Chat-1M: A Large-Scale Real-World LLM Conversation Dataset", "url": "https://openreview.net/forum?id=BOfDKxfwt0", "detail_url": "https://openreview.net/forum?id=BOfDKxfwt0", "authors": "Lianmin Zheng,Wei-Lin Chiang,Ying Sheng,Tianle Li,Siyuan Zhuang,Zhanghao Wu,Yonghao Zhuang,Zhuohan Li,Zi Lin,Eric Xing,Joseph E. Gonzalez,Ion Stoica,Hao Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "Studying how people interact with large language models (LLMs) in real-world scenarios is increasingly important due to their widespread use in various applications. In this paper, we introduce LMSYS-Chat-1M, a large-scale dataset containing one million real-world conversations with 25 state-of-the-art LLMs. This dataset is collected from 210K unique IP addresses in the wild on our Vicuna demo and Chatbot Arena website. We offer an overview of the dataset's content, including its curation process, basic statistics, and topic distribution, highlighting its diversity, originality, and scale. We demonstrate its versatility through four use cases: developing content moderation models that perform similarly to GPT-4, building a safety benchmark, training instruction-following models that perform similarly to Vicuna, and creating challenging benchmark questions. We believe that this dataset will serve as a valuable resource for understanding and advancing LLM capabilities. The dataset is publicly available at https://huggingface.co/datasets/lmsys/lmsys-chat-1m.", "pdf": "https://openreview.net/pdf/c90c03844055cc8a8e2b9c574d0d24cbeafe5034.pdf"} {"title": "EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Diffusion Models", "url": "https://openreview.net/forum?id=UmMa3UNDAz", "detail_url": "https://openreview.net/forum?id=UmMa3UNDAz", "authors": "Yefei He,Jing Liu,Weijia Wu,Hong Zhou,Bohan Zhuang", "tags": "ICLR 2024,Spotlight", "abstract": "Diffusion models have demonstrated remarkable capabilities in image synthesis and related generative tasks. Nevertheless, their practicality for low-latency real-world applications is constrained by substantial computational costs and latency issues. Quantization is a dominant way to compress and accelerate diffusion models, where post-training quantization (PTQ) and quantization-aware training (QAT) are two main approaches, each bearing its own properties. While PTQ exhibits efficiency in terms of both time and data usage, it may lead to diminished performance in low bit-width settings. On the other hand, QAT can help alleviate performance degradation but comes with substantial demands on computational and data resources. To capitalize on the advantages while avoiding their respective drawbacks, we introduce a data-free, quantization-aware and parameter-efficient fine-tuning framework for low-bit diffusion models, dubbed EfficientDM, to achieve QAT-level performance with PTQ-like efficiency. Specifically, we propose a quantization-aware variant of the low-rank adapter (QALoRA) that can be merged with model weights and jointly quantized to low bit-width. The fine-tuning process distills the denoising capabilities of the full-precision model into its quantized counterpart, eliminating the requirement for training data. To further enhance performance, we introduce scale-aware optimization to address ineffective learning of QALoRA due to variations in weight quantization scales across different layers. We also employ temporal learned step-size quantization to handle notable variations in activation distributions across denoising steps. Extensive experimental results demonstrate that our method significantly outperforms previous PTQ-based diffusion models while maintaining similar time and data efficiency. Specifically, there is only a marginal $0.05$ sFID increase when quantizing both weights and activations of LDM-4 to 4-bit on ImageNet $256\\times256$. Compared to QAT-based methods, our EfficientDM also boasts a $16.2\\times$ faster quantization speed with comparable generation quality, rendering it a compelling choice for practical applications.", "pdf": "https://openreview.net/pdf/ee376b0254ced2d4b2ed0b77eec1b7f0a73d0b01.pdf"} {"title": "BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models", "url": "https://openreview.net/forum?id=3TO3TtnOFl", "detail_url": "https://openreview.net/forum?id=3TO3TtnOFl", "authors": "Qingqing Cao,Sewon Min,Yizhong Wang,Hannaneh Hajishirzi", "tags": "ICLR 2024,Spotlight", "abstract": "Retrieval augmentation addresses many critical problems in large language models such as hallucination, staleness, and privacy leaks.\nHowever, running retrieval-augmented language models (LMs) is slow and difficult to scale due to processing large amounts of retrieved text. \nWe introduce binary token representations (BTR), which use 1-bit vectors to precompute every token in passages, significantly reducing computation during inference. \nDespite the potential loss of accuracy, our new calibration techniques and training objectives restore performance. Combined with offline and runtime compression, this only requires 127GB of disk space for encoding 3 billion tokens in Wikipedia.\nOur experiments show that on five knowledge-intensive NLP tasks, BTR accelerates state-of-the-art inference by up to 4x and reduces storage by over 100x while maintaining over 95% task performance. Our code is publicly available at https://github.com/csarron/BTR.", "pdf": "https://openreview.net/pdf/e8e8dce96bfa93af6f7cc3520631eb3db8e8d919.pdf"} {"title": "Frozen Transformers in Language Models Are Effective Visual Encoder Layers", "url": "https://openreview.net/forum?id=t0FI3Q66K5", "detail_url": "https://openreview.net/forum?id=t0FI3Q66K5", "authors": "Ziqi Pang,Ziyang Xie,Yunze Man,Yu-Xiong Wang", "tags": "ICLR 2024,Spotlight", "abstract": "This paper reveals that large language models (LLMs), despite being trained solely on text data, are surprisingly}strong encoders for purely visual tasks in the absence of language. Even more intriguingly, this can be achieved by a simple yet previously overlooked strategy -- employing a frozen transformer block from pre-trained LLMs as a constituent encoder layer to directly process visual tokens. Our work pushes the boundaries of leveraging LLMs for computer vision tasks, significantly departing from conventional practices that typically necessitate a multi-modal vision-language setup with associated language prompts, inputs, or outputs. We demonstrate that our approach consistently enhances performance across a diverse range of tasks} encompassing pure 2D or 3D visual recognition tasks (e.g., image and point cloud classification), temporal modeling tasks (e.g., action recognition), non-semantic tasks (e.g., motion forecasting), and multi-modal tasks (e.g., 2D/3D visual question answering and image-text retrieval). Such improvements are a general phenomenon, applicable to various types of LLMs (e.g., LLaMA and OPT) and different LLM transformer blocks. We additionally propose the information filtering hypothesis to explain the effectiveness of pre-trained LLMs in visual encoding -- the pre-trained LLM transformer blocks discern informative visual tokens and further amplify their effect. This hypothesis is empirically supported by the observation that the feature activation, after training with LLM transformer blocks, exhibits a stronger focus on relevant regions. We hope that our work inspires new perspectives on utilizing LLMs and deepening our understanding of their underlying mechanisms.", "pdf": "https://openreview.net/pdf/9c46b8dad8775df26907aed1cec97a61b8c7d014.pdf"} {"title": "SocioDojo: Building Lifelong Analytical Agents with Real-world Text and Time Series", "url": "https://openreview.net/forum?id=s9z0HzWJJp", "detail_url": "https://openreview.net/forum?id=s9z0HzWJJp", "authors": "Junyan Cheng,Peter Chin", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce SocioDojo, an open-ended lifelong learning environment for developing ready-to-deploy autonomous agents capable of performing human-like analysis and decision-making on societal topics such as economics, finance, politics, and culture. It consists of (1) information sources from news, social media, reports, etc., (2) a knowledge base built from books, journals, and encyclopedias, plus a toolbox of Internet and knowledge graph search interfaces, (3) 30K high-quality time series in finance, economy, society, and polls, which support a novel task called \"hyperportfolio\", that can reliably and scalably evaluate societal analysis and decision-making power of agents, inspired by portfolio optimization with time series as assets to \"invest\". We also propose a novel Analyst-Assistant-Actuator architecture for the hyperportfolio task, and a Hypothesis & Proof prompting for producing in-depth analyses on input news, articles, etc. to assist decision-making. We perform experiments and ablation studies to explore the factors that impact performance. The results show that our proposed method achieves improvements of 32.4% and 30.4% compared to the state-of-the-art method in the two experimental settings.", "pdf": "https://openreview.net/pdf/473beb76f232cb31d5e4d7594a17375a1347cef9.pdf"} {"title": "Learning Performance-Improving Code Edits", "url": "https://openreview.net/forum?id=ix7rLVHXyY", "detail_url": "https://openreview.net/forum?id=ix7rLVHXyY", "authors": "Alexander G Shypula,Aman Madaan,Yimeng Zeng,Uri Alon,Jacob R. Gardner,Yiming Yang,Milad Hashemi,Graham Neubig,Parthasarathy Ranganathan,Osbert Bastani,Amir Yazdanbakhsh", "tags": "ICLR 2024,Spotlight", "abstract": "With the decline of Moore's law, optimizing program performance has become a major focus of software research. However, high-level optimizations such as API and algorithm changes remain elusive due to the difficulty of understanding the semantics of code. Simultaneously, pretrained large language models (LLMs) have demonstrated strong capabilities at solving a wide range of programming tasks. To that end, we introduce a framework for adapting LLMs to high-level program optimization. First, we curate a dataset of performance-improving edits made by human programmers of over 77,000 competitive C++ programming submission pairs, accompanied by extensive unit tests. A major challenge is the significant variability of measuring performance on commodity hardware, which can lead to spurious \"improvements.\" To isolate and reliably evaluate the impact of program optimizations, we design an environment based on the gem5 full system simulator, the de facto simulator used in academia and industry. Next, we propose a broad range of adaptation strategies for code optimization; for prompting, these include retrieval-based few-shot prompting and chain-of-thought, and for finetuning, these include performance-conditioned generation and synthetic data augmentation based on self-play. A combination of these techniques achieves a mean speedup of 6.86$\\times$ with eight generations, higher than average optimizations from individual programmers (3.66$\\times$). Using our model's fastest generations, we set a new upper limit on the fastest speedup possible for our dataset at 9.64$\\times$ compared to using the fastest human submissions available (9.56$\\times$).", "pdf": "https://openreview.net/pdf/e94f139ce25197392e4252dc56ed557ff74094fd.pdf"} {"title": "Quasi-Monte Carlo for 3D Sliced Wasserstein", "url": "https://openreview.net/forum?id=Wd47f7HEXg", "detail_url": "https://openreview.net/forum?id=Wd47f7HEXg", "authors": "Khai Nguyen,Nicola Bariletto,Nhat Ho", "tags": "ICLR 2024,Spotlight", "abstract": "Monte Carlo (MC) integration has been employed as the standard approximation method for the Sliced Wasserstein (SW) distance, whose analytical expression involves an intractable expectation. However, MC integration is not optimal in terms of absolute approximation error. To provide a better class of empirical SW, we propose quasi-sliced Wasserstein (QSW) approximations that rely on Quasi-Monte Carlo (QMC) methods. For a comprehensive investigation of QMC for SW, we focus on the 3D setting, specifically computing the SW between probability measures in three dimensions. In greater detail, we empirically evaluate various methods to construct QMC point sets on the 3D unit-hypersphere, including the Gaussian-based and equal area mappings, generalized spiral points, and optimizing discrepancy energies. Furthermore, to obtain an unbiased estimator for stochastic optimization, we extend QSW to Randomized Quasi-Sliced Wasserstein (RQSW) by introducing randomness in the discussed point sets. Theoretically, we prove the asymptotic convergence of QSW and the unbiasedness of RQSW. Finally, we conduct experiments on various 3D tasks, such as point-cloud comparison, point-cloud interpolation, image style transfer, and training deep point-cloud autoencoders, to demonstrate the favorable performance of the proposed QSW and RQSW variants.", "pdf": "https://openreview.net/pdf/f68a8423256e1b343e90eb5d17ebb7636561a6fa.pdf"} {"title": "A Poincar\u00e9 Inequality and Consistency Results for Signal Sampling on Large Graphs", "url": "https://openreview.net/forum?id=l3qtSNsPvC", "detail_url": "https://openreview.net/forum?id=l3qtSNsPvC", "authors": "Thien Le,Luana Ruiz,Stefanie Jegelka", "tags": "ICLR 2024,Spotlight", "abstract": "Large-scale graph machine learning is challenging as the complexity of learning models scales with the graph size. Subsampling the graph is a viable alternative, but sampling on graphs is nontrivial as graphs are non-Euclidean. Existing graph sampling techniques require not only computing the spectra of large matrices but also repeating these computations when the graph changes, e.g., grows. In this paper, we introduce a signal sampling theory for a type of graph limit---the graphon. We prove a Poincar\u00e9 inequality for graphon signals and show that complements of node subsets satisfying this inequality are unique sampling sets for Paley-Wiener spaces of graphon signals. Exploiting connections with spectral clustering and Gaussian elimination, we prove that such sampling sets are consistent in the sense that unique sampling sets on a convergent graph sequence converge to unique sampling sets on the graphon. We then propose a related graphon signal sampling algorithm for large graphs, and demonstrate its good empirical performance on graph machine learning tasks.", "pdf": "https://openreview.net/pdf/787ddf1932fd5fa37f701cc1319cf90eb55d942b.pdf"} {"title": "Cascading Reinforcement Learning", "url": "https://openreview.net/forum?id=KjOAHlKMF5", "detail_url": "https://openreview.net/forum?id=KjOAHlKMF5", "authors": "Yihan Du,R. Srikant,Wei Chen", "tags": "ICLR 2024,Spotlight", "abstract": "Cascading bandits have gained popularity in recent years due to their applicability to recommendation systems and online advertising. In the cascading bandit model, at each timestep, an agent recommends an ordered subset of items (called an item list) from a pool of items, each associated with an unknown attraction probability. Then, the user examines the list, and clicks the first attractive item (if any), and after that, the agent receives a reward. The goal of the agent is to maximize the expected cumulative reward. However, the prior literature on cascading bandits ignores the influences of user states (e.g., historical behaviors) on recommendations and the change of states as the session proceeds. Motivated by this fact, we propose a generalized cascading RL framework, which considers the impact of user states and state transition into decisions. In cascading RL, we need to select items not only with large attraction probabilities but also leading to good successor states. This imposes a huge computational challenge due to the combinatorial action space. To tackle this challenge, we delve into the properties of value functions, and design an oracle BestPerm to efficiently find the optimal item list. Equipped with BestPerm, we develop two algorithms CascadingVI and CascadingBPI, which are both computationally-efficient and sample-efficient, and provide near-optimal regret and sample complexity guarantees. Furthermore, we present experiments to show the improved computational and sample efficiencies of our algorithms compared to straightforward adaptations of existing RL algorithms in practice.", "pdf": "https://openreview.net/pdf/831efaa7342d514569de4663496b0ff76a04015a.pdf"} {"title": "Complex priors and flexible inference in recurrent circuits with dendritic nonlinearities", "url": "https://openreview.net/forum?id=S5aUhpuyap", "detail_url": "https://openreview.net/forum?id=S5aUhpuyap", "authors": "Benjamin S. H. Lyo,Cristina Savin", "tags": "ICLR 2024,Spotlight", "abstract": "Despite many successful examples in which probabilistic inference can account for perception, we have little understanding of how the brain represents and uses structured priors that capture the complexity of natural input statistics. Here we construct a recurrent circuit model that can implicitly represent priors over latent variables, and combine them with sensory and contextual sources of information to encode task-specific posteriors. Inspired by the recent success of diffusion models as means of learning and using priors over images, our model uses dendritic nonlinearities optimized for denoising, and stochastic somatic integration with the degree of noise modulated by an oscillating global signal. Combining these elements into a recurrent network yields a stochastic dynamical system that samples from the prior at a rate prescribed by the period of the global oscillator. Additional inputs reflecting sensory or top-down contextual information alter these dynamics to generate samples from the corresponding posterior, with different input gating patterns selecting different inference tasks. We demonstrate that this architecture can sample from low dimensional nonlinear manifolds and multimodal posteriors. Overall, the model provides a new framework for circuit-level representation of probabilistic information, in a format that facilitates flexible inference.", "pdf": "https://openreview.net/pdf/70e72bd9cb947d915efc4ad99a0246f77307c5c0.pdf"} {"title": "On the hardness of learning under symmetries", "url": "https://openreview.net/forum?id=ARPrtuzAnQ", "detail_url": "https://openreview.net/forum?id=ARPrtuzAnQ", "authors": "Bobak Kiani,Thien Le,Hannah Lawrence,Stefanie Jegelka,Melanie Weber", "tags": "ICLR 2024,Spotlight", "abstract": "We study the problem of learning equivariant neural networks via gradient descent. The incorporation of known symmetries (\"equivariance\") into neural nets has empirically improved the performance of learning pipelines, in domains ranging from biology to computer vision. However, a rich yet separate line of learning theoretic research has demonstrated that actually learning shallow, fully-connected (i.e. non-symmetric) networks has exponential complexity in the correlational statistical query (CSQ) model, a framework encompassing gradient descent. In this work, we ask: are known problem symmetries sufficient to alleviate the fundamental hardness of learning neural nets with gradient descent? We answer this question in the negative. In particular, we give lower bounds for shallow graph neural networks, convolutional networks, invariant polynomials, and frame-averaged networks for permutation subgroups, which all scale either superpolynomially or exponentially in the relevant input dimension. Therefore, in spite of the significant inductive bias imparted via symmetry, actually learning the complete classes of functions represented by equivariant neural networks via gradient descent remains hard.", "pdf": "https://openreview.net/pdf/051e80193269101cefc8987b52aae50357e5f33d.pdf"} {"title": "An Image Is Worth 1000 Lies: Transferability of Adversarial Images across Prompts on Vision-Language Models", "url": "https://openreview.net/forum?id=nc5GgFAvtk", "detail_url": "https://openreview.net/forum?id=nc5GgFAvtk", "authors": "Haochen Luo,Jindong Gu,Fengyuan Liu,Philip Torr", "tags": "ICLR 2024,Spotlight", "abstract": "Different from traditional task-specific vision models, recent large VLMs can readily adapt to different vision tasks by simply using different textual instructions, i.e., prompts. However, a well-known concern about traditional task-specific vision models is that they can be misled by imperceptible adversarial perturbations. Furthermore, the concern is exacerbated by the phenomenon that the same adversarial perturbations can fool different task-specific models. Given that VLMs rely on prompts to adapt to different tasks, an intriguing question emerges: Can a single adversarial image mislead all predictions of VLMs when a thousand different prompts are given? This question essentially introduces a novel perspective on adversarial transferability: cross-prompt adversarial transferability. In this work, we propose the Cross-Prompt Attack (CroPA). This proposed method updates the visual adversarial perturbation with learnable textual prompts, which are designed to counteract the misleading effects of the adversarial image. By doing this, CroPA significantly improves the transferability of adversarial examples across prompts. Extensive experiments are conducted to verify the strong cross-prompt adversarial transferability of CroPA with prevalent VLMs including Flamingo, BLIP-2, and InstructBLIP in various different tasks.", "pdf": "https://openreview.net/pdf/67d91f245eccef09295bf6fe4b1cc4a142c6c4d5.pdf"} {"title": "One For All: Towards Training One Graph Model For All Classification Tasks", "url": "https://openreview.net/forum?id=4IT2pgc9v6", "detail_url": "https://openreview.net/forum?id=4IT2pgc9v6", "authors": "Hao Liu,Jiarui Feng,Lecheng Kong,Ningyue Liang,Dacheng Tao,Yixin Chen,Muhan Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "Designing a single model to address multiple tasks has been a long-standing objective in artificial intelligence. Recently, large language models have demonstrated exceptional capability in solving different tasks within the language domain. However, a unified model for various graph tasks remains underexplored, primarily due to the challenges unique to the graph learning domain. First, graph data from different areas carry distinct attributes and follow different distributions. Such discrepancy makes it hard to represent graphs in a single representation space. Second, tasks on graphs diversify into node, link, and graph tasks, requiring distinct embedding strategies. Finally, an appropriate graph prompting paradigm for in-context learning is unclear. We propose **One for All (OFA)**, the first general framework that can use a single graph model to address the above challenges. Specifically, OFA proposes text-attributed graphs to unify different graph data by describing nodes and edges with natural language and uses language models to encode the diverse and possibly cross-domain text attributes to feature vectors in the same embedding space. Furthermore, OFA introduces the concept of nodes-of-interest to standardize different tasks with a single task representation. For in-context learning on graphs, OFA introduces a novel graph prompting paradigm that appends prompting substructures to the input graph, which enables it to address varied tasks without fine-tuning. We train the OFA model using graph data from multiple domains (including citation networks, molecular graphs, knowledge graphs, etc.) simultaneously and evaluate its ability in supervised, few-shot, and zero-shot learning scenarios. OFA performs well across different tasks, making it the first general-purpose across-domains classification model on graphs.", "pdf": "https://openreview.net/pdf/390d857f1c69a3defd601a75fb78273100d01927.pdf"} {"title": "$\\texttt{NAISR}$: A 3D Neural Additive Model for Interpretable Shape Representation", "url": "https://openreview.net/forum?id=wg8NPfeMF9", "detail_url": "https://openreview.net/forum?id=wg8NPfeMF9", "authors": "Yining Jiao,Carlton Jude ZDANSKI,Julia S Kimbell,Andrew Prince,Cameron P Worden,Samuel Kirse,Christopher Rutter,Benjamin Shields,William Alexander Dunn,Jisan Mahmud,Marc Niethammer", "tags": "ICLR 2024,Spotlight", "abstract": "Deep implicit functions (DIFs) have emerged as a powerful paradigm for many computer vision tasks such as 3D shape reconstruction, generation, registration, completion, editing, and understanding. However, given a set of 3D shapes with associated covariates there is at present no shape representation method which allows to precisely represent the shapes while capturing the individual dependencies on each covariate. Such a method would be of high utility to researchers to discover knowledge hidden in a population of shapes. For scientific shape discovery purpose, we propose a 3D Neural Additive Model for Interpretable Shape Representation ($\\texttt{NAISR}$) which describes individual shapes by deforming a shape atlas in accordance to the effect of disentangled covariates. Our approach captures shape population trends and allows for patient-specific predictions through shape transfer. $\\texttt{NAISR}$ is the first approach to combine the benefits of deep implicit shape representations with an atlas deforming according to specified covariates. We evaluate $\\texttt{NAISR}$ with respect to shape reconstruction, shape disentanglement, shape evolution, and shape transfer on three datasets, i.e. 1) $\\textit{Starman}$, a simulated 2D shape dataset; 2) ADNI hippocampus 3D shape dataset; 3) pediatric airway 3D shape dataset. Our experiments demonstrate that $\\texttt{NAISR}$ achieves competitive shape reconstruction performance while retaining interpretability. Our code is available at https://github.com/uncbiag/NAISR.", "pdf": "https://openreview.net/pdf/4ac4d43849312f67923cbf9fd40290bebb68e6ca.pdf"} {"title": "Feature emergence via margin maximization: case studies in algebraic tasks", "url": "https://openreview.net/forum?id=i9wDX850jR", "detail_url": "https://openreview.net/forum?id=i9wDX850jR", "authors": "Depen Morwani,Benjamin L. Edelman,Costin-Andrei Oncescu,Rosie Zhao,Sham M. Kakade", "tags": "ICLR 2024,Spotlight", "abstract": "Understanding the internal representations learned by neural networks is a cornerstone challenge in the science of machine learning. While there have been significant recent strides in some cases towards understanding *how* neural networks implement specific target functions, this paper explores a complementary question -- *why* do networks arrive at particular computational strategies? \nOur inquiry focuses on the algebraic learning tasks of modular addition, sparse parities, and finite group operations. Our primary theoretical findings analytically characterize the features learned by stylized neural networks for these algebraic tasks. Notably, our main technique demonstrates how the principle of margin maximization alone can be used to fully specify the features learned by the network. \nSpecifically, we prove that the trained networks utilize Fourier features to perform modular addition and employ features corresponding to irreducible group-theoretic representations to perform compositions in general groups, aligning closely with the empirical observations of Nanda et al. (2023) and Chughtai et al. (2023). More generally, we hope our techniques can help to foster a deeper understanding of why neural networks adopt specific computational strategies.", "pdf": "https://openreview.net/pdf/fdd05f82a84d50198e42915f002f4df0aefac612.pdf"} {"title": "On the Stability of Iterative Retraining of Generative Models on their own Data", "url": "https://openreview.net/forum?id=JORAfH2xFd", "detail_url": "https://openreview.net/forum?id=JORAfH2xFd", "authors": "Quentin Bertrand,Joey Bose,Alexandre Duplessis,Marco Jiralerspong,Gauthier Gidel", "tags": "ICLR 2024,Spotlight", "abstract": "Deep generative models have made tremendous progress in modeling complex data, often exhibiting generation quality that surpasses a typical human's ability to discern the authenticity of samples. Undeniably, a key driver of this success is enabled by the massive amounts of web-scale data consumed by these models. Due to these models' striking performance and ease of availability, the web will inevitably be increasingly populated with synthetic content. Such a fact directly implies that future iterations of generative models will be trained on both clean and artificially generated data from past models. In this paper, we develop a framework to rigorously study the impact of training generative models on mixed datasets---from classical training on real data to self-consuming generative models trained on purely synthetic data. We first prove the stability of iterative training under the condition that the initial generative models approximate the data distribution well enough and the proportion of clean training data (w.r.t. synthetic data) is large enough. We empirically validate our theory on both synthetic and natural images by iteratively training normalizing flows and state-of-the-art diffusion models on CIFAR10 and FFHQ.", "pdf": "https://openreview.net/pdf/119c01132e3767a01a8777074185c4c7f8fb3824.pdf"} {"title": "Intriguing Properties of Generative Classifiers", "url": "https://openreview.net/forum?id=rmg0qMKYRQ", "detail_url": "https://openreview.net/forum?id=rmg0qMKYRQ", "authors": "Priyank Jaini,Kevin Clark,Robert Geirhos", "tags": "ICLR 2024,Spotlight", "abstract": "What is the best paradigm to recognize objects---discriminative inference (fast but potentially prone to shortcut learning) or using a generative model (slow but potentially more robust)? We build on recent advances in generative modeling that turn text-to-image models into classifiers. This allows us to study their behavior and to compare them against discriminative models and human psychophysical data.\nWe report four intriguing emergent properties of generative classifiers: they show a record-breaking human-like shape bias (99% for Imagen), near human-level out-of-distribution accuracy, state-of-the-art alignment with human classification errors, and they understand certain perceptual illusions. Our results indicate that while the current dominant paradigm for modeling human object recognition is discriminative inference, zero-shot generative models approximate human object recognition data surprisingly well.", "pdf": "https://openreview.net/pdf/2eb2ae04198b3791439cc178f097c96bc9aceb8a.pdf"} {"title": "Fast Imitation via Behavior Foundation Models", "url": "https://openreview.net/forum?id=qnWtw3l0jb", "detail_url": "https://openreview.net/forum?id=qnWtw3l0jb", "authors": "Matteo Pirotta,Andrea Tirinzoni,Ahmed Touati,Alessandro Lazaric,Yann Ollivier", "tags": "ICLR 2024,Spotlight", "abstract": "Imitation learning (IL) aims at producing agents that can imitate any behavior given a few expert demonstrations. Yet existing approaches require many demonstrations and/or running (online or offline) reinforcement learning (RL) algorithms for each new imitation task. Here we show that recent RL foundation models based on successor measures can imitate any expert behavior almost instantly with just a few demonstrations and no need for RL or fine-tuning, while accommodating several IL principles (behavioral cloning, feature matching, reward-based, and goal-based reductions). In our experiments, imitation via RL foundation models matches, and often surpasses, the performance of SOTA offline IL algorithms, and produces imitation policies from new demonstrations within seconds instead of hours.", "pdf": "https://openreview.net/pdf/5110a85131511658c1a9003bcf050afee966a8fc.pdf"} {"title": "Task Adaptation from Skills: Information Geometry, Disentanglement, and New Objectives for Unsupervised Reinforcement Learning", "url": "https://openreview.net/forum?id=zSxpnKh1yS", "detail_url": "https://openreview.net/forum?id=zSxpnKh1yS", "authors": "Yucheng Yang,Tianyi Zhou,Qiang He,Lei Han,Mykola Pechenizkiy,Meng Fang", "tags": "ICLR 2024,Spotlight", "abstract": "Unsupervised reinforcement learning (URL) aims to learn general skills for unseen downstream tasks. Mutual Information Skill Learning (MISL) addresses URL by maximizing the mutual information between states and skills but lacks sufficient theoretical analysis, e.g., how well its learned skills can initialize a downstream task's policy. Our new theoretical analysis shows that the diversity and separatability of learned skills are fundamentally critical to downstream task adaptation but MISL does not necessarily guarantee them. To improve MISL, we propose a novel disentanglement metric LSEPIN and build an information-geometric connection between LSEPIN and downstream task adaptation cost. For better geometric properties, we investigate a new strategy that replaces the KL divergence in information geometry with Wasserstein distance. We extend the geometric analysis to it, which leads to a novel skill-learning objective WSEP. It is theoretically justified to be helpful to task adaptation and it is capable of discovering more initial policies for downstream tasks than MISL. We further propose a Wasserstein distance-based algorithm PWSEP can theoretically discover all potentially optimal initial policies.", "pdf": "https://openreview.net/pdf/368c8ed5d5f3738709bdeefccb59dd4ed83ece35.pdf"} {"title": "NuwaDynamics: Discovering and Updating in Causal Spatio-Temporal Modeling", "url": "https://openreview.net/forum?id=sLdVl0q68X", "detail_url": "https://openreview.net/forum?id=sLdVl0q68X", "authors": "Kun Wang,Hao Wu,Yifan Duan,Guibin Zhang,Kai Wang,Xiaojiang Peng,Yu Zheng,Yuxuan Liang,Yang Wang", "tags": "ICLR 2024,Spotlight", "abstract": "Spatio-temporal (ST) prediction plays a pivotal role in earth sciences, such as meteorological prediction, urban computing. Adequate high-quality data, coupled with deep models capable of inference, are both indispensable and prerequisite for achieving meaningful results. However, the sparsity of data and the high costs associated with deploying sensors lead to significant data imbalances. Models that are overly tailored and lack causal relationships further compromise the generalizabilities of inference methods. Towards this end, we first establish a causal concept for ST predictions, named NuwaDynamics, which targets to identify causal regions in data and endow model with causal reasoning ability in a two-stage process. Concretely, we initially leverage upstream self-supervision to discern causal important patches, imbuing the model with generalized information and conducting informed interventions on complementary trivial patches to extrapolate potential test distributions. This phase is referred to as the discovery step. Advancing beyond discovery step, we transfer the data to downstream tasks for targeted ST objectives, aiding the model in recognizing a broader potential distribution and fostering its causal perceptual capabilities (refer as Update step). Our concept aligns seamlessly with the contemporary backdoor adjustment mechanism in causality theory. Extensive experiments on six real-world ST benchmarks showcase that models can gain outcomes upon the integration of the NuwaDynamics concept. NuwaDynamics also can significantly benefit a wide range of changeable ST tasks like extreme weather and long temporal step super-resolution predictions.", "pdf": "https://openreview.net/pdf/7c14ae2b67a7d27b177a742d38945097c4eda38c.pdf"} {"title": "Pre-Training and Fine-Tuning Generative Flow Networks", "url": "https://openreview.net/forum?id=ylhiMfpqkm", "detail_url": "https://openreview.net/forum?id=ylhiMfpqkm", "authors": "Ling Pan,Moksh Jain,Kanika Madan,Yoshua Bengio", "tags": "ICLR 2024,Spotlight", "abstract": "Generative Flow Networks (GFlowNets) are amortized samplers that learn stochastic policies to sequentially generate compositional objects from a given unnormalized reward distribution.\nThey can generate diverse sets of high-reward objects, which is an important consideration in scientific discovery tasks. However, as they are typically trained from a given extrinsic reward function, it remains an important open challenge about how to leverage the power of pre-training and train GFlowNets in an unsupervised fashion for efficient adaptation to downstream tasks.\nInspired by recent successes of unsupervised pre-training in various domains, we introduce a novel approach for reward-free pre-training of GFlowNets. By framing the training as a self-supervised problem, we propose an outcome-conditioned GFlowNet (OC-GFN) that learns to explore the candidate space. Specifically, OC-GFN learns to reach any targeted outcomes, akin to goal-conditioned policies in reinforcement learning. \nWe show that the pre-trained OC-GFN model can allow for a direct extraction of a policy capable of sampling from any new reward functions in downstream tasks.\nNonetheless, adapting OC-GFN on a downstream task-specific reward involves an intractable marginalization over possible outcomes. We propose a novel way to approximate this marginalization by learning an amortized predictor enabling efficient fine-tuning.\nExtensive experimental results validate the efficacy of our approach, demonstrating the effectiveness of pre-training the OC-GFN, and its ability to swiftly adapt to downstream tasks and discover modes more efficiently.\nThis work may serve as a foundation for further exploration of pre-training strategies in the context of GFlowNets.", "pdf": "https://openreview.net/pdf/4b4bd3338aba282409e92ec103c74997f2ab2c8f.pdf"} {"title": "CO2: Efficient Distributed Training with Full Communication-Computation Overlap", "url": "https://openreview.net/forum?id=ZO5cn4IfaN", "detail_url": "https://openreview.net/forum?id=ZO5cn4IfaN", "authors": "Weigao Sun,Zhen Qin,Weixuan Sun,Shidi Li,Dong Li,Xuyang Shen,Yu Qiao,Yiran Zhong", "tags": "ICLR 2024,Spotlight", "abstract": "The fundamental success of large language models hinges upon the efficacious implementation of large-scale distributed training techniques. Nevertheless, building a vast, high-performance cluster featuring high-speed communication interconnectivity is prohibitively costly, and accessible only to prominent entities. In this work, we aim to lower this barrier and democratize large-scale training with limited bandwidth clusters. We propose a new approach called CO2 that introduces local-updating and asynchronous communication to the distributed data-parallel training, thereby facilitating the full overlap of COmmunication with COmputation. CO2 is able to attain a high scalability even on extensive multi-node clusters constrained by very limited communication bandwidth. We further propose the staleness gap penalty and outer momentum clipping techniques together with CO2 to bolster its convergence and training stability. Besides, CO2 exhibits seamless integration with well-established ZeRO-series optimizers which mitigate memory consumption of model states with large model training. We also provide a mathematical proof of convergence, accompanied by the establishment of a stringent upper bound. Furthermore, we validate our findings through an extensive set of practical experiments encompassing a wide range of tasks in the fields of computer vision and natural language processing. These experiments serve to demonstrate the capabilities of CO2 in terms of convergence, generalization, and scalability when deployed across configurations comprising up to 128 A100 GPUs. The outcomes emphasize the outstanding capacity of CO2 to hugely improve scalability, no matter on clusters with 800Gbps RDMA or 80Gbps TCP/IP inter-node connections.", "pdf": "https://openreview.net/pdf/81103680cdf97ee7c0afa17e78b1b069b3facde6.pdf"} {"title": "CADS: Unleashing the Diversity of Diffusion Models through Condition-Annealed Sampling", "url": "https://openreview.net/forum?id=zMoNrajk2X", "detail_url": "https://openreview.net/forum?id=zMoNrajk2X", "authors": "Seyedmorteza Sadat,Jakob Buhmann,Derek Bradley,Otmar Hilliges,Romann M. Weber", "tags": "ICLR 2024,Spotlight", "abstract": "While conditional diffusion models are known to have good coverage of the data distribution, they still face limitations in output diversity, particularly when sampled with a high classifier-free guidance scale for optimal image quality or when trained on small datasets. We attribute this problem to the role of the conditioning signal in inference and offer an improved sampling strategy for diffusion models that can increase generation diversity, especially at high guidance scales, with minimal loss of sample quality. Our sampling strategy anneals the conditioning signal by adding scheduled, monotonically decreasing Gaussian noise to the conditioning vector during inference to balance diversity and condition alignment. Our Condition-Annealed Diffusion Sampler (CADS) can be used with any pretrained model and sampling algorithm, and we show that it boosts the diversity of diffusion models in various conditional generation tasks. Further, using an existing pretrained diffusion model, CADS achieves a new state-of-the-art FID of 1.70 and 2.31 for class-conditional ImageNet generation at 256$\\times$256 and 512$\\times$512 respectively.", "pdf": "https://openreview.net/pdf/7673e3515afc0d2540e2ff2de4f253aaa56742e7.pdf"} {"title": "Image Inpainting via Iteratively Decoupled Probabilistic Modeling", "url": "https://openreview.net/forum?id=rUf9G9k2im", "detail_url": "https://openreview.net/forum?id=rUf9G9k2im", "authors": "Wenbo Li,Xin Yu,Kun Zhou,Yibing Song,Zhe Lin", "tags": "ICLR 2024,Spotlight", "abstract": "Generative adversarial networks (GANs) have made great success in image inpainting yet still have difficulties tackling large missing regions. In contrast, iterative probabilistic algorithms, such as autoregressive and denoising diffusion models, have to be deployed with massive computing resources for decent effect. To achieve high-quality results with low computational cost, we present a novel pixel spread model (PSM) that iteratively employs decoupled probabilistic modeling, combining the optimization efficiency of GANs with the prediction tractability of probabilistic models. As a result, our model selectively spreads informative pixels throughout the image in a few iterations, largely enhancing the completion quality and efficiency. On multiple benchmarks, we achieve new state-of-the-art performance. Our code and models will be publicly available.", "pdf": "https://openreview.net/pdf/1e81620ef68cdb56dc5ca52bc4fa349c9b1ec33b.pdf"} {"title": "Image2Sentence based Asymmetrical Zero-shot Composed Image Retrieval", "url": "https://openreview.net/forum?id=5BXAXOpaWu", "detail_url": "https://openreview.net/forum?id=5BXAXOpaWu", "authors": "Yongchao Du,Min Wang,Wengang Zhou,Shuping Hui,Houqiang Li", "tags": "ICLR 2024,Spotlight", "abstract": "The task of composed image retrieval (CIR) aims to retrieve images based on the query image and the text describing the users' intent. \nExisting methods have made great progress with the advanced large vision-language (VL) model in CIR task, however, they generally suffer from two main issues: lack of labeled triplets for model training and difficulty of deployment on resource-restricted environments when deploying the large vision-language model. To tackle the above problems, we propose Image2Sentence based Asymmetric zero-shot composed image retrieval (ISA), which takes advantage of the VL model and only relies on unlabeled images for composition learning. In the framework, we propose a new adaptive token learner that maps an image to a sentence in the word embedding space of VL model. The sentence adaptively captures discriminative visual information and is further integrated with the text modifier. An asymmetric structure is devised for flexible deployment, in which the lightweight model is adopted for the query side while the large VL model is deployed on the gallery side. The global contrastive distillation and the local alignment regularization are adopted for the alignment between the light model and the VL model for CIR task. Our experiments demonstrate that the proposed ISA could better cope with the real retrieval scenarios and further improve retrieval accuracy and efficiency.", "pdf": "https://openreview.net/pdf/3668aaf0e113e66d0632d7fdc028f3f514a39db5.pdf"} {"title": "Bespoke Solvers for Generative Flow Models", "url": "https://openreview.net/forum?id=1PXEY7ofFX", "detail_url": "https://openreview.net/forum?id=1PXEY7ofFX", "authors": "Neta Shaul,Juan Perez,Ricky T. Q. Chen,Ali Thabet,Albert Pumarola,Yaron Lipman", "tags": "ICLR 2024,Spotlight", "abstract": "Diffusion or flow-based models are powerful generative paradigms that are notoriously hard to sample as samples are defined as solutions to high-dimensional Ordinary or Stochastic Differential Equations (ODEs/SDEs) which require a large Number of Function Evaluations (NFE) to approximate well. Existing methods to alleviate the costly sampling process include model distillation and designing dedicated ODE solvers. However, distillation is costly to train and sometimes can deteriorate quality, while dedicated solvers still require relatively large NFE to produce high quality samples. In this paper we introduce ``Bespoke solvers'', a novel framework for constructing custom ODE solvers tailored to the ODE of a given pre-trained flow model. Our approach optimizes an order consistent and parameter-efficient solver (e.g., with 80 learnable parameters), is trained for roughly 1\\% of the GPU time required for training the pre-trained model, and significantly improves approximation and generation quality compared to dedicated solvers. For example, a Bespoke solver for a CIFAR10 model produces samples with Fr\u00e9chet Inception Distance (FID) of 2.73 with 10 NFE, and gets to 1\\% of the Ground Truth (GT) FID (2.59) for this model with only 20 NFE. On the more challenging ImageNet-64$\\times$64, Bespoke samples at 2.2 FID with 10 NFE, and gets within 2\\% of GT FID (1.71) with 20 NFE.", "pdf": "https://openreview.net/pdf/aa7d0f5148157773b366fcd27b4a30eda6f077b9.pdf"} {"title": "Leveraging augmented-Lagrangian techniques for differentiating over infeasible quadratic programs in machine learning", "url": "https://openreview.net/forum?id=YCPDFfmkFr", "detail_url": "https://openreview.net/forum?id=YCPDFfmkFr", "authors": "Antoine Bambade,Fabian Schramm,Adrien Taylor,Justin Carpentier", "tags": "ICLR 2024,Spotlight", "abstract": "Optimization layers within neural network architectures have become increasingly popular for their ability to solve a wide range of machine learning tasks and to model domain-specific knowledge. However, designing optimization layers requires careful consideration as the underlying optimization problems might be infeasible during training. \nMotivated by applications in learning, control and robotics, this work focuses on convex quadratic programming (QP) layers. The specific structure of this type of optimization layer can be efficiently exploited for faster computations while still allowing rich modeling capabilities. We leverage primal-dual augmented Lagrangian techniques for computing derivatives of both feasible and infeasible QP solutions. \nMore precisely, we propose a unified approach which tackles the differentiability of the closest feasible QP solutions in a classical $\\ell_2$ sense. We then harness this approach to enrich the expressive capabilities of existing QP layers. More precisely, we show how differentiating through infeasible QPs during training enables to drive towards feasibility at test time a new range of QP layers. These layers notably demonstrate superior predictive performance in some conventional learning tasks. Additionally, we present alternative formulations that enhance numerical robustness, speed, and accuracy for training such layers. \nAlong with these contributions, we provide an open-source C++ software package called QPLayer for differentiating feasible and infeasible convex QPs and which can be interfaced with modern learning frameworks.", "pdf": "https://openreview.net/pdf/731de6c558997b87a5381292530348a26e31fb5f.pdf"} {"title": "ODEFormer: Symbolic Regression of Dynamical Systems with Transformers", "url": "https://openreview.net/forum?id=TzoHLiGVMo", "detail_url": "https://openreview.net/forum?id=TzoHLiGVMo", "authors": "St\u00e9phane d'Ascoli,S\u00f6ren Becker,Philippe Schwaller,Alexander Mathis,Niki Kilbertus", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce ODEFormer, the first transformer able to infer multidimensional ordinary differential equation (ODE) systems in symbolic form from the observation of a single solution trajectory. We perform extensive evaluations on two datasets: (i) the existing \u2018Strogatz\u2019 dataset featuring two-dimensional systems; (ii) ODEBench, a collection of one- to four-dimensional systems that we carefully curated from the literature to provide a more holistic benchmark. ODEFormer consistently outperforms existing methods while displaying substantially improved robustness to noisy and irregularly sampled observations, as well as faster inference. We release our code, model and benchmark at https://github.com/sdascoli/odeformer.", "pdf": "https://openreview.net/pdf/f91ca392113614803487b965a914d26be9910ccf.pdf"} {"title": "Convergence of Bayesian Bilevel Optimization", "url": "https://openreview.net/forum?id=fLXpXa7iiz", "detail_url": "https://openreview.net/forum?id=fLXpXa7iiz", "authors": "Shi Fu,Fengxiang He,Xinmei Tian,Dacheng Tao", "tags": "ICLR 2024,Spotlight", "abstract": "This paper presents the first theoretical guarantee for Bayesian bilevel optimization (BBO) that we term for the prevalent bilevel framework combining Bayesian optimization at the outer level to tune hyperparameters, and the inner-level stochastic gradient descent (SGD) for training the model. We prove sublinear regret bounds suggesting simultaneous convergence of the inner-level model parameters and outer-level hyperparameters to optimal configurations for generalization capability. A pivotal, technical novelty in the proofs is modeling the excess risk of the SGD-trained parameters as evaluation noise during Bayesian optimization. Our theory implies the inner unit horizon, defined as the number of SGD iterations, shapes the convergence behavior of BBO. This suggests practical guidance on configuring the inner unit horizon to enhance training efficiency and model performance.", "pdf": "https://openreview.net/pdf/209c3efe9f38c1f8a12c87cee69ec4513e0f8b8d.pdf"} {"title": "MovingParts: Motion-based 3D Part Discovery in Dynamic Radiance Field", "url": "https://openreview.net/forum?id=QQ6RgKYiQq", "detail_url": "https://openreview.net/forum?id=QQ6RgKYiQq", "authors": "Kaizhi Yang,Xiaoshuai Zhang,Zhiao Huang,Xuejin Chen,Zexiang Xu,Hao Su", "tags": "ICLR 2024,Spotlight", "abstract": "We present MovingParts, a NeRF-based method for dynamic scene reconstruction and part discovery. We consider motion as an important cue for identifying parts, that all particles on the same part share the common motion pattern. From the perspective of fluid simulation, existing deformation-based methods for dynamic NeRF can be seen as parameterizing the scene motion under the Eulerian view, i.e., focusing on specific locations in space through which the fluid flows as time passes. However, it is intractable to extract the motion of constituting objects or parts using the Eulerian view representation. In this work, we introduce the dual Lagrangian view and enforce representations under the Eulerian/Lagrangian views to be cycle-consistent. Under the Lagrangian view, we parameterize the scene motion by tracking the trajectory of particles on objects. The Lagrangian view makes it convenient to discover parts by factorizing the scene motion as a composition of part-level rigid motions. Experimentally, our method can achieve fast and high-quality dynamic scene reconstruction from even a single moving camera, and the induced part-based representation allows direct applications of part tracking, animation, 3D scene editing, etc.", "pdf": "https://openreview.net/pdf/5a09f0c81c6bf7abc3ea9cd5d6138c06f4812622.pdf"} {"title": "Equivariant Matrix Function Neural Networks", "url": "https://openreview.net/forum?id=yrgQdA5NkI", "detail_url": "https://openreview.net/forum?id=yrgQdA5NkI", "authors": "Ilyes Batatia,Lars Leon Schaaf,Gabor Csanyi,Christoph Ortner,Felix Andreas Faber", "tags": "ICLR 2024,Spotlight", "abstract": "Graph Neural Networks (GNNs), especially message-passing neural networks (MPNNs), have emerged as powerful architectures for learning on graphs in diverse applications. However, MPNNs face challenges when modeling non-local interactions in systems such as large conjugated molecules, metals, or amorphous materials.\nAlthough Spectral GNNs and traditional neural networks such as recurrent neural networks and transformers mitigate these challenges, they often lack extensivity, adaptability, generalizability, computational efficiency, or fail to capture detailed structural relationships or symmetries in the data. To address these concerns, we introduce Matrix Function Neural Networks (MFNs), a novel architecture that parameterizes non-local interactions through analytic matrix equivariant functions. Employing resolvent expansions offers a straightforward implementation and the potential for linear scaling with system size.\nThe MFN architecture achieves state-of-the-art performance in standard graph benchmarks, such as the ZINC and TU datasets, and is able to capture intricate non-local interactions in quantum systems. The code and the datasets will be made public.", "pdf": "https://openreview.net/pdf/f769ed973540097e6a721fc0b64e4271088eabe0.pdf"} {"title": "Multi-resolution HuBERT: Multi-resolution Speech Self-Supervised Learning with Masked Unit Prediction", "url": "https://openreview.net/forum?id=kUuKFW7DIF", "detail_url": "https://openreview.net/forum?id=kUuKFW7DIF", "authors": "Jiatong Shi,Hirofumi Inaguma,Xutai Ma,Ilia Kulikov,Anna Sun", "tags": "ICLR 2024,Spotlight", "abstract": "Existing Self-Supervised Learning (SSL) models for speech typically process speech signals at a fixed resolution of 20 milliseconds. This approach overlooks the varying informational content present at different resolutions in speech signals. In contrast, this paper aims to incorporate multi-resolution information into speech self-supervised representation learning. We introduce an SSL model that leverages a hierarchical Transformer architecture, complemented by HuBERT-style masked prediction objectives, to process speech at multiple resolutions. Experimental results indicate that the proposed model not only achieves more efficient inference but also exhibits superior or comparable performance to the original HuBERT model over various tasks. Specifically, significant performance improvements over the original HuBERT have been observed in fine-tuning experiments on the LibriSpeech speech recognition benchmark as well as in evaluations using the Speech Universal PERformance Benchmark (SUPERB) and Multilingual SUPERB (ML-SUPERB).", "pdf": "https://openreview.net/pdf/c0f81178cf7aa97eb93f18dd2672018c8ab232b3.pdf"} {"title": "Input-gradient space particle inference for neural network ensembles", "url": "https://openreview.net/forum?id=nLWiR5P3wr", "detail_url": "https://openreview.net/forum?id=nLWiR5P3wr", "authors": "Trung Trinh,Markus Heinonen,Luigi Acerbi,Samuel Kaski", "tags": "ICLR 2024,Spotlight", "abstract": "Deep Ensembles (DEs) demonstrate improved accuracy, calibration and robustness to perturbations over single neural networks partly due to their functional diversity. Particle-based variational inference (ParVI) methods enhance diversity by formalizing a repulsion term based on a network similarity kernel. However, weight-space repulsion is inefficient due to over-parameterization, while direct function-space repulsion has been found to produce little improvement over DEs. To sidestep these difficulties, we propose First-order Repulsive Deep Ensemble (FoRDE), an ensemble learning method based on ParVI, which performs repulsion in the space of first-order input gradients. As input gradients uniquely characterize a function up to translation and are much smaller in dimension than the weights, this method guarantees that ensemble members are functionally different. Intuitively, diversifying the input gradients encourages each network to learn different features, which is expected to improve the robustness of an ensemble. Experiments on image classification datasets and transfer learning tasks show that FoRDE significantly outperforms the gold-standard DEs and other ensemble methods in accuracy and calibration under covariate shift due to input perturbations.", "pdf": "https://openreview.net/pdf/bea7988b0afcf3076721719ce7a16e7b479e4de8.pdf"} {"title": "Prototypical Information Bottlenecking and Disentangling for Multimodal Cancer Survival Prediction", "url": "https://openreview.net/forum?id=otHZ8JAIgh", "detail_url": "https://openreview.net/forum?id=otHZ8JAIgh", "authors": "Yilan Zhang,Yingxue Xu,Jianqi Chen,Fengying Xie,Hao Chen", "tags": "ICLR 2024,Spotlight", "abstract": "Multimodal learning significantly benefits cancer survival prediction, especially the integration of pathological images and genomic data. Despite advantages of multimodal learning for cancer survival prediction, massive redundancy in multimodal data prevents it from extracting discriminative and compact information: (1) An extensive amount of intra-modal task-unrelated information blurs discriminability, especially for gigapixel whole slide images (WSIs) with many patches in pathology and thousands of pathways in genomic data, leading to an \"intra-modal redundancy\" issue. (2) Duplicated information among modalities dominates the representation of multimodal data, which makes modality-specific information prone to being ignored, resulting in an \"inter-modal redundancy\" issue. To address these, we propose a new framework, Prototypical Information Bottlenecking and Disentangling (PIBD), consisting of Prototypical Information Bottleneck (PIB) module for intra-modal redundancy and Prototypical Information Disentanglement (PID) module for inter-modal redundancy. Specifically, a variant of information bottleneck, PIB, is proposed to model prototypes approximating a bunch of instances for different risk levels, which can be used for selection of discriminative instances within modality. PID module decouples entangled multimodal data into compact distinct components: modality-common and modality-specific knowledge, under the guidance of the joint prototypical distribution. Extensive experiments on five cancer benchmark datasets demonstrated our superiority over other methods. The code is released.", "pdf": "https://openreview.net/pdf/3857619d5557d9869437c800e20c98370f49c9f8.pdf"} {"title": "MUSTARD: Mastering Uniform Synthesis of Theorem and Proof Data", "url": "https://openreview.net/forum?id=8xliOUg9EW", "detail_url": "https://openreview.net/forum?id=8xliOUg9EW", "authors": "Yinya Huang,Xiaohan Lin,Zhengying Liu,Qingxing Cao,Huajian Xin,Haiming Wang,Zhenguo Li,Linqi Song,Xiaodan Liang", "tags": "ICLR 2024,Spotlight", "abstract": "Recent large language models (LLMs) have witnessed significant advancement in various tasks, including mathematical reasoning and theorem proving. As these two tasks require strict and formal multi-step inference, they are appealing domains for exploring the reasoning ability of LLMs but still face important challenges. Previous studies such as Chain-of-Thought (CoT) have revealed the effectiveness of intermediate steps guidance. However, such step-wise annotation requires heavy labor, leading to insufficient training steps for current benchmarks. To fill this gap, this work introduces MUSTARD, a data generation framework that masters uniform synthesis of theorem and proof data of high quality and diversity. MUSTARD synthesizes data in three stages: (1) It samples a few mathematical concept seeds as the problem category. (2) Then, it prompts a generative language model with the sampled concepts to obtain both the problems and their step-wise formal solutions. (3) Lastly, the framework utilizes a proof assistant (e.g., Lean Prover) to filter the valid proofs. With the proposed MUSTARD, we present a theorem-and-proof benchmark MUSTARDSAUCE with 5,866 valid data points. Each data point contains an informal statement, an informal proof, and a translated formal proof that passes the prover validation. We perform extensive analysis and demonstrate that MUSTARD generates validated high-quality step-by-step data. We further apply the MUSTARDSAUCE for fine-tuning smaller language models. The fine-tuned Llama 2-7B achieves a 15.41% average relative performance gain in automated theorem proving, and 8.18% in math word problems. Codes and data are available at https://github.com/Eleanor-H/MUSTARD.", "pdf": "https://openreview.net/pdf/426fb3de115df70dc1a81aa777b86b7e5afb7a75.pdf"} {"title": "FLD: Fourier Latent Dynamics for Structured Motion Representation and Learning", "url": "https://openreview.net/forum?id=xsd2llWYSA", "detail_url": "https://openreview.net/forum?id=xsd2llWYSA", "authors": "Chenhao Li,Elijah Stanger-Jones,Steve Heim,Sang bae Kim", "tags": "ICLR 2024,Spotlight", "abstract": "Motion trajectories offer reliable references for physics-based motion learning but suffer from sparsity, particularly in regions that lack sufficient data coverage. To address this challenge, we introduce a self-supervised, structured representation and generation method that extracts spatial-temporal relationships in periodic or quasi-periodic motions. The motion dynamics in a continuously parameterized latent space enable our method to enhance the interpolation and generalization capabilities of motion learning algorithms. The motion learning controller, informed by the motion parameterization, operates online tracking of a wide range of motions, including targets unseen during training. With a fallback mechanism, the controller dynamically adapts its tracking strategy and automatically resorts to safe action execution when a potentially risky target is proposed. By leveraging the identified spatial-temporal structure, our work opens new possibilities for future advancements in general motion representation and learning algorithms.", "pdf": "https://openreview.net/pdf/85c1068390ec60fc360e6c2d790602c02f0502ed.pdf"} {"title": "Towards Reliable and Efficient Backdoor Trigger Inversion via Decoupling Benign Features", "url": "https://openreview.net/forum?id=Tw9wemV6cb", "detail_url": "https://openreview.net/forum?id=Tw9wemV6cb", "authors": "Xiong Xu,Kunzhe Huang,Yiming Li,Zhan Qin,Kui Ren", "tags": "ICLR 2024,Spotlight", "abstract": "Recent studies revealed that using third-party models may lead to backdoor threats, where adversaries can maliciously manipulate model predictions based on backdoors implanted during model training. Arguably, backdoor trigger inversion (BTI), which generates trigger patterns of given benign samples for a backdoored model, is the most critical module for backdoor defenses used in these scenarios. With BTI, defenders can remove backdoors by fine-tuning based on generated poisoned samples with ground-truth labels or deactivate backdoors by removing trigger patterns during the inference process. However, we find that existing BTI methods suffer from relatively poor performance, $i.e.$, their generated triggers are significantly different from the ones used by the adversaries even in the feature space. We argue that it is mostly because existing methods require to 'extract' backdoor features at first, while this task is very difficult since defenders have no information ($e.g.$, trigger pattern or target label) about poisoned samples. In this paper, we explore BTI from another perspective where we decouple benign features instead of decoupling backdoor features directly. Specifically, our method consists of two main steps, including \\textbf{(1)} decoupling benign features and \\textbf{(2)} trigger inversion by minimizing the differences between benign samples and their generated poisoned version in decoupled benign features while maximizing the differences in remaining backdoor features. In particular, our method is more efficient since it doesn't need to `scan' all classes to speculate the target label, as required by existing BTI. We also exploit our BTI module to further design backdoor-removal and pre-processing-based defenses. Extensive experiments on benchmark datasets demonstrate that our defenses can reach state-of-the-art performances.", "pdf": "https://openreview.net/pdf/9b2bab46886305e469d757ab239833eacb477014.pdf"} {"title": "Long-Term Typhoon Trajectory Prediction: A Physics-Conditioned Approach Without Reanalysis Data", "url": "https://openreview.net/forum?id=ziDFH8TPPK", "detail_url": "https://openreview.net/forum?id=ziDFH8TPPK", "authors": "Young-Jae Park,Minseok Seo,Doyi Kim,Hyeri Kim,Sanghoon Choi,Beomkyu Choi,Jeongwon Ryu,Sohee Son,Hae-Gon Jeon,Yeji Choi", "tags": "ICLR 2024,Spotlight", "abstract": "In the face of escalating climate changes, typhoon intensities and their ensuing damage have surged. Accurate trajectory prediction is crucial for effective damage control. Traditional physics-based models, while comprehensive, are computationally intensive and rely heavily on the expertise of forecasters. Contemporary data-driven methods often rely on reanalysis data, which can be considered to be the closest to the true representation of weather conditions. However, reanalysis data is not produced in real-time and requires time for adjustment since prediction models are calibrated with observational data. This reanalysis data, such as ERA5, falls short in challenging real-world situations. Optimal preparedness necessitates predictions at least 72 hours in advance, beyond the capabilities of standard physics models. In response to these constraints, we present an approach that harnesses real-time Unified Model (UM) data, sidestepping the limitations of reanalysis data. Our model provides predictions at 6-hour intervals for up to 72 hours in advance and outperforms both state-of-the-art data-driven methods and numerical weather prediction models. In line with our efforts to mitigate adversities inflicted by \\rthree{typhoons}, we release our preprocessed \\textit{PHYSICS TRACK} dataset, which includes ERA5 reanalysis data, typhoon best-track, and UM forecast data.", "pdf": "https://openreview.net/pdf/3f35815328b8f1ce4a016bc05e882d4f298a97f8.pdf"} {"title": "Poisoned Forgery Face: Towards Backdoor Attacks on Face Forgery Detection", "url": "https://openreview.net/forum?id=8iTpB4RNvP", "detail_url": "https://openreview.net/forum?id=8iTpB4RNvP", "authors": "Jiawei Liang,Siyuan Liang,Aishan Liu,Xiaojun Jia,Junhao Kuang,Xiaochun Cao", "tags": "ICLR 2024,Spotlight", "abstract": "The proliferation of face forgery techniques has raised significant concerns within society, thereby motivating the development of face forgery detection methods. These methods aim to distinguish forged faces from genuine ones and have proven effective in practical applications. However, this paper introduces a novel and previously unrecognized threat in face forgery detection scenarios caused by backdoor attack. By embedding backdoors into models and incorporating specific trigger patterns into the input, attackers can deceive detectors into producing erroneous predictions for forged faces. To achieve this goal, this paper proposes \\emph{Poisoned Forgery Face} framework, which enables clean-label backdoor attacks on face forgery detectors. Our approach involves constructing a scalable trigger generator and utilizing a novel convolving process to generate translation-sensitive trigger patterns. Moreover, we employ a relative embedding method based on landmark-based regions to enhance the stealthiness of the poisoned samples. Consequently, detectors trained on our poisoned samples are embedded with backdoors. Notably, our approach surpasses SoTA backdoor baselines with a significant improvement in attack success rate (+16.39\\% BD-AUC) and reduction in visibility (-12.65\\% $L_\\infty$). Furthermore, our attack exhibits promising performance against backdoor defenses. We anticipate that this paper will draw greater attention to the potential threats posed by backdoor attacks in face forgery detection scenarios. Our codes will be made available at \\url{https://github.com/JWLiang007/PFF}.", "pdf": "https://openreview.net/pdf/d2be569831b6c78668eb40760d2fcdfeb7ea4d51.pdf"} {"title": "Unified Human-Scene Interaction via Prompted Chain-of-Contacts", "url": "https://openreview.net/forum?id=1vCnDyQkjg", "detail_url": "https://openreview.net/forum?id=1vCnDyQkjg", "authors": "Zeqi Xiao,Tai Wang,Jingbo Wang,Jinkun Cao,Wenwei Zhang,Bo Dai,Dahua Lin,Jiangmiao Pang", "tags": "ICLR 2024,Spotlight", "abstract": "Human-Scene Interaction (HSI) is a vital component of fields like embodied AI and virtual reality. Despite advancements in motion quality and physical plausibility, two pivotal factors, versatile interaction control and the development of a user-friendly interface, require further exploration before the practical application of HSI. This paper presents a unified HSI framework, UniHSI, which supports unified control of diverse interactions through language commands. The framework defines interaction as ``Chain of Contacts (CoC)\", representing steps involving human joint-object part pairs. This concept is inspired by the strong correlation between interaction types and corresponding contact regions. Based on the definition, UniHSI constitutes a Large Language Model (LLM) Planner to translate language prompts into task plans in the form of CoC, and a Unified Controller that turns CoC into uniform task execution. To facilitate training and evaluation, we collect a new dataset named ScenePlan that encompasses thousands of task plans generated by LLMs based on diverse scenarios. Comprehensive experiments demonstrate the effectiveness of our framework in versatile task execution and generalizability to real scanned scenes.", "pdf": "https://openreview.net/pdf/ac7284562ef813e689f348d02334547591546d9b.pdf"} {"title": "PTaRL: Prototype-based Tabular Representation Learning via Space Calibration", "url": "https://openreview.net/forum?id=G32oY4Vnm8", "detail_url": "https://openreview.net/forum?id=G32oY4Vnm8", "authors": "Hangting Ye,Wei Fan,Xiaozhuang Song,Shun Zheng,He Zhao,Dan dan Guo,Yi Chang", "tags": "ICLR 2024,Spotlight", "abstract": "Tabular data have been playing a mostly important role in diverse real-world fields, such as healthcare, engineering, finance, etc.\nWith the recent success of deep learning, many tabular machine learning (ML) methods based on deep networks (e.g., Transformer, ResNet) have achieved competitive performance on tabular benchmarks. However, existing deep tabular ML methods suffer from the representation entanglement and localization, which largely hinders their prediction performance and leads to performance inconsistency on tabular tasks.\nTo overcome these problems, we explore a novel direction of applying prototype learning for tabular ML and propose a prototype-based tabular representation learning framework, PTaRL, for tabular prediction tasks. The core idea of PTaRL is to construct prototype-based projection space (P-Space) and learn the disentangled representation around global data prototypes. Specifically, PTaRL mainly involves two stages: (i) Prototype Generating, that constructs global prototypes as the basis vectors of P-Space for representation, and (ii) Prototype Projecting, that projects the data samples into P-Space and keeps the core global data information via Optimal Transport. Then, to further acquire the disentangled representations, we constrain PTaRL with two strategies: (i) to diversify the coordinates towards global prototypes of different representations within P-Space, we bring up a diversifying constraint for representation calibration; (ii) to avoid prototype entanglement in P-Space, we introduce a matrix orthogonalization constraint to ensure the independence of global prototypes. \nFinally, we conduct extensive experiments in PTaRL coupled with state-of-the-art deep tabular ML models on various tabular benchmarks and the results have shown our consistent superiority.", "pdf": "https://openreview.net/pdf/015d2b21c36da1eda54d0672c954af346b0c8ae1.pdf"} {"title": "Stable Neural Stochastic Differential Equations in Analyzing Irregular Time Series Data", "url": "https://openreview.net/forum?id=4VIgNuQ1pY", "detail_url": "https://openreview.net/forum?id=4VIgNuQ1pY", "authors": "YongKyung Oh,Dongyoung Lim,Sungil Kim", "tags": "ICLR 2024,Spotlight", "abstract": "Irregular sampling intervals and missing values in real-world time series data present challenges for conventional methods that assume consistent intervals and complete data. Neural Ordinary Differential Equations (Neural ODEs) offer an alternative approach, utilizing neural networks combined with ODE solvers to learn continuous latent representations through parameterized vector fields. Neural Stochastic Differential Equations (Neural SDEs) extend Neural ODEs by incorporating a diffusion term, although this addition is not trivial, particularly when addressing irregular intervals and missing values. Consequently, careful design of drift and diffusion functions is crucial for maintaining stability and enhancing performance, while incautious choices can result in adverse properties such as the absence of strong solutions, stochastic destabilization, or unstable Euler discretizations, significantly affecting Neural SDEs' performance. In this study, we propose three stable classes of Neural SDEs: Langevin-type SDE, Linear Noise SDE, and Geometric SDE. Then, we rigorously demonstrate their robustness in maintaining excellent performance under distribution shift, while effectively preventing overfitting. To assess the effectiveness of our approach, we conduct extensive experiments on four benchmark datasets for interpolation, forecasting, and classification tasks, and analyze the robustness of our methods with 30 public datasets under different missing rates. Our results demonstrate the efficacy of the proposed method in handling real-world irregular time series data.", "pdf": "https://openreview.net/pdf/999f7149f0b997016d06c2ea29b7f3da17d2ebc1.pdf"} {"title": "What does the Knowledge Neuron Thesis Have to do with Knowledge?", "url": "https://openreview.net/forum?id=2HJRwwbV3G", "detail_url": "https://openreview.net/forum?id=2HJRwwbV3G", "authors": "Jingcheng Niu,Andrew Liu,Zining Zhu,Gerald Penn", "tags": "ICLR 2024,Spotlight", "abstract": "We reassess the Knowledge Neuron (KN) Thesis: an interpretation of the mechanism underlying the ability of large language models to recall facts from a training corpus. This nascent thesis proposes that facts are recalled from the training corpus through the MLP weights in a manner resembling key-value memory, implying in effect that \"knowledge\" is stored in the network. Furthermore, by modifying the MLP modules, one can control the language model's generation of factual information. The plausibility of the KN thesis has been demonstrated by the success of KN-inspired model editing methods (Dai et al., 2022; Meng et al., 2022).\n\nWe find that this thesis is, at best, an oversimplification. Not only have we found that we can edit the expression of certain linguistic phenomena using the same model editing methods but, through a more comprehensive evaluation, we have found that the KN thesis does not adequately explain the process of factual expression. While it is possible to argue that the MLP weights store complex patterns that are interpretable both syntactically and semantically, these patterns do not constitute \"knowledge.\" To gain a more comprehensive understanding of the knowledge representation process, we must look beyond the MLP weights and explore recent models' complex layer structures and attention mechanisms.", "pdf": "https://openreview.net/pdf/04389528ee4dfce215a52f1dd0987190ad2adc1d.pdf"} {"title": "Point2SSM: Learning Morphological Variations of Anatomies from Point Clouds", "url": "https://openreview.net/forum?id=DqziS8DG4M", "detail_url": "https://openreview.net/forum?id=DqziS8DG4M", "authors": "Jadie Adams,Shireen Elhabian", "tags": "ICLR 2024,Spotlight", "abstract": "We present Point2SSM, a novel unsupervised learning approach for constructing correspondence-based statistical shape models (SSMs) directly from raw point clouds. SSM is crucial in clinical research, enabling population-level analysis of morphological variation in bones and organs. Traditional methods of SSM construction have limitations, including the requirement of noise-free surface meshes or binary volumes, reliance on assumptions or templates, and prolonged inference times due to simultaneous optimization of the entire cohort. Point2SSM overcomes these barriers by providing a data-driven solution that infers SSMs directly from raw point clouds, reducing inference burdens and increasing applicability as point clouds are more easily acquired. While deep learning on 3D point clouds has seen success in unsupervised representation learning and shape correspondence, its application to anatomical SSM construction is largely unexplored. We conduct a benchmark of state-of-the-art point cloud deep networks on the SSM task, revealing their limited robustness to clinical challenges such as noisy, sparse, or incomplete input and limited training data. Point2SSM addresses these issues through an attention-based module, providing effective correspondence mappings from learned point features. Our results demonstrate that the proposed method significantly outperforms existing networks in terms of accurate surface sampling and correspondence, better capturing population-level statistics. The source code is provided at https://github.com/jadie1/Point2SSM.", "pdf": "https://openreview.net/pdf/623e642ed0d9f91c3050e148c6b3074974e0fae4.pdf"} {"title": "Improving Domain Generalization with Domain Relations", "url": "https://openreview.net/forum?id=Dc4rXq3HIA", "detail_url": "https://openreview.net/forum?id=Dc4rXq3HIA", "authors": "Huaxiu Yao,Xinyu Yang,Xinyi Pan,Shengchao Liu,Pang Wei Koh,Chelsea Finn", "tags": "ICLR 2024,Spotlight", "abstract": "Distribution shift presents a significant challenge in machine learning, where models often underperform during the test stage when faced with a different distribution than the one they were trained on. In this paper, we focus on domain shifts, which occur when the model is applied to new domains that are different from the ones it was trained on, and propose a new approach called DG. Unlike previous approaches that aim to learn a single model that is domain invariant, DG leverages domain similarities based on domain metadata to learn domain-specific models. Concretely, DG learns a set of training-domain-specific functions during the training stage and reweights them based on domain relations during the test stage. These domain relations can be directly obtained and learned from domain metadata. Under mild assumptions, we theoretically prove that using domain relations to reweight training-domain-specific functions achieves stronger out-of-domain generalization compared to the conventional averaging approach. Empirically, we evaluate the effectiveness of DG using both toy and real-world datasets for tasks such as temperature regression, land use classification, and molecule-protein binding affinity prediction. Our results show that DG consistently outperforms state-of-the-art methods.", "pdf": "https://openreview.net/pdf/23ba11d88b46af7cce33aab0eac91aa8bb30ff30.pdf"} {"title": "Generating Images with 3D Annotations Using Diffusion Models", "url": "https://openreview.net/forum?id=XlkN11Xj6J", "detail_url": "https://openreview.net/forum?id=XlkN11Xj6J", "authors": "Wufei Ma,Qihao Liu,Jiahao Wang,Angtian Wang,Xiaoding Yuan,Yi Zhang,Zihao Xiao,Guofeng Zhang,Beijia Lu,Ruxiao Duan,Yongrui Qi,Adam Kortylewski,Yaoyao Liu,Alan Yuille", "tags": "ICLR 2024,Spotlight", "abstract": "Diffusion models have emerged as a powerful generative method, capable of producing stunning photo-realistic images from natural language descriptions. However, these models lack explicit control over the 3D structure in the generated images. Consequently, this hinders our ability to obtain detailed 3D annotations for the generated images or to craft instances with specific poses and distances. In this paper, we propose 3D Diffusion Style Transfer (3D-DST), which incorporates 3D geometry control into diffusion models. Our method exploits ControlNet, which extends diffusion models by using visual prompts in addition to text prompts. We generate images of the 3D objects taken from 3D shape repositories~(e.g., ShapeNet and Objaverse), render them from a variety of poses and viewing directions, compute the edge maps of the rendered images, and use these edge maps as visual prompts to generate realistic images. With explicit 3D geometry control, we can easily change the 3D structures of the objects in the generated images and obtain ground-truth 3D annotations automatically. This allows us to improve a wide range of vision tasks, e.g., classification and 3D pose estimation, in both in-distribution (ID) and out-of-distribution (OOD) settings. We demonstrate the effectiveness of our method through extensive experiments on ImageNet-100/200, ImageNet-R, PASCAL3D+, ObjectNet3D, and OOD-CV. The results show that our method significantly outperforms existing methods, e.g., 3.8 percentage points on ImageNet-100 using DeiT-B. Our code is available at ", "pdf": "https://openreview.net/pdf/c5afc0f3ba7f4ccee44efa2439b9ebf433502898.pdf"} {"title": "High-dimensional SGD aligns with emerging outlier eigenspaces", "url": "https://openreview.net/forum?id=MHjigVnI04", "detail_url": "https://openreview.net/forum?id=MHjigVnI04", "authors": "Gerard Ben Arous,Reza Gheissari,Jiaoyang Huang,Aukosh Jagannath", "tags": "ICLR 2024,Spotlight", "abstract": "We rigorously study the joint evolution of training dynamics via stochastic gradient descent (SGD) and the spectra of empirical Hessian and gradient matrices. We prove that in two canonical classification tasks for multi-class high-dimensional mixtures and either 1 or 2-layer neural networks, the SGD trajectory rapidly aligns with emerging low-rank outlier eigenspaces of the Hessian and gradient matrices. Moreover, in multi-layer settings this alignment occurs per layer, with the final layer's outlier eigenspace evolving over the course of training, and exhibiting rank deficiency when the SGD converges to sub-optimal classifiers. This establishes some of the rich predictions that have arisen from extensive numerical studies in the last decade about the spectra of Hessian and information matrices over the course of training in overparametrized networks.", "pdf": "https://openreview.net/pdf/c8dc4127b70cc66dd5a649a848ee32a1b553666a.pdf"} {"title": "$\\mathcal{B}$-Coder: Value-Based Deep Reinforcement Learning for Program Synthesis", "url": "https://openreview.net/forum?id=fLf589bx1f", "detail_url": "https://openreview.net/forum?id=fLf589bx1f", "authors": "Zishun Yu,Yunzhe Tao,Liyu Chen,Tao Sun,Hongxia Yang", "tags": "ICLR 2024,Spotlight", "abstract": "Program synthesis aims to create accurate, executable programs from problem specifications, specifically from natural language descriptions in our context. \nRecent studies have leveraged the power of reinforcement learning (RL) in conjunction with large language models (LLMs), significantly enhancing code generation capabilities. The application of RL focuses on directly optimizing for functional correctness, offering an advantage over conventional supervised methods. \nDespite policy-based RL methods dominating the literature on RL for program synthesis, the nature of program synthesis tasks hints at a natural alignment with value-based methods.\nThis stems from the rich collection of off-policy programs, including those developed by human programmers and also historical samples, coupled with the straightforward verification of generated programs through automated unit testing, meaning rewards are easy to obtain.\nDiverging from the dominant use of policy-based algorithms, our work explores the feasibility of value-based approaches, leading to the development of our $\\mathcal{B}$-Coder (pronounced Bellman coder).\nYet, training value-based methods presents challenges due to the enormous search space inherent to program synthesis. \nTo this end, we introduce an initialization protocol for RL agents utilizing pre-trained LMs and a conservative Bellman operator to reduce training complexities. \nMoreover, we demonstrate how to leverage the learned value functions as a dual strategy to post-process generated programs. \nOur empirical evaluations demonstrated $\\mathcal{B}$-Coder's capability in achieving state-of-the-art performance when compared to policy-based methods. \nRemarkably, this achievement is reached with minimal reward engineering effort, highlighting the effectiveness of value-based RL, independent of reward designs.", "pdf": "https://openreview.net/pdf/9d267a827e269389f34c62cd790e14e20c1e8a72.pdf"} {"title": "Adaptive Chameleon or Stubborn Sloth: Revealing the Behavior of Large Language Models in Knowledge Conflicts", "url": "https://openreview.net/forum?id=auKAUJZMO6", "detail_url": "https://openreview.net/forum?id=auKAUJZMO6", "authors": "Jian Xie,Kai Zhang,Jiangjie Chen,Renze Lou,Yu Su", "tags": "ICLR 2024,Spotlight", "abstract": "By providing external information to large language models (LLMs), tool augmentation (including retrieval augmentation) has emerged as a promising solution for addressing the limitations of LLMs' static parametric memory.\nHowever, how receptive are LLMs to such external evidence, especially when the evidence conflicts with their parametric memory? \nWe present the first comprehensive and controlled investigation into the behavior of LLMs when encountering knowledge conflicts.\nWe propose a systematic framework to elicit high-quality parametric memory from LLMs and construct the corresponding counter-memory, which enables us to conduct a series of controlled experiments.\nOur investigation reveals seemingly contradicting behaviors of LLMs.\nOn the one hand, different from prior wisdom, we find that LLMs can be highly receptive to external evidence even when that conflicts with their parametric memory, given that the external evidence is coherent and convincing.\nOn the other hand, LLMs also demonstrate a strong confirmation bias when the external evidence contains some information that is consistent with their parametric memory, despite being presented with conflicting evidence at the same time.\nThese results pose important implications that are worth careful consideration for the further development and deployment of tool- and retrieval-augmented LLMs.\nResources are available at https://github.com/OSU-NLP-Group/LLM-Knowledge-Conflict.", "pdf": "https://openreview.net/pdf/89c45c674a8193212ede350c2f72b29b272db207.pdf"} {"title": "A Hierarchical Bayesian Model for Few-Shot Meta Learning", "url": "https://openreview.net/forum?id=mQ72XRfYRZ", "detail_url": "https://openreview.net/forum?id=mQ72XRfYRZ", "authors": "Minyoung Kim,Timothy Hospedales", "tags": "ICLR 2024,Spotlight", "abstract": "We propose a novel hierarchical Bayesian model for the few-shot meta learning problem. We consider episode-wise random variables to model episode-specific generative processes, where these local random variables are governed by a higher-level global random variable. The global variable captures information shared across episodes, while controlling how much the model needs to be adapted to new episodes in a principled Bayesian manner. Within our framework, prediction on a novel episode/task can be seen as a Bayesian inference problem. For tractable training, we need to be able to relate each local episode-specific solution to the global higher-level parameters. We propose a Normal-Inverse-Wishart model, for which establishing this local-global relationship becomes feasible due to the approximate closed-form solutions for the local posterior distributions. The resulting algorithm is more attractive than the MAML in that it does not maintain a costly computational graph for the sequence of gradient descent steps in an episode. Our approach is also different from existing Bayesian meta learning methods in that rather than modeling a single random variable for all episodes, it leverages a hierarchical structure that exploits the local-global relationships desirable for principled Bayesian learning with many related tasks.", "pdf": "https://openreview.net/pdf/92f087e2b731b4e0629d1db29e70e1a90a9d076f.pdf"} {"title": "Faithful and Efficient Explanations for Neural Networks via Neural Tangent Kernel Surrogate Models", "url": "https://openreview.net/forum?id=yKksu38BpM", "detail_url": "https://openreview.net/forum?id=yKksu38BpM", "authors": "Andrew William Engel,Zhichao Wang,Natalie Frank,Ioana Dumitriu,Sutanay Choudhury,Anand Sarwate,Tony Chiang", "tags": "ICLR 2024,Spotlight", "abstract": "A recent trend in explainable AI research has focused on surrogate modeling, where neural networks are approximated as simpler ML algorithms such as kernel machines. A second trend has been to utilize kernel functions in various explain-by-example or data attribution tasks. In this work, we combine these two trends to analyze approximate empirical neural tangent kernels (eNTK) for data attribution. Approximation is critical for eNTK analysis due to the high computational cost to compute the eNTK. We define new approximate eNTK and perform novel analysis on how well the resulting kernel machine surrogate models correlate with the underlying neural network. We introduce two new random projection variants of approximate eNTK which allow users to tune the time and memory complexity of their calculation. We conclude that kernel machines using approximate neural tangent kernel as the kernel function are effective surrogate models, with the introduced trace NTK the most consistent performer.", "pdf": "https://openreview.net/pdf/8cda600c2a15702e918b77654b5d7db852c98879.pdf"} {"title": "Conformal Risk Control", "url": "https://openreview.net/forum?id=33XGfHLtZg", "detail_url": "https://openreview.net/forum?id=33XGfHLtZg", "authors": "Anastasios Nikolas Angelopoulos,Stephen Bates,Adam Fisch,Lihua Lei,Tal Schuster", "tags": "ICLR 2024,Spotlight", "abstract": "We extend conformal prediction to control the expected value of any monotone loss function. The algorithm generalizes split conformal prediction together with its coverage guarantee. Like conformal prediction, the conformal risk control procedure is tight up to an $\\mathcal{O}(1/n)$ factor. We also introduce extensions of the idea to distribution shift, quantile risk control, multiple and adversarial risk control, and expectations of U-statistics. Worked examples from computer vision and natural language processing demonstrate the usage of our algorithm to bound the false negative rate, graph distance, and token-level F1-score.", "pdf": "https://openreview.net/pdf/482dcb57a6da0d60fce0a3b533b80bcbe99b03f6.pdf"} {"title": "RetroBridge: Modeling Retrosynthesis with Markov Bridges", "url": "https://openreview.net/forum?id=770DetV8He", "detail_url": "https://openreview.net/forum?id=770DetV8He", "authors": "Ilia Igashov,Arne Schneuing,Marwin Segler,Michael M. Bronstein,Bruno Correia", "tags": "ICLR 2024,Spotlight", "abstract": "Retrosynthesis planning is a fundamental challenge in chemistry which aims at designing multi-step reaction pathways from commercially available starting materials to a target molecule. Each step in multi-step retrosynthesis planning requires accurate prediction of possible precursor molecules given the target molecule and confidence estimates to guide heuristic search algorithms. We model single-step retrosynthesis as a distribution learning problem in a discrete state space. First, we introduce the Markov Bridge Model, a generative framework aimed to approximate the dependency between two intractable discrete distributions accessible via a finite sample of coupled data points. Our framework is based on the concept of a Markov bridge, a Markov process pinned at its endpoints. Unlike diffusion-based methods, our Markov Bridge Model does not need a tractable noise distribution as a sampling proxy and directly operates on the input product molecules as samples from the intractable prior distribution. We then address the retrosynthesis planning problem with our novel framework and introduce RetroBridge, a template-free retrosynthesis modeling approach that achieves state-of-the-art results on standard evaluation benchmarks.", "pdf": "https://openreview.net/pdf/20efd425e667325385a150734d8f20307ab8ccd4.pdf"} {"title": "InstructScene: Instruction-Driven 3D Indoor Scene Synthesis with Semantic Graph Prior", "url": "https://openreview.net/forum?id=LtuRgL03pI", "detail_url": "https://openreview.net/forum?id=LtuRgL03pI", "authors": "Chenguo Lin,Yadong MU", "tags": "ICLR 2024,Spotlight", "abstract": "Comprehending natural language instructions is a charming property for 3D indoor scene synthesis systems. Existing methods directly model object joint distributions and express object relations implicitly within a scene, thereby hindering the controllability of generation. We introduce InstructScene, a novel generative framework that integrates a semantic graph prior and a layout decoder to improve controllability and fidelity for 3D scene synthesis. The proposed semantic graph prior jointly learns scene appearances and layout distributions, exhibiting versatility across various downstream tasks in a zero-shot manner. To facilitate the benchmarking for text-driven 3D scene synthesis, we curate a high-quality dataset of scene-instruction pairs with large language and multimodal models. Extensive experimental results reveal that the proposed method surpasses existing state-of-the-art approaches by a large margin. Thorough ablation studies confirm the efficacy of crucial design components. Project page: https://chenguolin.github.io/projects/InstructScene.", "pdf": "https://openreview.net/pdf/c50a222a8cd87a10d51df5f3c0245a363e6cf567.pdf"} {"title": "Single Motion Diffusion", "url": "https://openreview.net/forum?id=DrhZneqz4n", "detail_url": "https://openreview.net/forum?id=DrhZneqz4n", "authors": "Sigal Raab,Inbal Leibovitch,Guy Tevet,Moab Arar,Amit Haim Bermano,Daniel Cohen-Or", "tags": "ICLR 2024,Spotlight", "abstract": "Synthesizing realistic animations of humans, animals, and even imaginary creatures, has long been a goal for artists and computer graphics professionals. Compared to the imaging domain, which is rich with large available datasets, the number of data instances for the motion domain is limited, particularly for the animation of animals and exotic creatures (e.g., dragons), which have unique skeletons and motion patterns. In this work, we introduce SinMDM, a Single Motion Diffusion Model. It is designed to learn the internal motifs of a single motion sequence with arbitrary topology and synthesize a variety of motions of arbitrary length that remain faithful to the learned motifs. We harness the power of diffusion models and present a denoising network explicitly designed for the task of learning from a single input motion. SinMDM is crafted as a lightweight architecture, which avoids overfitting by using a shallow network with local attention layers that narrow the receptive field and encourage motion diversity. Our work applies to multiple contexts, including spatial and temporal in-betweening, motion expansion, style transfer, and crowd animation. Our results show that SinMDM outperforms existing methods both qualitatively and quantitatively. Moreover, while prior network-based approaches require additional training for different applications, SinMDM supports these applications during inference. Our project page, which includes links to the code and trained models, is accessible at https://sinmdm.github.io/SinMDM-page.", "pdf": "https://openreview.net/pdf/e0ae7f79fa711d623931b8987da94361b315e566.pdf"} {"title": "Physics-Regulated Deep Reinforcement Learning: Invariant Embeddings", "url": "https://openreview.net/forum?id=5Dwqu5urzs", "detail_url": "https://openreview.net/forum?id=5Dwqu5urzs", "authors": "Hongpeng Cao,Yanbing Mao,Lui Sha,Marco Caccamo", "tags": "ICLR 2024,Spotlight", "abstract": "This paper proposes the Phy-DRL: a physics-regulated deep reinforcement learning (DRL) framework for safety-critical autonomous systems. The Phy-DRL has three distinguished invariant-embedding designs: i) residual action policy (i.e., integrating data-driven-DRL action policy and physics-model-based action policy), ii) automatically constructed safety-embedded reward, and iii) physics-model-guided neural network (NN) editing, including link editing and activation editing. Theoretically, the Phy-DRL exhibits 1) a mathematically provable safety guarantee and 2) strict compliance of critic and actor networks with physics knowledge about the action-value function and action policy. Finally, we evaluate the Phy-DRL on a cart-pole system and a quadruped robot. The experiments validate our theoretical results and demonstrate that Phy-DRL features guaranteed safety compared to purely data-driven DRL and solely model-based design while offering remarkably fewer learning parameters and fast training towards safety guarantee.", "pdf": "https://openreview.net/pdf/da97cca0d38172e4f655bd23891b3c1509513545.pdf"} {"title": "BatteryML: An Open-source Platform for Machine Learning on Battery Degradation", "url": "https://openreview.net/forum?id=sxGugrYhP9", "detail_url": "https://openreview.net/forum?id=sxGugrYhP9", "authors": "Han Zhang,Xiaofan Gui,Shun Zheng,Ziheng Lu,Yuqi Li,Jiang Bian", "tags": "ICLR 2024,Spotlight", "abstract": "Battery degradation remains a pivotal concern in the energy storage domain, with machine learning emerging as a potent tool to drive forward insights and solutions. However, this intersection of electrochemical science and machine learning poses complex challenges. Machine learning experts often grapple with the intricacies of battery science, while battery researchers face hurdles in adapting intricate models tailored to specific datasets. Beyond this, a cohesive standard for battery degradation modeling, inclusive of data formats and evaluative benchmarks, is conspicuously absent. Recognizing these impediments, we present BatteryML\u2014a one-step, all-encompass, and open-source platform designed to unify data preprocessing, feature extraction, and the implementation of both traditional and state-of-the-art models. This streamlined approach promises to enhance the practicality and efficiency of research applications. BatteryML seeks to fill this void, fostering an environment where experts from diverse specializations can collaboratively contribute, thus elevating the collective understanding and advancement of battery research.", "pdf": "https://openreview.net/pdf/513de629c33aecbc871ce161ed98f8f39a4662e9.pdf"} {"title": "SaProt: Protein Language Modeling with Structure-aware Vocabulary", "url": "https://openreview.net/forum?id=6MRm3G4NiU", "detail_url": "https://openreview.net/forum?id=6MRm3G4NiU", "authors": "Jin Su,Chenchen Han,Yuyang Zhou,Junjie Shan,Xibin Zhou,Fajie Yuan", "tags": "ICLR 2024,Spotlight", "abstract": "Large-scale protein language models (PLMs), such as the ESM family, have achieved remarkable performance in various downstream tasks related to protein structure and function by undergoing unsupervised training on residue sequences. They have become essential tools for researchers and practitioners in biology. However, a limitation of vanilla PLMs is their lack of explicit consideration for protein structure information, which suggests the potential for further improvement. Motivated by this, we introduce the concept of a ``structure-aware vocabulary\" that integrates residue tokens with structure tokens. The structure tokens are derived by encoding the 3D structure of proteins using Foldseek. We then propose SaProt, a large-scale general-purpose PLM trained on an extensive dataset comprising approximately 40 million protein sequences and structures. Through extensive evaluation, our SaProt model surpasses well-established and renowned baselines across 10 significant downstream tasks, demonstrating its exceptional capacity and broad applicability. We have made the code, pre-trained model, and all relevant materials available at https://github.com/westlake-repl/SaProt.", "pdf": "https://openreview.net/pdf/5fccd3773c44a0809138e72b95c259790551303e.pdf"} {"title": "PixArt-$\\alpha$: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis", "url": "https://openreview.net/forum?id=eAKmQPe3m1", "detail_url": "https://openreview.net/forum?id=eAKmQPe3m1", "authors": "Junsong Chen,Jincheng YU,Chongjian GE,Lewei Yao,Enze Xie,Zhongdao Wang,James Kwok,Ping Luo,Huchuan Lu,Zhenguo Li", "tags": "ICLR 2024,Spotlight", "abstract": "The most advanced text-to-image (T2I) models require significant training costs (e.g., millions of GPU hours), seriously hindering the fundamental innovation for the AIGC community while increasing CO2 emissions. This paper introduces PixArt-$\\alpha$, a Transformer-based T2I diffusion model whose image generation quality is competitive with state-of-the-art image generators (e.g., Imagen, SDXL, and even Midjourney), reaching near-commercial application standards. Additionally, it supports high-resolution image synthesis up to 1024px resolution with low training cost, as shown in Figure 1 and 2. To achieve this goal, three core designs are proposed: (1) Training strategy decomposition: We devise three distinct training steps that separately optimize pixel dependency, text-image alignment, and image aesthetic quality; (2) Efficient T2I Transformer: We incorporate cross-attention modules into Diffusion Transformer (DiT) to inject text conditions and streamline the computation-intensive class-condition branch; (3) High-informative data: We emphasize the significance of concept density in text-image pairs and leverage a large Vision-Language model to auto-label dense pseudo-captions to assist text-image alignment learning. As a result, PixArt-$\\alpha$'s training speed markedly surpasses existing large-scale T2I models, e.g., PixArt-$\\alpha$ only takes 10.8% of Stable Diffusion v1.5's training time (~675 vs. ~6,250 A100 GPU days), saving nearly \\\\$300,000 (\\\\$26,000 vs. \\\\$320,000) and reducing 90% CO2 emissions. Moreover, compared with a larger SOTA model, RAPHAEL, our training cost is merely 1%. Extensive experiments demonstrate that PixArt-$\\alpha$ excels in image quality, artistry, and semantic control. We hope PixArt-$\\alpha$ will provide new insights to the AIGC community and startups to accelerate building their own high-quality yet low-cost generative models from scratch.", "pdf": "https://openreview.net/pdf/1b5209278e0321ca4b148bb7d58459bd314e8e2c.pdf"} {"title": "Sentence-level Prompts Benefit Composed Image Retrieval", "url": "https://openreview.net/forum?id=m3ch3kJL7q", "detail_url": "https://openreview.net/forum?id=m3ch3kJL7q", "authors": "Yang bai,Xinxing Xu,Yong Liu,Salman Khan,Fahad Khan,Wangmeng Zuo,Rick Siow Mong Goh,Chun-Mei Feng", "tags": "ICLR 2024,Spotlight", "abstract": "Composed image retrieval (CIR) is the task of retrieving specific images by using a query that involves both a reference image and a relative caption. Most existing CIR models adopt the late-fusion strategy to combine visual and language features. Besides, several approaches have also been suggested to generate a pseudo-word token from the reference image, which is further integrated into the relative caption for CIR. However, these pseudo-word-based prompting methods have limitations when target image encompasses complex changes on reference image, e.g., object removal and attribute modification. In this work, we demonstrate that learning an appropriate sentence-level prompt for the relative caption (SPRC) is sufficient for achieving effective composed image retrieval. Instead of relying on pseudo- word-based prompts, we propose to leverage pretrained V-L models, e.g., BLIP-2, to generate sentence-level prompts. By concatenating the learned sentence-level prompt with the relative caption, one can readily use existing text-based image retrieval models to enhance CIR performance. Furthermore, we introduce both image-text contrastive loss and text prompt alignment loss to enforce the learning of suitable sentence-level prompts. Experiments show that our proposed method performs favorably against the state-of-the-art CIR methods on the Fashion-IQ and CIRR datasets.", "pdf": "https://openreview.net/pdf/440012f67fd4222bd4ed875af49d99018d3f2b8c.pdf"} {"title": "Compositional Generative Inverse Design", "url": "https://openreview.net/forum?id=wmX0CqFSd7", "detail_url": "https://openreview.net/forum?id=wmX0CqFSd7", "authors": "Tailin Wu,Takashi Maruyama,Long Wei,Tao Zhang,Yilun Du,Gianluca Iaccarino,Jure Leskovec", "tags": "ICLR 2024,Spotlight", "abstract": "Inverse design, where we seek to design input variables in order to optimize an underlying objective function, is an important problem that arises across fields such as mechanical engineering to aerospace engineering. Inverse design is typically formulated as an optimization problem, with recent works leveraging optimization across learned dynamics models. However, as models are optimized they tend to fall into adversarial modes, preventing effective sampling. We illustrate that by instead optimizing over the learned energy function captured by the diffusion model, we can avoid such adversarial examples and significantly improve design performance. We further illustrate how such a design system is compositional, enabling us to combine multiple different diffusion models representing subcomponents of our desired system to design systems with every specified component. In an N-body interaction task and a challenging 2D multi-airfoil design task, we demonstrate that by composing the learned diffusion model at test time, our method allows us to design initial states and boundary shapes that are more complex than those in the training data. Our method generalizes to more objects for N-body dataset and discovers formation flying to minimize drag in the multi-airfoil design task. Project website and code can be found at https://github.com/AI4Science-WestlakeU/cindm.", "pdf": "https://openreview.net/pdf/ea28dd2d95cfceacb17820ff058ae38eec039540.pdf"} {"title": "What does automatic differentiation compute for neural networks?", "url": "https://openreview.net/forum?id=8vKknbgXxf", "detail_url": "https://openreview.net/forum?id=8vKknbgXxf", "authors": "Sejun Park,Sanghyuk Chun,Wonyeol Lee", "tags": "ICLR 2024,Spotlight", "abstract": "Forward- or reverse-mode automatic differentiation (AD) is a popular algorithm for computing the derivative of a function expressed by a program. AD always outputs the correct derivative if a program does not use any non-differentiable functions and control flows; however, it may return an arbitrary value otherwise. In this work, we investigate what AD computes for neural networks that may contain non-differentiable functions such as ReLU and maxpools. We first prove that AD always returns a generalized derivative called a Clarke subderivative for networks with pointwise activation functions, if the minibatch size is one and all non-differentiable neurons have distinct bias parameters. We show that the same conclusion does not hold otherwise, but does hold under some mild sufficient conditions. We also prove similar results for more general networks that can use maxpools and bias parameters shared across different neurons. We empirically check our sufficient conditions over popular network architectures and observe that AD almost always computes a Clarke subderivative in practical learning setups.", "pdf": "https://openreview.net/pdf/5dd1c7a6e5c5d0b42966e7e1500d634ee275c07b.pdf"} {"title": "OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models", "url": "https://openreview.net/forum?id=8Wuvhh0LYW", "detail_url": "https://openreview.net/forum?id=8Wuvhh0LYW", "authors": "Wenqi Shao,Mengzhao Chen,Zhaoyang Zhang,Peng Xu,Lirui Zhao,Zhiqian Li,Kaipeng Zhang,Peng Gao,Yu Qiao,Ping Luo", "tags": "ICLR 2024,Spotlight", "abstract": "Large language models (LLMs) have revolutionized natural language processing tasks. However, their practical deployment is hindered by their immense memory and computation requirements. Although recent post-training quantization (PTQ) methods are effective in reducing memory footprint and improving the computational efficiency of LLM, they hand-craft quantization parameters, leading to low performance, especially in extremely low-bit quantization. To tackle this issue, we introduce an Omnidirectionally calibrated Quantization ($\\textbf{OmniQuant}$) technique for LLMs, which achieves good performance in diverse quantization settings while maintaining the computational efficiency of PTQ by efficiently optimizing various quantization parameters. OmniQuant comprises two innovative components including Learnable Weight Clipping (LWC) and Learnable Equivalent Transformation (LET). LWC modulates the extreme values of weights by optimizing the clipping threshold. Meanwhile, LET tackles activation outliers by shifting the challenge of quantization from activations to weights. Operating within a differentiable framework using block-wise error minimization, OmniQuant can optimize the quantization process efficiently for both weight-only and weight-activation quantization. For instance, the LLaMA-2 model family size 7-70B can be processed with OmniQuant on a single A100-40G GPU within 1-16 hours using 128 samples. Extensive experiments validate OmniQuant's superior performance across diverse quantization configurations such as W4A4 (4-bit weight, 4-bit activation), W6A6, W4A16, W3A16, and W2A16. Additionally, OmniQuant demonstrates effectiveness in instruction-tuned models and delivers notable improvements in inference speed and memory reduction on real devices. Codes are available at \n\\url{https://github.com/OpenGVLab/OmniQuant}.", "pdf": "https://openreview.net/pdf/3eb6251031dde954f0cded606252c2de4b922e4d.pdf"} {"title": "Ferret: Refer and Ground Anything Anywhere at Any Granularity", "url": "https://openreview.net/forum?id=2msbbX3ydD", "detail_url": "https://openreview.net/forum?id=2msbbX3ydD", "authors": "Haoxuan You,Haotian Zhang,Zhe Gan,Xianzhi Du,Bowen Zhang,Zirui Wang,Liangliang Cao,Shih-Fu Chang,Yinfei Yang", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce Ferret, a new Multimodal Large Language Model (MLLM) capable of understanding spatial referring of any shape or granularity within an image and accurately grounding open-vocabulary descriptions. To unify referring and grounding in the LLM paradigm, Ferret employs a novel and powerful hybrid region representation that integrates discrete coordinates and continuous features jointly to represent a region in the image. To extract the continuous features of versatile regions, we propose a spatial-aware visual sampler, adept at handling varying sparsity across different shapes. Consequently, Ferret can accept diverse region inputs, such as points, bounding boxes, and free-form shapes. To bolster the desired capability of Ferret, we curate GRIT, a comprehensive refer-and-ground instruction tuning dataset including 1.1M samples that contain rich hierarchical spatial knowledge, with an additional 130K hard negative data to promote model robustness. The resulting model not only achieves superior performance in classical referring and grounding tasks, but also greatly outperforms existing MLLMs in region-based and localization-demanded multimodal chatting. Our evaluations also reveal a significantly improved capability of describing image details and a remarkable alleviation in object hallucination.", "pdf": "https://openreview.net/pdf/2334b28229d561c4b8239640b86f836796cdfa2f.pdf"} {"title": "SalUn: Empowering Machine Unlearning via Gradient-based Weight Saliency in Both Image Classification and Generation", "url": "https://openreview.net/forum?id=gn0mIhQGNM", "detail_url": "https://openreview.net/forum?id=gn0mIhQGNM", "authors": "Chongyu Fan,Jiancheng Liu,Yihua Zhang,Eric Wong,Dennis Wei,Sijia Liu", "tags": "ICLR 2024,Spotlight", "abstract": "With evolving data regulations, machine unlearning (MU) has become an important tool for fostering trust and safety in today's AI models. However, existing MU methods focusing on data and/or weight perspectives often suffer limitations in unlearning accuracy, stability, and cross-domain applicability. To address these challenges, we introduce the concept of 'weight saliency' for MU, drawing parallels with input saliency in model explanation. This innovation directs MU's attention toward specific model weights rather than the entire model, improving effectiveness and efficiency. The resultant method that we call saliency unlearning (SalUn) narrows the performance gap with 'exact' unlearning (model retraining from scratch after removing the forgetting data points). To the best of our knowledge, SalUn is the first principled MU approach that can effectively erase the influence of forgetting data, classes, or concepts in both image classification and generation tasks. As highlighted below, For example, SalUn yields a stability advantage in high-variance random data forgetting, e.g., with a 0.2% gap compared to exact unlearning on the CIFAR-10 dataset. Moreover, in preventing conditional diffusion models from generating harmful images, SalUn achieves nearly 100% unlearning accuracy, outperforming current state-of-the-art baselines like Erased Stable Diffusion and Forget-Me-Not. Codes are available at https://github.com/OPTML-Group/Unlearn-Saliency.\n\n**WARNING**: This paper contains model outputs that may be offensive in nature.", "pdf": "https://openreview.net/pdf/d1a26ff55c8290e1b5ca6daf635489a64bba1162.pdf"} {"title": "Retroformer: Retrospective Large Language Agents with Policy Gradient Optimization", "url": "https://openreview.net/forum?id=KOZu91CzbK", "detail_url": "https://openreview.net/forum?id=KOZu91CzbK", "authors": "Weiran Yao,Shelby Heinecke,Juan Carlos Niebles,Zhiwei Liu,Yihao Feng,Le Xue,Rithesh R N,Zeyuan Chen,Jianguo Zhang,Devansh Arpit,Ran Xu,Phil L Mui,Huan Wang,Caiming Xiong,Silvio Savarese", "tags": "ICLR 2024,Spotlight", "abstract": "Recent months have seen the emergence of a powerful new trend in which large language models (LLMs) are augmented to become autonomous language agents capable of performing objective oriented multi-step tasks on their own, rather than merely responding to queries from human users. Most existing language agents, however, are not optimized using environment-specific rewards. Although some agents enable iterative refinement through verbal feedback, they do not reason and plan in ways that are compatible with gradient-based learning from rewards. This paper introduces a principled framework for reinforcing large language agents by learning a retrospective model, which automatically tunes the language agent prompts from environment feedback through policy gradient. Specifically, our proposed agent architecture learns from rewards across multiple environments and tasks, for fine-tuning a pre-trained language model which refines the language agent prompt by summarizing the root cause of prior failed attempts and proposing action plans. Experimental results on various tasks demonstrate that the language agents improve over time and that our approach considerably outperforms baselines that do not properly leverage gradients from the environment.", "pdf": "https://openreview.net/pdf/d02f39256b41ec50ca8cd3a5b136065bc7a4caae.pdf"} {"title": "BECLR: Batch Enhanced Contrastive Few-Shot Learning", "url": "https://openreview.net/forum?id=k9SVcrmXL8", "detail_url": "https://openreview.net/forum?id=k9SVcrmXL8", "authors": "Stylianos Poulakakis-Daktylidis,Hadi Jamali-Rad", "tags": "ICLR 2024,Spotlight", "abstract": "Learning quickly from very few labeled samples is a fundamental attribute that separates machines and humans in the era of deep representation learning. Unsupervised few-shot learning (U-FSL) aspires to bridge this gap by discarding the reliance on annotations at training time. Intrigued by the success of contrastive learning approaches in the realm of U-FSL, we structurally approach their shortcomings in both pretraining and downstream inference stages. We propose a novel Dynamic Clustered mEmory (DyCE) module to promote a highly separable latent representation space for enhancing positive sampling at the pretraining phase and infusing implicit class-level insights into unsupervised contrastive learning. We then tackle the, somehow overlooked yet critical, issue of sample bias at the few-shot inference stage. We propose an iterative Optimal Transport-based distribution Alignment (OpTA) strategy and demonstrate that it efficiently addresses the problem, especially in low-shot scenarios where FSL approaches suffer the most from sample bias. We later on discuss that DyCE and OpTA are two intertwined pieces of a novel end-to-end approach (we coin as BECLR), constructively magnifying each other's impact. We then present a suite of extensive quantitative and qualitative experimentation to corroborate that BECLR sets a new state-of-the-art across ALL existing U-FSL benchmarks (to the best of our knowledge), and significantly outperforms the best of the current baselines (codebase available at https://github.com/stypoumic/BECLR).", "pdf": "https://openreview.net/pdf/4e0ce4a0dbf6fbf7d4ad5cc091ce7a0df0dc793b.pdf"} {"title": "How to Capture Higher-order Correlations? Generalizing Matrix Softmax Attention to Kronecker Computation", "url": "https://openreview.net/forum?id=v0zNCwwkaV", "detail_url": "https://openreview.net/forum?id=v0zNCwwkaV", "authors": "Josh Alman,Zhao Song", "tags": "ICLR 2024,Spotlight", "abstract": "In the classical transformer attention scheme, we are given three $n \\times d$ size matrices $Q, K, V$ (the query, key, and value tokens), and the goal is to compute a new $n \\times d$ size matrix $D^{-1} \\exp(QK^\\top) V$ where $D = \\mathrm{diag}( \\exp(QK^\\top) {\\bf 1}_n )$. Here, $\\exp()$ is applied entry-wise and ${\\bf 1}_n$ denotes a length-$n$ vector whose entries are all ones.\n\nIntuitively, attention computation captures pairwise information between words in a sentence, but not higher-order information. Indeed, recent work \\cite{sht23} has shown that attention units cannot solve simple problems about detecting triples of connected words.\n\nIn this work, we study a generalization of attention which captures triple-wise correlations. The generalization is based on computations involving tensors defined by tuples of words. More formally, given five $n \\times d$ size matrices $Q, K_1, K_2, V_1$ and $V_2$ (generalized query, key, and value tokens), our new goal is to compute an $n \\times d$ size matrix $D^{-1} \\exp( Q ( K_1 \\oslash K_2)^\\top ) (V_1 \\oslash V_2) $ where $D = \\mathrm{diag}( \\exp( Q ( K_1 \\oslash K_2)^\\top ) {\\bf 1}_{n^2} )$ and $K_1 \\oslash K_2 \\in \\mathbb{R}^{n^2 \\times d}$ denotes the column-wise Kronecker product of $K_1$ and $K_2$. This generalization is indeed able to solve problems about detecting triple-wise connections that were shown to be impossible for transformers.\n\nThe potential downside of this generalization is that it appears as though computations are even more difficult, since the straightforward algorithm requires cubic time in $n$. However, we show that in the bounded-entry setting (which arises in practice, and which is well-studied in both theory and practice), there is actually a near-linear time algorithm. More precisely, we show that bounded entries are both necessary and sufficient for quickly performing generalized computations:\n\n$\\bullet$ On the positive side, if all entries of the input matrices are bounded above by $o(\\sqrt[3]{\\log n})$ then we show how to approximate the ``tensor-type'' attention matrix in $n^{1+o(1)}$ time.\n\n$\\bullet$ On the negative side, we show that if the entries of the input matrices may be as large as $\\Omega(\\sqrt[3]{\\log n})$, then there is no algorithm that runs faster than $n^{3-o(1)}$ (assuming the Strong Exponential \nTime Hypothesis from fine-grained complexity theory).\n\n\nWe also show that our construction, algorithms, and lower bounds naturally generalize to higher-order tensors and correlations. Interestingly, the higher the order of the tensors, the lower the bound on the entries needs to be for an efficient algorithm. Our results thus yield a natural tradeoff between the boundedness of the entries, and order of the tensor one may use for more expressive, efficient attention computation.\n\nOur constructions make use of a novel connection with a higher-order variant on the kernel density estimation problem. They combine a number of technical tools, including the polynomial method, algebraic geometry codes, and multiparty Merlin-Arthur communication protocols.", "pdf": "https://openreview.net/pdf/c5ca1dc9f485832dce82dbf8d2569d405b9f37fb.pdf"} {"title": "DreamLLM: Synergistic Multimodal Comprehension and Creation", "url": "https://openreview.net/forum?id=y01KGvd9Bw", "detail_url": "https://openreview.net/forum?id=y01KGvd9Bw", "authors": "Runpei Dong,Chunrui Han,Yuang Peng,Zekun Qi,Zheng Ge,Jinrong Yang,Liang Zhao,Jianjian Sun,Hongyu Zhou,Haoran Wei,Xiangwen Kong,Xiangyu Zhang,Kaisheng Ma,Li Yi", "tags": "ICLR 2024,Spotlight", "abstract": "This paper presents DreamLLM, a learning framework that first achieves versatile Multimodal Large Language Models (MLLMs) empowered with frequently overlooked synergy between multimodal comprehension and creation. DreamLLM operates on two fundamental principles. The first focuses on the generative modeling of both language and image posteriors by direct sampling in the raw multimodal space. This approach circumvents the limitations and information loss inherent to external feature extractors like CLIP, and a more thorough multimodal understanding is obtained. Second, DreamLLM fosters the generation of raw, interleaved documents, modeling both text and image contents, along with unstructured layouts. This allows DreamLLM to learn all conditional, marginal, and joint multimodal distributions effectively. As a result, DreamLLM is the first MLLM capable of generating free-form interleaved content. Comprehensive experiments highlight DreamLLM's superior performance as a zero-shot multimodal generalist, reaping from the enhanced learning synergy. Project page: https://dreamllm.github.io.", "pdf": "https://openreview.net/pdf/90127f90e1f27ec4c2fa4cd59e1921ab9f5d0a31.pdf"} {"title": "Learning to Act from Actionless Videos through Dense Correspondences", "url": "https://openreview.net/forum?id=Mhb5fpA1T0", "detail_url": "https://openreview.net/forum?id=Mhb5fpA1T0", "authors": "Po-Chen Ko,Jiayuan Mao,Yilun Du,Shao-Hua Sun,Joshua B. Tenenbaum", "tags": "ICLR 2024,Spotlight", "abstract": "In this work, we present an approach to construct a video-based robot policy capable of reliably executing diverse tasks across different robots and environments from few video demonstrations without using any action annotations. Our method leverages images as a task-agnostic representation, encoding both the state and action information, and text as a general representation for specifying robot goals. By synthesizing videos that \"hallucinate\" robot executing actions and in combination with dense correspondences between frames, our approach can infer the closed-formed action to execute to an environment without the need of any explicit action labels. This unique capability allows us to train the policy solely based on RGB videos and deploy learned policies to various robotic tasks. We demonstrate the efficacy of our approach in learning policies on table-top manipulation and navigation tasks. Additionally, we contribute an open-source framework for efficient video modeling, enabling the training of high-fidelity policy models with four GPUs within a single day.", "pdf": "https://openreview.net/pdf/fff0f1d4e51a3d9d660e98a91daabc1b70364dbd.pdf"} {"title": "On Penalty Methods for Nonconvex Bilevel Optimization and First-Order Stochastic Approximation", "url": "https://openreview.net/forum?id=CvYBvgEUK9", "detail_url": "https://openreview.net/forum?id=CvYBvgEUK9", "authors": "Jeongyeol Kwon,Dohyun Kwon,Stephen Wright,Robert D Nowak", "tags": "ICLR 2024,Spotlight", "abstract": "In this work, we study first-order algorithms for solving Bilevel Optimization (BO) where the objective functions are smooth but possibly nonconvex in both levels and the variables are restricted to closed convex sets. As a first step, we study the landscape of BO through the lens of penalty methods, in which the upper- and lower-level objectives are combined in a weighted sum with penalty parameter $\\sigma > 0$. In particular, we establish a strong connection between the penalty function and the hyper-objective by explicitly characterizing the conditions under which the values and derivatives of the two must be $O(\\sigma)$-close. A by-product of our analysis is the explicit formula for the gradient of hyper-objective when the lower-level problem has multiple solutions under minimal conditions, which could be of independent interest. Next, viewing the penalty formulation as $O(\\sigma)$-approximation of the original BO, we propose first-order algorithms that find an $\\epsilon$-stationary solution by optimizing the penalty formulation with $\\sigma = O(\\epsilon)$. When the perturbed lower-level problem uniformly satisfies the {\\it small-error} proximal error-bound (EB) condition, we propose a first-order algorithm that converges to an $\\epsilon$-stationary point of the penalty function using in total $O(\\epsilon^{-7})$ accesses to first-order stochastic gradient oracles. Under an additional assumption on stochastic oracles, we show that the algorithm can be implemented in a fully {\\it single-loop} manner, {\\it i.e.,} with $O(1)$ samples per iteration, and achieves the improved oracle-complexity of $O(\\epsilon^{-5})$.", "pdf": "https://openreview.net/pdf/7e1da919067d21451ef10b9db63ead9823c7395f.pdf"} {"title": "Scaling Laws for Sparsely-Connected Foundation Models", "url": "https://openreview.net/forum?id=i9K2ZWkYIP", "detail_url": "https://openreview.net/forum?id=i9K2ZWkYIP", "authors": "Elias Frantar,Carlos Riquelme Ruiz,Neil Houlsby,Dan Alistarh,Utku Evci", "tags": "ICLR 2024,Spotlight", "abstract": "We explore the impact of parameter sparsity on the scaling behavior of Transformers trained on massive datasets (i.e., \"foundation models\"), in both vision and language domains. In this setting, we identify the first scaling law describing the relationship between weight sparsity, number of non-zero parameters, and amount of training data, which we validate empirically across model and data scales; on ViT/JFT-4B and T5/C4. These results allow us to characterize the \"optimal sparsity\", the sparsity level which yields the best performance for a given effective model size and training budget. For a fixed number of non-zero parameters, we identify that the optimal sparsity increases with the amount of data used for training. We also extend our study to different sparsity structures (such as the hardware-friendly n:m pattern) and strategies (such as starting from a pretrained dense model). Our findings shed light on the power and limitations of weight sparsity across various parameter and computational settings, offering both theoretical understanding and practical implications for leveraging sparsity towards computational efficiency improvements. We provide pruning and scaling law fitting code at: github.com/google-research/jaxpruner/tree/main/jaxpruner/projects/bigsparse.", "pdf": "https://openreview.net/pdf/07a48d9ea45fa58df98ef989ab0046935ac248cf.pdf"} {"title": "Nearly $d$-Linear Convergence Bounds for Diffusion Models via Stochastic Localization", "url": "https://openreview.net/forum?id=r5njV3BsuD", "detail_url": "https://openreview.net/forum?id=r5njV3BsuD", "authors": "Joe Benton,Valentin De Bortoli,Arnaud Doucet,George Deligiannidis", "tags": "ICLR 2024,Spotlight", "abstract": "Denoising diffusions are a powerful method to generate approximate samples from high-dimensional data distributions. Recent results provide polynomial bounds on their convergence rate, assuming $L^2$-accurate scores. Until now, the tightest bounds were either superlinear in the data dimension or required strong smoothness assumptions. We provide the first convergence bounds which are linear in the data dimension (up to logarithmic factors) assuming only finite second moments of the data distribution. We show that diffusion models require at most $\\tilde O(\\frac{d \\log^2(1/\\delta)}{\\varepsilon^2})$ steps to approximate an arbitrary distribution on $\\mathbb{R}^d$ corrupted with Gaussian noise of variance $\\delta$ to within $\\varepsilon^2$ in KL divergence. Our proof extends the Girsanov-based methods of previous works. We introduce a refined treatment of the error from discretizing the reverse SDE inspired by stochastic localization.", "pdf": "https://openreview.net/pdf/b3be4a4ee32009fa1997a1c54f00e4168ebc1980.pdf"} {"title": "DragonDiffusion: Enabling Drag-style Manipulation on Diffusion Models", "url": "https://openreview.net/forum?id=OEL4FJMg1b", "detail_url": "https://openreview.net/forum?id=OEL4FJMg1b", "authors": "Chong Mou,Xintao Wang,Jiechong Song,Ying Shan,Jian Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "Despite the ability of text-to-image (T2I) diffusion models to generate high-quality images, transferring this ability to accurate image editing remains a challenge. In this paper, we propose a novel image editing method, DragonDiffusion, enabling Drag-style manipulation on Diffusion models. Specifically, we treat image editing as the change of feature correspondence in a pre-trained diffusion model. By leveraging feature correspondence, we develop energy functions that align with the editing target, transforming image editing operations into gradient guidance. Based on this guidance approach, we also construct multi-scale guidance that considers both semantic and geometric alignment. Furthermore, we incorporate a visual cross-attention strategy based on a memory bank design to ensure consistency between the edited result and original image. Benefiting from these efficient designs, all content editing and consistency operations come from the feature correspondence without extra model fine-tuning. Extensive experiments demonstrate that our method has promising performance on various image editing tasks, including within a single image (e.g., object moving, resizing, and content dragging) or across images (e.g., appearance replacing and object pasting). Code is available at https://github.com/MC-E/DragonDiffusion.", "pdf": "https://openreview.net/pdf/d1d88ddeb5844da8d6140ed188e076244748489e.pdf"} {"title": "Uni3D: Exploring Unified 3D Representation at Scale", "url": "https://openreview.net/forum?id=wcaE4Dfgt8", "detail_url": "https://openreview.net/forum?id=wcaE4Dfgt8", "authors": "Junsheng Zhou,Jinsheng Wang,Baorui Ma,Yu-Shen Liu,Tiejun Huang,Xinlong Wang", "tags": "ICLR 2024,Spotlight", "abstract": "Scaling up representations for images or text has been extensively investigated in the past few years and has led to revolutions in learning vision and language. However, scalable representation for 3D objects and scenes is relatively unexplored. In this work, we present Uni3D, a 3D foundation model to explore the unified 3D representation at scale. Uni3D uses a 2D initialized ViT end-to-end pretrained to align the 3D point cloud features with the image-text aligned features. Via the simple architecture and pretext task, Uni3D can leverage abundant 2D pretrained models as initialization and image-text aligned models as the target, unlocking the great potential of 2D model zoos and scaling-up strategies to the 3D world. We efficiently scale up Uni3D to one billion parameters, and set new records on a broad range of 3D tasks, such as zero-shot classification, few-shot classification, open-world understanding and zero-shot part segmentation. We show that the strong Uni3D representation also enables applications such as 3D painting and retrieval in the wild. We believe that Uni3D provides a new direction for exploring both scaling up and efficiency of the representation in 3D domain.", "pdf": "https://openreview.net/pdf/88048b7b0b8f24d97f795d4218b905bbe24a2a9e.pdf"} {"title": "CivRealm: A Learning and Reasoning Odyssey in Civilization for Decision-Making Agents", "url": "https://openreview.net/forum?id=UBVNwD3hPN", "detail_url": "https://openreview.net/forum?id=UBVNwD3hPN", "authors": "Siyuan Qi,Shuo Chen,Yexin Li,Xiangyu Kong,Junqi Wang,Bangcheng Yang,Pring Wong,Yifan Zhong,Xiaoyuan Zhang,Zhaowei Zhang,Nian Liu,Yaodong Yang,Song-Chun Zhu", "tags": "ICLR 2024,Spotlight", "abstract": "The generalization of decision-making agents encompasses two fundamental elements: learning from past experiences and reasoning in novel contexts. However, the predominant emphasis in most interactive environments is on learning, often at the expense of complexity in reasoning. In this paper, we introduce CivRealm, an environment inspired by the Civilization game. Civilization\u2019s profound alignment with human society requires sophisticated learning and prior knowledge, while its ever-changing space and action space demand robust reasoning for generalization. Particularly, CivRealm sets up an imperfect-information general-sum game with a changing number of players; it presents a plethora of complex features, challenging the agent to deal with open-ended stochastic environments that require diplomacy and negotiation skills. Within CivRealm, we provide interfaces for two typical agent types: tensor-based agents that focus on learning, and language-based agents that emphasize reasoning. To catalyze further research, we present initial results for both paradigms. The canonical RL-based agents exhibit reasonable performance in mini-games, whereas both RL- and LLM-based agents struggle to make substantial progress in the full game. Overall, CivRealm stands as a unique learning and reasoning challenge for decision-making agents. The code is available at https://github.com/bigai-ai/civrealm.", "pdf": "https://openreview.net/pdf/8e9b4884f752883a17e58942608d9cc769610a01.pdf"} {"title": "Open-ended VQA benchmarking of Vision-Language models by exploiting Classification datasets and their semantic hierarchy", "url": "https://openreview.net/forum?id=EXitynZhYn", "detail_url": "https://openreview.net/forum?id=EXitynZhYn", "authors": "Simon Ging,Maria Alejandra Bravo,Thomas Brox", "tags": "ICLR 2024,Spotlight", "abstract": "The evaluation of text-generative vision-language models is a challenging yet crucial endeavor. By addressing the limitations of existing Visual Question Answering (VQA) benchmarks and proposing innovative evaluation methodologies, our research seeks to advance our understanding of these models\u2019 capabilities. We propose a novel VQA benchmark based on well-known visual classification datasets which allows a granular evaluation of text-generative vision-language models and their comparison with discriminative vision-language models. To improve the assessment of coarse answers on fine-grained classification tasks, we suggest using the semantic hierarchy of the label space to ask automatically generated follow-up questions about the ground-truth category. Finally, we compare traditional NLP and LLM-based metrics for the problem of evaluating model predictions given ground-truth answers. We perform a human evaluation study upon which we base our decision on the final metric. We apply our benchmark to a suite of vision-language models and show a detailed comparison of their abilities on object, action, and attribute classification. Our contributions aim to lay the foundation for more precise and meaningful assessments, facilitating targeted progress in the exciting field of vision-language modeling.", "pdf": "https://openreview.net/pdf/aa20ca1d2b8b77f839621a27a8e079ff29b639d0.pdf"} {"title": "GIM: Learning Generalizable Image Matcher From Internet Videos", "url": "https://openreview.net/forum?id=NYN1b8GRGS", "detail_url": "https://openreview.net/forum?id=NYN1b8GRGS", "authors": "Xuelun Shen,zhipeng cai,Wei Yin,Matthias M\u00fcller,Zijun Li,Kaixuan Wang,Xiaozhi Chen,Cheng Wang", "tags": "ICLR 2024,Spotlight", "abstract": "Image matching is a fundamental computer vision problem. While learning-based methods achieve state-of-the-art performance on existing benchmarks, they generalize poorly to in-the-wild images. Such methods typically need to train separate models for different scene types (e.g., indoor vs. outdoor) and are impractical when the scene type is unknown in advance. One of the underlying problems is the limited scalability of existing data construction pipelines, which limits the diversity of standard image matching datasets. To address this problem, we propose GIM, a self-training framework for learning a single generalizable model based on any image matching architecture using internet videos, an abundant and diverse data source. Given an architecture, GIM first trains it on standard domain-specific datasets and then combines it with complementary matching methods to create dense labels on nearby frames of novel videos. These labels are filtered by robust fitting, and then enhanced by propagating them to distant frames. The final model is trained on propagated data with strong augmentations. Not relying on complex 3D reconstruction makes GIM much more efficient and less likely to fail than standard SfM-and-MVS based frameworks. We also propose ZEB, the first zero-shot evaluation benchmark for image matching. By mixing data from diverse domains, ZEB can thoroughly assess the cross-domain generalization performance of different methods. Experiments demonstrate the effectiveness and generality of GIM. Applying GIM consistently improves the zero-shot performance of 3 state-of-the-art image matching architectures as the number of downloaded videos increases (Fig. 1 (a)); with 50 hours of YouTube videos, the relative zero-shot performance improves by 6.9% \u2212 18.1%. GIM also enables generalization to extreme cross-domain data such as Bird Eye View (BEV) images of projected 3D point clouds (Fig. 1 (c)). More importantly, our single zero-shot model consistently outperforms domain-specific baselines when evaluated on downstream tasks inherent to their respective domains. The code will be released upon acceptance.", "pdf": "https://openreview.net/pdf/c51ae05af771bb0017771606dc8f8cc514003720.pdf"} {"title": "SyncDreamer: Generating Multiview-consistent Images from a Single-view Image", "url": "https://openreview.net/forum?id=MN3yH2ovHb", "detail_url": "https://openreview.net/forum?id=MN3yH2ovHb", "authors": "Yuan Liu,Cheng Lin,Zijiao Zeng,Xiaoxiao Long,Lingjie Liu,Taku Komura,Wenping Wang", "tags": "ICLR 2024,Spotlight", "abstract": "In this paper, we present a novel diffusion model called SyncDreamer that generates multiview-consistent images from a single-view image. Using pretrained large-scale 2D diffusion models, recent work Zero123 demonstrates the ability to generate plausible novel views from a single-view image of an object. However, maintaining consistency in geometry and colors for the generated images remains a challenge. To address this issue, we propose a synchronized multiview diffusion model that models the joint probability distribution of multiview images, enabling the generation of multiview-consistent images in a single reverse process. SyncDreamer synchronizes the intermediate states of all the generated images at every step of the reverse process through a 3D-aware feature attention mechanism that correlates the corresponding features across different views. Experiments show that SyncDreamer generates images with high consistency across different views, thus making it well-suited for various 3D generation tasks such as novel-view-synthesis, text-to-3D, and image-to-3D. Project page: https://liuyuan-pal.github.io/SyncDreamer/.", "pdf": "https://openreview.net/pdf/da8460ef21ce4f5cdf6d529397c3d451785545be.pdf"} {"title": "Finite-State Autoregressive Entropy Coding for Efficient Learned Lossless Compression", "url": "https://openreview.net/forum?id=D5mJSNtUtv", "detail_url": "https://openreview.net/forum?id=D5mJSNtUtv", "authors": "Yufeng Zhang,Hang Yu,Jianguo Li,Weiyao Lin", "tags": "ICLR 2024,Spotlight", "abstract": "Learned lossless data compression has garnered significant attention recently due to its superior compression ratios compared to traditional compressors. However, the computational efficiency of these models jeopardizes their practicality. This paper proposes a novel system for improving the compression ratio while maintaining computational efficiency for learned lossless data compression. Our approach incorporates two essential innovations. First, we propose the Finite-State AutoRegressive (FSAR) entropy coder, an efficient autoregressive Markov model based entropy coder that utilizes a lookup table to expedite autoregressive entropy coding. Next, we present a Straight-Through Hardmax Quantization (STHQ) scheme to enhance the optimization of discrete latent space. Our experiments show that the proposed lossless compression method could improve the compression ratio by up to 6\\% compared to the baseline, with negligible extra computational time. Our work provides valuable insights into enhancing the computational efficiency of learned lossless data compression, which can have practical applications in various fields. Code is available at https://github.com/alipay/Finite_State_Autoregressive_Entropy_Coding.", "pdf": "https://openreview.net/pdf/5594d916eebbbc22170cfdf6b5c25177191835fc.pdf"} {"title": "ToolLLM: Facilitating Large Language Models to Master 16000+ Real-world APIs", "url": "https://openreview.net/forum?id=dHng2O0Jjr", "detail_url": "https://openreview.net/forum?id=dHng2O0Jjr", "authors": "Yujia Qin,Shihao Liang,Yining Ye,Kunlun Zhu,Lan Yan,Yaxi Lu,Yankai Lin,Xin Cong,Xiangru Tang,Bill Qian,Sihan Zhao,Lauren Hong,Runchu Tian,Ruobing Xie,Jie Zhou,Mark Gerstein,dahai li,Zhiyuan Liu,Maosong Sun", "tags": "ICLR 2024,Spotlight", "abstract": "Despite the advancements of open-source large language models (LLMs), e.g., LLaMA, they remain significantly limited in tool-use capabilities, i.e., using external tools (APIs) to fulfill human instructions. The reason is that current instruction tuning largely focuses on basic language tasks but ignores the tool-use domain. This is in contrast to the excellent tool-use capabilities of state-of-the-art (SOTA) closed-source LLMs, e.g., ChatGPT. To bridge this gap, we introduce ToolLLM, a general tool-use framework encompassing data construction, model training, and evaluation. We first present ToolBench, an instruction-tuning dataset for tool use, which is constructed automatically using ChatGPT. Specifically, the construction can be divided into three stages: (i) API collection: we collect 16,464 real-world RESTful APIs spanning 49 categories from RapidAPI Hub; (ii) instruction generation: we prompt ChatGPT to generate diverse instructions involving these APIs, covering both single-tool and multi-tool scenarios; (iii) solution path annotation: we use ChatGPT to search for a valid solution path (chain of API calls) for each instruction. To enhance the reasoning capabilities of LLMs, we develop a novel depth-first search-based decision tree algorithm. It enables LLMs to evaluate multiple reasoning traces and expand the search space. Moreover, to evaluate the tool-use capabilities of LLMs, we develop an automatic evaluator: ToolEval. Based on ToolBench, we fine-tune LLaMA to obtain an LLM ToolLLaMA, and equip it with a neural API retriever to recommend appropriate APIs for each instruction. Experiments show that ToolLLaMA demonstrates a remarkable ability to execute complex instructions and generalize to unseen APIs, and exhibits comparable performance to ChatGPT. Our ToolLLaMA also demonstrates strong zero-shot generalization ability in an out-of-distribution tool-use dataset: APIBench.", "pdf": "https://openreview.net/pdf/ccd998057a57b2ae3f5615915e57554028f1ae35.pdf"} {"title": "Enhanced Face Recognition using Intra-class Incoherence Constraint", "url": "https://openreview.net/forum?id=uELjxVbrqG", "detail_url": "https://openreview.net/forum?id=uELjxVbrqG", "authors": "Yuanqing Huang,Yinggui Wang,Le Yang,Lei Wang", "tags": "ICLR 2024,Spotlight", "abstract": "The current face recognition (FR) algorithms has achieved a high level of accuracy, making further improvements increasingly challenging. While existing FR algorithms primarily focus on optimizing margins and loss functions, limited attention has been given to exploring the feature representation space. Therefore, this paper endeavors to improve FR performance in the view of feature representation space. Firstly, we consider two FR models that exhibit distinct performance discrepancies, where one model exhibits superior recognition accuracy compared to the other. We implement orthogonal decomposition on the features from the superior model along those from the inferior model and obtain two sub-features. Surprisingly, we find the sub-feature perpendicular to the inferior still possesses a certain level of face distinguishability. We adjust the modulus of the sub-features and recombine them through vector addition. Experiments demonstrate this recombination is likely to contribute to an improved facial feature representation, even better than features from the original superior model. Motivated by this discovery, we further consider how to improve FR accuracy when there is only one FR model available. Inspired by knowledge distillation, we incorporate the intra-class incoherence constraint (IIC) to solve the problem. Experiments on various FR benchmarks show the existing state-of-the-art method with IIC can be further improved, highlighting its potential to further enhance FR performance.", "pdf": "https://openreview.net/pdf/64f49baaba5e3d14bb727d898b334143acfde5f7.pdf"} {"title": "Entropy is not Enough for Test-Time Adaptation: From the Perspective of Disentangled Factors", "url": "https://openreview.net/forum?id=9w3iw8wDuE", "detail_url": "https://openreview.net/forum?id=9w3iw8wDuE", "authors": "Jonghyun Lee,Dahuin Jung,Saehyung Lee,Junsung Park,Juhyeon Shin,Uiwon Hwang,Sungroh Yoon", "tags": "ICLR 2024,Spotlight", "abstract": "Test-time adaptation (TTA) fine-tunes pre-trained deep neural networks for unseen test data. The primary challenge of TTA is limited access to the entire test dataset during online updates, causing error accumulation. To mitigate it, TTA methods have utilized the model output's entropy as a confidence metric that aims to determine which samples have a lower likelihood of causing error. Through experimental studies, however, we observed the unreliability of entropy as a confidence metric for TTA under biased scenarios and theoretically revealed that it stems from the neglect of the influence of latent disentangled factors of data on predictions. Building upon these findings, we introduce a novel TTA method named Destroy Your Object (DeYO), which leverages a newly proposed confidence metric named Pseudo-Label Probability Difference (PLPD). PLPD quantifies the influence of the shape of an object on prediction by measuring the difference between predictions before and after applying an object-destructive transformation. DeYO consists of sample selection and sample weighting, which employ entropy and PLPD concurrently. For robust adaptation, DeYO prioritizes samples that dominantly incorporate shape information when making predictions. Our extensive experiments demonstrate the consistent superiority of DeYO over baseline methods across various scenarios, including biased and wild. Project page is publicly available at https://whitesnowdrop.github.io/DeYO/.", "pdf": "https://openreview.net/pdf/efba9a12bfc53499b73044f7ba478887cccc7d77.pdf"} {"title": "SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution", "url": "https://openreview.net/forum?id=CGlczSBBSj", "detail_url": "https://openreview.net/forum?id=CGlczSBBSj", "authors": "Wenlong Zhang,Xiaohui Li,Xiangyu Chen,Xiaoyun Zhang,Yu Qiao,Xiao-Ming Wu,Chao Dong", "tags": "ICLR 2024,Spotlight", "abstract": "Real-world Super-Resolution (Real-SR) methods focus on dealing with diverse real-world images and have attracted increasing attention in recent years. The key idea is to use a complex and high-order degradation model to mimic real-world degradations. \nAlthough they have achieved impressive results in various scenarios, they are faced with the obstacle of evaluation. Currently, these methods are only assessed by their average performance on a small set of degradation cases randomly selected from a large space, which fails to provide a comprehensive understanding of their overall performance and often yields inconsistent and potentially misleading results.\nTo overcome the limitation in evaluation, we propose SEAL, a framework for systematic evaluation of real-SR. In particular, we cluster the extensive degradation space to create a set of representative degradation cases, which serves as a comprehensive test set. Next, we propose a coarse-to-fine evaluation protocol to measure the distributed and relative performance of real-SR methods on the test set. The protocol incorporates two new metrics: acceptance rate (AR) and relative performance ratio (RPR), derived from acceptance and excellence lines. Under SEAL, we benchmark existing real-SR methods, obtain new observations and insights into their performance, and develop a new strong baseline. We consider SEAL as the first step towards creating an unbiased and comprehensive real-SR evaluation platform, which can promote the development of real-SR.", "pdf": "https://openreview.net/pdf/ef2c19059bfaeab85c449243ad90b25a62910da9.pdf"} {"title": "Learning Energy-Based Models by Cooperative Diffusion Recovery Likelihood", "url": "https://openreview.net/forum?id=AyzkDpuqcl", "detail_url": "https://openreview.net/forum?id=AyzkDpuqcl", "authors": "Yaxuan Zhu,Jianwen Xie,Ying Nian Wu,Ruiqi Gao", "tags": "ICLR 2024,Spotlight", "abstract": "Training energy-based models (EBMs) on high-dimensional data can be both challenging and time-consuming, and there exists a noticeable gap in sample quality between EBMs and other generative frameworks like GANs and diffusion models. To close this gap, inspired by the recent efforts of learning EBMs by maximimizing diffusion recovery likelihood (DRL), we propose cooperative diffusion recovery likelihood (CDRL), an effective approach to tractably learn and sample from a series of EBMs defined on increasingly noisy versons of a dataset, paired with an initializer model for each EBM. At each noise level, the two models are jointly estimated within a cooperative training framework: Samples from the initializer serve as starting points that are refined by a few MCMC sampling steps from the EBM. The EBM is then optimized by maximizing recovery likelihood, while the initializer model is optimized by learning from the difference between the refined samples and the initial samples. In addition, we made several practical designs for EBM training to further improve the sample quality. Combining these advances, we significantly boost the generation performance compared to existing EBM methods on CIFAR-10 and ImageNet 32x32. And we have shown that CDRL has great potential to largely reduce the sampling time. We also demonstrate the effectiveness of our models for several downstream tasks, including classifier-free guided generation, compositional generation, image inpainting and out-of-distribution detection.", "pdf": "https://openreview.net/pdf/2e17cdd8ba006f1c205e1971f050d6cb5e8d572d.pdf"} {"title": "MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning", "url": "https://openreview.net/forum?id=yLClGs770I", "detail_url": "https://openreview.net/forum?id=yLClGs770I", "authors": "Xiang Yue,Xingwei Qu,Ge Zhang,Yao Fu,Wenhao Huang,Huan Sun,Yu Su,Wenhu Chen", "tags": "ICLR 2024,Spotlight", "abstract": "We introduce MAmmoTH, a series of open-source large language models (LLMs) specifically tailored for general math problem-solving. The MAmmoTH models are trained on MathInstruct, our meticulously curated instruction tuning dataset. MathInstruct is compiled from 13 math datasets with intermediate rationales, six of which have rationales newly curated by us. It presents a unique hybrid of chain-of-thought (CoT) and program-of-thought (PoT) rationales, and also ensures extensive coverage of diverse fields in math. The hybrid of CoT and PoT not only unleashes the potential of tool use but also allows different thought processes for different math problems. As a result, the MAmmoTH series substantially outperform existing open-source models on nine mathematical reasoning datasets across all scales with an average accuracy gain between 16% and 32%. Remarkably, our MAmmoTH-7B model reaches 33% on MATH (a competition-level dataset), which exceeds the best open-source 7B model (WizardMath) by 23%, and the MAmmoTH-34B model achieves 44% accuracy on MATH, even surpassing GPT-4\u2019s CoT result. Our work underscores the importance of diverse problem coverage and the use of hybrid rationales in developing superior math generalist models.", "pdf": "https://openreview.net/pdf/25429cff694fecfd301ec230f0fd6fb30e9c7373.pdf"} {"title": "Time Travel in LLMs: Tracing Data Contamination in Large Language Models", "url": "https://openreview.net/forum?id=2Rwq6c3tvr", "detail_url": "https://openreview.net/forum?id=2Rwq6c3tvr", "authors": "Shahriar Golchin,Mihai Surdeanu", "tags": "ICLR 2024,Spotlight", "abstract": "Data contamination, i.e., the presence of test data from downstream tasks in the training data of large language models (LLMs), is a potential major issue in measuring LLMs' real effectiveness on other tasks. We propose a straightforward yet effective method for identifying data contamination within LLMs. At its core, our approach starts by identifying potential contamination at the instance level; using this information, our approach then assesses wider contamination at the partition level. To estimate contamination of individual instances, we employ \"guided instruction:\" a prompt consisting of the dataset name, partition type, and the random-length initial segment of a reference instance, asking the LLM to complete it. An instance is flagged as contaminated if the LLM's output either exactly or nearly matches the latter segment of the reference. To understand if an entire partition is contaminated, we propose two ideas. The first idea marks a dataset partition as contaminated if the average overlap score with the reference instances (as measured by ROUGE-L or BLEURT) is statistically significantly better with the completions from guided instruction compared to a \"general instruction\" that does not include the dataset and partition name. The second idea marks a dataset partition as contaminated if a classifier based on GPT-4 with few-shot in-context learning prompt marks multiple generated completions as exact/near-exact matches of the corresponding reference instances. Our best method achieves an accuracy between 92% and 100% in detecting if an LLM is contaminated with seven datasets, containing train and test/validation partitions, when contrasted with manual evaluation by human experts. Further, our findings indicate that GPT-4 is contaminated with AG News, WNLI, and XSum datasets.", "pdf": "https://openreview.net/pdf/468dbd68de8bbc08d54772aa59abfabac15acc8c.pdf"} {"title": "Variational Inference for SDEs Driven by Fractional Noise", "url": "https://openreview.net/forum?id=rtx8B94JMS", "detail_url": "https://openreview.net/forum?id=rtx8B94JMS", "authors": "Rembert Daems,Manfred Opper,Guillaume Crevecoeur,Tolga Birdal", "tags": "ICLR 2024,Spotlight", "abstract": "We present a novel variational framework for performing inference in (neural) stochastic differential equations (SDEs) driven by Markov-approximate fractional Brownian motion (fBM). SDEs offer a versatile tool for modeling real-world continuous-time dynamic systems with inherent noise and randomness. Combining SDEs with the powerful inference capabilities of variational methods, enables the learning of representative distributions through stochastic gradient descent. However, conventional SDEs typically assume the underlying noise to follow a Brownian motion (BM), which hinders their ability to capture long-term dependencies. In contrast, fractional Brownian motion (fBM) extends BM to encompass non-Markovian dynamics, but existing methods for inferring fBM parameters are either computationally demanding or statistically inefficient. \n\nIn this paper, building upon the Markov approximation of fBM, we derive the evidence lower bound essential for efficient variational inference of posterior path measures, drawing from the well-established field of stochastic analysis. Additionally, we provide a closed-form expression for optimal approximation coefficients and propose to use neural networks to learn the drift, diffusion and control terms within our variational posterior, leading to the variational training of neural-SDEs. In this framework, we also optimize the Hurst index, governing the nature of our fractional noise. Beyond validation on synthetic data, we contribute a novel architecture for variational latent video prediction,\u2014an approach that, to the best of our knowledge, enables the first variational neural-SDE application to video perception.", "pdf": "https://openreview.net/pdf/d5ed1bbdd5e5bd4a9dd5fa78f017dc8eed8f7fbc.pdf"} {"title": "Implicit regularization of deep residual networks towards neural ODEs", "url": "https://openreview.net/forum?id=AbXGwqb5Ht", "detail_url": "https://openreview.net/forum?id=AbXGwqb5Ht", "authors": "Pierre Marion,Yu-Han Wu,Michael Eli Sander,G\u00e9rard Biau", "tags": "ICLR 2024,Spotlight", "abstract": "Residual neural networks are state-of-the-art deep learning models. Their continuous-depth analog, neural ordinary differential equations (ODEs), are also widely used. Despite their success, the link between the discrete and continuous models still lacks a solid mathematical foundation. In this article, we take a step in this direction by establishing an implicit regularization of deep residual networks towards neural ODEs, for nonlinear networks trained with gradient flow. We prove that if the network is initialized as a discretization of a neural ODE, then such a discretization holds throughout training. Our results are valid for a finite training time, and also as the training time tends to infinity provided that the network satisfies a Polyak-\u0141ojasiewicz condition. Importantly, this condition holds for a family of residual networks where the residuals are two-layer perceptrons with an overparameterization in width that is only linear, and implies the convergence of gradient flow to a global minimum. Numerical experiments illustrate our results.", "pdf": "https://openreview.net/pdf/e6ce44da0d154a08f419bcc961a66debd0c7bdf7.pdf"} {"title": "NetInfoF Framework: Measuring and Exploiting Network Usable Information", "url": "https://openreview.net/forum?id=KY8ZNcljVU", "detail_url": "https://openreview.net/forum?id=KY8ZNcljVU", "authors": "Meng-Chieh Lee,Haiyang Yu,Jian Zhang,Vassilis N. Ioannidis,Xiang song,Soji Adeshina,Da Zheng,Christos Faloutsos", "tags": "ICLR 2024,Spotlight", "abstract": "Given a node-attributed graph, and a graph task (link prediction or node classification), can we tell if a graph neural network (GNN) will perform well? More specifically, do the graph structure and the node features carry enough usable information for the task? Our goals are\n(1) to develop a fast tool to measure how much information is in the graph structure and in the node features, and\n(2) to exploit the information to solve the task, if there is enough.\nWe propose NetInfoF, a framework including NetInfoF_Probe and NetInfoF_Act, for the measurement and the exploitation of network usable information (NUI), respectively. Given a graph data, NetInfoF_Probe measures NUI without any model training, and NetInfoF_Act solves link prediction and node classification, while two modules share the same backbone.\nIn summary, NetInfoF has following notable advantages:\n(a) General, handling both link prediction and node classification;\n(b) Principled, with theoretical guarantee and closed-form solution;\n(c) Effective, thanks to the proposed adjustment to node similarity;\n(d) Scalable, scaling linearly with the input size.\nIn our carefully designed synthetic datasets, NetInfoF correctly identifies the ground truth of NUI and is the only method being robust to all graph scenarios. Applied on real-world datasets, NetInfoF wins in 11 out of 12 times on link prediction compared to general GNN baselines.", "pdf": "https://openreview.net/pdf/472fc3640b561e46ab9bb9189f262b42f5ef4114.pdf"} {"title": "BarLeRIa: An Efficient Tuning Framework for Referring Image Segmentation", "url": "https://openreview.net/forum?id=wHLDHRkmEu", "detail_url": "https://openreview.net/forum?id=wHLDHRkmEu", "authors": "Yaoming Wang,Jin Li,XIAOPENG ZHANG,Bowen Shi,Chenglin Li,Wenrui Dai,Hongkai Xiong,Qi Tian", "tags": "ICLR 2024,Spotlight", "abstract": "Pre-training followed by full fine-tuning has gradually been substituted by Parameter-Efficient Tuning (PET) in the field of computer vision. PET has gained popularity, especially in the context of large-scale models, due to its ability to reduce transfer learning costs and conserve hardware resources. However, existing PET approaches primarily focus on recognition tasks and typically support uni-modal optimization, while neglecting dense prediction tasks and vision language interactions. To address this limitation, we propose a novel PET framework called **B**i-direction**a**l Inte**r**twined Vision **L**anguage Effici**e**nt Tuning for **R**eferring **I**mage Segment**a**tion (**BarLeRIa**), which leverages bi-directional intertwined vision language adapters to fully exploit the frozen pre-trained models' potential in cross-modal dense prediction tasks. In BarLeRIa, two different tuning modules are employed for efficient attention, one for global, and the other for local, along with an intertwined vision language tuning module for efficient modal fusion.\nExtensive experiments conducted on RIS benchmarks demonstrate the superiority of BarLeRIa over prior PET methods with a significant margin, i.e., achieving an average improvement of 5.6\\%. Remarkably, without requiring additional training datasets, BarLeRIa even surpasses SOTA full fine-tuning approaches. The code is available at https://github.com/NastrondAd/BarLeRIa.", "pdf": "https://openreview.net/pdf/48b914767a0a6be82afb6de0bda746780fa23312.pdf"} {"title": "Local Search GFlowNets", "url": "https://openreview.net/forum?id=6cFcw1Rxww", "detail_url": "https://openreview.net/forum?id=6cFcw1Rxww", "authors": "Minsu Kim,Taeyoung Yun,Emmanuel Bengio,Dinghuai Zhang,Yoshua Bengio,Sungsoo Ahn,Jinkyoo Park", "tags": "ICLR 2024,Spotlight", "abstract": "Generative Flow Networks (GFlowNets) are amortized sampling methods that learn a distribution over discrete objects proportional to their rewards. GFlowNets exhibit a remarkable ability to generate diverse samples, yet occasionally struggle to consistently produce samples with high rewards due to over-exploration on wide sample space. \nThis paper proposes to train GFlowNets with local search, which focuses on exploiting high-rewarded sample space to resolve this issue. Our main idea is to explore the local neighborhood via backtracking and reconstruction guided by backward and forward policies, respectively. This allows biasing the samples toward high-reward solutions, which is not possible for a typical GFlowNet solution generation scheme, which uses the forward policy to generate the solution from scratch. Extensive experiments demonstrate a remarkable performance improvement in several biochemical tasks. Source code is available: \\url{https://github.com/dbsxodud-11/ls_gfn}.", "pdf": "https://openreview.net/pdf/b6a0d0454e4fa4256c84c2e635908eef4469ccb2.pdf"} {"title": "Enabling Efficient Equivariant Operations in the Fourier Basis via Gaunt Tensor Products", "url": "https://openreview.net/forum?id=mhyQXJ6JsK", "detail_url": "https://openreview.net/forum?id=mhyQXJ6JsK", "authors": "Shengjie Luo,Tianlang Chen,Aditi S. Krishnapriyan", "tags": "ICLR 2024,Spotlight", "abstract": "Developing equivariant neural networks for the E(3) group plays an important role in modeling 3D data across real-world applications. Enforcing this equivariance primarily involves the tensor products of irreducible representations (irreps). However, the computational complexity of such operations increases significantly as higher-order tensors are used. In this work, we propose a systematic approach to substantially accelerate the computation of the tensor products of irreps. We mathematically connect the commonly used Clebsch-Gordan coefficients to the Gaunt coefficients, which are integrals of products of three spherical harmonics. Through Gaunt coefficients, the tensor product of irreps becomes equivalent to the multiplication between spherical functions represented by spherical harmonics. This perspective further allows us to change the basis for the equivariant operations from spherical harmonics to a 2D Fourier basis. Consequently, the multiplication between spherical functions represented by a 2D Fourier basis can be efficiently computed via the convolution theorem and Fast Fourier Transforms. This transformation reduces the complexity of full tensor products of irreps from $\\mathcal{O}(L^6)$ to $\\mathcal{O}(L^3)$, where $L$ is the max degree of irreps. Leveraging this approach, we introduce the Gaunt Tensor Product, which serves as a new method to construct efficient equivariant operations across different model architectures. Our experiments on the Open Catalyst Project and 3BPA datasets demonstrate both the increased efficiency and improved performance of our approach. The code and models will be made publicly available at https://github.com/lsj2408/Gaunt-Tensor-Product.", "pdf": "https://openreview.net/pdf/02f7b65d108e84bedf6eff36f0c47e0923014ad5.pdf"} {"title": "Idempotence and Perceptual Image Compression", "url": "https://openreview.net/forum?id=Cy5v64DqEF", "detail_url": "https://openreview.net/forum?id=Cy5v64DqEF", "authors": "Tongda Xu,Ziran Zhu,Dailan He,Yanghao Li,Lina Guo,Yuanyuan Wang,Zhe Wang,Hongwei Qin,Yan Wang,Jingjing Liu,Ya-Qin Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "Idempotence is the stability of image codec to re-compression. At the first glance, it is unrelated to perceptual image compression. However, we find that theoretically: 1) Conditional generative model-based perceptual codec satisfies idempotence; 2) Unconditional generative model with idempotence constraint is equivalent to conditional generative codec. Based on this newfound equivalence, we propose a new paradigm of perceptual image codec by inverting unconditional generative model with idempotence constraints. Our codec is theoretically equivalent to conditional generative codec, and it does not require training new models. Instead, it only requires a pre-trained mean-square-error codec and unconditional generative model. Empirically, we show that our proposed approach outperforms state-of-the-art methods such as HiFiC and ILLM, in terms of Fr\u00e9chet Inception Distance (FID). The source code is provided in https://github.com/tongdaxu/Idempotence-and-Perceptual-Image-Compression.", "pdf": "https://openreview.net/pdf/64ab88c07b072dd26084986657e2d1311fae2331.pdf"} {"title": "Forward $\\chi^2$ Divergence Based Variational Importance Sampling", "url": "https://openreview.net/forum?id=HD5Y7M8Xdk", "detail_url": "https://openreview.net/forum?id=HD5Y7M8Xdk", "authors": "Chengrui Li,Yule Wang,Weihan Li,Anqi Wu", "tags": "ICLR 2024,Spotlight", "abstract": "Maximizing the marginal log-likelihood is a crucial aspect of learning latent variable models, and variational inference (VI) stands as the commonly adopted method. However, VI can encounter challenges in achieving a high marginal log-likelihood when dealing with complicated posterior distributions. In response to this limitation, we introduce a novel variational importance sampling (VIS) approach that directly estimates and maximizes the marginal log-likelihood. VIS leverages the optimal proposal distribution, achieved by minimizing the forward $\\chi^2$ divergence, to enhance marginal log-likelihood estimation. We apply VIS to various popular latent variable models, including mixture models, variational auto-encoders, and partially observable generalized linear models. Results demonstrate that our approach consistently outperforms state-of-the-art baselines, in terms of both log-likelihood and model parameter estimation. Code: \\url{https://github.com/JerrySoybean/vis}.", "pdf": "https://openreview.net/pdf/e8661138b9bebb99452c4310f1eb71e825b8ab04.pdf"} {"title": "Noisy Interpolation Learning with Shallow Univariate ReLU Networks", "url": "https://openreview.net/forum?id=GTUoTJXPBf", "detail_url": "https://openreview.net/forum?id=GTUoTJXPBf", "authors": "Nirmit Joshi,Gal Vardi,Nathan Srebro", "tags": "ICLR 2024,Spotlight", "abstract": "Understanding how overparameterized neural networks generalize despite perfect interpolation of noisy training data is a fundamental question. Mallinar et. al. (2022) noted that neural networks seem to often exhibit ``tempered overfitting'', wherein the population risk does not converge to the Bayes optimal error, but neither does it approach infinity, yielding non-trivial generalization. However, this has not been studied rigorously. We provide the first rigorous analysis of the overfiting behaviour of regression with minimum norm ($\\ell_2$ of weights), focusing on univariate two-layer ReLU networks. We show overfitting is tempered (with high probability) when measured with respect to the $L_1$ loss, but also show that the situation is more complex than suggested by Mallinar et. al., and overfitting is catastrophic with respect to the $L_2$ loss, or when taking an expectation over the training set.", "pdf": "https://openreview.net/pdf/aa43dd315f355d322a924dea13a754935635b4b7.pdf"} {"title": "Initializing Models with Larger Ones", "url": "https://openreview.net/forum?id=dyrGMhicMw", "detail_url": "https://openreview.net/forum?id=dyrGMhicMw", "authors": "Zhiqiu Xu,Yanjie Chen,Kirill Vishniakov,Yida Yin,Zhiqiang Shen,Trevor Darrell,Lingjie Liu,Zhuang Liu", "tags": "ICLR 2024,Spotlight", "abstract": "Weight initialization plays an important role in neural network training. Widely used initialization methods are proposed and evaluated for networks that are trained from scratch. However, the growing number of pretrained models now offers new opportunities for tackling this classical problem of weight initialization. In this work, we introduce weight selection, a method for initializing smaller models by selecting a subset of weights from a pretrained larger model. This enables the transfer of knowledge from pretrained weights to smaller models. Our experiments demonstrate that weight selection can significantly enhance the performance of small models and reduce their training time. Notably, it can also be used together with knowledge distillation. Weight selection offers a new approach to leverage the power of pretrained models in resource-constrained settings, and we hope it can be a useful tool for training small models in the large-model era.", "pdf": "https://openreview.net/pdf/d6512714812d4eaa4cfab7e7cb04a9809c574f79.pdf"} {"title": "DMV3D: Denoising Multi-view Diffusion Using 3D Large Reconstruction Model", "url": "https://openreview.net/forum?id=H4yQefeXhp", "detail_url": "https://openreview.net/forum?id=H4yQefeXhp", "authors": "Yinghao Xu,Hao Tan,Fujun Luan,Sai Bi,Peng Wang,Jiahao Li,Zifan Shi,Kalyan Sunkavalli,Gordon Wetzstein,Zexiang Xu,Kai Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "We propose DMV3D, a novel 3D generation approach that uses a transformer-based 3D large reconstruction model to denoise multi-view diffusion. Our reconstruction model incorporates a triplane NeRF representation and, functioning as a denoiser, can denoise noisy multi-view images via 3D NeRF reconstruction and rendering, achieving single-stage 3D generation in the 2D diffusion denoising process. We train DMV3D on large-scale multi-view image datasets of extremely diverse objects using only image reconstruction losses, without accessing 3D assets. We demonstrate state-of-the-art results for the single-image reconstruction problem where probabilistic modeling of unseen object parts is required for generating diverse reconstructions with sharp textures. We also show high-quality text-to-3D generation results outperforming previous 3D diffusion models. Our project website is at: https://dmv3d.github.io/.", "pdf": "https://openreview.net/pdf/dda54a3f50dec9a3cde3428a8330614a5c258d9a.pdf"} {"title": "Influencer Backdoor Attack on Semantic Segmentation", "url": "https://openreview.net/forum?id=VmGRoNDQgJ", "detail_url": "https://openreview.net/forum?id=VmGRoNDQgJ", "authors": "Haoheng Lan,Jindong Gu,Philip Torr,Hengshuang Zhao", "tags": "ICLR 2024,Spotlight", "abstract": "When a small number of poisoned samples are injected into the training dataset of a deep neural network, the network can be induced to exhibit malicious behavior during inferences, which poses potential threats to real-world applications. While they have been intensively studied in classification, backdoor attacks on semantic segmentation have been largely overlooked. Unlike classification, semantic segmentation aims to classify every pixel within a given image. In this work, we explore backdoor attacks on segmentation models to misclassify all pixels of a victim class by injecting a specific trigger on non-victim pixels during inferences, which is dubbed Influencer Backdoor Attack (IBA). IBA is expected to maintain the classification accuracy of non-victim pixels and mislead classifications of all victim pixels in every single inference and could be easily applied to real-world scenes. Based on the context aggregation ability of segmentation models, we proposed a simple, yet effective, Nearest-Neighbor trigger injection strategy. We also introduce an innovative Pixel Random Labeling strategy which maintains optimal performance even when the trigger is placed far from the victim pixels. Our extensive experiments reveal that current segmentation models do suffer from backdoor attacks, demonstrate IBA real-world applicability, and show that our proposed techniques can further increase attack performance.", "pdf": "https://openreview.net/pdf/81df12528640b4d855f26ec8e8c61fb38da33dcd.pdf"} {"title": "PF-LRM: Pose-Free Large Reconstruction Model for Joint Pose and Shape Prediction", "url": "https://openreview.net/forum?id=noe76eRcPC", "detail_url": "https://openreview.net/forum?id=noe76eRcPC", "authors": "Peng Wang,Hao Tan,Sai Bi,Yinghao Xu,Fujun Luan,Kalyan Sunkavalli,Wenping Wang,Zexiang Xu,Kai Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "We propose a Pose-Free Large Reconstruction Model (PF-LRM) for reconstructing a 3D object from a few unposed images even with little visual overlap, while simultaneously estimating the relative camera poses in ~1.3 seconds on a single A100 GPU. PF-LRM is a highly scalable method utilizing self-attention blocks to exchange information between 3D object tokens and 2D image tokens; we predict a coarse point cloud for each view, and then use a differentiable Perspective-n-Point (PnP) solver to obtain camera poses. When trained on a huge amount of multi-view posed data of ~1M objects, PF-LRM shows strong cross-dataset generalization ability, and outperforms baseline methods by a large margin in terms of pose prediction accuracy and 3D reconstruction quality on various unseen evaluation datasets. We also demonstrate our model's applicability in downstream text/image-to-3D task with fast feed-forward inference. Our project website is at: https://totoro97.github.io/pf-lrm.", "pdf": "https://openreview.net/pdf/edf0c386092f3593a9e749732dcd13300f2d0ea8.pdf"} {"title": "Procedural Fairness Through Decoupling Objectionable Data Generating Components", "url": "https://openreview.net/forum?id=cxfPefbu1s", "detail_url": "https://openreview.net/forum?id=cxfPefbu1s", "authors": "Zeyu Tang,Jialu Wang,Yang Liu,Peter Spirtes,Kun Zhang", "tags": "ICLR 2024,Spotlight", "abstract": "We reveal and address the frequently overlooked yet important issue of _disguised procedural unfairness_, namely, the potentially inadvertent alterations on the behavior of neutral (i.e., not problematic) aspects of data generating process, and/or the lack of procedural assurance of the greatest benefit of the least advantaged individuals. Inspired by John Rawls's advocacy for _pure procedural justice_ (Rawls, 1971; 2001), we view automated decision-making as a microcosm of social institutions, and consider how the data generating process itself can satisfy the requirements of procedural fairness. We propose a framework that decouples the objectionable data generating components from the neutral ones by utilizing reference points and the associated value instantiation rule. Our findings highlight the necessity of preventing _disguised procedural unfairness_, drawing attention not only to the objectionable data generating components that we aim to mitigate, but also more importantly, to the neutral components that we intend to keep unaffected.", "pdf": "https://openreview.net/pdf/722f3994df5037dded9effb23cabfc87c0824ce6.pdf"} {"title": "Vision-Language Foundation Models as Effective Robot Imitators", "url": "https://openreview.net/forum?id=lFYj0oibGR", "detail_url": "https://openreview.net/forum?id=lFYj0oibGR", "authors": "Xinghang Li,Minghuan Liu,Hanbo Zhang,Cunjun Yu,Jie Xu,Hongtao Wu,Chilam Cheang,Ya Jing,Weinan Zhang,Huaping Liu,Hang Li,Tao Kong", "tags": "ICLR 2024,Spotlight", "abstract": "Recent progress in vision language foundation models has shown their ability to understand multimodal data and resolve complicated vision language tasks, including robotics manipulation. We seek a straightforward way of making use of existing vision-language models (VLMs) with simple fine-tuning on robotics data.\nTo this end, we derive a simple and novel vision-language manipulation framework, dubbed RoboFlamingo, built upon the open-source VLMs, OpenFlamingo. Unlike prior works, RoboFlamingo utilizes pre-trained VLMs for single-step vision-language comprehension, models sequential history information with an explicit policy head, and is slightly fine-tuned by imitation learning only on language-conditioned manipulation datasets. Such a decomposition provides RoboFlamingo the flexibility for open-loop control and deployment on low-performance platforms. By exceeding the state-of-the-art performance with a large margin on the tested benchmark, we show RoboFlamingo can be an effective and competitive alternative to adapt VLMs to robot control.\nOur extensive experimental results also reveal several interesting conclusions regarding the behavior of different pre-trained VLMs on manipulation tasks. We believe RoboFlamingo has the potential to be a cost-effective and easy-to-use solution for robotics manipulation, empowering everyone with the ability to fine-tune their own robotics policy. Our code will be made public upon acceptance.", "pdf": "https://openreview.net/pdf/a1af30de31af35eb9ee853f1020a7e7041aa3e9a.pdf"} {"title": "OctoPack: Instruction Tuning Code Large Language Models", "url": "https://openreview.net/forum?id=mw1PWNSWZP", "detail_url": "https://openreview.net/forum?id=mw1PWNSWZP", "authors": "Niklas Muennighoff,Qian Liu,Armel Randy Zebaze,Qinkai Zheng,Binyuan Hui,Terry Yue Zhuo,Swayam Singh,Xiangru Tang,Leandro Von Werra,Shayne Longpre", "tags": "ICLR 2024,Spotlight", "abstract": "Finetuning large language models (LLMs) on instructions leads to vast performance improvements on natural language tasks. We apply instruction tuning using code, leveraging the natural structure of Git commits, which pair code changes with human instructions. We compile CommitPack: 4 terabytes of Git commits across 350 programming languages. We benchmark CommitPack against other natural and synthetic code instructions (xP3x, Self-Instruct, OASST) on the 16B parameter StarCoder model, and achieve state-of-the-art performance among models not trained on OpenAI outputs, on the HumanEval Python benchmark (46.2% pass@1). We further introduce HumanEvalPack, expanding the HumanEval benchmark to a total of 3 coding tasks (Code Repair, Code Explanation, Code Synthesis) across 6 languages (Python, JavaScript, Java, Go, C++, Rust). Our models, OctoCoder and OctoGeeX, achieve the best performance across HumanEvalPack among all permissive models, demonstrating CommitPack's benefits in generalizing to a wider set of languages and natural coding tasks. Code, models and data are freely available at https://github.com/bigcode-project/octopack.", "pdf": "https://openreview.net/pdf/2b332f4f4e9406870d019ed23d22f771fefbe70f.pdf"} {"title": "Q-Bench: A Benchmark for General-Purpose Foundation Models on Low-level Vision", "url": "https://openreview.net/forum?id=0V5TVt9bk0", "detail_url": "https://openreview.net/forum?id=0V5TVt9bk0", "authors": "Haoning Wu,Zicheng Zhang,Erli Zhang,Chaofeng Chen,Liang Liao,Annan Wang,Chunyi Li,Wenxiu Sun,Qiong Yan,Guangtao Zhai,Weisi Lin", "tags": "ICLR 2024,Spotlight", "abstract": "The rapid evolution of Multi-modality Large Language Models (MLLMs) has catalyzed a shift in computer vision from specialized models to general-purpose foundation models. Nevertheless, there is still an inadequacy in assessing the abilities of MLLMs on **low-level visual perception and understanding**. To address this gap, we present **Q-Bench**, a holistic benchmark crafted to systematically evaluate potential abilities of MLLMs on three realms: low-level visual perception, low-level visual description, and overall visual quality assessment. **_a)_** To evaluate the low-level **_perception_** ability, we construct the **LLVisionQA** dataset, consisting of 2,990 diverse-sourced images, each equipped with a human-asked question focusing on its low-level attributes. We then measure the correctness of MLLMs on answering these questions. **_b)_** To examine the **_description_** ability of MLLMs on low-level information, we propose the **LLDescribe** dataset consisting of long expert-labelled *golden* low-level text descriptions on 499 images, and a GPT-involved comparison pipeline between outputs of MLLMs and the *golden* descriptions. **_c)_** Besides these two tasks, we further measure their visual quality **_assessment_** ability to align with human opinion scores. Specifically, we design a softmax-based strategy that enables MLLMs to predict *quantifiable* quality scores, and evaluate them on various existing image quality assessment (IQA) datasets. Our evaluation across the three abilities confirms that MLLMs possess preliminary low-level visual skills. However, these skills are still unstable and relatively imprecise, indicating the need for specific enhancements on MLLMs towards these abilities. We hope that our benchmark can encourage the research community to delve deeper to discover and enhance these untapped potentials of MLLMs.", "pdf": "https://openreview.net/pdf/5cb0fc8befab83d5b0fb1f3257587eb3a4243387.pdf"} {"title": "iTransformer: Inverted Transformers Are Effective for Time Series Forecasting", "url": "https://openreview.net/forum?id=JePfAI8fah", "detail_url": "https://openreview.net/forum?id=JePfAI8fah", "authors": "Yong Liu,Tengge Hu,Haoran Zhang,Haixu Wu,Shiyu Wang,Lintao Ma,Mingsheng Long", "tags": "ICLR 2024,Spotlight", "abstract": "The recent boom of linear forecasting models questions the ongoing passion for architectural modifications of Transformer-based forecasters. These forecasters leverage Transformers to model the global dependencies over temporal tokens of time series, with each token formed by multiple variates of the same timestamp. However, Transformers are challenged in forecasting series with larger lookback windows due to performance degradation and computation explosion. Besides, the embedding for each temporal token fuses multiple variates that represent potential delayed events and distinct physical measurements, which may fail in learning variate-centric representations and result in meaningless attention maps. In this work, we reflect on the competent duties of Transformer components and repurpose the Transformer architecture without any modification to the basic components. We propose iTransformer that simply applies the attention and feed-forward network on the inverted dimensions. Specifically, the time points of individual series are embedded into variate tokens which are utilized by the attention mechanism to capture multivariate correlations; meanwhile, the feed-forward network is applied for each variate token to learn nonlinear representations. The iTransformer model achieves state-of-the-art on challenging real-world datasets, which further empowers the Transformer family with promoted performance, generalization ability across different variates, and better utilization of arbitrary lookback windows, making it a nice alternative as the fundamental backbone of time series forecasting. Code is available at this repository: https://github.com/thuml/iTransformer.", "pdf": "https://openreview.net/pdf/2dff92321d132ee1a70ba42c0046c99ee9bcd972.pdf"} {"title": "De novo Protein Design Using Geometric Vector Field Networks", "url": "https://openreview.net/forum?id=9UIGyJJpay", "detail_url": "https://openreview.net/forum?id=9UIGyJJpay", "authors": "Weian Mao,Muzhi Zhu,Zheng Sun,Shuaike Shen,Lin Yuanbo Wu,Hao Chen,Chunhua Shen", "tags": "ICLR 2024,Spotlight", "abstract": "Advances like protein diffusion have marked revolutionary progress in $\\textit{de novo}$ protein design, a central topic in life science. These methods typically depend on protein structure encoders to model residue backbone frames, where atoms do not exist. Most prior encoders rely on atom-wise features, such as angles and distances between atoms, which are not available in this context. Only a few basic encoders, like IPA, have been proposed for this scenario, exposing the frame modeling as a bottleneck. In this work, we introduce the Vector Field Network (VFN), that enables network layers to perform learnable vector computations between coordinates of frame-anchored virtual atoms, thus achieving a higher capability for modeling frames. The vector computation operates in a manner similar to a linear layer, with each input channel receiving 3D virtual atom coordinates instead of scalar values. The multiple feature vectors output by the vector computation are then used to update the residue representations and virtual atom coordinates via attention aggregation. Remarkably, VFN also excels in modeling both frames and atoms, as the real atoms can be treated as the virtual atoms for modeling, positioning VFN as a potential $\\textit{universal encoder}$. In protein diffusion (frame modeling), VFN exhibits a impressive performance advantage over IPA, excelling in terms of both designability ($\\textbf{67.04}$\\% vs. 53.58\\%) and diversity ($\\textbf{66.54}$\\% vs. 51.98\\%). In inverse folding(frame and atom modeling), VFN outperforms the previous SoTA model, PiFold ($\\textbf{54.7}$\\% vs. 51.66\\%), on sequence recovery rate; we also propose a method of equipping VFN with the ESM model, which significantly surpasses the previous ESM-based SoTA ($\\textbf{62.67}$\\% vs. 55.65\\%), LM-Design, by a substantial margin. Code is available at https://github.com/aim-uofa/VFN", "pdf": "https://openreview.net/pdf/88687a41b251ea3e5018907ebe06860c15e31ff3.pdf"} {"title": "Prompt Gradient Projection for Continual Learning", "url": "https://openreview.net/forum?id=EH2O3h7sBI", "detail_url": "https://openreview.net/forum?id=EH2O3h7sBI", "authors": "Jingyang Qiao,zhizhong zhang,Xin Tan,Chengwei Chen,Yanyun Qu,Yong Peng,Yuan Xie", "tags": "ICLR 2024,Spotlight", "abstract": "Prompt-tuning has demonstrated impressive performance in continual learning by querying relevant prompts for each input instance, which can avoid the introduction of task identifier. Its forgetting is therefore reduced as this instance-wise query mechanism enables us to select and update only relevant prompts. In this paper, we further integrate prompt-tuning with gradient projection approach. Our observation is: prompt-tuning releases the necessity of task identifier for gradient projection method; and gradient projection provides theoretical guarantees against forgetting for prompt-tuning. This inspires a new prompt gradient projection approach (PGP) for continual learning. In PGP, we deduce that reaching the orthogonal condition for prompt gradient can effectively prevent forgetting via the self-attention mechanism in vision-transformer. The condition equations are then realized by conducting Singular Value Decomposition (SVD) on an element-wise sum space between input space and prompt space. We validate our method on diverse datasets and experiments demonstrate the efficiency of reducing forgetting both in class incremental, online class incremental, and task incremental settings. The code is available at https://github.com/JingyangQiao/prompt-gradient-projection.", "pdf": "https://openreview.net/pdf/1328fe645bf366bf46fb88b51ebb5b08bc5e76f4.pdf"} {"title": "R-EDL: Relaxing Nonessential Settings of Evidential Deep Learning", "url": "https://openreview.net/forum?id=Si3YFA641c", "detail_url": "https://openreview.net/forum?id=Si3YFA641c", "authors": "Mengyuan Chen,Junyu Gao,Changsheng Xu", "tags": "ICLR 2024,Spotlight", "abstract": "A newly-arising uncertainty estimation method named Evidential Deep Learning (EDL), which can obtain reliable predictive uncertainty in a single forward pass, has garnered increasing interest. Guided by the subjective logic theory, EDL obtains Dirichlet concentration parameters from deep neural networks, thus constructing a Dirichlet probability density function (PDF) to model the distribution of class probabilities. Despite its great success, we argue that EDL keeps nonessential settings in both stages of model construction and optimization.\nIn this work, our analysis indicates that (1) in the construction of the Dirichlet PDF, a commonly ignored parameter termed prior weight governs the balance between leveraging the proportion of evidence and its magnitude in deriving predictive scores, and (2) in model optimization, a variance-minimized regularization term adopted by traditional EDL encourages the Dirichlet PDF to approach a Dirac delta function, potentially exacerbating overconfidence. Therefore, we propose the R-EDL (Relaxed-EDL) method by relaxing these nonessential settings. Specifically, R-EDL treats the prior weight as an adjustable hyper-parameter instead of a fixed scalar, and directly optimizes the expectation of the Dirichlet PDF provided to deprecate the variance-minimized regularization term. Extensive experiments and SOTA performances demonstrate the effectiveness of our method. Source codes are provided in Appendix E.", "pdf": "https://openreview.net/pdf/34d74bb1c927a5db5a344d0e49273f94e775cacd.pdf"} {"title": "Neuron Activation Coverage: Rethinking Out-of-distribution Detection and Generalization", "url": "https://openreview.net/forum?id=SNGXbZtK6Q", "detail_url": "https://openreview.net/forum?id=SNGXbZtK6Q", "authors": "Yibing Liu,Chris XING TIAN,Haoliang Li,Lei Ma,Shiqi Wang", "tags": "ICLR 2024,Spotlight", "abstract": "The out-of-distribution (OOD) problem generally arises when neural networks encounter data that significantly deviates from the training data distribution, i.e., in-distribution (InD). In this paper, we study the OOD problem from a neuron activation view. We first formulate neuron activation states by considering both the neuron output and its influence on model decisions. Then, to characterize the relationship between neurons and OOD issues, we introduce the *neuron activation coverage* (NAC) -- a simple measure for neuron behaviors under InD data. Leveraging our NAC, we show that 1) InD and OOD inputs can be largely separated based on the neuron behavior, which significantly eases the OOD detection problem and beats the 21 previous methods over three benchmarks (CIFAR-10, CIFAR-100, and ImageNet-1K). 2) a positive correlation between NAC and model generalization ability consistently holds across architectures and datasets, which enables a NAC-based criterion for evaluating model robustness. Compared to prevalent InD validation criteria, we show that NAC not only can select more robust models, but also has a stronger correlation with OOD test performance.", "pdf": "https://openreview.net/pdf/4c57d3fc275b61826d7f559c23288a82feebda62.pdf"} {"title": "ResFields: Residual Neural Fields for Spatiotemporal Signals", "url": "https://openreview.net/forum?id=EHrvRNs2Y0", "detail_url": "https://openreview.net/forum?id=EHrvRNs2Y0", "authors": "Marko Mihajlovic,Sergey Prokudin,Marc Pollefeys,Siyu Tang", "tags": "ICLR 2024,Spotlight", "abstract": "Neural fields, a category of neural networks trained to represent high-frequency signals, have gained significant attention in recent years due to their impressive performance in modeling complex 3D data, such as signed distance (SDFs) or radiance fields (NeRFs), via a single multi-layer perceptron (MLP). However, despite the power and simplicity of representing signals with an MLP, these methods still face challenges when modeling large and complex temporal signals due to the limited capacity of MLPs. In this paper, we propose an effective approach to address this limitation by incorporating temporal residual layers into neural fields, dubbed ResFields. It is a novel class of networks specifically designed to effectively represent complex temporal signals. We conduct a comprehensive analysis of the properties of ResFields and propose a matrix factorization technique to reduce the number of trainable parameters and enhance generalization capabilities. Importantly, our formulation seamlessly integrates with existing MLP-based neural fields and consistently improves results across various challenging tasks: 2D video approximation, dynamic shape modeling via temporal SDFs, and dynamic NeRF reconstruction. Lastly, we demonstrate the practical utility of ResFields by showcasing its effectiveness in capturing dynamic 3D scenes from sparse RGBD cameras of a lightweight capture system.", "pdf": "https://openreview.net/pdf/f19ba6990ec3698aae30dc2d3ca992a323a90cc3.pdf"} {"title": "TD-MPC2: Scalable, Robust World Models for Continuous Control", "url": "https://openreview.net/forum?id=Oxh5CstDJU", "detail_url": "https://openreview.net/forum?id=Oxh5CstDJU", "authors": "Nicklas Hansen,Hao Su,Xiaolong Wang", "tags": "ICLR 2024,Spotlight", "abstract": "TD-MPC is a model-based reinforcement learning (RL) algorithm that performs local trajectory optimization in the latent space of a learned implicit (decoder-free) world model. In this work, we present TD-MPC2: a series of improvements upon the TD-MPC algorithm. We demonstrate that TD-MPC2 improves significantly over baselines across 104 online RL tasks spanning 4 diverse task domains, achieving consistently strong results with a single set of hyperparameters. We further show that agent capabilities increase with model and data size, and successfully train a single 317M parameter agent to perform 80 tasks across multiple task domains, embodiments, and action spaces. We conclude with an account of lessons, opportunities, and risks associated with large TD-MPC2 agents.\n\nExplore videos, models, data, code, and more at https://tdmpc2.com", "pdf": "https://openreview.net/pdf/d7962b8e764953c9c4db96aa85307e9e9b689881.pdf"} {"title": "Stochastic Controlled Averaging for Federated Learning with Communication Compression", "url": "https://openreview.net/forum?id=jj5ZjZsWJe", "detail_url": "https://openreview.net/forum?id=jj5ZjZsWJe", "authors": "Xinmeng Huang,Ping Li,Xiaoyun Li", "tags": "ICLR 2024,Spotlight", "abstract": "Communication compression has been an important topic in Federated Learning (FL) for alleviating the communication overhead. However, communication compression brings forth new challenges in FL due to the interplay of compression-incurred information distortion and inherent characteristics of FL such as partial participation and data heterogeneity. Despite the recent development, the existing approaches either cannot accommodate arbitrary data heterogeneity or partial participation, or require stringent conditions on compression. In this paper, we revisit the seminal stochastic controlled averaging method by proposing an equivalent but more efficient/simplified formulation with halved uplink communication costs, building upon which we propose two compressed FL algorithms, SCALLION and SCAFCOM, to support unbiased and biased compression, respectively. Both the proposed methods outperform the existing compressed FL methods in terms of communication and computation complexities. Moreover,SCALLION and SCAFCOM attain fast convergence rates under arbitrary data heterogeneity without any additional assumptions on compression errors. Experiments show that \\scallion and \\scafcom outperform recent compressed FL methods under the same communication budget.", "pdf": "https://openreview.net/pdf/b8db6edbb504d97145e9868b1a1f677abcd2f776.pdf"} {"title": "AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning", "url": "https://openreview.net/forum?id=Fx2SbBgcte", "detail_url": "https://openreview.net/forum?id=Fx2SbBgcte", "authors": "Yuwei Guo,Ceyuan Yang,Anyi Rao,Zhengyang Liang,Yaohui Wang,Yu Qiao,Maneesh Agrawala,Dahua Lin,Bo Dai", "tags": "ICLR 2024,Spotlight", "abstract": "With the advance of text-to-image (T2I) diffusion models (e.g., Stable Diffusion) and corresponding personalization techniques such as DreamBooth and LoRA, everyone can manifest their imagination into high-quality images at an affordable cost. However, adding motion dynamics to existing high-quality personalized T2Is and enabling them to generate animations remains an open challenge. In this paper, we present AnimateDiff, a practical framework for animating personalized T2I models without requiring model-specific tuning. At the core of our framework is a plug-and-play motion module that can be trained once and seamlessly integrated into any personalized T2Is originating from the same base T2I. Through our proposed training strategy, the motion module effectively learns transferable motion priors from real-world videos. Once trained, the motion module can be inserted into a personalized T2I model to form a personalized animation generator. We further propose MotionLoRA, a lightweight fine-tuning technique for AnimateDiff that enables a pre-trained motion module to adapt to new motion patterns, such as different shot types, at a low training and data collection cost. We evaluate AnimateDiff and MotionLoRA on several public representative personalized T2I models collected from the community. The results demonstrate that our approaches help these models generate temporally smooth animation clips while preserving the visual quality and motion diversity. Codes and pre-trained weights are available at https://github.com/guoyww/AnimateDiff.", "pdf": "https://openreview.net/pdf/f6573e09993ba1372e3b282b7d2b9d1b19cef04f.pdf"} {"title": "Guiding Instruction-based Image Editing via Multimodal Large Language Models", "url": "https://openreview.net/forum?id=S1RKWSyZ2Y", "detail_url": "https://openreview.net/forum?id=S1RKWSyZ2Y", "authors": "Tsu-Jui Fu,Wenze Hu,Xianzhi Du,William Yang Wang,Yinfei Yang,Zhe Gan", "tags": "ICLR 2024,Spotlight", "abstract": "Instruction-based image editing improves the controllability and flexibility of image manipulation via natural commands without elaborate descriptions or regional masks. However, human instructions are sometimes too brief for current methods to capture and follow. Multimodal large language models (MLLMs) show promising capabilities in cross-modal understanding and visual-aware response generation via LMs. We investigate how MLLMs facilitate edit instructions and present MLLM-Guided Image Editing (MGIE). MGIE learns to derive expressive instructions and provides explicit guidance. The editing model jointly captures this visual imagination and performs manipulation through end-to-end training. We evaluate various aspects of Photoshop-style modification, global photo optimization, and local editing. Extensive experimental results demonstrate that expressive instructions are crucial to instruction-based image editing, and our MGIE can lead to a notable improvement in automatic metrics and human evaluation while maintaining competitive inference efficiency.", "pdf": "https://openreview.net/pdf/859bb6596377ab0632e753f95d8701a768c8a514.pdf"} {"title": "Tuning LayerNorm in Attention: Towards Efficient Multi-Modal LLM Finetuning", "url": "https://openreview.net/forum?id=YR3ETaElNK", "detail_url": "https://openreview.net/forum?id=YR3ETaElNK", "authors": "Bingchen Zhao,Haoqin Tu,Chen Wei,Jieru Mei,Cihang Xie", "tags": "ICLR 2024,Spotlight", "abstract": "This paper introduces an efficient strategy to transform Large Language Models (LLMs) into Multi-Modal Large Language Models. \nBy conceptualizing this transformation as a domain adaptation process, \\ie, transitioning from text understanding to embracing multiple modalities, we intriguingly note that, within each attention block, tuning LayerNorm suffices to yield strong performance. \nMoreover, when benchmarked against other tuning approaches like full parameter finetuning or LoRA, its benefits on efficiency are substantial.\nFor example, when compared to LoRA on a 13B model scale, performance can be enhanced by an average of over 20\\% across five multi-modal tasks, and meanwhile, \nresults in a significant reduction of trainable parameters by 41.9\\% and a decrease in GPU memory usage by 17.6\\%. On top of this LayerNorm strategy, we showcase that selectively tuning only with conversational data can improve efficiency further. \nBeyond these empirical outcomes, we provide a comprehensive analysis to explore the role of LayerNorm in adapting LLMs to the multi-modal domain and improving the expressive power of the model.", "pdf": "https://openreview.net/pdf/9ac467b9b8ce5db3006b34be33aadcefa332ff07.pdf"} {"title": "Universal Humanoid Motion Representations for Physics-Based Control", "url": "https://openreview.net/forum?id=OrOd8PxOO2", "detail_url": "https://openreview.net/forum?id=OrOd8PxOO2", "authors": "Zhengyi Luo,Jinkun Cao,Josh Merel,Alexander Winkler,Jing Huang,Kris M. Kitani,Weipeng Xu", "tags": "ICLR 2024,Spotlight", "abstract": "We present a universal motion representation that encompasses a comprehensive range of motor skills for physics-based humanoid control. Due to the high dimensionality of humanoids and the inherent difficulties in reinforcement learning, prior methods have focused on learning skill embeddings for a narrow range of movement styles (e.g. locomotion, game characters) from specialized motion datasets. This limited scope hampers their applicability in complex tasks. We close this gap by significantly increasing the coverage of our motion representation space. To achieve this, we first learn a motion imitator that can imitate all of human motion from a large, unstructured motion dataset. We then create our motion representation by distilling skills directly from the imitator. This is achieved by using an encoder-decoder structure with a variational information bottleneck. Additionally, we jointly learn a prior conditioned on proprioception (humanoid's own pose and velocities) to improve model expressiveness and sampling efficiency for downstream tasks. By sampling from the prior, we can generate long, stable, and diverse human motions. Using this latent space for hierarchical RL, we show that our policies solve tasks using human-like behavior. We demonstrate the effectiveness of our motion representation by solving generative tasks (e.g. strike, terrain traversal) and motion tracking using VR controllers.", "pdf": "https://openreview.net/pdf/233f936dc4663dc0d31d5b53ea52bdc601f0b2f0.pdf"} {"title": "Adaptive Rational Activations to Boost Deep Reinforcement Learning", "url": "https://openreview.net/forum?id=g90ysX1sVs", "detail_url": "https://openreview.net/forum?id=g90ysX1sVs", "authors": "Quentin Delfosse,Patrick Schramowski,Martin Mundt,Alejandro Molina,Kristian Kersting", "tags": "ICLR 2024,Spotlight", "abstract": "Latest insights from biology show that intelligence not only emerges from the connections between neurons, but that individual neurons shoulder more computational responsibility than previously anticipated. Specifically, neural plasticity should be critical in the context of constantly changing reinforcement learning (RL) environments, yet current approaches still primarily employ static activation functions. In this work, we motivate the use of adaptable activation functions in RL and show that rational activation functions are particularly suitable for augmenting plasticity. Inspired by residual networks, we derive a condition under which rational units are closed under residual connections and formulate a naturally regularised version. The proposed joint-rational activation allows for desirable degrees of flexibility, yet regularises plasticity to an extent that avoids overfitting by leveraging a mutual set of activation function parameters across layers. We demonstrate that equipping popular algorithms with (joint) rational activations leads to consistent improvements on different games from the Atari Learning Environment benchmark, notably making DQN competitive to DDQN and Rainbow.", "pdf": "https://openreview.net/pdf/0ecc8520a8a62e463d88c1c7abc28fb1fc02ca9a.pdf"} {"title": "Learning No-Regret Sparse Generalized Linear Models with Varying Observation(s)", "url": "https://openreview.net/forum?id=wISvONp3Kq", "detail_url": "https://openreview.net/forum?id=wISvONp3Kq", "authors": "Diyang Li,Charles Ling,zhiqiang xu,Huan Xiong,Bin Gu", "tags": "ICLR 2024,Spotlight", "abstract": "Generalized Linear Models (GLMs) encompass a wide array of regression and classification models, where prediction is a function of a linear combination of the input variables. Often in real-world scenarios, a number of observations would be added into or removed from the existing training dataset, necessitating the development of learning systems that can efficiently train optimal models with varying observations in an online (sequential) manner instead of retraining from scratch. Despite the significance of data-varying scenarios, most existing approaches to sparse GLMs concentrate on offline batch updates, leaving online solutions largely underexplored. In this work, we present the first algorithm without compromising accuracy for GLMs regularized by sparsity-enforcing penalties trained on varying observations. Our methodology is capable of handling the addition and deletion of observations simultaneously, while adaptively updating data-dependent regularization parameters to ensure the best statistical performance. Specifically, we recast sparse GLMs as a bilevel optimization objective upon varying observations and characterize it as an explicit gradient flow in the underlying space for the inner and outer subproblems we are optimizing over, respectively. We further derive a set of rules to ensure a proper transition at regions of non-smoothness, and establish the guarantees of theoretical consistency and finite convergence. Encouraging results are exhibited on real-world benchmarks.", "pdf": "https://openreview.net/pdf/8a85e78845868ba7ca6a019bc980cb4360741379.pdf"} {"title": "Tensor Trust: Interpretable Prompt Injection Attacks from an Online Game", "url": "https://openreview.net/forum?id=fsW7wJGLBd", "detail_url": "https://openreview.net/forum?id=fsW7wJGLBd", "authors": "Sam Toyer,Olivia Watkins,Ethan Adrian Mendes,Justin Svegliato,Luke Bailey,Tiffany Wang,Isaac Ong,Karim Elmaaroufi,Pieter Abbeel,Trevor Darrell,Alan Ritter,Stuart Russell", "tags": "ICLR 2024,Spotlight", "abstract": "While Large Language Models (LLMs) are increasingly being used in real-world applications, they remain vulnerable to *prompt injection attacks*: malicious third party prompts that subvert the intent of the system designer. To help researchers study this problem, we present a dataset of over 563,000 prompt injection attacks and 118,000 prompt-based \"defenses\" against prompt injection, all created by players of an online game called Tensor Trust. To the best of our knowledge, this is the first dataset that includes both human-generated attacks and defenses for instruction-following LLMs. The attacks in our dataset have easily interpretable structure, and shed light on the weaknesses of LLMs. We also use the dataset to create a benchmark for resistance to two types of prompt injection, which we refer to as *prompt extraction* and *prompt hijacking*. Our benchmark results show that many models are vulnerable to the attack strategies in the Tensor Trust dataset. Furthermore, we show that some attack strategies from the dataset generalize to deployed LLM-based applications, even though they have a very different set of constraints to the game. We release data and code at [tensortrust.ai/paper](https://tensortrust.ai/paper)", "pdf": "https://openreview.net/pdf/abc69ff28e9d7de9a838f7c5d380e33bef63bc80.pdf"} {"title": "Leveraging Unpaired Data for Vision-Language Generative Models via Cycle Consistency", "url": "https://openreview.net/forum?id=kNjrhD67LP", "detail_url": "https://openreview.net/forum?id=kNjrhD67LP", "authors": "Tianhong Li,Sangnie Bhardwaj,Yonglong Tian,Han Zhang,Jarred Barber,Dina Katabi,Guillaume Lajoie,Huiwen Chang,Dilip Krishnan", "tags": "ICLR 2024,Spotlight", "abstract": "Current vision-language generative models rely on expansive corpora of $\\textit{paired}$ image-text data to attain optimal performance and generalization capabilities. However, automatically collecting such data (e.g. via large-scale web scraping) leads to low quality and poor image-text correlation, while human annotation is more accurate but requires significant manual effort and expense. We introduce $\\textbf{ITIT}$ ($\\textbf{I}$n$\\textbf{T}$egrating $\\textbf{I}$mage $\\textbf{T}$ext): an innovative training paradigm grounded in the concept of cycle consistency which allows vision-language training on $\\textit{unpaired}$ image and text data. ITIT is comprised of a joint image-text encoder with disjoint image and text decoders that enable bidirectional image-to-text and text-to-image generation in a single framework. During training, ITIT leverages a small set of paired image-text data to ensure its output matches the input reasonably well in both directions. Simultaneously, the model is also trained on much larger datasets containing only images or texts. This is achieved by enforcing cycle consistency between the original unpaired samples and the cycle-generated counterparts. For instance, it generates a caption for a given input image and then uses the caption to create an output image, and enforces similarity between the input and output images. Our experiments show that ITIT with unpaired datasets exhibits similar scaling behavior as using high-quality paired data. We demonstrate image generation and captioning performance on par with state-of-the-art text-to-image and image-to-text models with orders of magnitude fewer (only 3M) paired image-text data. Code will be released at https://github.com/LTH14/itit.", "pdf": "https://openreview.net/pdf/3962f9f3be59d06e85977b838e506fc5374735bf.pdf"} {"title": "Learning the greatest common divisor: explaining transformer predictions", "url": "https://openreview.net/forum?id=cmcD05NPKa", "detail_url": "https://openreview.net/forum?id=cmcD05NPKa", "authors": "Francois Charton", "tags": "ICLR 2024,Spotlight", "abstract": "The predictions of small transformers, trained to calculate the greatest common divisor (GCD) of two positive integers, can be fully characterized by looking at model inputs and outputs.\nAs training proceeds, the model learns a list $\\mathcal D$ of integers, products of divisors of the base used to represent integers and small primes, and predicts the largest element of $\\mathcal D$ that divides both inputs. \nTraining distributions impact performance. Models trained from uniform operands only learn a handful of GCD (up to $38$ GCD $\\leq100$). Log-uniform operands boost performance to $73$ GCD $\\leq 100$, and a log-uniform distribution of outcomes (i.e. GCD) to $91$. However, training from uniform (balanced) GCD breaks explainability.", "pdf": "https://openreview.net/pdf/d05e583ed331a78daf927ca6dbe517768d804505.pdf"} {"title": "Space and time continuous physics simulation from partial observations", "url": "https://openreview.net/forum?id=4yaFQ7181M", "detail_url": "https://openreview.net/forum?id=4yaFQ7181M", "authors": "Steeven JANNY,Madiha Nadri,Julie Digne,Christian Wolf", "tags": "ICLR 2024,Spotlight", "abstract": "Modern techniques for physical simulations rely on numerical schemes and mesh-refinement methods to address trade-offs between precision and complexity, but these handcrafted solutions are tedious and require high computational power. Data-driven methods based on large-scale machine learning promise high adaptivity by integrating long-range dependencies more directly and efficiently. In this work, we focus on computational fluid dynamics and address the shortcomings of a large part of the literature, which are based on fixed support for computations and predictions in the form of regular or irregular grids. We propose a novel setup to perform predictions in a continuous spatial and temporal domain while being trained on sparse observations. We formulate the task as a double observation problem and propose a solution with two interlinked dynamical systems defined on, respectively, the sparse positions and the continuous domain, which allows to forecast and interpolate a solution from the initial condition. Our practical implementation involves recurrent GNNs and a spatio-temporal attention observer capable of interpolating the solution at arbitrary locations. Our model not only generalizes to new initial conditions (as standard auto-regressive models do) but also performs evaluation at arbitrary space and time locations. We evaluate on three standard datasets in fluid dynamics and compare to strong baselines, which are outperformed in classical settings and the extended new task requiring continuous predictions.", "pdf": "https://openreview.net/pdf/1d8bb2f00597f5c08519af2931e6d36c1c80ddcf.pdf"} {"title": "GROOT: Learning to Follow Instructions by Watching Gameplay Videos", "url": "https://openreview.net/forum?id=uleDLeiaT3", "detail_url": "https://openreview.net/forum?id=uleDLeiaT3", "authors": "Shaofei Cai,Bowei Zhang,Zihao Wang,Xiaojian Ma,Anji Liu,Yitao Liang", "tags": "ICLR 2024,Spotlight", "abstract": "We study the problem of building a controller that can follow open-ended instructions in open-world environments. We propose to follow reference videos as instructions, which offer expressive goal specifications while eliminating the need for expensive text-gameplay annotations. A new learning framework is derived to allow learning such instruction-following controllers from gameplay videos while producing a video instruction encoder that induces a structured goal space. We implement our agent GROOT in a simple yet effective encoder-decoder architecture based on causal transformers. We evaluate GROOT against open-world counterparts and human players on a proposed Minecraft SkillForge benchmark. The Elo ratings clearly show that GROOT is closing the human-machine gap as well as exhibiting a 70% winning rate over the best generalist agent baseline. Qualitative analysis of the induced goal space further demonstrates some interesting emergent properties, including the goal composition and complex gameplay behavior synthesis.", "pdf": "https://openreview.net/pdf/4da4eff07073a157b856755f8dfedb2c2e607c6b.pdf"} {"title": "Mask-Based Modeling for Neural Radiance Fields", "url": "https://openreview.net/forum?id=SEiuSzlD1d", "detail_url": "https://openreview.net/forum?id=SEiuSzlD1d", "authors": "Ganlin Yang,Guoqiang Wei,Zhizheng Zhang,Yan Lu,Dong Liu", "tags": "ICLR 2024,Spotlight", "abstract": "Most Neural Radiance Fields (NeRFs) exhibit limited generalization capabilities,which restrict their applicability in representing multiple scenes using a single model. To address this problem, existing generalizable NeRF methods simply condition the model on image features. These methods still struggle to learn precise global representations over diverse scenes since they lack an effective mechanism for interacting among different points and views. In this work, we unveil that 3D implicit representation learning can be significantly improved by mask-based modeling. Specifically, we propose **m**asked **r**ay and **v**iew **m**odeling for generalizable **NeRF** (**MRVM-NeRF**), which is a self-supervised pretraining target to predict complete scene representations from partially masked features along each ray. With this pretraining target, MRVM-NeRF enables better use of correlations across different rays and views as the geometry priors, which thereby strengthens the capability of capturing intricate details within the scenes and boosts the generalization capability across different scenes. Extensive experiments demonstrate the effectiveness of our proposed MRVM-NeRF on both synthetic and real-world datasets, qualitatively and quantitatively. Besides, we also conduct experiments to show the compatibility of our proposed method with various backbones and its superiority under few-shot cases.", "pdf": "https://openreview.net/pdf/60aa67a90ce56bdf66d121eaf8fa35163fc77a76.pdf"} {"title": "Large Language Models Are Not Robust Multiple Choice Selectors", "url": "https://openreview.net/forum?id=shr9PXz7T0", "detail_url": "https://openreview.net/forum?id=shr9PXz7T0", "authors": "Chujie Zheng,Hao Zhou,Fandong Meng,Jie Zhou,Minlie Huang", "tags": "ICLR 2024,Spotlight", "abstract": "Multiple choice questions (MCQs) serve as a common yet important task format in the evaluation of large language models (LLMs). This work shows that modern LLMs are vulnerable to option position changes in MCQs due to their inherent \u201cselection bias\u201d, namely, they prefer to select specific option IDs as answers (like \u201cOption A\u201d). Through extensive empirical analyses with 20 LLMs on three benchmarks, we pinpoint that this behavioral bias primarily stems from LLMs\u2019 token bias, where the model a priori assigns more probabilistic mass to specific option ID tokens (e.g., A/B/C/D) when predicting answers from the option IDs. To mitigate selection bias, we propose a label-free, inference-time debiasing method, called PriDe, which separates the model\u2019s prior bias for option IDs from the overall prediction distribution. PriDe first estimates the prior by permutating option contents on a small number of test samples, and then applies the estimated prior to debias the remaining samples. We demonstrate that it achieves interpretable and transferable debiasing with high computational efficiency. We hope this work can draw broader research attention to the bias and robustness of modern LLMs.", "pdf": "https://openreview.net/pdf/e44715dd5477d90ebbb069b1ec6c7e18756295f5.pdf"} {"title": "Robust Adversarial Reinforcement Learning via Bounded Rationality Curricula", "url": "https://openreview.net/forum?id=pFOoOdaiue", "detail_url": "https://openreview.net/forum?id=pFOoOdaiue", "authors": "Aryaman Reddi,Maximilian T\u00f6lle,Jan Peters,Georgia Chalvatzaki,Carlo D'Eramo", "tags": "ICLR 2024,Spotlight", "abstract": "Robustness against adversarial attacks and distribution shifts is a long-standing goal of Reinforcement Learning (RL). To this end, Robust Adversarial Reinforcement Learning (RARL) trains a protagonist against destabilizing forces exercised by an adversary in a competitive zero-sum Markov game, whose optimal solution, i.e., rational strategy, corresponds to a Nash equilibrium. However, finding Nash equilibria requires facing complex saddle point optimization problems, which can be prohibitive to solve, especially for high-dimensional control. In this paper, we propose a novel approach for adversarial RL based on entropy regularization to ease the complexity of the saddle point optimization problem. We show that the solution of this entropy-regularized problem corresponds to a Quantal Response Equilibrium (QRE), a generalization of Nash equilibria that accounts for bounded rationality, i.e., agents sometimes play random actions instead of optimal ones. Crucially, the connection between the entropy-regularized objective and QRE enables free modulation of the rationality of the agents by simply tuning the temperature coefficient. We leverage this insight to propose our novel algorithm, Quantal Adversarial RL (QARL), which gradually increases the rationality of the adversary in a curriculum fashion until it is fully rational, easing the complexity of the optimization problem while retaining robustness. We provide extensive evidence of QARL outperforming RARL and recent baselines across several MuJoCo locomotion and navigation problems in overall performance and robustness.", "pdf": "https://openreview.net/pdf/8a7e2e74cb77bfa87eaf438766767c4f2fc26a18.pdf"} {"title": "Fine-tuning Multimodal LLMs to Follow Zero-shot Demonstrative Instructions", "url": "https://openreview.net/forum?id=BXY6fe7q31", "detail_url": "https://openreview.net/forum?id=BXY6fe7q31", "authors": "Juncheng Li,Kaihang Pan,Zhiqi Ge,Minghe Gao,Wei Ji,Wenqiao Zhang,Tat-Seng Chua,Siliang Tang,Hanwang Zhang,Yueting Zhuang", "tags": "ICLR 2024,Spotlight", "abstract": "Recent advancements in Multimodal Large Language Models (MLLMs) have been utilizing Visual Prompt Generators (VPGs) to convert visual features into tokens that LLMs can recognize. This is achieved by training the VPGs on millions of image-caption pairs, where the VPG-generated tokens of images are fed into a frozen LLM to generate the corresponding captions. However, this image-captioning based training objective inherently biases the VPG to concentrate solely on the primary visual contents sufficient for caption generation, often neglecting other visual details. This shortcoming results in MLLMs\u2019 underperformance in comprehending demonstrative instructions consisting of multiple, interleaved, and multimodal instructions that demonstrate the required context to complete a task. To address this issue, we introduce a generic and lightweight Visual Prompt Generator Complete module (VPG-C), which can infer and complete the missing details essential for comprehending demonstrative instructions. Further, we propose a synthetic discriminative training strategy to fine-tune VPG-C, eliminating the need for supervised demonstrative instructions. As for evaluation, we build DEMON, a comprehensive benchmark for demonstrative instruction understanding. Synthetically trained with the proposed strategy, VPG-C achieves significantly stronger zero-shot performance across all tasks of DEMON. Further evaluation on the MME and OwlEval benchmarks also demonstrate the superiority of VPG-C. The code and models are available at https://github.com/DCDmllm/Cheetah.", "pdf": "https://openreview.net/pdf/2c2260fc62d6a180e12e943725968e430205fe0a.pdf"} {"title": "CLAP: Collaborative Adaptation for Patchwork Learning", "url": "https://openreview.net/forum?id=8EyRkd3Qj2", "detail_url": "https://openreview.net/forum?id=8EyRkd3Qj2", "authors": "Sen Cui,Abudukelimu Wuerkaixi,Weishen Pan,Jian Liang,Lei Fang,Changshui Zhang,Fei Wang", "tags": "ICLR 2024,Spotlight", "abstract": "In this paper, we investigate a new practical learning scenario, where the data distributed in different sources/clients are typically generated with various modalities. Existing research on learning from multi-source data mostly assume that each client owns the data of all modalities, which may largely limit its practicability. In light of the expensiveness and sparsity of multimodal data, we propose patchwork learning to jointly learn from fragmented multimodal data in distributed clients. Considering the concerns on data privacy, patchwork learning aims to impute incomplete multimodal data for diverse downstream tasks without accessing the raw data directly. Local clients could miss different modality combinations. Due to the statistical heterogeneity induced by non-i.i.d. data, the imputation is more challenging since the learned dependencies fail to adapt to the imputation of other clients. In this paper, we provide a novel imputation framework to tackle modality combination heterogeneity and statistical heterogeneity simultaneously, called ``collaborative adaptation''. In particular, for two observed modality combinations from two clients, we learn the transformations between their maximal intersection and other modalities by proposing a novel ELBO. We improve the worst-performing required transformations through a Pareto min-max optimization framework. In extensive experiments, we demonstrate the superiority of the proposed method compared to existing related methods on benchmark data sets and a real-world clinical data set.", "pdf": "https://openreview.net/pdf/aef23787e155a9467c38a494226d5c7a499c0fec.pdf"} {"title": "Towards Energy Efficient Spiking Neural Networks: An Unstructured Pruning Framework", "url": "https://openreview.net/forum?id=eoSeaK4QJo", "detail_url": "https://openreview.net/forum?id=eoSeaK4QJo", "authors": "Xinyu Shi,Jianhao Ding,Zecheng Hao,Zhaofei Yu", "tags": "ICLR 2024,Spotlight", "abstract": "Spiking Neural Networks (SNNs) have emerged as energy-efficient alternatives to Artificial Neural Networks (ANNs) when deployed on neuromorphic chips. While recent studies have demonstrated the impressive performance of deep SNNs on challenging tasks, their energy efficiency advantage has been diminished. Existing methods targeting energy consumption reduction do not fully exploit sparsity, whereas powerful pruning methods can achieve high sparsity but are not directly targeted at energy efficiency, limiting their effectiveness in energy saving. Furthermore, none of these works fully exploit the sparsity of neurons or the potential for unstructured neuron pruning in SNNs. In this paper, we propose a novel pruning framework that combines unstructured weight pruning with unstructured neuron pruning to maximize the utilization of the sparsity of neuromorphic computing, thereby enhancing energy efficiency. To the best of our knowledge, this is the first application of unstructured neuron pruning to deep SNNs. Experimental results demonstrate that our method achieves impressive energy efficiency gains. The sparse network pruned by our method with only 0.63\\% remaining connections can achieve a remarkable 91 times increase in energy efficiency compared to the original dense network, requiring only 8.5M SOPs for inference, with merely 2.19\\% accuracy loss on the CIFAR-10 dataset. Our work suggests that deep and dense SNNs exhibit high redundancy in energy consumption, highlighting the potential for targeted SNN sparsification to save energy.", "pdf": "https://openreview.net/pdf/0d2fdc7d0fd2120025beecf584d1633bbeebad5e.pdf"} {"title": "Online Stabilization of Spiking Neural Networks", "url": "https://openreview.net/forum?id=CIj1CVbkpr", "detail_url": "https://openreview.net/forum?id=CIj1CVbkpr", "authors": "Yaoyu Zhu,Jianhao Ding,Tiejun Huang,Xiaodong Xie,Zhaofei Yu", "tags": "ICLR 2024,Spotlight", "abstract": "Spiking neural networks (SNNs), attributed to the binary, event-driven nature of spikes, possess heightened biological plausibility and enhanced energy efficiency on neuromorphic hardware compared to analog neural networks (ANNs). Mainstream SNN training schemes apply backpropagation-through-time (BPTT) with surrogate gradients to replace the non-differentiable spike emitting process during backpropagation. While achieving competitive performance, the requirement for storing intermediate information at all time-steps incurs higher memory consumption and fails to fulfill the online property crucial to biological brains. \nOur work focuses on online training techniques, aiming for memory efficiency while preserving biological plausibility. \nThe limitation of not having access to future information in early time steps in online training has constrained previous efforts to incorporate advantageous modules such as batch normalization. \nTo address this problem, we propose Online Spiking Renormalization (OSR) to ensure consistent parameters between testing and training, and Online Threshold Stabilizer (OTS) to stabilize neuron firing rates across time steps. Furthermore, we design a novel online approach to compute the sample mean and variance over time for OSR. Experiments conducted on various datasets demonstrate the proposed method's superior performance among SNN online training algorithms.\nOur code is available at https://github.com/zhuyaoyu/SNN-online-normalization.", "pdf": "https://openreview.net/pdf/661b6685d85268b6527ce98af4c631b5df5b121d.pdf"} {"title": "CrIBo: Self-Supervised Learning via Cross-Image Object-Level Bootstrapping", "url": "https://openreview.net/forum?id=3M0GXoUEzP", "detail_url": "https://openreview.net/forum?id=3M0GXoUEzP", "authors": "Tim Lebailly,Thomas Stegm\u00fcller,Behzad Bozorgtabar,Jean-Philippe Thiran,Tinne Tuytelaars", "tags": "ICLR 2024,Spotlight", "abstract": "Leveraging nearest neighbor retrieval for self-supervised representation learning has proven beneficial with object-centric images. However, this approach faces limitations when applied to scene-centric datasets, where multiple objects within an image are only implicitly captured in the global representation. Such global bootstrapping can lead to undesirable entanglement of object representations. Furthermore, even object-centric datasets stand to benefit from a finer-grained bootstrapping approach. In response to these challenges, we introduce a novel $\\textbf{Cr}$oss-$\\textbf{I}$mage Object-Level $\\textbf{Bo}$otstrapping method tailored to enhance dense visual representation learning. By employing object-level nearest neighbor bootstrapping throughout the training, CrIBo emerges as a notably strong and adequate candidate for in-context learning, leveraging nearest neighbor retrieval at test time. CrIBo shows state-of-the-art performance on the latter task while being highly competitive in more standard downstream segmentation tasks. Our code and pretrained models are publicly available at https://github.com/tileb1/CrIBo.", "pdf": "https://openreview.net/pdf/90f302e1380ba1015e1c1a8508286a9ec74ca3b1.pdf"} {"title": "Real3D-Portrait: One-shot Realistic 3D Talking Portrait Synthesis", "url": "https://openreview.net/forum?id=7ERQPyR2eb", "detail_url": "https://openreview.net/forum?id=7ERQPyR2eb", "authors": "Zhenhui Ye,Tianyun Zhong,Yi Ren,Jiaqi Yang,Weichuang Li,Jiawei Huang,Ziyue Jiang,Jinzheng He,Rongjie Huang,Jinglin Liu,Chen Zhang,Xiang Yin,Zejun MA,Zhou Zhao", "tags": "ICLR 2024,Spotlight", "abstract": "One-shot 3D talking portrait generation aims to reconstruct a 3D avatar from an unseen image, and then animate it with a reference video or audio to generate a talking portrait video. The existing methods fail to simultaneously achieve the goals of accurate 3D avatar reconstruction and stable talking face animation. Besides, while the existing works mainly focus on synthesizing the head part, it is also vital to generate natural torso and background segments to obtain a realistic talking portrait video. To address these limitations, we present Real3D-Potrait, a framework that (1) improves the one-shot 3D reconstruction power with a large image-to-plane model that distills 3D prior knowledge from a 3D face generative model; (2) facilitates accurate motion-conditioned animation with an efficient motion adapter; (3) synthesizes realistic video with natural torso movement and switchable background using a head-torso-background super-resolution model; and (4) supports one-shot audio-driven talking face generation with a generalizable audio-to-motion model. Extensive experiments show that Real3D-Portrait generalizes well to unseen identities and generates more realistic talking portrait videos compared to previous methods. Video samples are available at https://real3dportrait.github.io.", "pdf": "https://openreview.net/pdf/c59a7b868258573cafe7e9a84243e3f096008b51.pdf"} {"title": "TabR: Tabular Deep Learning Meets Nearest Neighbors", "url": "https://openreview.net/forum?id=rhgIgTSSxW", "detail_url": "https://openreview.net/forum?id=rhgIgTSSxW", "authors": "Yury Gorishniy,Ivan Rubachev,Nikolay Kartashev,Daniil Shlenskii,Akim Kotelnikov,Artem Babenko", "tags": "ICLR 2024,Poster", "abstract": "Deep learning (DL) models for tabular data problems (e.g. classification, regression) are currently receiving increasingly more attention from researchers.\nHowever, despite the recent efforts, the non-DL algorithms based on gradient-boosted decision trees (GBDT) remain a strong go-to solution for these problems.\nOne of the research directions aimed at improving the position of tabular DL involves designing so-called retrieval-augmented models.\nFor a target object, such models retrieve other objects (e.g. the nearest neighbors) from the available training data and use their features and labels to make a better prediction.\n\nIn this work, we present TabR -- essentially, a feed-forward network with a custom k-Nearest-Neighbors-like component in the middle.\nOn a set of public benchmarks with datasets up to several million objects, TabR marks a big step forward for tabular DL: it demonstrates the best average performance among tabular DL models, becomes the new state-of-the-art on several datasets, and even outperforms GBDT models on the recently proposed \"GBDT-friendly\" benchmark (see Figure 1).\nAmong the important findings and technical details powering TabR, the main ones lie in the attention-like mechanism that is responsible for retrieving the nearest neighbors and extracting valuable signal from them.\nIn addition to the higher performance, TabR is simple and significantly more efficient compared to prior retrieval-based tabular DL models.", "pdf": "https://openreview.net/pdf/178e173a880d7872c0a79d88e005426c20501329.pdf"} {"title": "Boosting of Thoughts: Trial-and-Error Problem Solving with Large Language Models", "url": "https://openreview.net/forum?id=qBL04XXex6", "detail_url": "https://openreview.net/forum?id=qBL04XXex6", "authors": "Sijia Chen,Baochun Li,Di Niu", "tags": "ICLR 2024,Poster", "abstract": "The reasoning performance of Large Language Models (LLMs) on a wide range of problems critically relies on chain-of-thought prompting, which involves providing a few chain of thought demonstrations as exemplars in prompts. Recent work, e.g., Tree of Thoughts, has pointed out the importance of exploration and self-evaluation in reasoning step selection for complex problem solving. In this paper, we present Boosting of Thoughts (BoT), an automated prompting framework for problem solving with LLMs by iteratively exploring and self-evaluating many trees of thoughts in order to acquire an ensemble of trial-and-error reasoning experiences, which will serve as a new form of prompting to solve the complex problem. Starting from a simple prompt without requiring examples, BoT iteratively explores and evaluates a large collection of reasoning steps, and more importantly, uses error analysis obtained from the LLM on them to explicitly revise prompting, which in turn enhances reasoning step generation, until a final answer is attained. Our experiments with GPT-4 and Llama2 across extensive complex mathematical problems demonstrate that BoT consistently achieves higher or comparable problem-solving rates than other advanced prompting approaches.", "pdf": "https://openreview.net/pdf/a30673a601700226be14c851d887cc7181f4c78f.pdf"} {"title": "Locality Sensitive Sparse Encoding for Learning World Models Online", "url": "https://openreview.net/forum?id=i8PjQT3Uig", "detail_url": "https://openreview.net/forum?id=i8PjQT3Uig", "authors": "Zichen Liu,Chao Du,Wee Sun Lee,Min Lin", "tags": "ICLR 2024,Poster", "abstract": "Acquiring an accurate world model $\\textit{online}$ for model-based reinforcement learning (MBRL) is challenging due to data nonstationarity, which typically causes catastrophic forgetting for neural networks (NNs). From the online learning perspective, a Follow-The-Leader (FTL) world model is desirable, which optimally fits all previous experiences at each round. Unfortunately, NN-based models need re-training on all accumulated data at every interaction step to achieve FTL, which is computationally expensive for lifelong agents. In this paper, we revisit models that can achieve FTL with incremental updates. Specifically, our world model is a linear regression model supported by nonlinear random features. The linear part ensures efficient FTL update while the nonlinear random feature empowers the fitting of complex environments. To best trade off model capacity and computation efficiency, we introduce a locality sensitive sparse encoding, which allows us to conduct efficient sparse updates even with very high dimensional nonlinear features. We validate the representation power of our encoding and verify that it allows efficient online learning under data covariate shift. We also show, in the Dyna MBRL setting, that our world models learned online using a $\\textit{single pass}$ of trajectory data either surpass or match the performance of deep world models trained with replay and other continual learning methods.", "pdf": "https://openreview.net/pdf/0149a804bccf18bec867875f9e9fb664c4288f4b.pdf"} {"title": "Enhancing Neural Subset Selection: Integrating Background Information into Set Representations", "url": "https://openreview.net/forum?id=eepoE7iLpL", "detail_url": "https://openreview.net/forum?id=eepoE7iLpL", "authors": "Binghui Xie,Yatao Bian,Kaiwen Zhou,Yongqiang Chen,Peilin Zhao,Bo Han,Wei Meng,James Cheng", "tags": "ICLR 2024,Poster", "abstract": "Learning neural subset selection tasks, such as compound selection in AI-aided drug discovery, have become increasingly pivotal across diverse applications. The existing methodologies in the field primarily concentrate on constructing models that capture the relationship between utility function values and subsets within their respective supersets. However, these approaches tend to overlook the valuable information contained within the superset when utilizing neural networks to model set functions. In this work, we address this oversight by adopting a probabilistic perspective. Our theoretical findings demonstrate that when the target value is conditioned on both the input set and subset, it is essential to incorporate an invariant sufficient statistic of the superset into the subset of interest for effective learning. This ensures that the output value remains invariant to permutations of the subset and its corresponding superset, enabling identification of the specific superset from which the subset originated. Motivated by these insights, we propose a simple yet effective information aggregation module designed to merge the representations of subsets and supersets from a permutation invariance perspective. Comprehensive empirical evaluations across diverse tasks and datasets validate the enhanced efficacy of our approach over conventional methods, underscoring the practicality and potency of our proposed strategies in real-world contexts.", "pdf": "https://openreview.net/pdf/c23dcae9832d01a0cbe616d94729f6b4a7c8365c.pdf"} {"title": "Bridging Vision and Language Spaces with Assignment Prediction", "url": "https://openreview.net/forum?id=lK2V2E2MNv", "detail_url": "https://openreview.net/forum?id=lK2V2E2MNv", "authors": "Jungin Park,Jiyoung Lee,Kwanghoon Sohn", "tags": "ICLR 2024,Poster", "abstract": "This paper introduces VLAP, a novel approach that bridges pretrained vision models and large language models (LLMs) to make frozen LLMs understand the visual world. VLAP transforms the embedding space of pretrained vision models into the LLMs' word embedding space using a single linear layer for efficient and general-purpose visual and language understanding. Specifically, we harness well-established word embeddings to bridge two modality embedding spaces. The visual and text representations are simultaneously assigned to a set of word embeddings within pretrained LLMs by formulating the assigning procedure as an optimal transport problem. We predict the assignment of one modality from the representation of another modality data, enforcing consistent assignments for paired multimodal data. This allows vision and language representations to contain the same information, grounding the frozen LLMs' word embedding space in visual data. Moreover, a robust semantic taxonomy of LLMs can be preserved with visual data since the LLMs interpret and reason linguistic information from correlations between word embeddings. Experimental results show that VLAP achieves substantial improvements over the previous linear transformation-based approaches across a range of vision-language tasks, including image captioning, visual question answering, and cross-modal retrieval. We also demonstrate the learned visual representations hold a semantic taxonomy of LLMs, making visual semantic arithmetic possible.", "pdf": "https://openreview.net/pdf/f92c528b5ee0113caeb927bb8305b6afaaff0004.pdf"} {"title": "Generative Judge for Evaluating Alignment", "url": "https://openreview.net/forum?id=gtkFw6sZGS", "detail_url": "https://openreview.net/forum?id=gtkFw6sZGS", "authors": "Junlong Li,Shichao Sun,Weizhe Yuan,Run-Ze Fan,hai zhao,Pengfei Liu", "tags": "ICLR 2024,Poster", "abstract": "The rapid development of Large Language Models (LLMs) has substantially expanded the range of tasks they can address. In the field of Natural Language Processing (NLP), researchers have shifted their focus from conventional NLP tasks (e.g., sequence tagging and parsing) towards tasks that revolve around aligning with human needs (e.g., brainstorming and email writing). This shift in task distribution imposes new requirements on evaluating these aligned models regarding *generality* (i.e., assessing performance across diverse scenarios), *flexibility* (i.e., examining under different protocols), and *interpretability* (i.e., scrutinizing models with explanations). In this paper, we propose a generative judge with 13B parameters, **Auto-J**, designed to address these challenges. Our model is trained on user queries and LLM-generated responses under massive real-world scenarios and accommodates diverse evaluation protocols (e.g., pairwise response comparison and single-response evaluation) with well-structured natural language critiques. To demonstrate the efficacy of our approach, we construct a new testbed covering 58 different scenarios. Experimentally, **Auto-J** outperforms a series of strong competitors, including both open-source and closed-source models, by a large margin. We also provide detailed analysis and case studies to further reveal the potential of our method and make a variety of resources public at https://github.com/GAIR-NLP/auto-j.", "pdf": "https://openreview.net/pdf/7fe3087c6257d9121061bc6f3cb0571abebf9277.pdf"} {"title": "Rethinking and Extending the Probabilistic Inference Capacity of GNNs", "url": "https://openreview.net/forum?id=7vVWiCrFnd", "detail_url": "https://openreview.net/forum?id=7vVWiCrFnd", "authors": "Tuo Xu,Lei Zou", "tags": "ICLR 2024,Poster", "abstract": "Designing expressive Graph Neural Networks (GNNs) is an important topic in graph machine learning fields. Despite the existence of numerous approaches proposed to enhance GNNs based on Weisfeiler-Lehman (WL) tests, what GNNs can and cannot learn still lacks a deeper understanding. This paper adopts a fundamentally different approach to examine the expressive power of GNNs from a probabilistic perspective. By establishing connections between GNNs' predictions and the central inference problems of probabilistic graphical models (PGMs), we can analyze previous GNN variants with a novel hierarchical framework and gain new insights into their node-level and link-level behaviors. Additionally, we introduce novel methods that can provably enhance GNNs' ability to capture complex dependencies and make complex predictions. Experiments on both synthetic and real-world datasets demonstrate the effectiveness of our approaches.", "pdf": "https://openreview.net/pdf/20a211772e90fa923b10b62f56acff93f0b25cef.pdf"} {"title": "Learning model uncertainty as variance-minimizing instance weights", "url": "https://openreview.net/forum?id=bDWXhzZT40", "detail_url": "https://openreview.net/forum?id=bDWXhzZT40", "authors": "Nishant Jain,Karthikeyan Shanmugam,Pradeep Shenoy", "tags": "ICLR 2024,Poster", "abstract": "Predictive uncertainty--a model\u2019s self-awareness regarding its accuracy on an input--is key for both building robust models via training interventions and for test-time applications such as selective classification. We propose a novel instance-conditional reweighting approach that captures predictive uncertainty using an auxiliary network, and unifies these train- and test-time applications. The auxiliary network is trained using a meta-objective in a bilevel optimization framework. A key contribution of our proposal is the meta-objective of minimizing dropout variance, an approximation of Bayesian predictive uncertainty, We show in controlled experiments that we effectively capture diverse specific notions of uncertainty through this meta-objective, while previous approaches only capture certain aspects. These results translate to significant gains in real-world settings\u2013selective classification, label noise, domain adaptation, calibration\u2013and across datasets\u2013Imagenet, Cifar100, diabetic retinopathy, Camelyon, WILDs, Imagenet-C,-A,-R, Clothing-1.6M, etc. For Diabetic Retinopathy, we see upto 3.4\\%/3.3\\% accuracy & AUC gains over SOTA in selective classification. We also improve upon large-scale pretrained models such as PLEX.", "pdf": "https://openreview.net/pdf/ebdd64cc233d279eff550647eb57b1baf57fd9eb.pdf"} {"title": "Forward Learning with Top-Down Feedback: Empirical and Analytical Characterization", "url": "https://openreview.net/forum?id=My7lkRNnL9", "detail_url": "https://openreview.net/forum?id=My7lkRNnL9", "authors": "Ravi Francesco Srinivasan,Francesca Mignacco,Martino Sorbaro,Maria Refinetti,Avi Cooper,Gabriel Kreiman,Giorgia Dellaferrera", "tags": "ICLR 2024,Poster", "abstract": "\"Forward-only\" algorithms, which train neural networks while avoiding a backward pass, have recently gained attention as a way of solving the biologically unrealistic aspects of backpropagation. Here, we first address compelling challenges related to the \"forward-only\" rules, which include reducing the performance gap with backpropagation and providing an analytical understanding of their dynamics. To this end, we show that the forward-only algorithm with top-down feedback is well-approximated by an \"adaptive-feedback-alignment\" algorithm, and we analytically track its performance during learning in a prototype high-dimensional setting. Then, we compare different versions of forward-only algorithms, focusing on the Forward-Forward and PEPITA frameworks, and we show that they share the same learning principles. Overall, our work unveils the connections between three key neuro-inspired learning rules, providing a link between \"forward-only\" algorithms, i.e., Forward-Forward and PEPITA, and an approximation of backpropagation, i.e., Feedback Alignment.", "pdf": "https://openreview.net/pdf/d2749ae05700db2eaae75a159874c2e229ae8222.pdf"} {"title": "Towards Cross Domain Generalization of Hamiltonian Representation via Meta Learning", "url": "https://openreview.net/forum?id=AZGIwqCyYY", "detail_url": "https://openreview.net/forum?id=AZGIwqCyYY", "authors": "Yeongwoo Song,Hawoong Jeong", "tags": "ICLR 2024,Poster", "abstract": "Recent advances in deep learning for physics have focused on discovering shared representations of target systems by incorporating physics priors or inductive biases into neural networks. While effective, these methods are limited to the system domain, where the type of system remains consistent and thus cannot ensure the adaptation to new, or unseen physical systems governed by different laws. For instance, a neural network trained on a mass-spring system cannot guarantee accurate predictions for the behavior of a two-body system or any other system with different physical laws.\nIn this work, we take a significant leap forward by targeting cross domain generalization within the field of Hamiltonian dynamics. \nWe model our system with a graph neural network (GNN) and employ a meta learning algorithm to enable the model to gain experience over a distribution of systems and make it adapt to new physics. Our approach aims to learn a unified Hamiltonian representation that is generalizable across multiple system domains, thereby overcoming the limitations of system-specific models. \nWe demonstrate that the meta-trained model captures the generalized Hamiltonian representation that is consistent across different physical domains.\nOverall, through the use of meta learning, we offer a framework that achieves cross domain generalization, providing a step towards a unified model for understanding a wide array of dynamical systems via deep learning.", "pdf": "https://openreview.net/pdf/b18ec7715410bb144074f86aecfa3acfa91b52d9.pdf"} {"title": "What Makes Good Data for Alignment? A Comprehensive Study of Automatic Data Selection in Instruction Tuning", "url": "https://openreview.net/forum?id=BTKAeLqLMw", "detail_url": "https://openreview.net/forum?id=BTKAeLqLMw", "authors": "Wei Liu,Weihao Zeng,Keqing He,Yong Jiang,Junxian He", "tags": "ICLR 2024,Poster", "abstract": "Instruction tuning is a standard technique employed to align large language models to end tasks and user preferences after the initial pretraining phase. Recent research indicates the critical role of data engineering in instruction tuning -- when appropriately selected, only limited data is necessary to achieve superior performance. However, we still lack a principled understanding of what makes good instruction tuning data for alignment, and how we should select data automatically and effectively. In this work, we delve deeply into automatic data selection strategies for alignment. We start with controlled studies to measure data across three dimensions: complexity, quality, and diversity, along which we examine existing methods and introduce novel techniques for enhanced data measurement. Subsequently, we propose a simple strategy to select data samples based on the measurement. We present Deita (short for Data-Efficient Instruction Tuning for Alignment), a series of models fine-tuned from LLaMA models using data samples automatically selected with our proposed approach. When assessed through both automatic metrics and human evaluation, Deita performs better or on par with the state-of-the-art open-source alignment models such as Vicuna and WizardLM with only 6K training data samples -- 10x less than the data used in the baselines. We anticipate this work to provide clear guidelines and tools on automatic data selection, aiding researchers and practitioners in achieving data-efficient alignment.", "pdf": "https://openreview.net/pdf/a617374cea846fefdd6511494d4f9722b000a237.pdf"} {"title": "Causality-Inspired Spatial-Temporal Explanations for Dynamic Graph Neural Networks", "url": "https://openreview.net/forum?id=AJBkfwXh3u", "detail_url": "https://openreview.net/forum?id=AJBkfwXh3u", "authors": "Kesen Zhao,Liang Zhang", "tags": "ICLR 2024,Poster", "abstract": "Dynamic Graph Neural Networks (DyGNNs) have gained significant popularity in the research of dynamic graphs, but are limited by the low transparency, such that human-understandable insights can hardly be drawn from their predictions. Although a number of existing research have been devoted to investigating the interpretability of graph neural networks (GNNs), achieving the interpretability of DyGNNs is pivotally challenging due to the complex spatial-temporal correlations in dynamic graphs. To this end, we propose an innovative causality-inspired generative model based on structural causal model (SCM), which explores the underlying philosophies of DyGNN predictions by identifying the trivial, static, and dynamic causal relationships. To reach this goal, two critical tasks need to be accomplished including (1) disentangling the complex causal relationships, and (2) fitting the spatial-temporal explanations of DyGNNs in the SCM architecture. To tackle these challenges, the proposed method incorporates a contrastive learning module to disentangle trivial and causal relationships, and a dynamic correlating module to disentangle dynamic and static causal relationships, respectively. A dynamic VGAE-based framework is further developed, which generates causal-and-dynamic masks for spatial interpretability, and recognizes dynamic relationships along the time horizon through causal invention for temporal interpretability. Comprehensive experiments have been conducted on both synthetic and real-world datasets, where our approach yields substantial improvements, thereby demonstrating significant superiority.", "pdf": "https://openreview.net/pdf/da061144568aad672fd37e90edc5b87a39b91ba9.pdf"} {"title": "Dissecting learning and forgetting in language model finetuning", "url": "https://openreview.net/forum?id=tmsqb6WpLz", "detail_url": "https://openreview.net/forum?id=tmsqb6WpLz", "authors": "Xiao Zhang,Ji Wu", "tags": "ICLR 2024,Poster", "abstract": "Finetuning language models on domain-specific corpus is a common approach to enhance their domain knowledge and capability. While improving performance on domain tasks, it often brings a side-effect of forgetting of the model's general abilities. In this study, we analyze the effects of finetuning on language models by dissecting its impacts on the modeling of topic, style, and factual knowledge in text. Our method uses instruction-following LLMs such as ChatGPT to auto-generate controlled-variable text examples which we use to probe the model. Our findings reveal that finetuning results in significant shifts in the language model's topic and style priors, while actual knowledge learning only contributes to a small fraction of the total probability change. Analysis shows that the adaptation of topic and style priors behave akin to learning simple features: they are learned rapidly and require little model capacity. They are also learned independently and primarily at the beginning of a text sequence. In contrast, factual knowledge is learned stably but slowly and requires significant model capacity to learn. The research offers insights and understanding into the finer dynamics of learning and forgetting in language models, and can potentially inform future research on improving domain adaptation and addressing the challenges of forgetting in continual learning of language models.", "pdf": "https://openreview.net/pdf/11ec6e4f64662107e73511797d5b3135c9385ef5.pdf"} {"title": "Test-time Adaptation against Multi-modal Reliability Bias", "url": "https://openreview.net/forum?id=TPZRq4FALB", "detail_url": "https://openreview.net/forum?id=TPZRq4FALB", "authors": "Mouxing Yang,Yunfan Li,Changqing Zhang,Peng Hu,Xi Peng", "tags": "ICLR 2024,Poster", "abstract": "Test-time adaptation (TTA) has emerged as a new paradigm for reconciling distribution shifts across domains without accessing source data. However, existing TTA methods mainly concentrate on uni-modal tasks, overlooking the complexity of multi-modal scenarios.\nIn this paper, we delve into the multi-modal test-time adaptation and reveal a new challenge named reliability bias. Different from the definition of traditional distribution shifts, reliability bias refers to the information discrepancies across different modalities derived from intra-modal distribution shifts. To solve the challenge, we propose a novel method, dubbed REliable fusion and robust ADaptation (READ). On the one hand, unlike the existing TTA paradigm that mainly repurposes the normalization layers, READ employs a new paradigm that modulates the attention between modalities in a self-adaptive way, supporting reliable fusion against reliability bias. On the other hand, READ adopts a novel objective function for robust multi-modal adaptation, where the contributions of confident predictions could be amplified and the negative impacts of noisy predictions could be mitigated. Moreover, we introduce two new benchmarks to facilitate comprehensive evaluations of multi-modal TTA under reliability bias. Extensive experiments on the benchmarks verify the effectiveness of our method against multi-modal reliability bias. The code and benchmarks are available at https://github.com/XLearning-SCU/2024-ICLR-READ.", "pdf": "https://openreview.net/pdf/b4995bcc3ee4d649705147591c205b904f4a9d13.pdf"} {"title": "Mirage: Model-agnostic Graph Distillation for Graph Classification", "url": "https://openreview.net/forum?id=78iGZdqxYY", "detail_url": "https://openreview.net/forum?id=78iGZdqxYY", "authors": "Mridul Gupta,Sahil Manchanda,HARIPRASAD KODAMANA,Sayan Ranu", "tags": "ICLR 2024,Poster", "abstract": "GNNs, like other deep learning models, are data and computation hungry. There is a pressing need to scale training of GNNs on large datasets to enable their usage on low-resource environments. Graph distillation is an effort in that direction with the aim to construct a smaller synthetic training set from the original training data without significantly compromising model performance. While initial efforts are promising, this work is motivated by two key observations: (1) Existing graph distillation algorithms themselves rely on training with the full dataset, which undermines the very premise of graph distillation. (2) The distillation process is specific to the target GNN architecture and hyper-parameters and thus not robust to changes in the modeling pipeline. We circumvent these limitations by designing a distillation algorithm called MIRAGE for graph classification. MIRAGE is built on the insight that a message-passing GNN decomposes the input graph into a multiset of computation trees. Furthermore, the frequency distribution of computation trees is often skewed in nature, enabling us to condense this data into a concise distilled summary. By compressing the computation data itself, as opposed to emulating gradient flows on the original training set\u2014a prevalent approach to date\u2014MIRAGE transforms into an unsupervised and architecture-agnostic distillation algorithm. Extensive benchmarking on real-world datasets underscores MIRAGE\u2019s superiority, showcasing enhanced generalization accuracy, data compression, and distillation efficiency when compared to state-of-the-art baselines.", "pdf": "https://openreview.net/pdf/6734ad0e165c7017f3c6c0f400183a1bae684654.pdf"} {"title": "On the Learnability of Watermarks for Language Models", "url": "https://openreview.net/forum?id=9k0krNzvlV", "detail_url": "https://openreview.net/forum?id=9k0krNzvlV", "authors": "Chenchen Gu,Xiang Lisa Li,Percy Liang,Tatsunori Hashimoto", "tags": "ICLR 2024,Poster", "abstract": "Watermarking of language model outputs enables statistical detection of model-generated text, which can mitigate harms and misuses of language models. Existing watermarking strategies operate by altering the decoder of an existing language model. In this paper, we ask whether language models can directly learn to generate watermarked text, which would have significant implications for the real-world deployment of watermarks. First, learned watermarks could be used to build open models that naturally generate watermarked text, enabling watermarking for open models, where users can control the decoding procedure. Second, if watermarking is used to determine the provenance of generated text, an adversary can hurt the reputation of a victim model by spoofing its watermark and generating damaging watermarked text. To investigate the learnability of watermarks, we propose watermark distillation, which trains a student model to behave like a teacher model that uses decoding-based watermarking. We test our approach on three decoding-based watermarking strategies and various hyperparameter settings, finding that models can learn to generate watermarked text with high detectability. We also find limitations to learnability, including the loss of watermarking capabilities under fine-tuning on normal text and high sample complexity when learning low-distortion watermarks.", "pdf": "https://openreview.net/pdf/218fcaae4d72e8c622881df1415344813b0aa3f0.pdf"} {"title": "Bellman Optimal Stepsize Straightening of Flow-Matching Models", "url": "https://openreview.net/forum?id=Iyve2ycvGZ", "detail_url": "https://openreview.net/forum?id=Iyve2ycvGZ", "authors": "Bao Nguyen,Binh Nguyen,Viet Anh Nguyen", "tags": "ICLR 2024,Poster", "abstract": "Flow matching is a powerful framework for generating high-quality samples in various applications, especially image synthesis. However, the intensive computational demands of these models, especially during the finetuning process and sampling processes, pose significant challenges for low-resource scenarios. This paper introduces Bellman Optimal Stepsize Straightening (BOSS) technique for distilling flow-matching generative models: it aims specifically for a few-step efficient image sampling while adhering to a computational budget constraint. First, this technique involves a dynamic programming algorithm that optimizes the stepsizes of the pretrained network. Then, it refines the velocity network to match the optimal step sizes, aiming to straighten the generation paths. Extensive experimental evaluations across image generation tasks demonstrate the efficacy of BOSS in terms of both resource utilization and image quality. Our results reveal that BOSS achieves substantial gains in efficiency while maintaining competitive sample quality, effectively bridging the gap between low-resource constraints and the demanding requirements of flow-matching generative models. Our paper also fortifies the responsible development of artificial intelligence, offering a more sustainable generative model that reduces computational costs and environmental footprints. Our code can be found at https://github.com/nguyenngocbaocmt02/BOSS.", "pdf": "https://openreview.net/pdf/49149ffea0255c53a80d017478e79f91b0a6c402.pdf"} {"title": "Enhancing Tail Performance in Extreme Classifiers by Label Variance Reduction", "url": "https://openreview.net/forum?id=6ARlSgun7J", "detail_url": "https://openreview.net/forum?id=6ARlSgun7J", "authors": "Anirudh Buvanesh,Rahul Chand,Jatin Prakash,Bhawna Paliwal,Mudit Dhawan,Neelabh Madan,Deepesh Hada,Vidit Jain,SONU MEHTA,Yashoteja Prabhu,Manish Gupta,Ramachandran Ramjee,Manik Varma", "tags": "ICLR 2024,Poster", "abstract": "Extreme Classification (XC) architectures, which utilize a massive One-vs-All (OvA) classifier layer at the output, have demonstrated remarkable performance on problems with large label sets. Nonetheless, these architectures falter on tail labels with few representative samples. This phenomenon has been attributed to factors such as classifier over-fitting and missing label bias, and solutions involving regularization and loss re-calibration have been developed. This paper explores the impact of label variance - a previously unexamined factor - on the tail performance in extreme classifiers. It also develops a method to systematically reduce label variance in XC by transferring the knowledge from a specialized tail-robust teacher model to the OvA classifiers. For this purpose, it proposes a principled knowledge distillation framework, LEVER, which enhances the tail performance in extreme classifiers with formal guarantees on generalization. Comprehensive experiments are conducted on a diverse set of XC datasets, demonstrating that LEVER can enhance tail performance by around 5\\% and 6\\% points in PSP and coverage metrics, respectively, when integrated with leading extreme classifiers. Moreover, it establishes a new state-of-the-art when added to the top-performing Renee classifier. Extensive ablations and analyses substantiate the efficacy of our design choices. Another significant contribution is the release of two new XC datasets that are different from and more challenging than the available benchmark datasets, thereby encouraging more rigorous algorithmic evaluation in the future. Code for LEVER is available at: aka.ms/lever.", "pdf": "https://openreview.net/pdf/a20d3764489727d5956f74f1b28ba95a1d848575.pdf"} {"title": "Is This the Subspace You Are Looking for? An Interpretability Illusion for Subspace Activation Patching", "url": "https://openreview.net/forum?id=Ebt7JgMHv1", "detail_url": "https://openreview.net/forum?id=Ebt7JgMHv1", "authors": "Aleksandar Makelov,Georg Lange,Atticus Geiger,Neel Nanda", "tags": "ICLR 2024,Poster", "abstract": "Mechanistic interpretability aims to attribute high-level model behaviors to specific, interpretable learned features. It is hypothesized that these features manifest as directions or low-dimensional subspaces within activation space. Accordingly, recent studies have explored the identification and manipulation of such subspaces to reverse-engineer computations, employing methods such as activation patching. In this work, we demonstrate that na\u00efve approaches to subspace interventions can give rise to interpretability illusions.\n\nSpecifically, even if patching along a subspace has the intended end-to-end causal effect on model behavior, this effect may be achieved by activating \\emph{a dormant parallel pathway} using a component that is \\textit{causally disconnected} from the model output.\nWe demonstrate this in a mathematical example, realize the example empirically in two different settings (the Indirect Object Identification (IOI) task and factual recall), and argue that activating dormant pathways ought to be prevalent in practice.\nIn the context of factual recall, we further show that the illusion is related to rank-1 fact editing, providing a mechanistic explanation for previous work observing an inconsistency between fact editing performance and fact localisation.\n\nHowever, this does not imply that activation patching of subspaces is intrinsically unfit for interpretability.\nTo contextualize our findings, we also show what a success case looks like in a task (IOI) where prior manual circuit analysis allows an understanding of the location of the ground truth feature. We explore the additional evidence needed to argue that a patched subspace is faithful.", "pdf": "https://openreview.net/pdf/41767a08ea94dbfa362b4807a25ed21035e4dc0e.pdf"} {"title": "Demonstration-Regularized RL", "url": "https://openreview.net/forum?id=lF2aip4Scn", "detail_url": "https://openreview.net/forum?id=lF2aip4Scn", "authors": "Daniil Tiapkin,Denis Belomestny,Daniele Calandriello,Eric Moulines,Alexey Naumov,Pierre Perrault,Michal Valko,Pierre Menard", "tags": "ICLR 2024,Poster", "abstract": "Incorporating expert demonstrations has empirically helped to improve the sample efficiency of reinforcement learning (RL). This paper quantifies theoretically to what extent this extra information reduces RL's sample complexity. In particular, we study the demonstration-regularized reinforcement learning framework that leverages the expert demonstrations by $\\mathrm{KL}$-regularization for a policy learned by behavior cloning. Our findings reveal that using $N^{\\mathrm{E}}$ expert demonstrations enables the identification of an optimal policy at a sample complexity of order $\\widetilde{\\mathcal{O}}(\\mathrm{Poly}(S,A,H)/(\\varepsilon^2 N^{\\mathrm{E}}))$ in finite and $\\widetilde{\\mathcal{O}}(\\mathrm{Poly}(d,H)/(\\varepsilon^2 N^{\\mathrm{E}}))$ in linear Markov decision processes, where $\\varepsilon$is the target precision, $H$ the horizon, $A$ the number of action, $S$ the number of states in the finite case and $d$ the dimension of the feature space in the linear case. As a by-product, we provide tight convergence guarantees for the behavior cloning procedure under general assumptions on the policy classes. Additionally, we establish that demonstration-regularized methods are provably efficient for reinforcement learning from human feedback (RLHF). In this respect, we provide theoretical evidence showing the benefits of KL-regularization for RLHF in tabular and linear MDPs. \nInterestingly, we avoid pessimism injection by employing computationally feasible regularization to handle reward estimation uncertainty, thus setting our approach apart from the prior works.", "pdf": "https://openreview.net/pdf/7f93971543d7b1a26e607380315994e7f7ad3051.pdf"} {"title": "Multilingual Jailbreak Challenges in Large Language Models", "url": "https://openreview.net/forum?id=vESNKdEMGp", "detail_url": "https://openreview.net/forum?id=vESNKdEMGp", "authors": "Yue Deng,Wenxuan Zhang,Sinno Jialin Pan,Lidong Bing", "tags": "ICLR 2024,Poster", "abstract": "While large language models (LLMs) exhibit remarkable capabilities across a wide range of tasks, they pose potential safety concerns, such as the ``jailbreak'' problem, wherein malicious instructions can manipulate LLMs to exhibit undesirable behavior. Although several preventive measures have been developed to mitigate the potential risks associated with LLMs, they have primarily focused on English. In this study, we reveal the presence of multilingual jailbreak challenges within LLMs and consider two potential risky scenarios: unintentional and intentional. The unintentional scenario involves users querying LLMs using non-English prompts and inadvertently bypassing the safety mechanisms, while the intentional scenario concerns malicious users combining malicious instructions with multilingual prompts to deliberately attack LLMs. The experimental results reveal that in the unintentional scenario, the rate of unsafe content increases as the availability of languages decreases. Specifically, low-resource languages exhibit about three times the likelihood of encountering harmful content compared to high-resource languages, with both ChatGPT and GPT-4. In the intentional scenario, multilingual prompts can exacerbate the negative impact of malicious instructions, with astonishingly high rates of unsafe output: 80.92\\% for ChatGPT and 40.71\\% for GPT-4. To handle such a challenge in the multilingual context, we propose a novel \\textsc{Self-Defense} framework that automatically generates multilingual training data for safety fine-tuning. Experimental results show that ChatGPT fine-tuned with such data can achieve a substantial reduction in unsafe content generation. Data is available at \\url{https://github.com/DAMO-NLP-SG/multilingual-safety-for-LLMs}.", "pdf": "https://openreview.net/pdf/5755fd1159bae5a7cf71a490847938b7f047a0b3.pdf"} {"title": "$t^3$-Variational Autoencoder: Learning Heavy-tailed Data with Student's t and Power Divergence", "url": "https://openreview.net/forum?id=RzNlECeoOB", "detail_url": "https://openreview.net/forum?id=RzNlECeoOB", "authors": "Juno Kim,Jaehyuk Kwon,Mincheol Cho,Hyunjong Lee,Joong-Ho Won", "tags": "ICLR 2024,Poster", "abstract": "The variational autoencoder (VAE) typically employs a standard normal prior as a regularizer for the probabilistic latent encoder. However, the Gaussian tail often decays too quickly to effectively accommodate the encoded points, failing to preserve crucial structures hidden in the data. In this paper, we explore the use of heavy-tailed models to combat over-regularization. Drawing upon insights from information geometry, we propose $t^3$VAE, a modified VAE framework that incorporates Student's t-distributions for the prior, encoder, and decoder. This results in a joint model distribution of a power form which we argue can better fit real-world datasets. We derive a new objective by reformulating the evidence lower bound as joint optimization of KL divergence between two statistical manifolds and replacing with $\\gamma$-power divergence, a natural alternative for power families. $t^3$VAE demonstrates superior generation of low-density regions when trained on heavy-tailed synthetic data. Furthermore, we show that $t^3$VAE significantly outperforms other models on CelebA and imbalanced CIFAR-100 datasets.", "pdf": "https://openreview.net/pdf/456160753ce7ca91fd5d14c5ba07b1183322d395.pdf"} {"title": "Rethinking the Power of Graph Canonization in Graph Representation Learning with Stability", "url": "https://openreview.net/forum?id=nTwb2vBLOV", "detail_url": "https://openreview.net/forum?id=nTwb2vBLOV", "authors": "Zehao Dong,Muhan Zhang,Philip Payne,Michael A Province,Carlos Cruchaga,Tianyu Zhao,Fuhai Li,Yixin Chen", "tags": "ICLR 2024,Poster", "abstract": "The expressivity of Graph Neural Networks (GNNs) has been studied broadly in recent years to reveal the design principles for more powerful GNNs. Graph canonization is known as a typical approach to distinguish non-isomorphic graphs, yet rarely adopted when developing expressive GNNs. This paper proposes to maximize the expressivity of GNNs by graph canonization, then the power of such GNNs is studies from the perspective of model stability. A stable GNN will map similar graphs to close graph representations in the vectorial space, and the stability of GNNs is critical to generalize their performance to unseen graphs. We theoretically reveal the trade-off of expressivity and stability in graph-canonization-enhanced GNNs. Then we introduce a notion of universal graph canonization as the general solution to address the trade-off and characterize a widely applicable sufficient condition to solve the universal graph canonization. A comprehensive set of experiments demonstrates the effectiveness of the proposed method. In many popular graph benchmark datasets, graph canonization successfully enhances GNNs and provides highly competitive performance, indicating the capability and great potential of proposed method in general graph representation learning. In graph datasets where the sufficient condition holds, GNNs enhanced by universal graph canonization consistently outperform GNN baselines and successfully improve the SOTA performance up to $31$%, providing the optimal solution to numerous challenging real-world graph analytical tasks like gene network representation learning in bioinformatics.", "pdf": "https://openreview.net/pdf/cbad74b981d4d1bb7c06f5c746926121d2dc638f.pdf"} {"title": "Gradual Optimization Learning for Conformational Energy Minimization", "url": "https://openreview.net/forum?id=FMMF1a9ifL", "detail_url": "https://openreview.net/forum?id=FMMF1a9ifL", "authors": "Artem Tsypin,Leonid Anatolievich Ugadiarov,Kuzma Khrabrov,Alexander Telepov,Egor Rumiantsev,Alexey Skrynnik,Aleksandr Panov,Dmitry P. Vetrov,Elena Tutubalina,Artur Kadurin", "tags": "ICLR 2024,Poster", "abstract": "Molecular conformation optimization is crucial to computer-aided drug discovery and materials design.\nTraditional energy minimization techniques rely on iterative optimization methods that use molecular forces calculated by a physical simulator (oracle) as anti-gradients.\nHowever, this is a computationally expensive approach that requires many interactions with a physical simulator.\nOne way to accelerate this procedure is to replace the physical simulator with a neural network.\nDespite recent progress in neural networks for molecular conformation energy prediction, such models are prone to errors due to distribution shift, leading to inaccurate energy minimization.\nWe find that the quality of energy minimization with neural networks can be improved by providing optimization trajectories as additional training data.\nStill, obtaining complete optimization trajectories demands a lot of additional computations.\nTo reduce the required additional data, we present the Gradual Optimization Learning Framework (GOLF) for energy minimization with neural networks.\nThe framework consists of an efficient data-collecting scheme and an external optimizer.\nThe external optimizer utilizes gradients from the energy prediction model to generate optimization trajectories, and the data-collecting scheme selects additional training data to be processed by the physical simulator. \nOur results demonstrate that the neural network trained with GOLF performs \\textit{on par} with the oracle on a benchmark of diverse drug-like molecules using significantly less additional data.", "pdf": "https://openreview.net/pdf/37b8c4267928b3f02b1f1f1c9df21956eb2c653c.pdf"} {"title": "AnomalyCLIP: Object-agnostic Prompt Learning for Zero-shot Anomaly Detection", "url": "https://openreview.net/forum?id=buC4E91xZE", "detail_url": "https://openreview.net/forum?id=buC4E91xZE", "authors": "Qihang Zhou,Guansong Pang,Yu Tian,Shibo He,Jiming Chen", "tags": "ICLR 2024,Poster", "abstract": "Zero-shot anomaly detection (ZSAD) requires detection models trained using auxiliary\ndata to detect anomalies without any training sample in a target dataset. It\nis a crucial task when training data is not accessible due to various concerns, e.g.,\ndata privacy, yet it is challenging since the models need to generalize to anomalies\nacross different domains where the appearance of foreground objects, abnormal\nregions, and background features, such as defects/tumors on different products/\norgans, can vary significantly. Recently large pre-trained vision-language\nmodels (VLMs), such as CLIP, have demonstrated strong zero-shot recognition\nability in various vision tasks, including anomaly detection. However, their ZSAD\nperformance is weak since the VLMs focus more on modeling the class semantics\nof the foreground objects rather than the abnormality/normality in the images. In\nthis paper we introduce a novel approach, namely AnomalyCLIP, to adapt CLIP\nfor accurate ZSAD across different domains. The key insight of AnomalyCLIP\nis to learn object-agnostic text prompts that capture generic normality and abnormality\nin an image regardless of its foreground objects. This allows our model to\nfocus on the abnormal image regions rather than the object semantics, enabling\ngeneralized normality and abnormality recognition on diverse types of objects.\nLarge-scale experiments on 17 real-world anomaly detection datasets show that\nAnomalyCLIP achieves superior zero-shot performance of detecting and segmenting\nanomalies in datasets of highly diverse class semantics from various defect\ninspection and medical imaging domains. Code will be made available at https://github.com/zqhang/AnomalyCLIP.", "pdf": "https://openreview.net/pdf/c8c456c139a7b7f9cbf659ae062d7e3f9cba1aff.pdf"} {"title": "Towards Faithful Explanations: Boosting Rationalization with Shortcuts Discovery", "url": "https://openreview.net/forum?id=uGtfk2OphU", "detail_url": "https://openreview.net/forum?id=uGtfk2OphU", "authors": "Linan Yue,Qi Liu,Yichao Du,Li Wang,Weibo Gao,Yanqing An", "tags": "ICLR 2024,Poster", "abstract": "The remarkable success in neural networks provokes the selective rationalization. It explains the prediction results by identifying a small subset of the inputs sufficient to support them. Since existing methods still suffer from adopting the shortcuts in data to compose rationales and limited large-scale annotated rationales by human, in this paper, we propose a Shortcuts-fused Selective Rationalization (SSR) method, which boosts the rationalization by discovering and exploiting potential shortcuts. Specifically, SSR first designs a shortcuts discovery approach to detect several potential shortcuts. Then, by introducing the identified shortcuts, we propose two strategies to mitigate the problem of utilizing shortcuts to compose rationales. Finally, we develop two data augmentations methods to close the gap in the number of annotated rationales. Extensive experimental results on real-world datasets clearly validate the effectiveness of our proposed method.", "pdf": "https://openreview.net/pdf/c8c102f26cb7fc4cddf6ff889fe2a63454b563c9.pdf"} {"title": "CORN: Contact-based Object Representation for Nonprehensile Manipulation of General Unseen Objects", "url": "https://openreview.net/forum?id=KTtEICH4TO", "detail_url": "https://openreview.net/forum?id=KTtEICH4TO", "authors": "Yoonyoung Cho,Junhyek Han,Yoontae Cho,Beomjoon Kim", "tags": "ICLR 2024,Poster", "abstract": "Nonprehensile manipulation is essential for manipulating objects that are too thin, large, or otherwise ungraspable in the wild. To sidestep the difficulty of contact modeling in conventional modeling-based approaches, reinforcement learning (RL) has recently emerged as a promising alternative. However, previous RL approaches either lack the ability to generalize over diverse object shapes, or use simple action primitives that limit the diversity of robot motions. Furthermore, using RL over diverse object geometry is challenging due to the high cost of training a policy that takes in high-dimensional sensory inputs. We propose a novel contact-based object representation and pretraining pipeline to tackle this. To enable massively parallel training, we leverage a lightweight patch-based transformer architecture for our encoder that processes point clouds, thus scaling our training across thousands of environments. Compared to learning from scratch, or other shape representation baselines, our representation facilitates both time- and data-efficient learning. We validate the efficacy of our overall system by zero-shot transferring the trained policy to novel real-world objects. We highly recommend the video attached in the supplementary material. Code and videos are available at \\url{https://sites.google.com/view/contact-non-prehensile}.", "pdf": "https://openreview.net/pdf/be6d29e6e7d18c8ea0250289f353011374d395b1.pdf"} {"title": "An Intuitive Multi-Frequency Feature Representation for SO(3)-Equivariant Networks", "url": "https://openreview.net/forum?id=5JWAOLBxwp", "detail_url": "https://openreview.net/forum?id=5JWAOLBxwp", "authors": "Dongwon Son,Jaehyung Kim,Sanghyeon Son,Beomjoon Kim", "tags": "ICLR 2024,Poster", "abstract": "The usage of 3D vision algorithms, such as shape reconstruction, remains limited because they require inputs to be at a fixed canonical rotation. Recently, a simple equivariant network, Vector Neuron (VN) has been proposed that can be easily used with the state-of-the-art 3D neural network (NN) architectures. However, its performance is limited because it is designed to use only three-dimensional features, which is insufficient to capture the details present in 3D data. In this paper, we introduce an equivariant feature representation for mapping a 3D point to a high-dimensional feature space. Our feature can discern multiple frequencies present in 3D data, which, as shown by Tancik et al. (2020), is the key to designing an expressive feature for 3D vision tasks. Our representation can be used as an input to VNs, and the results demonstrate that with our feature representation, VN captures more details, overcoming the limitation raised in its original paper.", "pdf": "https://openreview.net/pdf/ae84b2b3de8b202f110cedf0a834de33c23e665d.pdf"} {"title": "KoLA: Carefully Benchmarking World Knowledge of Large Language Models", "url": "https://openreview.net/forum?id=AqN23oqraW", "detail_url": "https://openreview.net/forum?id=AqN23oqraW", "authors": "Jifan Yu,Xiaozhi Wang,Shangqing Tu,Shulin Cao,Daniel Zhang-Li,Xin Lv,Hao Peng,Zijun Yao,Xiaohan Zhang,Hanming Li,Chunyang Li,Zheyuan Zhang,Yushi Bai,Yantao Liu,Amy Xin,Kaifeng Yun,Linlu GONG,Nianyi Lin,Jianhui Chen,Zhili Wu,Yunjia Qi,Weikai Li,Yong Guan,Kaisheng Zeng,Ji Qi,Hailong Jin,Jinxin Liu,Yu Gu,Yuan Yao,Ning Ding,Lei Hou,Zhiyuan Liu,Xu Bin,Jie Tang,Juanzi Li", "tags": "ICLR 2024,Poster", "abstract": "The unprecedented performance of large language models (LLMs) necessitates improvements in evaluations. Rather than merely exploring the breadth of LLM abilities, we believe meticulous and thoughtful designs are essential to thorough, unbiased, and applicable evaluations. Given the importance of world knowledge to LLMs, we construct a Knowledge-oriented LLM Assessment benchmark (KoLA), in which we carefully design three crucial factors: (1) For ability modeling, we mimic human cognition to form a four-level taxonomy of knowledge-related abilities, covering 19 tasks. (2) For data, to ensure fair comparisons, we use both Wikipedia, a corpus prevalently pre-trained by LLMs, along with continuously collected emerging corpora, aiming to evaluate the capacity to handle unseen data and evolving knowledge. (3) For evaluation criteria, we adopt a contrastive system, including overall standard scores for better numerical comparability across tasks and models, and a unique self-contrast metric for automatically evaluating knowledge-creating ability. We evaluate 21 open-source and commercial LLMs and obtain some intriguing findings. The KoLA dataset will be updated every three months to provide timely references for developing LLMs and knowledge-related systems.", "pdf": "https://openreview.net/pdf/5b10764ea102a63e929b8b1179abdca8c1903c0a.pdf"} {"title": "Graph Parsing Networks", "url": "https://openreview.net/forum?id=hv3SklibkL", "detail_url": "https://openreview.net/forum?id=hv3SklibkL", "authors": "Yunchong Song,Siyuan Huang,Xinbing Wang,Chenghu Zhou,Zhouhan Lin", "tags": "ICLR 2024,Poster", "abstract": "Graph pooling compresses graph information into a compact representation. State-of-the-art graph pooling methods follow a hierarchical approach, which reduces the graph size step-by-step. These methods must balance memory efficiency with preserving node information, depending on whether they use node dropping or node clustering. Additionally, fixed pooling ratios or numbers of pooling layers are predefined for all graphs, which prevents personalized pooling structures from being captured for each individual graph. In this work, inspired by bottom-up grammar induction, we propose an efficient graph parsing algorithm to infer the pooling structure, which then drives graph pooling. The resulting Graph Parsing Network (GPN) adaptively learns personalized pooling structure for each individual graph. GPN benefits from the discrete assignments generated by the graph parsing algorithm, allowing good memory efficiency while preserving node information intact. Experimental results on standard benchmarks demonstrate that GPN outperforms state-of-the-art graph pooling methods in graph classification tasks while being able to achieve competitive performance in node classification tasks. We also conduct a graph reconstruction task to show GPN's ability to preserve node information and measure both memory and time efficiency through relevant tests.", "pdf": "https://openreview.net/pdf/d9ce08bd60cb1e4ccd9fa59eade5645174e6c379.pdf"} {"title": "TESTAM: A Time-Enhanced Spatio-Temporal Attention Model with Mixture of Experts", "url": "https://openreview.net/forum?id=N0nTk5BSvO", "detail_url": "https://openreview.net/forum?id=N0nTk5BSvO", "authors": "Hyunwook Lee,Sungahn Ko", "tags": "ICLR 2024,Poster", "abstract": "Accurate traffic forecasting is challenging due to the complex dependency on road networks, various types of roads, and the abrupt speed change due to the events. Recent works mainly focus on dynamic spatial modeling with adaptive graph embedding or graph attention having less consideration for temporal characteristics and in-situ modeling. In this paper, we propose a novel deep learning model named TESTAM, which individually models recurring and non-recurring traffic patterns by a mixture-of-experts model with three experts on temporal modeling, spatio-temporal modeling with static graph, and dynamic spatio-temporal dependency modeling with dynamic graph. By introducing different experts and properly routing them, TESTAM could better model various circumstances, including spatially isolated nodes, highly related nodes, and recurring and non-recurring events. For the proper routing, we reformulate a gating problem into a classification problem with pseudo labels. Experimental results on three public traffic network datasets, METR-LA, PEMS-BAY, and EXPY-TKY, demonstrate that TESTAM achieves a better indication and modeling of recurring and non-recurring traffic.", "pdf": "https://openreview.net/pdf/75248a8c97bcbac15b26c02b68788bf7ddb56175.pdf"} {"title": "Learning From Simplicial Data Based on Random Walks and 1D Convolutions", "url": "https://openreview.net/forum?id=OsGUnYOzii", "detail_url": "https://openreview.net/forum?id=OsGUnYOzii", "authors": "Florian Frantzen,Michael T Schaub", "tags": "ICLR 2024,Poster", "abstract": "Triggered by limitations of graph-based deep learning methods in terms of computational expressivity and model flexibility, recent years have seen a surge of interest in computational models that operate on higher-order topological domains such as hypergraphs and simplicial complexes. While the increased expressivity of these models can indeed lead to a better classification performance and a more faithful representation of the underlying system, the computational cost of these higher-order models can increase dramatically. To this end, we here explore a simplicial complex neural network learning architecture based on random walks and fast 1D convolutions (SCRaWl), in which we can adjust the increase in computational cost by varying the length and number of random walks considered while accounting for higher-order relationships. Importantly, due to the random walk-based design, the expressivity of the proposed architecture is provably incomparable to that of existing message-passing simplicial neural networks. We empirically evaluate SCRaWl on real-world datasets and show that it outperforms other simplicial neural networks.", "pdf": "https://openreview.net/pdf/f7f8f6134614bab4c34752266fafd849070821c0.pdf"} {"title": "LUM-ViT: Learnable Under-sampling Mask Vision Transformer for Bandwidth Limited Optical Signal Acquisition", "url": "https://openreview.net/forum?id=wkbeqr5XhC", "detail_url": "https://openreview.net/forum?id=wkbeqr5XhC", "authors": "Lingfeng Liu,Dong Ni,Hangjie Yuan", "tags": "ICLR 2024,Poster", "abstract": "Bandwidth constraints during signal acquisition frequently impede real-time detection applications. Hyperspectral data is a notable example, whose vast volume compromises real-time hyperspectral detection. To tackle this hurdle, we introduce a novel approach leveraging pre-acquisition modulation to reduce the acquisition volume. This modulation process is governed by a deep learning model, utilizing prior information. Central to our approach is LUM-ViT, a Vision Transformer variant. Uniquely, LUM-ViT incorporates a learnable under-sampling mask tailored for pre-acquisition modulation. To further optimize for optical calculations, we propose a kernel-level weight binarization technique and a three-stage fine-tuning strategy. Our evaluations reveal that, by sampling a mere 10\\% of the original image pixels, LUM-ViT maintains the accuracy loss within 1.8\\% on the ImageNet classification task. The method sustains near-original accuracy when implemented on real-world optical hardware, demonstrating its practicality. Code will be available at [https://github.com/MaxLLF/LUM-ViT](https://github.com/MaxLLF/LUM-ViT).", "pdf": "https://openreview.net/pdf/e727a272852fd9b151a8c6fee5946d926aa8f97c.pdf"} {"title": "Social-Transmotion: Promptable Human Trajectory Prediction", "url": "https://openreview.net/forum?id=SQpnEfv9WH", "detail_url": "https://openreview.net/forum?id=SQpnEfv9WH", "authors": "Saeed Saadatnejad,Yang Gao,Kaouther Messaoud,Alexandre Alahi", "tags": "ICLR 2024,Poster", "abstract": "Accurate human trajectory prediction is crucial for applications such as autonomous vehicles, robotics, and surveillance systems. Yet, existing models often fail to fully leverage the non-verbal social cues human subconsciously communicate when navigating the space.\nTo address this, we introduce *Social-Transmotion*, a generic Transformer-based model that exploits diverse and numerous visual cues to predict human behavior. We translate the idea of a prompt from Natural Language Processing (NLP) to the task of human trajectory prediction, where a prompt can be a sequence of x-y coordinates on the ground, bounding boxes in the image plane, or body pose keypoints in either 2D or 3D. This, in turn, augments trajectory data, leading to enhanced human trajectory prediction.\nUsing masking technique, our model exhibits flexibility and adaptability by capturing spatiotemporal interactions between agents based on the available visual cues.\nWe delve into the merits of using 2D versus 3D poses, and a limited set of poses. Additionally, we investigate the spatial and temporal attention map to identify which keypoints and time-steps in the sequence are vital for optimizing human trajectory prediction.\nOur approach is validated on multiple datasets, including JTA, JRDB, Pedestrians and Cyclists in Road Traffic, and ETH-UCY.\nThe code is publicly available: [https://github.com/vita-epfl/social-transmotion](https://github.com/vita-epfl/social-transmotion).", "pdf": "https://openreview.net/pdf/5dcf47a66eb9a88f6cda4a59dfb2dd17df12bc68.pdf"} {"title": "Robust Classification via Regression for Learning with Noisy Labels", "url": "https://openreview.net/forum?id=wfgZc3IMqo", "detail_url": "https://openreview.net/forum?id=wfgZc3IMqo", "authors": "Erik Englesson,Hossein Azizpour", "tags": "ICLR 2024,Poster", "abstract": "Deep neural networks and large-scale datasets have revolutionized the field of machine learning. However, these large networks are susceptible to overfitting to label noise, resulting in reduced generalization. To address this challenge, two promising approaches have emerged: i) loss reweighting, which reduces the influence of noisy examples on the training loss, and ii) label correction that replaces noisy labels with estimated true labels. These directions have been pursued separately or combined as independent methods, lacking a unified approach. In this work, we present a unified method that seamlessly combines loss reweighting and label correction to enhance robustness against label noise in classification tasks. Specifically, by leveraging ideas from compositional data analysis in statistics, we frame the problem as a regression task, where loss reweighting and label correction can naturally be achieved with a shifted Gaussian label noise model. Our unified approach achieves strong performance compared to recent baselines on several noisy labelled datasets. We believe this work is a promising step towards robust deep learning in the presence of label noise. Our code is available at: https://github.com/ErikEnglesson/SGN.", "pdf": "https://openreview.net/pdf/dc6d5771bf99d6ea8d806a95a0283ab492a6448c.pdf"} {"title": "Learning to Reject with a Fixed Predictor: Application to Decontextualization", "url": "https://openreview.net/forum?id=dCHbFDsCZz", "detail_url": "https://openreview.net/forum?id=dCHbFDsCZz", "authors": "Christopher Mohri,Daniel Andor,Eunsol Choi,Michael Collins,Anqi Mao,Yutao Zhong", "tags": "ICLR 2024,Poster", "abstract": "We study the problem of classification with a reject option for a fixed predictor, crucial to natural language processing. We introduce a new problem formulation for this scenario, and an algorithm minimizing a new surrogate loss function. We provide a complete theoretical analysis of the surrogate loss function with a strong $H$-consistency guarantee. For evaluation, we choose the \\textit{decontextualization} task, and provide a manually-labelled dataset of $2\\mathord,000$ examples. Our algorithm significantly outperforms the baselines considered, with a $\\sim 25$% improvement in coverage when halving the error rate, which is only $\\sim 3$% away from the theoretical limit.", "pdf": "https://openreview.net/pdf/66caaecc6922bf5ef4e221d3e575e0c06e70075a.pdf"} {"title": "Dynamics-Informed Protein Design with Structure Conditioning", "url": "https://openreview.net/forum?id=jZPqf2G9Sw", "detail_url": "https://openreview.net/forum?id=jZPqf2G9Sw", "authors": "Urszula Julia Komorowska,Simon V Mathis,Kieran Didi,Francisco Vargas,Pietro Lio,Mateja Jamnik", "tags": "ICLR 2024,Poster", "abstract": "Current protein generative models are able to design novel backbones with desired shapes or functional motifs. However, despite the importance of a protein\u2019s dynamical properties for its function, conditioning on dynamical properties remains elusive. We present a new approach to protein generative modeling by leveraging Normal Mode Analysis that enables us to capture dynamical properties too. We introduce a method for conditioning the diffusion probabilistic models on protein dynamics, specifically on the lowest non-trivial normal mode of oscillation. Our method, similar to the classifier guidance conditioning, formulates the sampling process as being driven by conditional and unconditional terms. However, unlike previous works, we approximate the conditional term with a simple analytical function rather than an external neural network, thus making the eigenvector calculations approachable. We present the corresponding SDE theory as a formal justification of our approach. We extend our framework to conditioning on structure and dynamics at the same time, enabling scaffolding of the dynamical motifs. We demonstrate the empirical effectiveness of our method by turning the open-source unconditional protein diffusion model Genie into the conditional model with no retraining. Generated proteins exhibit the desired dynamical and structural properties while still being biologically plausible. Our work represents a first step towards incorporating dynamical behaviour in protein design and may open the door to designing more flexible and functional proteins in the future.", "pdf": "https://openreview.net/pdf/0ffbbb38174d8802162c0bef4914451f67318f38.pdf"} {"title": "Partitioning Message Passing for Graph Fraud Detection", "url": "https://openreview.net/forum?id=tEgrUrUuwA", "detail_url": "https://openreview.net/forum?id=tEgrUrUuwA", "authors": "Wei Zhuo,Zemin Liu,Bryan Hooi,Bingsheng He,Guang Tan,Rizal Fathony,Jia Chen", "tags": "ICLR 2024,Poster", "abstract": "Label imbalance and homophily-heterophily mixture are the fundamental problems encountered when applying Graph Neural Networks (GNNs) to Graph Fraud Detection (GFD) tasks. Existing GNN-based GFD models are designed to augment graph structure to accommodate the inductive bias of GNNs towards homophily, by excluding heterophilic neighbors during message passing. In our work, we argue that the key to applying GNNs for GFD is not to exclude but to {\\em distinguish} neighbors with different labels. Grounded in this perspective, we introduce Partitioning Message Passing (PMP), an intuitive yet effective message passing paradigm expressly crafted for GFD. Specifically, in the neighbor aggregation stage of PMP, neighbors with different classes are aggregated with distinct node-specific aggregation functions. By this means, the center node can adaptively adjust the information aggregated from its heterophilic and homophilic neighbors, thus avoiding the model gradient being dominated by benign nodes which occupy the majority of the population. We theoretically establish a connection between the spatial formulation of PMP and spectral analysis to characterize that PMP operates an adaptive node-specific spectral graph filter, which demonstrates the capability of PMP to handle heterophily-homophily mixed graphs. Extensive experimental results show that PMP can significantly boost the performance on GFD tasks.", "pdf": "https://openreview.net/pdf/7abe6053f04ef235ce8ddd5996dd27e266c6e058.pdf"} {"title": "Self-contradictory Hallucinations of Large Language Models: Evaluation, Detection and Mitigation", "url": "https://openreview.net/forum?id=EmQSOi1X2f", "detail_url": "https://openreview.net/forum?id=EmQSOi1X2f", "authors": "Niels M\u00fcndler,Jingxuan He,Slobodan Jenko,Martin Vechev", "tags": "ICLR 2024,Poster", "abstract": "Large language models (large LMs) are susceptible to producing text that contains hallucinated content. An important instance of this problem is self-contradiction, where the LM generates two contradictory sentences within the same context. In this work, we present a comprehensive investigation into self-contradiction for various instruction-tuned LMs, covering evaluation, detection, and mitigation. Our primary evaluation task is open-domain text generation, but we also demonstrate the applicability of our approach to shorter question answering. Our analysis reveals the prevalence of self-contradictions, e.g., in 17.7% of all sentences produced by ChatGPT. We then propose a novel prompting-based framework designed to effectively detect and mitigate self-contradictions. Our detector achieves high accuracy, e.g., around 80% F1 score when prompting ChatGPT. The mitigation algorithm iteratively refines the generated text to remove contradictory information while preserving text fluency and informativeness. Importantly, our entire framework is applicable to black-box LMs and does not require retrieval of external knowledge. Rather, our method complements retrieval-based methods, as a large portion of self-contradictions (e.g., 35.2% for ChatGPT) cannot be verified using online text. Our approach is practically effective and has been released as a push-button tool to benefit the public at https://chatprotect.ai/.", "pdf": "https://openreview.net/pdf/9087c7d18bb7707f5f8b47046499f922b5c47af3.pdf"} {"title": "Symmetric Neural-Collapse Representations with Supervised Contrastive Loss: The Impact of ReLU and Batching", "url": "https://openreview.net/forum?id=AyXIDfvYg8", "detail_url": "https://openreview.net/forum?id=AyXIDfvYg8", "authors": "Ganesh Ramachandra Kini,Vala Vakilian,Tina Behnia,Jaidev Gill,Christos Thrampoulidis", "tags": "ICLR 2024,Poster", "abstract": "Supervised contrastive loss (SCL) is a competitive and often superior alternative to the cross-entropy loss for classification. While prior studies have demonstrated that both losses yield symmetric training representations under balanced data, this symmetry breaks under class imbalances. This paper presents an intriguing discovery: the introduction of a ReLU activation at the final layer effectively restores the symmetry in SCL-learned representations. We arrive at this finding analytically, by establishing that the global minimizers of an unconstrained features model with SCL loss and entry-wise non-negativity constraints form an orthogonal frame. Extensive experiments conducted across various datasets, architectures, and imbalance scenarios corroborate our finding. Importantly, our experiments reveal that the inclusion of the ReLU activation restores symmetry without compromising test accuracy. This constitutes the first geometry characterization of SCL under imbalances. Additionally, our analysis and experiments underscore the pivotal role of batch selection strategies in representation geometry. By proving necessary and sufficient conditions for mini-batch choices that ensure invariant symmetric representations, we introduce batch-binding as an efficient strategy that guarantees these conditions hold.", "pdf": "https://openreview.net/pdf/b40ab759531e71a4137172d4dca6019ea210d8d1.pdf"} {"title": "Manipulating dropout reveals an optimal balance of efficiency and robustness in biological and machine visual systems", "url": "https://openreview.net/forum?id=ADDCErFzev", "detail_url": "https://openreview.net/forum?id=ADDCErFzev", "authors": "Jacob S. Prince,Gabriel Fajardo,George A. Alvarez,Talia Konkle", "tags": "ICLR 2024,Poster", "abstract": "According to the efficient coding hypothesis, neural populations encode information optimally when representations are high-dimensional and uncorrelated. However, such codes may carry a cost in terms of generalization and robustness. Past empirical studies of early visual cortex (V1) in rodents have suggested that this tradeoff indeed constrains sensory representations. However, it remains unclear whether these insights generalize across the hierarchy of the human visual system, and particularly to object representations in high-level occipitotemporal cortex (OTC). To gain new empirical clarity, here we develop a family of object recognition models with parametrically varying dropout proportion $p$, which induces systematically varying dimensionality of internal responses (while controlling all other inductive biases). We find that increasing dropout produces an increasingly smooth, low-dimensional representational space. Optimal robustness to lesioning is observed at around 70% dropout, after which both accuracy and robustness decline. Representational comparison to large-scale 7T fMRI data from occipitotemporal cortex in the Natural Scenes Dataset reveals that this optimal degree of dropout is also associated with maximal emergent neural predictivity. Finally, using new techniques for achieving denoised estimates of the eigenspectrum of human fMRI responses, we compare the rate of eigenspectrum decay between model and brain feature spaces. We observe that the match between model and brain representations is associated with a common balance between efficiency and robustness in the representational space. These results suggest that varying dropout may reveal an optimal point of balance between the efficiency of high-dimensional codes and the robustness of low dimensional codes in hierarchical vision systems.", "pdf": "https://openreview.net/pdf/7f10cc682acaa93ecdd9fa4c23e343be124115dd.pdf"} {"title": "DDMI: Domain-agnostic Latent Diffusion Models for Synthesizing High-Quality Implicit Neural Representations", "url": "https://openreview.net/forum?id=327tbF3S65", "detail_url": "https://openreview.net/forum?id=327tbF3S65", "authors": "Dogyun Park,Sihyeon Kim,Sojin Lee,Hyunwoo J. Kim", "tags": "ICLR 2024,Poster", "abstract": "Recent studies have introduced a new class of generative models for synthesizing implicit neural representations (INRs) that capture arbitrary continuous signals in various domains. These models opened the door for domain-agnostic generative models, but they often fail to achieve high-quality generation. We observed that the existing methods generate the weights of neural networks to parameterize INRs and evaluate the network with fixed positional embeddings (PEs). Arguably, this architecture limits the expressive power of generative models and results in low-quality INR generation. To address this limitation, we propose Domain-agnostic Latent Diffusion Model for INRs (DDMI) that generates adaptive positional embeddings instead of neural networks' weights. Specifically, we develop a Discrete-to-continuous space Variational AutoEncoder (D2C-VAE) that seamlessly connects discrete data and continuous signal functions in the shared latent space. Additionally, we introduce a novel conditioning mechanism for evaluating INRs with the hierarchically decomposed PEs to further enhance expressive power. Extensive experiments across four modalities, \\eg, 2D images, 3D shapes, Neural Radiance Fields, and videos, with seven benchmark datasets, demonstrate the versatility of DDMI and its superior performance compared to the existing INR generative models. Code is available at \\href{https://github.com/mlvlab/DDMI}{https://github.com/mlvlab/DDMI}.", "pdf": "https://openreview.net/pdf/716a67cdede3543daa6309c204a066e517decc41.pdf"} {"title": "Bayesian Coreset Optimization for Personalized Federated Learning", "url": "https://openreview.net/forum?id=uz7d2N2zul", "detail_url": "https://openreview.net/forum?id=uz7d2N2zul", "authors": "Prateek Chanda,Shrey Modi,Ganesh Ramakrishnan", "tags": "ICLR 2024,Poster", "abstract": "In a distributed machine learning setting like Federated Learning where there are multiple clients involved which update their individual weights to a single central server, often training on the entire individual client's dataset for each client becomes cumbersome. To address this issue we propose CORESET-PFEDBAYES : a personalized coreset weighted federated learning setup where the training updates for each individual clients are forwarded to the central server based on only individual client coreset based representative data points instead of the entire client data. Through theoretical analysis we present how the average generalization error is minimax optimal up to logarithm bounds (upper bounded by $\\mathcal{O}(n_k^{-\\frac{2 \\beta}{2 \\beta+\\boldsymbol{\\Lambda}}} \\log ^{2 \\delta^{\\prime}}(n_k))$) and lower bounds of $\\mathcal{O}(n_k^{-\\frac{2 \\beta}{2 \\beta+\\boldsymbol{\\Lambda}}})$, and how the overall generalization error on the data likelihood differs from a vanilla Federated Learning setup as a closed form function ${\\boldsymbol{\\Im}}(\\boldsymbol{w}, n_k)$ of the coreset weights $\\boldsymbol{w}$ and coreset sample size $n_k$. \nOur experiments on different benchmark datasets based on a variety of recent personalized federated learning architectures show significant gains as compared to random sampling on the training data followed by federated learning, thereby indicating how intelligently selecting such training samples can help in performance. Additionally, through experiments on medical datasets our proposed method showcases some gains as compared to other submodular optimization based approaches used for subset selection on client's data.", "pdf": "https://openreview.net/pdf/cd650d7bda1c2da0c78084f2b993c15c2ec0925f.pdf"} {"title": "In-context Autoencoder for Context Compression in a Large Language Model", "url": "https://openreview.net/forum?id=uREj4ZuGJE", "detail_url": "https://openreview.net/forum?id=uREj4ZuGJE", "authors": "Tao Ge,Hu Jing,Lei Wang,Xun Wang,Si-Qing Chen,Furu Wei", "tags": "ICLR 2024,Poster", "abstract": "We propose the In-context Autoencoder (ICAE), leveraging the power of a large language model (LLM) to compress a long context into short compact memory slots that can be directly conditioned on by the LLM for various purposes. ICAE is first pretrained using both autoencoding and language modeling objectives on massive text data, enabling it to generate memory slots that accurately and comprehensively represent the original context. Then, it is fine-tuned on instruction data for producing desirable responses to various prompts. Experiments demonstrate that our lightweight ICAE, introducing about 1% additional parameters, effectively achieves $4\\times$ context compression based on Llama, offering advantages in both improved latency and GPU memory cost during inference, and showing an interesting insight in memorization as well as potential for scalability. These promising results imply a novel perspective on the connection between working memory in cognitive science and representation learning in LLMs, revealing ICAE's significant implications in addressing the long context problem and suggesting further research in LLM context management. Our data, code and models are available at https://github.com/getao/icae.", "pdf": "https://openreview.net/pdf/0cb80bedd6a43e1383e8fe4642b39105d27be261.pdf"} {"title": "Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning", "url": "https://openreview.net/forum?id=J44HfH4JCg", "detail_url": "https://openreview.net/forum?id=J44HfH4JCg", "authors": "Fuxiao Liu,Kevin Lin,Linjie Li,Jianfeng Wang,Yaser Yacoob,Lijuan Wang", "tags": "ICLR 2024,Poster", "abstract": "Despite the promising progress in multi-modal tasks, current large multi-modal models (LMMs) are prone to hallucinating inconsistent descriptions with respect to the associated image and human instructions. This paper addresses this issue by introducing the first large and diverse visual instruction tuning dataset, named Large-scale Robust Visual (LRV)-Instruction. Our dataset comprises 400k visual\ninstructions generated by GPT4, covering 16 vision-and-language tasks with open-ended instructions and answers. Unlike existing studies that primarily focus on positive instruction samples, we design LRV-Instruction to include both positive and negative instructions for more robust visual instruction tuning. Our negative instructions are designed at three semantic levels: (i) Nonexistent Object Manipulation, (ii) Existent Object Manipulation and (iii) Knowledge Manipulation. To efficiently measure the hallucination generated by LMMs, we propose GPT4-Assisted Visual Instruction Evaluation (GAVIE), a stable approach to evaluate visual instruction tuning like human experts. GAVIE does not require human-annotated groundtruth answers and can adapt to diverse instruction formats. We conduct comprehensive experiments to investigate the hallucination of LMMs. Our results demonstrate existing LMMs exhibit significant hallucinations when presented with our negative instructions, particularly Existent Object and Knowledge Manipulation instructions. Moreover, we successfully mitigate hallucination by finetuning MiniGPT4 and mPLUG-Owl on LRV-Instruction while improving performance on several public\ndatasets compared to state-of-the-art methods. Additionally, we observed that a balanced ratio of positive and negative instances in the training data leads to a more robust model. Code and data will be released upon publication.", "pdf": "https://openreview.net/pdf/a7dc45dbafdb5c8be70b9ae7f987f44a08b62885.pdf"} {"title": "Multimarginal Generative Modeling with Stochastic Interpolants", "url": "https://openreview.net/forum?id=FHqAzWl2wE", "detail_url": "https://openreview.net/forum?id=FHqAzWl2wE", "authors": "Michael Samuel Albergo,Nicholas Matthew Boffi,Michael Lindsey,Eric Vanden-Eijnden", "tags": "ICLR 2024,Poster", "abstract": "Given a set of $K$ probability densities, we consider the multimarginal generative modeling problem of learning a joint distribution that recovers these densities as marginals. The structure of this joint distribution should identify multi-way correspondences among the prescribed marginals. We formalize an approach to this task within a generalization of the stochastic interpolant framework, leading to efficient learning algorithms built upon dynamical transport of measure. Our generative models are defined by velocity and score fields that can be characterized as the minimizers of simple quadratic objectives, and they are defined on a simplex that generalizes the time variable in the usual dynamical transport framework. The resulting transport on the simplex is influenced by all marginals, and we show that multi-way correspondences can be extracted. The identification of such correspondences has applications to style transfer, algorithmic fairness, and data decorruption. In addition, the multimarginal perspective enables an efficient algorithm for optimizing the dynamical transport cost in the ordinary two-marginal setting. We demonstrate these capacities with several numerical examples.", "pdf": "https://openreview.net/pdf/ffbe0129551a2886a81d49db461b05400f637dee.pdf"} {"title": "Rethinking Channel Dependence for Multivariate Time Series Forecasting: Learning from Leading Indicators", "url": "https://openreview.net/forum?id=JiTVtCUOpS", "detail_url": "https://openreview.net/forum?id=JiTVtCUOpS", "authors": "Lifan Zhao,Yanyan Shen", "tags": "ICLR 2024,Poster", "abstract": "Recently, channel-independent methods have achieved state-of-the-art performance in multivariate time series (MTS) forecasting. Despite reducing overfitting risks, these methods miss potential opportunities in utilizing channel dependence for accurate predictions. We argue that there exist locally stationary lead-lag relationships between variates, i.e., some lagged variates may follow the leading indicators within a short time period. Exploiting such channel dependence is beneficial since leading indicators offer advance information that can be used to reduce the forecasting difficulty of the lagged variates. In this paper, we propose a new method named LIFT that first efficiently estimates leading indicators and their leading steps at each time step and then judiciously allows the lagged variates to utilize the advance information from leading indicators. LIFT plays as a plugin that can be seamlessly collaborated with arbitrary time series forecasting methods. Extensive experiments on six real-world datasets demonstrate that LIFT improves the state-of-the-art methods by 5.4% in average forecasting performance. Our code is available at https://github.com/SJTU-Quant/LIFT.", "pdf": "https://openreview.net/pdf/d434e3f2213d54e8999f551b44bbc2454529e8fb.pdf"} {"title": "RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval", "url": "https://openreview.net/forum?id=GN921JHCRw", "detail_url": "https://openreview.net/forum?id=GN921JHCRw", "authors": "Parth Sarthi,Salman Abdullah,Aditi Tuli,Shubh Khanna,Anna Goldie,Christopher D Manning", "tags": "ICLR 2024,Poster", "abstract": "Retrieval-augmented language models can better adapt to changes in world state and incorporate long-tail knowledge. However, most existing methods retrieve only short contiguous chunks from a retrieval corpus, limiting holistic understanding of the overall document context. We introduce the novel approach of recursively embedding, clustering, and summarizing chunks of text, constructing a tree with differing levels of summarization from the bottom up. At inference time, our RAPTOR model retrieves from this tree, integrating information across lengthy documents at different levels of abstraction. Controlled experiments show that retrieval with recursive summaries offers significant improvements over traditional retrieval-augmented LMs on several tasks. On question-answering tasks that involve complex, multi-step reasoning, we show state-of-the-art results; for example, by coupling RAPTOR retrieval with the use of GPT-4, we can improve the best performance on the QuALITY benchmark by 20\\% in absolute accuracy.", "pdf": "https://openreview.net/pdf/1414c23eb8af2e30daa25a4f53471f69091f48d9.pdf"} {"title": "Fair and Efficient Contribution Valuation for Vertical Federated Learning", "url": "https://openreview.net/forum?id=sLQb8q0sUi", "detail_url": "https://openreview.net/forum?id=sLQb8q0sUi", "authors": "Zhenan Fan,Huang Fang,Xinglu Wang,Zirui Zhou,Jian Pei,Michael Friedlander,Yong Zhang", "tags": "ICLR 2024,Poster", "abstract": "Federated learning is an emerging technology for training machine learning models across decentralized data sources without sharing data. Vertical federated learning, also known as feature-based federated learning, applies to scenarios where data sources have the same sample IDs but different feature sets. To ensure fairness among data owners, it is critical to objectively assess the contributions from different data sources and compensate the corresponding data owners accordingly. The Shapley value is a provably fair contribution valuation metric originating from cooperative game theory. However, its straight-forward computation requires extensively retraining a model on each potential combination of data sources, leading to prohibitively high communication and computation overheads due to multiple rounds of federated learning. To tackle this challenge, we propose a contribution valuation metric called vertical federated Shapley value (VerFedSV) based on the classic Shapley value. We show that VerFedSV not only satisfies many desirable properties of fairness but is also efficient to compute. Moreover, VerFedSV can be adapted to both synchronous and asynchronous vertical federated learning algorithms. Both theoretical analysis and extensive experimental results demonstrate the fairness, efficiency, adaptability, and effectiveness of VerFedSV.", "pdf": "https://openreview.net/pdf/6a7f95c7baec41005e195a24ebc51afc7ef5acbf.pdf"} {"title": "In-Context Learning through the Bayesian Prism", "url": "https://openreview.net/forum?id=HX5ujdsSon", "detail_url": "https://openreview.net/forum?id=HX5ujdsSon", "authors": "Madhur Panwar,Kabir Ahuja,Navin Goyal", "tags": "ICLR 2024,Poster", "abstract": "In-context learning (ICL) is one of the surprising and useful features of large language models and subject of intense research. Recently, stylized meta-learning-like ICL setups have been devised that train transformers on sequences of input-output pairs $(x, f(x))$. The function $f$ comes from a function class and generalization is checked by evaluating on sequences generated from unseen functions from the same class. One of the main discoveries in this line of research has been that for several function classes, such as linear regression, transformers successfully generalize to new functions in the class. However, the inductive biases of these models resulting in this behavior are not clearly understood. A model with unlimited training data and compute is a Bayesian predictor: it learns the pretraining distribution.\nIn this paper we empirically examine how far this Bayesian perspective can help us understand ICL. To this end, we generalize the previous meta-ICL setup to hierarchical meta-ICL setup which involve unions of multiple task families. We instantiate this setup on a diverse range of linear and nonlinear function families and find that transformers can do ICL in this setting as well. Where Bayesian inference is tractable, we find evidence that high-capacity transformers mimic the Bayesian predictor. The Bayesian perspective provides insights into the inductive bias of ICL and how transformers perform a particular task when they are trained on multiple tasks. We also find that transformers can learn to generalize to new function classes that were not seen during pretraining. This involves deviation from the Bayesian predictor. We examine these deviations in more depth offering new insights and hypotheses.", "pdf": "https://openreview.net/pdf/fc9aa29ea37339217577f61679622246ebfce078.pdf"} {"title": "RingAttention with Blockwise Transformers for Near-Infinite Context", "url": "https://openreview.net/forum?id=WsRHpHH4s0", "detail_url": "https://openreview.net/forum?id=WsRHpHH4s0", "authors": "Hao Liu,Matei Zaharia,Pieter Abbeel", "tags": "ICLR 2024,Poster", "abstract": "Transformers have emerged as the architecture of choice for many state-of-the-art AI models, showcasing exceptional performance across a wide range of AI applications. However, the memory demands imposed by Transformers limit their ability to handle long sequences, thereby posing challenges in utilizing videos, actions, and other long-form sequences and modalities in complex environments. We present a novel approach, Blockwise RingAttention, which leverages blockwise computation of self-attention and feedforward to distribute long sequences across multiple devices while fully overlapping the communication of key-value blocks with the computation of blockwise attention. Our approach enables training and inference of sequences that are up to device count times longer than those achievable by prior memory-efficient Transformers, without resorting to approximations or incurring additional communication and computation overheads. Extensive experiments on language modeling and reinforcement learning tasks demonstrate the effectiveness of our approach in allowing millions of tokens context size and improving performance.", "pdf": "https://openreview.net/pdf/46002918e58387fc0091aa342ec23ebe66fd93e4.pdf"} {"title": "Chain of Hindsight aligns Language Models with Feedback", "url": "https://openreview.net/forum?id=6xfe4IVcOu", "detail_url": "https://openreview.net/forum?id=6xfe4IVcOu", "authors": "Hao Liu,Carmelo Sferrazza,Pieter Abbeel", "tags": "ICLR 2024,Poster", "abstract": "Learning from human preferences is important for language models to match human needs and to align with human and social values. \nPrior works have achieved remarkable successes by learning from human feedback to understand and follow instructions. Nonetheless, these methods are either founded on hand-picked model generations that are favored by human annotators, rendering them inefficient in terms of data utilization and challenging to apply in general, or they depend on reinforcement learning, which often suffers from imperfect reward functions and relies on extremely challenging optimizations. In this work, we propose a novel technique, Chain of Hindsight, that is easy to optimize and can learn from any form of feedback, regardless of its polarity. Our idea is inspired by how humans learn from extensive feedback presented in the form of languages. We convert all types of feedback into sequences of sentences, which are then used to fine-tune the model, allowing us to take advantage of the language comprehension capabilities of language models.\nWe condition the model on a sequence of model generations paired with feedback. By doing so, the model is trained to generate outputs based on feedback, while learning to identify and correct negative attributes or errors. Applying our method to large language models, we observed that Chain of Hindsight significantly surpasses previous methods in aligning language models with human preferences. We report significant improvements on summarization and dialogue benchmarks, with our approach markedly preferred in human evaluations.", "pdf": "https://openreview.net/pdf/6dc18121c96221c694b6c583053368af3aa30ace.pdf"} {"title": "GraphChef: Decision-Tree Recipes to Explain Graph Neural Networks", "url": "https://openreview.net/forum?id=IjMUGuUmBI", "detail_url": "https://openreview.net/forum?id=IjMUGuUmBI", "authors": "Peter M\u00fcller,Lukas Faber,Karolis Martinkus,Roger Wattenhofer", "tags": "ICLR 2024,Poster", "abstract": "We propose a new self-explainable Graph Neural Network (GNN) model: GraphChef. GraphChef integrates decision trees into the GNN message passing framework. Given a dataset, GraphChef returns a set of rules (a recipe) that explains each class in the dataset unlike existing GNNs and explanation methods that reason on individual graphs. Thanks to the decision trees, GraphChef recipes are human understandable. We also present a new pruning method to produce small and easy to digest trees. Experiments demonstrate that GraphChef reaches comparable accuracy to not self-explainable GNNs and produced decision trees are indeed small. We further validate the correctness of the discovered recipes on datasets where explanation ground truth is available: Reddit-Binary, MUTAG, BA-2Motifs, BA-Shapes, Tree-Cycle, and Tree-Grid.", "pdf": "https://openreview.net/pdf/0e5d8259f045fd1291bc9817e8008716c5c83cb3.pdf"} {"title": "Safe Collaborative Filtering", "url": "https://openreview.net/forum?id=yarUvgEXq3", "detail_url": "https://openreview.net/forum?id=yarUvgEXq3", "authors": "Riku Togashi,Tatsushi Oka,Naoto Ohsaka,Tetsuro Morimura", "tags": "ICLR 2024,Poster", "abstract": "Excellent tail performance is crucial for modern machine learning tasks, such as algorithmic fairness, class imbalance, and risk-sensitive decision making, as it ensures the effective handling of challenging samples within a dataset. Tail performance is also a vital determinant of success for personalized recommender systems to reduce the risk of losing users with low satisfaction. This study introduces a \"safe\" collaborative filtering method that prioritizes recommendation quality for less-satisfied users rather than focusing on the average performance. Our approach minimizes the conditional value at risk (CVaR), which represents the average risk over the tails of users' loss. To overcome computational challenges for web-scale recommender systems, we develop a robust yet practical algorithm that extends the most scalable method, implicit alternating least squares (iALS). Empirical evaluation on real-world datasets demonstrates the excellent tail performance of our approach while maintaining competitive computational efficiency.", "pdf": "https://openreview.net/pdf/51efd839f7a6eb9312e011822450ded36a856126.pdf"} {"title": "On Representation Complexity of Model-based and Model-free Reinforcement Learning", "url": "https://openreview.net/forum?id=3K3s9qxSn7", "detail_url": "https://openreview.net/forum?id=3K3s9qxSn7", "authors": "Hanlin Zhu,Baihe Huang,Stuart Russell", "tags": "ICLR 2024,Poster", "abstract": "We study the representation complexity of model-based and model-free reinforcement learning (RL) in the context of circuit complexity. We prove theoretically that there exists a broad class of MDPs such that their underlying transition and reward functions can be represented by constant depth circuits with polynomial size, while the optimal $Q$-function suffers an exponential circuit complexity in constant-depth circuits. By drawing attention to the approximation errors and building connections to complexity theory, our theory provides unique insights into why model-based algorithms usually enjoy better sample complexity than model-free algorithms from a novel representation complexity perspective: in some cases, the ground-truth rule (model) of the environment is simple to represent, while other quantities, such as $Q$-function, appear complex. We empirically corroborate our theory by comparing the approximation error of the transition kernel, reward function, and optimal $Q$-function in various Mujoco environments, which demonstrates that the approximation errors of the transition kernel and reward function are consistently lower than those of the optimal $Q$-function. To the best of our knowledge, this work is the first to study the circuit complexity of RL, which also provides a rigorous framework for future research.", "pdf": "https://openreview.net/pdf/94c3b9655119f9778dfab41fe3cb7661151ae6a3.pdf"} {"title": "Addressing Loss of Plasticity and Catastrophic Forgetting in Continual Learning", "url": "https://openreview.net/forum?id=sKPzAXoylB", "detail_url": "https://openreview.net/forum?id=sKPzAXoylB", "authors": "Mohamed Elsayed,A. Rupam Mahmood", "tags": "ICLR 2024,Poster", "abstract": "Deep representation learning methods struggle with continual learning, suffering from both catastrophic forgetting of useful units and loss of plasticity, often due to rigid and unuseful units. While many methods address these two issues separately, only a few currently deal with both simultaneously. In this paper, we introduce Utility-based Perturbed Gradient Descent (UPGD) as a novel approach for the continual learning of representations. UPGD combines gradient updates with perturbations, where it applies smaller modifications to more useful units, protecting them from forgetting, and larger modifications to less useful units, rejuvenating their plasticity. We use a challenging streaming learning setup where continual learning problems have hundreds of non-stationarities and unknown task boundaries. We show that many existing methods suffer from at least one of the issues, predominantly manifested by their decreasing accuracy over tasks. On the other hand, UPGD continues to improve performance and surpasses or is competitive with all methods in all problems. Finally, in extended reinforcement learning experiments with PPO, we show that while Adam exhibits a performance drop after initial learning, UPGD avoids it by addressing both continual learning issues.", "pdf": "https://openreview.net/pdf/7eb81f1c1c4fea7fa434ebe26bbf3145d56b032f.pdf"} {"title": "A Good Learner can Teach Better: Teacher-Student Collaborative Knowledge Distillation", "url": "https://openreview.net/forum?id=Ixi4j6LtdX", "detail_url": "https://openreview.net/forum?id=Ixi4j6LtdX", "authors": "Ayan Sengupta,Shantanu Dixit,Md Shad Akhtar,Tanmoy Chakraborty", "tags": "ICLR 2024,Poster", "abstract": "Knowledge distillation (KD) is a technique used to transfer knowledge from a larger ''teacher'' model into a smaller ''student'' model. Recent advancements in meta-learning-based knowledge distillation (MetaKD) emphasize that the fine-tuning of teacher models should be aware of the student's need to achieve better knowledge distillation. However, existing MetaKD methods often lack incentives for the teacher model to improve itself. In this study, we introduce MPDistil, a meta-policy distillation technique, that utilizes novel optimization strategies to foster both *collaboration* and *competition* during the fine-tuning of the teacher model in the meta-learning step. Additionally, we propose a curriculum learning framework for the student model in a competitive setup, in which the student model aims to outperform the teacher model by self-training on various tasks. Exhaustive experiments on SuperGLUE and GLUE benchmarks demonstrate the efficacy of MPDistil compared to $20$ conventional KD and advanced MetaKD baselines, showing significant performance enhancements in the student model -- e.g., a distilled 6-layer BERT model outperforms a 12-layer BERT model on five out of six SuperGLUE tasks. Furthermore, MPDistil, while applied to a large language teacher model (DeBERTa-v2-xxlarge), significantly narrows the performance gap of its smaller student counterpart (DeBERTa-12) by just $4.6$% on SuperGLUE. We further demonstrate how higher rewards and customized training curricula strengthen the student model and enhance generalizability.", "pdf": "https://openreview.net/pdf/79548908914d91f19b250084cc53384846d7ddbb.pdf"} {"title": "Diagnosing Transformers: Illuminating Feature Spaces for Clinical Decision-Making", "url": "https://openreview.net/forum?id=k581sTMyPt", "detail_url": "https://openreview.net/forum?id=k581sTMyPt", "authors": "Aliyah R. Hsu,Yeshwanth Cherapanamjeri,Briton Park,Tristan Naumann,Anobel Odisho,Bin Yu", "tags": "ICLR 2024,Poster", "abstract": "Pre-trained transformers are often fine-tuned to aid clinical decision-making using limited clinical notes. Model interpretability is crucial, especially in high-stakes domains like medicine, to establish trust and ensure safety, which requires human engagement. We introduce SUFO, a systematic framework that enhances interpretability of fine-tuned transformer feature spaces. SUFO utilizes a range of analytic and visualization techniques, including Supervised probing, Unsupervised similarity analysis, Feature dynamics, and Outlier analysis to address key questions about model trust and interpretability (e.g. model suitability for a task, feature space evolution during fine-tuning, and interpretation of fine-tuned features and failure modes). We conduct a case study investigating the impact of pre-training data where we focus on real-world pathology classification tasks, and validate our findings on MedNLI. We evaluate five 110M-sized pre-trained transformer models, categorized into general-domain (BERT, TNLR), mixed-domain (BioBERT, Clinical BioBERT), and domain-specific (PubMedBERT) groups. Our SUFO analyses reveal that: (1) while PubMedBERT, the domain-specific model, contains valuable information for fine-tuning, it can overfit to minority classes when class imbalances exist. In contrast, mixed-domain models exhibit greater resistance to overfitting, suggesting potential improvements in domain-specific model robustness; (2) in-domain pre-training accelerates feature disambiguation during fine-tuning; and (3) feature spaces undergo significant sparsification during this process, enabling clinicians to identify common outlier modes among fine-tuned models as demonstrated in this paper. These findings showcase the utility of SUFO in enhancing trust and safety when using transformers in medicine, and we believe SUFO can aid practitioners in evaluating fine-tuned language models (LMs) for other applications in medicine and in more critical domains.", "pdf": "https://openreview.net/pdf/4396dc3858433d144c7809bad60e3be5f5a5ebae.pdf"} {"title": "Mechanistically analyzing the effects of fine-tuning on procedurally defined tasks", "url": "https://openreview.net/forum?id=A0HKeKl4Nl", "detail_url": "https://openreview.net/forum?id=A0HKeKl4Nl", "authors": "Samyak Jain,Robert Kirk,Ekdeep Singh Lubana,Robert P. Dick,Hidenori Tanaka,Tim Rockt\u00e4schel,Edward Grefenstette,David Krueger", "tags": "ICLR 2024,Poster", "abstract": "Fine-tuning large pre-trained models has become the de facto strategy for developing both task-specific and general-purpose machine learning systems, including developing models that are safe to deploy. Despite its clear importance, there has been minimal work that explains how fine-tuning alters the underlying capabilities learned by a model during pretraining: does fine-tuning yield entirely novel capabilities or does it just modulate existing ones? We address this question empirically in synthetic, controlled settings where we can use mechanistic interpretability tools (e.g., network pruning and probing) to understand how the model's underlying capabilities are changing. We perform an extensive analysis of the effects of fine-tuning in these settings, and show that: (i) fine-tuning rarely alters the underlying model capabilities; (ii) a minimal transformation, which we call a `wrapper', is typically learned on top of the underlying model capabilities, creating the illusion that they have been modified; and (iii) further fine-tuning on a task where such ``wrapped capabilities'' are relevant leads to sample-efficient revival of the capability, i.e., the model begins reusing these capabilities after only a few gradient steps. This indicates that practitioners can unintentionally remove a model's safety wrapper merely by fine-tuning it on a, e.g., superficially unrelated, downstream task. We additionally perform analysis on language models trained on the TinyStories dataset to support our claims in a more realistic setup.", "pdf": "https://openreview.net/pdf/b46dce52717c9ba5bb6dc047b34dd04064101c1d.pdf"} {"title": "RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems", "url": "https://openreview.net/forum?id=pPjZIOuQuF", "detail_url": "https://openreview.net/forum?id=pPjZIOuQuF", "authors": "Tianyang Liu,Canwen Xu,Julian McAuley", "tags": "ICLR 2024,Poster", "abstract": "Large Language Models (LLMs) have greatly advanced code auto-completion systems, with a potential for substantial productivity enhancements for developers. However, current benchmarks mainly focus on single-file tasks, leaving an assessment gap for more complex, real-world, multi-file programming scenarios. To fill this gap, we introduce RepoBench, a new benchmark specifically designed for evaluating repository-level code auto-completion systems. RepoBench consists of three interconnected evaluation tasks: RepoBench-R (Retrieval), RepoBench-C (Code Completion), and RepoBench-P (Pipeline). Each task respectively measures the system's ability to retrieve the most relevant code snippets from other files as cross-file context, predict the next line of code with cross-file and in-file context, and handle complex tasks that require a combination of both retrieval and next-line prediction. RepoBench aims to facilitate a more complete comparison of performance and encouraging continuous improvement in auto-completion systems. RepoBench is actively maintained with the latest code, serving as a live benchmark publicly available at https://github.com/Leolty/repobench.", "pdf": "https://openreview.net/pdf/904b94b516d638671ae5c0877f71de8c576853cb.pdf"} {"title": "Seeking Neural Nuggets: Knowledge Transfer in Large Language Models from a Parametric Perspective", "url": "https://openreview.net/forum?id=mIEHIcHGOo", "detail_url": "https://openreview.net/forum?id=mIEHIcHGOo", "authors": "Ming Zhong,Chenxin An,Weizhu Chen,Jiawei Han,Pengcheng He", "tags": "ICLR 2024,Poster", "abstract": "Large Language Models (LLMs) inherently encode a wealth of knowledge within their parameters through pre-training on extensive corpora. While prior research has delved into operations on these parameters to manipulate the underlying implicit knowledge \u2014 encompassing detection, editing, and merging \u2014 there remains an ambiguous understanding regarding their transferability across models with varying scales. In this paper, we seek to empirically investigate knowledge transfer from larger to smaller models through a parametric perspective. To achieve this, we employ sensitivity-based techniques to extract and align knowledge-specific parameters between different LLMs. Moreover, the LoRA module is used as the intermediary mechanism for injecting the extracted knowledge into smaller models. Evaluations across four benchmarks validate the efficacy of our proposed method. Our findings highlight the critical factors contributing to the process of parametric knowledge transfer, underscoring the transferability of model parameters across LLMs of different scales. Project website: https://maszhongming.github.io/ParaKnowTransfer.", "pdf": "https://openreview.net/pdf/403d80fb4f79d7b3e0026bbe47d1dbef35d9b3a4.pdf"} {"title": "Harnessing Explanations: LLM-to-LM Interpreter for Enhanced Text-Attributed Graph Representation Learning", "url": "https://openreview.net/forum?id=RXFVcynVe1", "detail_url": "https://openreview.net/forum?id=RXFVcynVe1", "authors": "Xiaoxin He,Xavier Bresson,Thomas Laurent,Adam Perold,Yann LeCun,Bryan Hooi", "tags": "ICLR 2024,Poster", "abstract": "Representation learning on text-attributed graphs (TAGs) has become a critical research problem in recent years. A typical example of a TAG is a paper citation graph, where the text of each paper serves as node attributes. Initial graph neural network (GNN) pipelines handled these text attributes by transforming them into shallow or hand-crafted features, such as skip-gram or bag-of-words features. Recent efforts have focused on enhancing these pipelines with language models (LMs), which typically demand intricate designs and substantial computational resources. With the advent of powerful large language models (LLMs) such as GPT or Llama2, which demonstrate an ability to reason and to utilize general knowledge, there is a growing need for techniques which combine the textual modelling abilities of LLMs with the structural learning capabilities of GNNs. Hence, in this work, we focus on leveraging LLMs to capture textual information as features, which can be used to boost GNN performance on downstream tasks. A key innovation is our use of \\emph{explanations as features}: we prompt an LLM to perform zero-shot classification, request textual explanations for its decision-making process, and design an \\emph{LLM-to-LM interpreter} to translate these explanations into informative features for downstream GNNs. Our experiments demonstrate that our method achieves state-of-the-art results on well-established TAG datasets, including \\texttt{Cora}, \\texttt{PubMed}, \\texttt{ogbn-arxiv}, as well as our newly introduced dataset, \\texttt{tape-arxiv23}. Furthermore, our method significantly speeds up training, achieving a 2.88 times improvement over the closest baseline on \\texttt{ogbn-arxiv}. Lastly, we believe the versatility of the proposed method extends beyond TAGs and holds the potential to enhance other tasks involving graph-text data~\\footnote{Our codes and datasets are available at: \\url{https://github.com/XiaoxinHe/TAPE}}.", "pdf": "https://openreview.net/pdf/0db04867c257dc081f5a8f03268da344deb07417.pdf"} {"title": "SuRe: Summarizing Retrievals using Answer Candidates for Open-domain QA of LLMs", "url": "https://openreview.net/forum?id=w4DW6qkRmt", "detail_url": "https://openreview.net/forum?id=w4DW6qkRmt", "authors": "Jaehyung Kim,Jaehyun Nam,Sangwoo Mo,Jongjin Park,Sang-Woo Lee,Minjoon Seo,Jung-Woo Ha,Jinwoo Shin", "tags": "ICLR 2024,Poster", "abstract": "Large language models (LLMs) have made significant advancements in various natural language processing tasks, including question answering (QA) tasks. While incorporating new information with the retrieval of relevant passages is a promising way to improve QA with LLMs, the existing methods often require additional fine-tuning which becomes infeasible with recent LLMs. Augmenting retrieved passages via prompting has the potential to address this limitation, but this direction has been limitedly explored. To this end, we design a simple yet effective framework to enhance open-domain QA (ODQA) with LLMs, based on the summarized retrieval (SuRe). SuRe helps LLMs predict more accurate answers for a given question, which are well-supported by the summarized retrieval that could be viewed as an explicit rationale extracted from the retrieved passages. Specifically, SuRe first constructs summaries of the retrieved passages for each of the multiple answer candidates. Then, SuRe confirms the most plausible answer from the candidate set by evaluating the validity and ranking of the generated summaries. Experimental results on diverse ODQA benchmarks demonstrate the superiority of SuRe, with improvements of up to 4.6\\% in exact match (EM) and 4.0\\% in F1 score over standard prompting approaches. SuRe also can be integrated with a broad range of retrieval methods and LLMs. Finally, the generated summaries from SuRe show additional advantages to measure the importance of retrieved passages and serve as more preferred rationales by models and humans.", "pdf": "https://openreview.net/pdf/9192639fe8e3dbb64d5431c85984894b9e1b089d.pdf"} {"title": "Retrieval meets Long Context Large Language Models", "url": "https://openreview.net/forum?id=xw5nxFWMlo", "detail_url": "https://openreview.net/forum?id=xw5nxFWMlo", "authors": "Peng Xu,Wei Ping,Xianchao Wu,Lawrence McAfee,Chen Zhu,Zihan Liu,Sandeep Subramanian,Evelina Bakhturina,Mohammad Shoeybi,Bryan Catanzaro", "tags": "ICLR 2024,Poster", "abstract": "Extending the context window of large language models (LLMs) is getting popular recently, while the solution of augmenting LLMs with retrieval has existed for years. The natural questions are: i) Retrieval-augmentation versus long context window, which one is better for downstream tasks? ii) Can both methods be combined to get the best of both worlds? In this work, we answer these questions by studying both solutions using two state-of-the-art pretrained LLMs, i.e., a proprietary 43B GPT and Llama2-70B. Perhaps surprisingly, we find that LLM with 4K context window using simple retrieval-augmentation at generation can achieve comparable performance to finetuned LLM with 16K context window via positional interpolation on long context tasks, while taking much less computation. More importantly, we demonstrate that retrieval can significantly improve the performance of LLMs regardless of their extended context window sizes. Our best model, retrieval-augmented Llama2-70B with 32K context window, outperforms GPT-3.5-turbo-16k and Davinci003 in terms of average score on nine long context tasks including question answering, query-based summarization, and in-context few-shot learning tasks. It also outperforms its non-retrieval Llama2-70B-32k baseline by a margin, while being much faster at generation. Our study provides general insights on the choice of retrieval-augmentation versus long context extension of LLM for practitioners.", "pdf": "https://openreview.net/pdf/4cd5150fe82d2f05e1ac91ccde87cca2e5f6d8e2.pdf"} {"title": "Neural Spectral Methods: Self-supervised learning in the spectral domain", "url": "https://openreview.net/forum?id=2DbVeuoa6a", "detail_url": "https://openreview.net/forum?id=2DbVeuoa6a", "authors": "Yiheng Du,Nithin Chalapathi,Aditi S. Krishnapriyan", "tags": "ICLR 2024,Poster", "abstract": "We present Neural Spectral Methods, a technique to solve parametric Partial Differential Equations (PDEs), grounded in classical spectral methods. Our method uses orthogonal bases to learn PDE solutions as mappings between spectral coefficients, instantiating a spectral-based neural operator. In contrast to current machine learning approaches which enforce PDE constraints by minimizing the numerical quadrature of the residuals in the spatiotemporal domain, we leverage Parseval's identity and introduce a new training strategy through a spectral loss. Our spectral loss enables more efficient differentiation through the neural network, and substantially reduces training complexity. At inference time, the computational cost of our method remains constant, regardless of the spatiotemporal resolution of the domain. Our experimental results demonstrate that our method significantly outperforms previous machine learning approaches in terms of speed and accuracy by one to two orders of magnitude on multiple different problems, including reaction-diffusion, and forced and unforced Navier-Stokes equations. When compared to numerical solvers of the same accuracy, our method demonstrates a $10\\times$ increase in performance speed. Our source code is publicly available at https://github.com/ASK-Berkeley/Neural-Spectral-Methods.", "pdf": "https://openreview.net/pdf/9aa74e5bf7d501d1a636aee71ec751a621b15eee.pdf"} {"title": "Kosmos-G: Generating Images in Context with Multimodal Large Language Models", "url": "https://openreview.net/forum?id=he6mX9LTyE", "detail_url": "https://openreview.net/forum?id=he6mX9LTyE", "authors": "Xichen Pan,Li Dong,Shaohan Huang,Zhiliang Peng,Wenhu Chen,Furu Wei", "tags": "ICLR 2024,Poster", "abstract": "Recent advancements in subject-driven image generation have made significant strides. However, current methods still fall short in diverse application scenarios, as they require test-time tuning and cannot accept interleaved multi-image and text input. These limitations keep them far from the ultimate goal of \"image as a foreign language in image generation.\" This paper presents Kosmos-G, a model that leverages the advanced multimodal perception capabilities of Multimodal Large Language Models (MLLMs) to tackle the aforementioned challenge. Our approach aligns the output space of MLLM with CLIP using the textual modality as an anchor and performs compositional instruction tuning on curated data. Kosmos-G demonstrates an impressive capability of zero-shot subject-driven generation with interleaved multi-image and text input. Notably, the score distillation instruction tuning requires no modifications to the image decoder. This allows for a seamless substitution of CLIP and effortless integration with a myriad of U-Net techniques ranging from fine-grained controls to personalized image decoder variants. We posit Kosmos-G as an initial attempt towards the goal of \"image as a foreign language in image generation.\"", "pdf": "https://openreview.net/pdf/8ffe0cc3c3fb4e1f945894b836d9a97b0dbaf9b5.pdf"} {"title": "Assessing Uncertainty in Similarity Scoring: Performance & Fairness in Face Recognition", "url": "https://openreview.net/forum?id=lAhQCHuANV", "detail_url": "https://openreview.net/forum?id=lAhQCHuANV", "authors": "Jean-R\u00e9my Conti,Stephan Cl\u00e9men\u00e7on", "tags": "ICLR 2024,Poster", "abstract": "The ROC curve is the major tool for assessing not only the performance but also the fairness properties of a similarity scoring function. In order to draw reliable conclusions based on empirical ROC analysis, accurately evaluating the uncertainty level related to statistical versions of the ROC curves of interest is absolutely necessary, especially for applications with considerable societal impact such as Face Recognition. In this article, we prove asymptotic guarantees for empirical ROC curves of similarity functions as well as for by-product metrics useful to assess fairness. We also explain that, because the false acceptance/rejection rates are of the form of U-statistics in the case of similarity scoring, the naive bootstrap approach may jeopardize the assessment procedure. A dedicated recentering technique must be used instead. Beyond the theoretical analysis carried out, various experiments using real face image datasets provide strong empirical evidence of the practical relevance of the methods promoted here, when applied to several ROC-based measures such as popular fairness metrics.", "pdf": "https://openreview.net/pdf/2aa88e74521ae92488dd1b2c23c8c9f5996dc778.pdf"} {"title": "LitCab: Lightweight Language Model Calibration over Short- and Long-form Responses", "url": "https://openreview.net/forum?id=jH67LHVOIO", "detail_url": "https://openreview.net/forum?id=jH67LHVOIO", "authors": "Xin Liu,Muhammad Khalifa,Lu Wang", "tags": "ICLR 2024,Poster", "abstract": "A model is considered well-calibrated when its probability estimate aligns with the actual likelihood of the output being correct. Calibrating language models (LMs) is crucial, as it plays a vital role in detecting and mitigating hallucinations of LMs as well as building more trustworthy models. However, standard calibration techniques may not be suited for LM calibration. For instance, post-processing methods such as temperature scaling do not reorder the candidate generations. On the other hand, training-based methods require fine-tuning the entire model, which is impractical for LMs of large scale. We present LitCab, a lightweight calibration mechanism consisting of a single linear layer that takes the input text representation and predicts a bias term, which is then added to the LM output logits. LitCab improves model calibration by only adding < 2% of the original model parameters. For evaluation, we construct CaT, a benchmark consisting of eight text generation tasks, covering responses ranging from short phrases to paragraphs. We test LitCab with Llama2-7B, where it improves calibration across all tasks, reducing the average ECE score by as large as 30%. We further conduct a comprehensive evaluation with multiple popular open-sourced LMs from GPT and LLaMA families, yielding the following key findings: (i) Larger models within the same family exhibit better calibration on tasks with short generation tasks, but not necessarily for longer ones. (ii) GPT-family models show superior calibration compared to LLaMA, Llama2, and Vicuna models, despite having much fewer parameters. (iii) Fine-tuning pretrained model (e.g., LLaMA) with samples of limited purpose (e.g., conversations) may lead to worse calibration, highlighting the importance of fine-tuning setups for calibrating LMs.", "pdf": "https://openreview.net/pdf/060aede7175d70a3fe37974ccb7cb976fcfa6486.pdf"} {"title": "Chain-of-Knowledge: Grounding Large Language Models via Dynamic Knowledge Adapting over Heterogeneous Sources", "url": "https://openreview.net/forum?id=cPgh4gWZlz", "detail_url": "https://openreview.net/forum?id=cPgh4gWZlz", "authors": "Xingxuan Li,Ruochen Zhao,Yew Ken Chia,Bosheng Ding,Shafiq Joty,Soujanya Poria,Lidong Bing", "tags": "ICLR 2024,Poster", "abstract": "We present chain-of-knowledge (CoK), a novel framework that augments large language models (LLMs) by dynamically incorporating grounding information from heterogeneous sources. It results in more factual rationales and reduced hallucination in generation. \nSpecifically, CoK consists of three stages: reasoning preparation, dynamic knowledge adapting, and answer consolidation. \nGiven a knowledge-intensive question, CoK first prepares several preliminary rationales and answers while identifying the relevant knowledge domains.\nIf there is no majority consensus among the answers from samples, CoK corrects the rationales step by step by adapting knowledge from the identified domains.\nThese corrected rationales can plausibly serve as a better foundation for the final answer consolidation.\nUnlike prior studies that primarily use unstructured data, CoK also leverages structured knowledge sources such as Wikidata and tables that provide more reliable factual information.\nTo access both unstructured and structured knowledge sources in the dynamic knowledge adapting stage, we propose an adaptive query generator that allows the generation of queries for various types of query languages, including SPARQL, SQL, and natural sentences. Moreover, to minimize error propagation between rationales, CoK corrects the rationales progressively using preceding corrected rationales to generate and correct subsequent rationales.\nExtensive experiments show that CoK consistently improves the performance of LLMs on knowledge-intensive tasks across different domains.", "pdf": "https://openreview.net/pdf/99bf7907ce66cffb067fbb21e933967471cbfdb7.pdf"} {"title": "Policy Rehearsing: Training Generalizable Policies for Reinforcement Learning", "url": "https://openreview.net/forum?id=m3xVPaZp6Z", "detail_url": "https://openreview.net/forum?id=m3xVPaZp6Z", "authors": "Chengxing Jia,Chenxiao Gao,Hao Yin,Fuxiang Zhang,Xiong-Hui Chen,Tian Xu,Lei Yuan,Zongzhang Zhang,Zhi-Hua Zhou,Yang Yu", "tags": "ICLR 2024,Poster", "abstract": "Human beings can make adaptive decisions in a preparatory manner, i.e., by making preparations in advance, which offers significant advantages in scenarios where both online and offline experiences are expensive and limited. Meanwhile, current reinforcement learning methods commonly rely on numerous environment interactions but hardly obtain generalizable policies. In this paper, we introduce the idea of \\textit{rehearsal} into policy optimization, where the agent plans for all possible outcomes in mind and acts adaptively according to actual responses from the environment. To effectively rehearse, we propose ReDM, an algorithm that generates a diverse and eligible set of dynamics models and then rehearse the policy via adaptive training on the generated model set. Rehearsal enables the policy to make decision plans for various hypothetical dynamics and to naturally generalize to previously unseen environments. Our experimental results demonstrate that ReDM is capable of learning a valid policy solely through rehearsal, even with \\emph{zero} interaction data. We further extend ReDM to scenarios where limited or mismatched interaction data is available, and our experimental results reveal that ReDM produces high-performing policies compared to other offline RL baselines.", "pdf": "https://openreview.net/pdf/b983c88da15dde7d91c48aee3b97aa22087d7cc0.pdf"} {"title": "Energy-based Automated Model Evaluation", "url": "https://openreview.net/forum?id=CHGcP6lVWd", "detail_url": "https://openreview.net/forum?id=CHGcP6lVWd", "authors": "Ru Peng,Heming Zou,Haobo Wang,Yawen Zeng,Zenan Huang,Junbo Zhao", "tags": "ICLR 2024,Poster", "abstract": "The conventional evaluation protocols on machine learning models rely heavily on a labeled, i.i.d-assumed testing dataset, which is not often present in real-world applications.\nThe Automated Model Evaluation (AutoEval) shows an alternative to this traditional workflow, by forming a proximal prediction pipeline of the testing performance without the presence of ground-truth labels.\nDespite its recent successes, the AutoEval frameworks still suffer from an overconfidence issue, substantial storage and computational cost.\nIn that regard, we propose a novel measure --- Meta-Distribution Energy (MDE) that allows the AutoEval framework to be both more efficient and effective.\nThe core of the MDE is to establish a meta-distribution statistic, on the information (energy) associated with individual samples, then offer a smoother representation enabled by energy-based learning.\nWe further provide our theoretical insights by connecting the MDE with the classification loss.\nWe provide extensive experiments across modalities, datasets and different architectural backbones to validate MDE's validity, together with its superiority compared with prior approaches.\nWe also prove MDE's versatility by showing its seamless integration with large-scale models, and easy adaption to learning scenarios with noisy- or imbalanced- labels.", "pdf": "https://openreview.net/pdf/b67cd952981636bb89569bc035666d70b30d02bc.pdf"} {"title": "Deceptive Fairness Attacks on Graphs via Meta Learning", "url": "https://openreview.net/forum?id=iS5ADHNg2A", "detail_url": "https://openreview.net/forum?id=iS5ADHNg2A", "authors": "Jian Kang,Yinglong Xia,Ross Maciejewski,Jiebo Luo,Hanghang Tong", "tags": "ICLR 2024,Poster", "abstract": "We study deceptive fairness attacks on graphs to answer the following question: How can we achieve poisoning attacks on a graph learning model to exacerbate the bias deceptively? We answer this question via a bi-level optimization problem and propose a meta learning-based framework named FATE. FATE is broadly applicable with respect to various fairness definitions and graph learning models, as well as arbitrary choices of manipulation operations. We further instantiate FATE to attack statistical parity or individual fairness on graph neural networks. We conduct extensive experimental evaluations on real-world datasets in the task of semi-supervised node classification. The experimental results demonstrate that FATE could amplify the bias of graph neural networks with or without fairness consideration while maintaining the utility on the downstream task. We hope this paper provides insights into the adversarial robustness of fair graph learning and can shed light on designing robust and fair graph learning in future studies.", "pdf": "https://openreview.net/pdf/ec8c9c4fa87c072aa9815fe7601e5b37a8b938b2.pdf"} {"title": "What Matters to You? Towards Visual Representation Alignment for Robot Learning", "url": "https://openreview.net/forum?id=CTlUHIKF71", "detail_url": "https://openreview.net/forum?id=CTlUHIKF71", "authors": "Thomas Tian,Chenfeng Xu,Masayoshi Tomizuka,Jitendra Malik,Andrea Bajcsy", "tags": "ICLR 2024,Poster", "abstract": "When operating in service of people, robots need to optimize rewards aligned with end-user preferences. Since robots will rely on raw perceptual inputs, their rewards will inevitably use visual representations. Recently there has been excitement in using representations from pre-trained visual models, but key to making these work in robotics is fine-tuning, which is typically done via proxy tasks like dynamics prediction or enforcing temporal cycle-consistency. However, all these proxy tasks bypass the human\u2019s input on what matters to them, exacerbating spurious correlations and ultimately leading to behaviors that are misaligned with user preferences. In this work, we propose that robots should leverage human feedback to align their visual representations with the end-user and disentangle what matters for the task. We propose Representation-Aligned Preference-based Learning (RAPL), a method for solving the visual representation alignment problem and visual reward learning problem through the lens of preference-based learning and optimal transport. Across experiments in X MAGICAL and in robotic manipulation, we find that RAPL\u2019s reward consistently generates preferred robot behaviors with high sample efficiency, and shows strong zero-shot generalization when the visual representation is learned from a different embodiment than the robot\u2019s.", "pdf": "https://openreview.net/pdf/b089a2d0ee33f551f8ee252854a9a7630830fe59.pdf"} {"title": "FedDA: Faster Adaptive Gradient Methods for Federated Constrained Optimization", "url": "https://openreview.net/forum?id=kjn99xFUF3", "detail_url": "https://openreview.net/forum?id=kjn99xFUF3", "authors": "Junyi Li,Feihu Huang,Heng Huang", "tags": "ICLR 2024,Poster", "abstract": "Federated learning (FL) is an emerging learning paradigm where a set of distributed clients learns a task under the coordination of a server. The FedAvg algorithm is one of the most widely used methods in FL. In FedAvg, the learning rate is a constant rather than changing adaptively. Adaptive gradient methods have demonstrated superior performance over the constant learning rate schedules in non-distributed settings, and they have recently been adapted to FL. However, the majority of these methods are designed for unconstrained settings. Meanwhile, many crucial FL applications, like disease diagnosis and biomarker identification, often rely on constrained formulations such as Lasso and group Lasso. It remains an open question as to whether adaptive gradient methods can be effectively applied to FL problems with constrains. In this work, we introduce \\textbf{FedDA}, a novel adaptive gradient framework for FL. This framework utilizes a restarted dual averaging technique and is compatible with a range of gradient estimation methods and adaptive learning rate schedules. Specifically, an instantiation of our framework FedDA-MVR achieves sample complexity $\\tilde{O}(K^{-1}\\epsilon^{-1.5})$ and communication complexity $\\tilde{O}(K^{-0.25}\\epsilon^{-1.25})$ for finding a stationary point $\\epsilon$ in the constrained setting with $K$ be the number of clients. We conduct experiments over both constrained and unconstrained tasks to confirm the effectiveness of our approach.", "pdf": "https://openreview.net/pdf/db0f03c6b9016274933d2407dddcaef05789c2f2.pdf"} {"title": "Extending Power of Nature from Binary to Real-Valued Graph Learning in Real World", "url": "https://openreview.net/forum?id=qT7DXUmX7j", "detail_url": "https://openreview.net/forum?id=qT7DXUmX7j", "authors": "Chunshu Wu,Ruibing Song,Chuan Liu,Yunan Yang,Ang Li,Michael Huang,Tong Geng", "tags": "ICLR 2024,Poster", "abstract": "Nature performs complex computations constantly at clearly lower cost and higher performance than digital computers. It is crucial to understand how to harness the unique computational power of nature in Machine Learning (ML). In the past decade, besides the development of Neural Networks (NNs), the community has also relentlessly explored nature-powered ML paradigms. Although most of them are still predominantly theoretical, a new practical paradigm enabled by the recent advent of CMOS-compatible room-temperature nature-based computers has emerged. By harnessing a dynamical system's intrinsic behavior of chasing the lowest energy state, this paradigm can solve some simple binary problems delivering considerable speedup and energy savings compared with NNs, while maintaining comparable accuracy. Regrettably, its values to the real world are highly constrained by its binary nature. A clear pathway to its extension to real-valued problems remains elusive. This paper aims to unleash this pathway by proposing a novel end-to-end Nature-Powered Graph Learning (NP-GL) framework. Specifically, through a three-dimensional co-design, NP-GL can leverage the spontaneous energy decrease in nature to efficiently solve real-valued graph learning problems. Experimental results across 4 real-world applications with 6 datasets demonstrate that NP-GL delivers, on average, $6.97\\times 10^3$ speedup and $10^5$ energy consumption reduction with comparable or even higher accuracy than Graph Neural Networks (GNNs).", "pdf": "https://openreview.net/pdf/c3a5eccec09e9fea31f8e2a25c42986e31463191.pdf"} {"title": "Meta-VBO: Utilizing Prior Tasks in Optimizing Risk Measures with Gaussian Processes", "url": "https://openreview.net/forum?id=ElykcDu5YK", "detail_url": "https://openreview.net/forum?id=ElykcDu5YK", "authors": "Quoc Phong Nguyen,Bryan Kian Hsiang Low,Patrick Jaillet", "tags": "ICLR 2024,Poster", "abstract": "Research on optimizing the risk measure of a blackbox function using Gaussian processes, especially Bayesian optimization (BO) of risk measures, has become increasingly important due to the inevitable presence of uncontrollable variables in real-world applications. Nevertheless, existing works on BO of risk measures start the optimization from scratch for every new task without considering the results of prior tasks. In contrast, its vanilla BO counterpart has received a thorough investigation on utilizing prior tasks to speed up the current task through the body of works on meta-BO which, however, have not considered risk measures. To bridge this gap, this paper presents the first algorithm for meta-BO of risk measures (i.e., value-at-risk (VaR) and the conditional VaR), namely meta-VBO, by introducing a novel adjustment to the upper confidence bound acquisition function. Our proposed algorithm exhibits two desirable properties: (i) invariance to scaling and vertical shifting of the blackbox function and (ii) robustness to prior harmful tasks. We provide a theoretical performance guarantee for our algorithm and empirically demonstrate its performance using several synthetic function benchmarks and real-world objective functions.", "pdf": "https://openreview.net/pdf/438a0598bfd4e1457d158d87209039d77cfd4c53.pdf"} {"title": "Data Debugging with Shapley Importance over Machine Learning Pipelines", "url": "https://openreview.net/forum?id=qxGXjWxabq", "detail_url": "https://openreview.net/forum?id=qxGXjWxabq", "authors": "Bojan Karla\u0161,David Dao,Matteo Interlandi,Sebastian Schelter,Wentao Wu,Ce Zhang", "tags": "ICLR 2024,Poster", "abstract": "When a machine learning (ML) model exhibits poor quality (e.g., poor accuracy or fairness), the problem can often be traced back to errors in the training data. Being able to discover the data examples that are the most likely culprits is a fundamental concern that has received a lot of attention recently. One prominent way to measure \"data importance\" with respect to model quality is the Shapley value. Unfortunately, existing methods only focus on the ML model in isolation, without considering the broader ML pipeline for data preparation and feature extraction, which appears in the majority of real-world ML code. This presents a major limitation to applying existing methods in practical settings. In this paper, we propose Datascope, a method for efficiently computing Shapley-based data importance over ML pipelines. We introduce several approximations that lead to dramatic improvements in terms of computational speed. Finally, our experimental evaluation demonstrates that our methods are capable of data error discovery that is as effective as existing Monte Carlo baselines, and in some cases even outperform them. We release our code as an open-source data debugging library available at https://github.com/easeml/datascope.", "pdf": "https://openreview.net/pdf/102cae20a16e15e1c544b446d7ec05ad7b8f036a.pdf"} {"title": "Pessimistic Nonlinear Least-Squares Value Iteration for Offline Reinforcement Learning", "url": "https://openreview.net/forum?id=4kLVvIh8cp", "detail_url": "https://openreview.net/forum?id=4kLVvIh8cp", "authors": "Qiwei Di,Heyang Zhao,Jiafan He,Quanquan Gu", "tags": "ICLR 2024,Poster", "abstract": "Offline reinforcement learning (RL), where the agent aims to learn the optimal policy based on the data collected by a behavior policy, has attracted increasing attention in recent years. While offline RL with linear function approximation has been extensively studied with optimal results achieved under certain assumptions, many works shift their interest to offline RL with non-linear function approximation.\nHowever, limited works on offline RL with non-linear function approximation have instance-dependent regret guarantees.\n In this paper, we propose an oracle-efficient algorithm, dubbed Pessimistic Nonlinear Least-Square Value Iteration (PNLSVI), for offline RL with non-linear function approximation. Our algorithmic design comprises three innovative components: (1) a variance-based weighted regression scheme that can be applied to a wide range of function classes, (2) a subroutine for variance estimation, and (3) a planning phase that utilizes a pessimistic value iteration approach. Our algorithm enjoys a regret bound that has a tight dependency on the function class complexity and achieves minimax optimal instance-dependent regret when specialized to linear function approximation. Our work extends the previous instance-dependent results within simpler function classes, such as linear and differentiable function to a more general framework. To the best of our knowledge, this is the first statistically optimal algorithm for nonlinear offline RL.", "pdf": "https://openreview.net/pdf/eccb6a13ad39f4a83f32bd2e74f2359e20086e81.pdf"} {"title": "SKILL-MIX: a Flexible and Expandable Family of Evaluations for AI Models", "url": "https://openreview.net/forum?id=Jf5gplvglq", "detail_url": "https://openreview.net/forum?id=Jf5gplvglq", "authors": "Dingli Yu,Simran Kaur,Arushi Gupta,Jonah Brown-Cohen,Anirudh Goyal,Sanjeev Arora", "tags": "ICLR 2024,Poster", "abstract": "With LLMs shifting their role from statistical modeling of language to serving as general-purpose AI agents, how should LLM evaluations change? Arguably, a key ability of an AI agent is to flexibly combine, as needed, the basic skills it has learned. The capability to combine skills plays an important role in (human) pedagogy and also in a paper on emergence phenomena (Arora & Goyal, 2023).\n\nThis work introduces SKILL-MIX, a new evaluation to measure ability to combine skills. Using a list of $N$ skills the evaluator repeatedly picks random subsets of $k$ skills and asks the LLM to produce text combining that subset of skills. Since the number of subsets grows like $N^k$, for even modest $k$ this evaluation will, with high probability, require the LLM to produce text significantly different from any text in the training set. \nThe paper develops a methodology for (a) designing and administering such an evaluation, and (b) automatic grading (plus spot-checking by humans) of the results using GPT-4 as well as the open LLaMA-2 70B model. \n\nAdministering a version of SKILL-MIX to popular chatbots gave results that, while generally in line with prior expectations, contained surprises. Sizeable differences exist among model capabilities that are not captured by their ranking on popular LLM leaderboards (\"cramming for the leaderboard\"). Furthermore, simple probability calculations indicate that GPT-4's reasonable performance on $k=5$ is suggestive of going beyond \"stochastic parrot\" behavior (Bender et al., 2021), i.e., it combines skills in ways that it had not seen during training.\n\nWe sketch how the methodology can lead to a SKILL-MIX based eco-system of open evaluations for AI capabilities of future models. We maintain a leaderboard of SKILL-MIX at [https://skill-mix.github.io](https://skill-mix.github.io).", "pdf": "https://openreview.net/pdf/2f6d42f9e2ebcdc95ea494a075e5cd3fc7e5e119.pdf"} {"title": "A Quadratic Synchronization Rule for Distributed Deep Learning", "url": "https://openreview.net/forum?id=yroyhkhWS6", "detail_url": "https://openreview.net/forum?id=yroyhkhWS6", "authors": "Xinran Gu,Kaifeng Lyu,Sanjeev Arora,Jingzhao Zhang,Longbo Huang", "tags": "ICLR 2024,Poster", "abstract": "In distributed deep learning with data parallelism, synchronizing gradients at each training step can cause a huge communication overhead, especially when many nodes work together to train large models.\n Local gradient methods, such as Local SGD, address this issue by allowing workers to compute locally for $H$ steps without synchronizing with others, hence reducing communication frequency.\n While $H$ has been viewed as a hyperparameter to trade optimization efficiency for communication cost, recent research indicates that setting a proper $H$ value can lead to generalization improvement. Yet, selecting a proper $H$ is elusive. This work proposes a theory-grounded method for determining $H$, named the Quadratic Synchronization Rule (QSR), which recommends dynamically setting $H$ in proportion to $\\frac{1}{\\eta^2}$ as the learning rate $\\eta$ decays over time.\n Extensive ImageNet experiments on ResNet and ViT show that local gradient methods with QSR consistently improve the test accuracy over other synchronization strategies. Compared to the standard data parallel training, QSR enables Local AdamW to cut the training time on 16 or 64 GPUs down from 26.7 to 20.2 hours or from 8.6 to 5.5 hours and, at the same time, achieves 1.16% or 0.84% higher top-1 validation accuracy.", "pdf": "https://openreview.net/pdf/d7d0725ff6973ff72e22785c8490bab873333e1c.pdf"} {"title": "ArchLock: Locking DNN Transferability at the Architecture Level with a Zero-Cost Binary Predictor", "url": "https://openreview.net/forum?id=e2YOVTenU9", "detail_url": "https://openreview.net/forum?id=e2YOVTenU9", "authors": "Tong Zhou,Shaolei Ren,Xiaolin Xu", "tags": "ICLR 2024,Poster", "abstract": "Deep neural network (DNN) models, despite their impressive performance, are vulnerable to exploitation by attackers who attempt to transfer them to other tasks for their own benefit. Current defense strategies mainly address this vulnerability at the model parameter level, leaving the potential of architectural-level defense largely unexplored. This paper, for the first time, addresses the issue of model protection by reducing transferability at the architecture level. Specifically, we present a novel neural architecture search (NAS)-enabled algorithm that employs zero-cost proxies and evolutionary search, to explore model architectures with low transferability. Our method, namely ArchLock, aims to achieve high performance on the source task, while degrading the performance on potential target tasks, i.e., locking the transferability of a DNN model. To achieve efficient cross-task search without accurately knowing the training data owned by the attackers, we utilize zero-cost proxies to speed up architecture evaluation and simulate potential target task embeddings to assist cross-task search with a binary performance predictor. Extensive experiments on NAS-Bench-201 and TransNAS-Bench-101 demonstrate that ArchLock reduces transferability by up to 30% and 50%, respectively, with negligible performance degradation on source tasks (<2%). The code is available at https://github.com/Tongzhou0101/ArchLock.", "pdf": "https://openreview.net/pdf/c801da252c8a18ba94f5b374ffd9515a915c541d.pdf"} {"title": "Leftover Lunch: Advantage-based Offline Reinforcement Learning for Language Models", "url": "https://openreview.net/forum?id=ZDGKPbF0VQ", "detail_url": "https://openreview.net/forum?id=ZDGKPbF0VQ", "authors": "Ashutosh Baheti,Ximing Lu,Faeze Brahman,Ronan Le Bras,Maarten Sap,Mark Riedl", "tags": "ICLR 2024,Poster", "abstract": "Reinforcement Learning with Human Feedback (RLHF) is the most prominent method for Language Model (LM) alignment. However, RLHF is an unstable and data-hungry process that continually requires new high-quality LM-generated data for finetuning. We introduce Advantage-Leftover Lunch RL (A-LoL), a new class of offline policy gradient algorithms that enable RL training on any pre-existing data. By assuming the entire LM output sequence as a single action, A-LoL allows incorporating sequence-level classifiers or human-designed scoring functions as\nrewards. Subsequently, by using LM\u2019s value estimate, A-LoL only trains on positive advantage (leftover) data points, making it resilient to noise. Overall, A-LoL is an easy-to-implement, sample-efficient, and stable LM training recipe.\n\nWe demonstrate the effectiveness of A-LoL and its variants with a set of four different language generation tasks. We compare against both online RL (PPO) and recent preference-based (DPO, PRO) and reward-based (GOLD) offline RL baselines. On the commonly-used RLHF benchmark, Helpful and Harmless Assistant (HHA), LMs trained with A-LoL methods achieve the highest diversity while also being rated more safe and helpful than the baselines according to humans. Additionally, in the remaining three tasks, A-LoL could optimize multiple distinct reward functions even when using noisy or suboptimal training data.", "pdf": "https://openreview.net/pdf/12ee80d1f10cff1b285c1198d93c7a1190f0dac3.pdf"} {"title": "RECOMP: Improving Retrieval-Augmented LMs with Context Compression and Selective Augmentation", "url": "https://openreview.net/forum?id=mlJLVigNHp", "detail_url": "https://openreview.net/forum?id=mlJLVigNHp", "authors": "Fangyuan Xu,Weijia Shi,Eunsol Choi", "tags": "ICLR 2024,Poster", "abstract": "Retrieval-augmented language models improve language models (LMs) by retrieving documents and prepending them in-context.\nHowever, these documents, often spanning hundreds of words, make inference substantially less efficient. We propose compressing the retrieved documents into textual summaries prior to in-context integration. This not only reduces the computational costs but also relieve the burden of LMs to identify relevant information in long retrieved documents. We present two compressors -- an extractive compressor which selects useful sentences from retrieved documents and an abstractive compressor which generates summary by synthesizing information from multiple documents. Both are trained to achieve performance gain in LMs when we prepend the generated summary from the compressor to LMs' input, while minimizing the summary length. When retrieved documents are irrelevant to the input or offer no additional information to LM, our compressors output an empty string, enabling selective augmentation. We evaluate our approach on the language modeling task and open domain question answering task. We achieve a compression rate of as low as 6% with minimal loss in performance for both tasks, significantly outperforming the off-the-shelf summarization models. We show that our compressors trained for one LM can transfer to other LMs on the language modeling task and provide a summary largely faithful to the retrieved documents.", "pdf": "https://openreview.net/pdf/43938c98697c89512480fceb61ff554001727889.pdf"} {"title": "Gen-Z: Generative Zero-Shot Text Classification with Contextualized Label Descriptions", "url": "https://openreview.net/forum?id=rkplYfqUr0", "detail_url": "https://openreview.net/forum?id=rkplYfqUr0", "authors": "Sachin Kumar,Chan Young Park,Yulia Tsvetkov", "tags": "ICLR 2024,Poster", "abstract": "Language model (LM) prompting\u2014a popular paradigm for solving NLP tasks\u2014has been shown to be susceptible to miscalibration and brittleness to slight prompt variations, caused by its discriminative prompting approach, i.e., predicting the label given the input. To address these issues, we propose Gen-Z\u2014a generative prompting framework for zero-shot text classification. GEN-Z is generative, as it measures the LM likelihood of input text, conditioned on natural language descriptions of labels. The framework is multivariate, as label descriptions allow us to seamlessly integrate additional contextual information about the labels to improve task performance. On various standard classification benchmarks, with six open-source LM families, we show that zero-shot classification with simple contextualization of the data source of the evaluation set consistently outperforms both zero-shot and few-shot baselines while improving robustness to prompt variations. Further, our approach enables personalizing classification in a zero-shot manner by incorporating author, subject, or reader information in the label descriptions.", "pdf": "https://openreview.net/pdf/2c0276dcc674aef018e1899f39fed7767c226540.pdf"} {"title": "In-Context Learning Dynamics with Random Binary Sequences", "url": "https://openreview.net/forum?id=62K7mALO2q", "detail_url": "https://openreview.net/forum?id=62K7mALO2q", "authors": "Eric J Bigelow,Ekdeep Singh Lubana,Robert P. Dick,Hidenori Tanaka,Tomer Ullman", "tags": "ICLR 2024,Poster", "abstract": "Large language models (LLMs) trained on huge text datasets demonstrate intriguing capabilities, achieving state-of-the-art performance on tasks they were not explicitly trained for. The precise nature of LLM capabilities is often mysterious, and different prompts can elicit different capabilities through in-context learning. We propose a framework that enables us to analyze in-context learning dynamics to understand latent concepts underlying LLMs\u2019 behavioral patterns. This provides a more nuanced understanding than success-or-failure evaluation benchmarks, but does not require observing internal activations as a mechanistic interpretation of circuits would. Inspired by the cognitive science of human randomness perception, we use random binary sequences as context and study dynamics of in-context learning by manipulating properties of context data, such as sequence length. In the latest GPT-3.5+ models, we find emergent abilities to generate seemingly random numbers and learn basic formal languages, with striking in-context learning dynamics where model outputs transition sharply from seemingly random behaviors to deterministic repetition.", "pdf": "https://openreview.net/pdf/b3b2da0b2a8f41b25ce90b041f1fad321dd98835.pdf"} {"title": "Dichotomy of Early and Late Phase Implicit Biases Can Provably Induce Grokking", "url": "https://openreview.net/forum?id=XsHqr9dEGH", "detail_url": "https://openreview.net/forum?id=XsHqr9dEGH", "authors": "Kaifeng Lyu,Jikai Jin,Zhiyuan Li,Simon Shaolei Du,Jason D. Lee,Wei Hu", "tags": "ICLR 2024,Poster", "abstract": "Recent work by Power et al. (2022) highlighted a surprising \"grokking\" phenomenon in learning arithmetic tasks: a neural net first \"memorizes\" the training set, resulting in perfect training accuracy but near-random test accuracy, and after training for sufficiently longer, it suddenly transitions to perfect test accuracy. This paper studies the grokking phenomenon in theoretical setups and shows that it can be induced by a dichotomy of early and late phase implicit biases. Specifically, when training homogeneous neural nets with large initialization and small weight decay on both classification and regression tasks, we prove that the training process gets trapped at a solution corresponding to a kernel predictor for a long time, and then a very sharp transition to min-norm/max-margin predictors occurs, leading to a dramatic change in test accuracy.", "pdf": "https://openreview.net/pdf/a44b7ffbd53a80fbd4d969f2e39aa59edf5c8012.pdf"} {"title": "Be Aware of the Neighborhood Effect: Modeling Selection Bias under Interference", "url": "https://openreview.net/forum?id=52fz5sUAy2", "detail_url": "https://openreview.net/forum?id=52fz5sUAy2", "authors": "Haoxuan Li,Chunyuan Zheng,Sihao Ding,Peng Wu,Zhi Geng,Fuli Feng,Xiangnan He", "tags": "ICLR 2024,Poster", "abstract": "Selection bias in recommender system arises from the recommendation process of system filtering and the interactive process of user selection. Many previous studies have focused on addressing selection bias to achieve unbiased learning of the prediction model, but ignore the fact that potential outcomes for a given user-item pair may vary with the treatments assigned to other user-item pairs, named neighborhood effect. To fill the gap, this paper formally formulates the neighborhood effect as an interference problem from the perspective of causal inference, and introduces a treatment representation to capture the neighborhood effect. On this basis, we propose a novel ideal loss that can be used to deal with selection bias in the presence of neighborhood effect. We further develop two new estimators for estimating the proposed ideal loss. We theoretically establish the connection between the proposed and previous debiasing methods ignoring the neighborhood effect, showing that the proposed methods can achieve unbiased learning when both selection bias and neighborhood effects are present, while the existing methods are biased. Extensive semi-synthetic and real-world experiments are conducted to demonstrate the effectiveness of the proposed methods.", "pdf": "https://openreview.net/pdf/9205f9cf9861ea57ce78d3007b54e1bcec2e5df6.pdf"} {"title": "PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization", "url": "https://openreview.net/forum?id=22pyNMuIoa", "detail_url": "https://openreview.net/forum?id=22pyNMuIoa", "authors": "Xinyuan Wang,Chenxi Li,Zhen Wang,Fan Bai,Haotian Luo,Jiayou Zhang,Nebojsa Jojic,Eric Xing,Zhiting Hu", "tags": "ICLR 2024,Poster", "abstract": "Expert-level prompts, carefully engineered by human experts who have a deep understanding of both large language models (LLMs) and domain knowledge, are the future of prompting and pivotal to harnessing the full power of advanced LLMs. Discovering such prompts with an automated process remains a sought-after and unresolved challenge. Existing prompt optimization techniques, though automated through iterative sampling, often fall short in injecting domain knowledge and exploring the vast prompt space for complex expert-level prompts efficiently. To address this pressing need and achieve expert-level prompting, we introduce PromptAgent, which autonomously discovers prompts equivalent in quality to those handcrafted by experts. At its core, PromptAgent views prompt optimization as a strategic planning problem and employs a principled planning algorithm (rooted in Monte Carlo Tree Search) to strategically explore the vast expert-level prompt space. PromptAgent interacts with the LLM in a human-like trial-and-error manner during the planning, and injects expert-level knowledge by reflecting on model errors and generating insightful error feedback. This novel formulation allows it to iteratively evaluate intermediate prompts, refine them based on errors, simulate future rewards, and search for high-reward paths leading to expert-level prompts. We apply PromptAgent to 12 tasks spanning three practical domains: BIG-Bench Hard (BBH), domain-expert, and general NLU tasks, showing PromptAgent consistently outperforms strong prompting and prompt optimization baselines by great margins. Our qualitative analysis further emphasizes PromptAgent's capability to distill insightful errors into expert-level prompts.", "pdf": "https://openreview.net/pdf/d4dbdee0d105cd3020bffdc2f56d99c33429d49c.pdf"} {"title": "Repeated Random Sampling for Minimizing the Time-to-Accuracy of Learning", "url": "https://openreview.net/forum?id=JnRStoIuTe", "detail_url": "https://openreview.net/forum?id=JnRStoIuTe", "authors": "Patrik Okanovic,Roger Waleffe,Vasilis Mageirakos,Konstantinos Nikolakakis,Amin Karbasi,Dionysios Kalogerias,Nezihe Merve G\u00fcrel,Theodoros Rekatsinas", "tags": "ICLR 2024,Poster", "abstract": "Methods for carefully selecting or generating a small set of training data to learn from, i.e., data pruning, coreset selection, and dataset distillation, have been shown to be effective in reducing the ever-increasing cost of training neural networks. Behind this success are rigorously designed, yet expensive, strategies for identifying the most informative training examples out of large datasets. In this work, we revisit these methods to understand if the additional computational costs associated with such strategies are justified from the perspective of time-to-accuracy, which has become a critical efficiency measure of deep neural network training over large datasets. Surprisingly, we find that many of the recently proposed methods underperform what we call Repeated Sampling of Random Subsets (RSRS or RS2), a powerful yet overlooked extension of the standard random baseline that learns from repeatedly sampled data throughout training instead of a fixed random subset. We test RS2 against thirty-two state-of-the-art data pruning and distillation methods across four datasets including ImageNet. Our results demonstrate that RS2 significantly reduces time-to-accuracy, particularly in practical regimes where accuracy, but not runtime, is similar to that of training on full dataset. For example, when training ResNet-18 on ImageNet, with 10\\% of the dataset each epoch RS2 reaches an accuracy of 66\\% versus 69\\% when training with the full dataset. The best competing method achieves only 55\\% while training 1.6$\\times$ slower than RS2. Beyond the above meta-study, we discuss the theoretical properties of RS2 such as its convergence rate and generalization error. Our primary goal is to highlight that future works that aim to minimize total training cost by using subset selection, need to consider 1) the total computation cost (including preparing the subset) and 2) should aim to outperform a simple extension of random sampling (i.e., RS2).", "pdf": "https://openreview.net/pdf/d7e4c9daefbd1d5624c0875538fb1fb95b9a2ce9.pdf"} {"title": "Bias Runs Deep: Implicit Reasoning Biases in Persona-Assigned LLMs", "url": "https://openreview.net/forum?id=kGteeZ18Ir", "detail_url": "https://openreview.net/forum?id=kGteeZ18Ir", "authors": "Shashank Gupta,Vaishnavi Shrivastava,Ameet Deshpande,Ashwin Kalyan,Peter Clark,Ashish Sabharwal,Tushar Khot", "tags": "ICLR 2024,Poster", "abstract": "Recent works have showcased the ability of large-scale language models (LLMs) to embody diverse personas in their responses, exemplified by prompts like \u2018_You are Yoda. Explain the Theory of Relativity._\u2019 While this ability allows personalization of LLMs and enables human behavior simulation, its effect on LLMs\u2019 capabilities remains unclear. To fill this gap, we present the first extensive study of the unintended side-effects of persona assignment on the ability of LLMs to perform _basic reasoning tasks_. Our study covers 24 reasoning datasets (spanning mathematics, law, medicine, morals, and more), 4 LLMs (2 versions of ChatGPT-3.5, GPT-4-Turbo, and Llama-2-70b-chat), and 19 diverse personas (e.g., \u2018an Asian person\u2019) spanning 5 socio-demographic groups: race, gender, religion, disability, and political affiliation. Our experiments unveil that LLMs harbor deep rooted bias against various socio-demographics underneath a veneer of fairness. While they overtly reject stereotypes when explicitly asked (\u2018_Are Black people less skilled at mathematics?_\u2019), they manifest stereotypical and often erroneous presumptions when prompted to answer questions while adopting a persona. These can be observed as abstentions in the model\u2019s response, e.g., \u2018_As a Black person, I am unable to answer this question as it requires math knowledge_\u2019, and generally result in a substantial drop in performance on reasoning tasks. Our experiments with ChatGPT-3.5 show that this bias is _ubiquitous_—80% of our personas demonstrate bias; it is _significant_—some datasets show performance drops of 70%+; and can be especially _harmful for certain groups_—some personas suffer statistically significant drops on 80%+ of the datasets. Overall, all four LLMs exhibit persona-induced bias to varying extents, with GPT-4-Turbo showing the least but still a problematic amount of bias (evident in 42% of the personas). Further analysis shows that these persona-induced errors can be hard-to-discern as they do not always manifest as explicit abstentions, and can also be hard-to-avoid—we find de-biasing prompts to have minimal to no effect. Our findings serve as a cautionary tale that the practice of assigning personas to LLMs—a trend on the rise—can surface their deep-rooted biases and have unforeseeable and detrimental side-effects.", "pdf": "https://openreview.net/pdf/80ad8992d5c1096ee5f775cfb3ce54c4de41a376.pdf"} {"title": "Enhancing Instance-Level Image Classification with Set-Level Labels", "url": "https://openreview.net/forum?id=AZW3qlCGTe", "detail_url": "https://openreview.net/forum?id=AZW3qlCGTe", "authors": "Renyu Zhang,Aly A Khan,Yuxin Chen,Robert L. Grossman", "tags": "ICLR 2024,Poster", "abstract": "Instance-level image classification tasks have traditionally relied on single-instance labels to train models, e.g., few-shot learning and transfer learning. However, set-level coarse-grained labels that capture relationships among instances can provide richer information in real-world scenarios. In this paper, we present a novel approach to enhance instance-level image classification by leveraging set-level labels. We provide a theoretical analysis of the proposed method, including recognition conditions for fast excess risk rate, shedding light on the theoretical foundations of our approach. We conducted experiments on two distinct categories of datasets: natural image datasets and histopathology image datasets. Our experimental results demonstrate the effectiveness of our approach, showcasing improved classification performance compared to traditional single-instance label-based methods. Notably, our algorithm achieves 13\\% improvement in classification accuracy compared to the strongest baseline on the histopathology image classification benchmarks. Importantly, our experimental findings align with the theoretical analysis, reinforcing the robustness and reliability of our proposed method. This work bridges the gap between instance-level and set-level image classification, offering a promising avenue for advancing the capabilities of image classification models with set-level coarse-grained labels.", "pdf": "https://openreview.net/pdf/c61a439433658edb2a1f929a97ec263da9c31fea.pdf"} {"title": "Pushing Boundaries: Mixup's Influence on Neural Collapse", "url": "https://openreview.net/forum?id=jTSKkcbEsj", "detail_url": "https://openreview.net/forum?id=jTSKkcbEsj", "authors": "Quinn LeBlanc Fisher,Haoming Meng,Vardan Papyan", "tags": "ICLR 2024,Poster", "abstract": "Mixup is a data augmentation strategy that employs convex combinations of training instances and their respective labels to improve the robustness and calibration of deep neural networks. Despite its widespread adoption, the nuanced mechanisms that underpin its success are not entirely understood. The observed phenomenon of Neural Collapse, where the last-layer activations and classifier of deep networks converge to a simplex equiangular tight frame (ETF), provides a compelling motivation to explore whether mixup induces alternative geometric configurations and whether those could explain its success. In this study, we delve into the last-layer activations of training data for deep networks subjected to mixup, aiming to uncover insights into its operational efficacy. Our investigation, spanning various architectures and dataset pairs, reveals that mixup's last-layer activations predominantly converge to a distinctive configuration different than one might expect. In this configuration, activations from mixed-up examples of identical classes align with the classifier, while those from different classes delineate channels along the decision boundary. These findings are unexpected, as mixed-up features are not simple convex combinations of feature class means (as one might get, for example, by training mixup with the mean squared error loss). By analyzing this distinctive geometric configuration, we elucidate the mechanisms by which mixup enhances model calibration. To further validate our empirical observations, we conduct a theoretical analysis under the assumption of an unconstrained features model, utilizing the mixup loss. Through this, we characterize and derive the optimal last-layer features under the assumption that the classifier forms a simplex ETF.", "pdf": "https://openreview.net/pdf/fc3d9e6d2cf3f29c440ba403f9ad95a4ef697601.pdf"} {"title": "sRGB Real Noise Modeling via Noise-Aware Sampling with Normalizing Flows", "url": "https://openreview.net/forum?id=2XBBumBGeP", "detail_url": "https://openreview.net/forum?id=2XBBumBGeP", "authors": "Dongjin Kim,Donggoo Jung,Sungyong Baik,Tae Hyun Kim", "tags": "ICLR 2024,Poster", "abstract": "Noise poses a widespread challenge in signal processing, particularly when it comes to denoising images. Although convolutional neural networks (CNNs) have exhibited remarkable success in this field, they are predicated upon the belief that noise follows established distributions, which restricts their practicality when dealing with real-world noise. To overcome this limitation, several efforts have been taken to collect noisy image datasets from the real world. Generative methods, employing techniques such as generative adversarial networks (GANs) and normalizing flows (NFs), have emerged as a solution for generating realistic noisy images. Recent works model noise using camera metadata, however requiring metadata even for sampling phase. In contrast, in this work, we aim to estimate the underlying camera settings, enabling us to improve noise modeling and generate diverse noise distributions. To this end, we introduce a new NF framework that allows us to both classify noise based on camera settings and generate various noisy images. Through experimental results, our model demonstrates exceptional noise quality and leads in denoising performance on benchmark datasets.", "pdf": "https://openreview.net/pdf/4e4c68a8b09ae4ef7e4d0ff1f101a225642f3723.pdf"} {"title": "Uncertainty-aware Graph-based Hyperspectral Image Classification", "url": "https://openreview.net/forum?id=8dN7gApKm3", "detail_url": "https://openreview.net/forum?id=8dN7gApKm3", "authors": "Linlin Yu,Yifei Lou,Feng Chen", "tags": "ICLR 2024,Poster", "abstract": "Hyperspectral imaging (HSI) technology captures spectral information across a broad wavelength range, providing richer pixel features compared to traditional color images with only three channels. Although pixel classification in HSI has been extensively studied, especially using graph convolution neural networks (GCNs), quantifying epistemic and aleatoric uncertainties associated with the HSI classification (HSIC) results remains an unexplored area. These two uncertainties are effective for out-of-distribution (OOD) and misclassification detection, respectively. In this paper, we adapt two advanced uncertainty quantification models, evidential GCNs (EGCN) and graph posterior networks (GPN), designed for node classifications in graphs, into the realm of HSIC. We first reveal theoretically that a popular uncertainty cross-entropy (UCE) loss function is insufficient to produce good epistemic uncertainty when learning EGCNs. To mitigate the limitations, we propose two regularization terms. One leverages the inherent property of HSI data where each feature vector is a linear combination of the spectra signatures of the confounding materials, while the other is the total variation (TV) regularization to enforce the spatial smoothness of the evidence with edge-preserving. We demonstrate the effectiveness of the proposed regularization terms on both EGCN and GPN on three real-world HSIC datasets for OOD and misclassification detection tasks. The code is available at GitHub.", "pdf": "https://openreview.net/pdf/5cb7dbbaad37d6d8ad2e4be0826caf667a69732a.pdf"} {"title": "Generative Adversarial Equilibrium Solvers", "url": "https://openreview.net/forum?id=TlyiaPXaVN", "detail_url": "https://openreview.net/forum?id=TlyiaPXaVN", "authors": "Denizalp Goktas,David C. Parkes,Ian Gemp,Luke Marris,Georgios Piliouras,Romuald Elie,Guy Lever,Andrea Tacchetti", "tags": "ICLR 2024,Poster", "abstract": "We introduce the use of generative adversarial learning to compute equilibria in general game-theoretic settings, specifically the generalized Nash equilibrium (GNE) in pseudo-games, and its specific instantiation as the competitive equilibrium (CE) in Arrow-Debreu competitive economies. Pseudo-games are a generalization of games in which players' actions affect not only the payoffs of other players but also their feasible action spaces. Although the computation of GNE and CE is intractable in the worst-case, i.e., PPAD-hard, in practice, many applications only require solutions with high accuracy in expectation over a distribution of problem instances. We introduce Generative Adversarial Equilibrium Solvers (GAES): a family of generative adversarial neural networks that can learn GNE and CE from only a sample of problem instances. We provide computational and sample complexity bounds for Lipschitz-smooth function approximators in a large class of concave pseudo-games, and apply the framework to finding Nash equilibria in normal-form games, CE in Arrow-Debreu competitive economies, and GNE in an environmental economic model of the Kyoto mechanism.", "pdf": "https://openreview.net/pdf/9b540423b50ab612d08d9f12ea387e4bb85a7477.pdf"} {"title": "Graph Transformers on EHRs: Better Representation Improves Downstream Performance", "url": "https://openreview.net/forum?id=pe0Vdv7rsL", "detail_url": "https://openreview.net/forum?id=pe0Vdv7rsL", "authors": "Raphael Poulain,Rahmatollah Beheshti", "tags": "ICLR 2024,Poster", "abstract": "Following the success of transformer-based methods across various machine learning applications, their adoption for healthcare predictive tasks using electronic health records (EHRs) has also expanded extensively. Similarly, graph-based methods have been shown to be very effective in capturing inherent graph-type relationships in EHRs, leading to improved downstream performance. Although integrating these two families of approaches seems like a natural next step, in practice, creating such a design is challenging and has not been done. This is partly due to known EHR problems, such as high sparsity, making extracting meaningful temporal representations of medical visits challenging. In this study, we propose GT-BEHRT, a new approach that leverages temporal visit embeddings extracted from a graph transformer and uses a BERT-based model to obtain more robust patient representations, especially on longer EHR sequences. The graph-based approach allows GT-BEHRT to implicitly capture the intrinsic graphical relationships between medical observations, while the BERT model extracts the temporal relationships between visits, loosely mimicking the clinicians' decision-making process. As part of our method, we also present a two-step pre-training strategy for learning better graphical and temporal representations. Our proposed method achieves state-of-the-art performance in a variety of standard medical predictive tasks, demonstrating the versatility of our approach.", "pdf": "https://openreview.net/pdf/cc3e4cf0f1122fe0c256b5e62246f05026012d2a.pdf"} {"title": "On the Scalability and Memory Efficiency of Semidefinite Programs for Lipschitz Constant Estimation of Neural Networks", "url": "https://openreview.net/forum?id=dwzLn78jq7", "detail_url": "https://openreview.net/forum?id=dwzLn78jq7", "authors": "Zi Wang,Bin Hu,Aaron J Havens,Alexandre Araujo,Yang Zheng,Yudong Chen,Somesh Jha", "tags": "ICLR 2024,Poster", "abstract": "Lipschitz constant estimation plays an important role in understanding generalization, robustness, and fairness in deep learning. Unlike naive bounds based on the network weight norm product, semidefinite programs (SDPs) have shown great promise in providing less conservative Lipschitz bounds with polynomial-time complexity guarantees. However, due to the memory consumption and running speed, standard SDP algorithms cannot scale to modern neural network architectures. In this paper, we transform the SDPs for Lipschitz constant estimation into an eigenvalue optimization problem, which aligns with the modern large-scale optimization paradigms based on first-order methods. This is amenable to autodiff frameworks such as PyTorch and TensorFlow, requiring significantly less memory than standard SDP algorithms. The transformation also allows us to leverage various existing numerical techniques for eigenvalue optimization, opening the way for further memory improvement and computational speedup. The essential technique of our eigenvalue-problem transformation is to introduce redundant quadratic constraints and then utilize both Lagrangian and Shor's SDP relaxations under a certain trace constraint. Notably, our numerical study successfully scales the SDP-based Lipschitz constant estimation to address large neural networks on ImageNet. Our numerical examples on CIFAR10 and ImageNet demonstrate that our technique is more scalable than existing approaches. Our code is available at https://github.com/z1w/LipDiff.", "pdf": "https://openreview.net/pdf/333a8145de80d282eac48f6722bab292ed03b563.pdf"} {"title": "Large Language Models as Automated Aligners for benchmarking Vision-Language Models", "url": "https://openreview.net/forum?id=kZEXgtMNNo", "detail_url": "https://openreview.net/forum?id=kZEXgtMNNo", "authors": "Yuanfeng Ji,Chongjian GE,Weikai Kong,Enze Xie,Zhengying Liu,Zhenguo Li,Ping Luo", "tags": "ICLR 2024,Poster", "abstract": "With the advancements in Large Language Models (LLMs), Vision-Language Models (VLMs) have reached a new level of sophistication, showing notable competence in executing intricate cognition and reasoning tasks. However, existing evaluation benchmarks, primarily relying on rigid, hand-crafted datasets to measure task-specific performance, face significant limitations in assessing the alignment of these increasingly anthropomorphic models with human intelligence. In this work, we address the limitations via Auto-Bench, which delves into exploring LLMs as proficient aligners, measuring the alignment between VLMs and human intelligence and value through automatic data curation and assessment. Specifically, for data curation, Auto-Bench utilizes LLMs (e.g., GPT-4) to automatically generate a vast set of question-answer-reasoning triplets via prompting on visual symbolic representations (e.g., captions, object locations, instance relationships, and etc. The curated data closely matches human intent, owing to the extensive world knowledge embedded in LLMs. Through this pipeline, a total of 28.5K human-verified and 3,504K unfiltered question-answer-reasoning triplets have been curated, covering 4 primary abilities and 16 sub-abilities. We subsequently engage LLMs like GPT-3.5 to serve as judges, implementing the quantitative and qualitative automated assessments to facilitate a comprehensive evaluation of VLMs. Our validation results reveal that LLMs are proficient in both evaluation data curation and model assessment, achieving an average agreement rate of 85%. We envision Auto-Bench as a flexible, scalable, and comprehensive benchmark for evaluating the evolving sophisticated VLMs.", "pdf": "https://openreview.net/pdf/fec2f0e416c0a90d47240e5522b34b70940223f4.pdf"} {"title": "CLaM-TTS: Improving Neural Codec Language Model for Zero-Shot Text-to-Speech", "url": "https://openreview.net/forum?id=ofzeypWosV", "detail_url": "https://openreview.net/forum?id=ofzeypWosV", "authors": "Jaehyeon Kim,Keon Lee,Seungjun Chung,Jaewoong Cho", "tags": "ICLR 2024,Poster", "abstract": "With the emergence of neural audio codecs, which encode multiple streams of discrete tokens from audio, large language models have recently gained attention as a promising approach for zero-shot Text-to-Speech (TTS) synthesis. Despite the ongoing rush towards scaling paradigms, audio tokenization ironically amplifies the scalability challenge, stemming from its long sequence length and the complexity of modelling the multiple sequences. To mitigate these issues, we present CLaM-TTS that employs a probabilistic residual vector quantization to (1) achieve superior compression in the token length, and (2) allow a language model to generate multiple tokens at once, thereby eliminating the need for cascaded modeling to handle the number of token streams. Our experimental results demonstrate that CLaM-TTS is better than or comparable to state-of-the-art neural codec-based TTS models regarding naturalness, intelligibility, speaker similarity, and inference speed. In addition, we examine the impact of the pretraining extent of the language models and their text tokenization strategies on performances.", "pdf": "https://openreview.net/pdf/4aa68c6552ace824647a0f32a7f3b5ff97a6cd58.pdf"} {"title": "Unveiling the Unseen: Identifiable Clusters in Trained Depthwise Convolutional Kernels", "url": "https://openreview.net/forum?id=4VgBjsOC8k", "detail_url": "https://openreview.net/forum?id=4VgBjsOC8k", "authors": "Zahra Babaiee,Peyman Kiasari,Daniela Rus,Radu Grosu", "tags": "ICLR 2024,Poster", "abstract": "Recent advances in depthwise-separable convolutional neural networks (DS-CNNs) have led to novel architectures, that surpass the performance of classical CNNs, by a considerable scalability and accuracy margin. This paper reveals another striking property of DS-CNN architectures: discernible and explainable patterns emerge in their trained depthwise convolutional kernels in all layers. Through an extensive analysis of millions of trained filters, with different sizes and from various models, we employed unsupervised clustering with autoencoders, to categorize these filters. Astonishingly, the patterns converged into a few main clusters, each resembling the difference of Gaussian (DoG) functions, and their first and second-order derivatives. Notably, we classify over 95\\% and 90\\% of the filters from state-of-the-art ConvNeXtV2 and ConvNeXt models, respectively. This finding is not merely a technological curiosity; it echoes the foundational models neuroscientists have long proposed for the vision systems of mammals. Our results thus deepen our understanding of the emergent properties of trained DS-CNNs and provide a bridge between artificial and biological visual processing systems. More broadly, they pave the way for more interpretable and biologically-inspired neural network designs in the future.", "pdf": "https://openreview.net/pdf/3c890643bdfe86705d7ed24e33e7edf242f989d7.pdf"} {"title": "UNR-Explainer: Counterfactual Explanations for Unsupervised Node Representation Learning Models", "url": "https://openreview.net/forum?id=0j9ZDzMPqr", "detail_url": "https://openreview.net/forum?id=0j9ZDzMPqr", "authors": "Hyunju Kang,Geonhee Han,Hogun Park", "tags": "ICLR 2024,Poster", "abstract": "Node representation learning, such as Graph Neural Networks (GNNs), has become one of the important learning methods in machine learning, and the demand for reliable explanation generation is growing. Despite extensive research on explanation generation for supervised node representation learning, explaining unsupervised models has been less explored. To address this gap, we propose a method for generating counterfactual (CF) explanations in unsupervised node representation learning, aiming to identify the most important subgraphs that cause a significant change in the $k$-nearest neighbors of a node of interest in the learned embedding space upon perturbation. The $k$-nearest neighbor-based CF explanation method provides simple, yet pivotal, information for understanding unsupervised downstream tasks, such as top-$k$ link prediction and clustering. Furthermore, we introduce a Monte Carlo Tree Search (MCTS)-based explainability method for generating expressive CF explanations for **U**nsupervised **N**ode **R**epresentation learning methods, which we call **UNR-Explainer**. The proposed method demonstrates improved performance on six datasets for both unsupervised GraphSAGE and DGI.", "pdf": "https://openreview.net/pdf/790d3e0525600daa0b02aecf21fda646b3197859.pdf"} {"title": "Are Bert Family Good Instruction Followers? A Study on Their Potential And Limitations", "url": "https://openreview.net/forum?id=x8VNtpCu1I", "detail_url": "https://openreview.net/forum?id=x8VNtpCu1I", "authors": "yisheng xiao,Juntao Li,Zechen Sun,Zechang Li,Qingrong Xia,Xinyu Duan,Zhefeng Wang,Min Zhang", "tags": "ICLR 2024,Poster", "abstract": "Language modeling at scale has proven very effective and brought unprecedented success to natural language models. Many typical representatives, especially decoder-only models, e.g., BLOOM and LLaMA, and encoder-decoder models, e.g., Flan-T5 and AlexaTM, have exhibited incredible instruction-following capabilities while keeping strong task completion ability. These large language models can achieve superior performance in various tasks and even yield emergent capabilities, e.g., reasoning and universal generalization. Though the above two paradigms are mainstream and well explored, the potential of the BERT family, which are encoder-only based models and have ever been one of the most representative pre-trained models, also deserves attention, at least should be discussed. In this work, we adopt XML-R to explore the effectiveness of the BERT family for instruction following and zero-shot learning. We first design a simple yet effective strategy to utilize the encoder-only models for generation tasks and then conduct multi-task instruction tuning. Experimental results demonstrate that our fine-tuned model, Instruct-XMLR, outperforms Bloomz on all evaluation tasks and achieves comparable performance with mT0 on most tasks. Surprisingly, Instruct-XMLR also possesses strong task and language generalization abilities, indicating that Instruct-XMLR can also serve as a good instruction follower and zero-shot learner. Besides, Instruct-XMLR can accelerate decoding due to its non-autoregressive generation manner, achieving around 3 times speedup compared with current autoregressive large language models. Although we also witnessed several limitations through our experiments, such as the performance decline in long-generation tasks and the shortcoming of length prediction, Instruct-XMLR can still become a good member of the family of current large language models.", "pdf": "https://openreview.net/pdf/e58804ca5c30798461a4aa73b0cc89f9836c6880.pdf"} {"title": "Exploring the Promise and Limits of Real-Time Recurrent Learning", "url": "https://openreview.net/forum?id=V2cBKtdC3a", "detail_url": "https://openreview.net/forum?id=V2cBKtdC3a", "authors": "Kazuki Irie,Anand Gopalakrishnan,J\u00fcrgen Schmidhuber", "tags": "ICLR 2024,Poster", "abstract": "Real-time recurrent learning (RTRL) for sequence-processing recurrent neural networks (RNNs) offers certain conceptual advantages over backpropagation through time (BPTT). RTRL requires neither caching past activations nor truncating context, and enables online learning. However, RTRL's time and space complexity make it impractical. To overcome this problem, most recent work on RTRL focuses on approximation theories, while experiments are often limited to diagnostic settings. Here we explore the practical promise of RTRL in more realistic settings. We study actor-critic methods that combine RTRL and policy gradients, and test them in several subsets of DMLab-30, ProcGen, and Atari-2600 environments. On DMLab memory tasks, our system trained on fewer than 1.2B environmental frames is competitive with or outperforms well-known IMPALA and R2D2 baselines trained on 10B frames. To scale to such challenging tasks, we focus on certain well-known neural architectures with element-wise recurrence, allowing for tractable RTRL without approximation. Importantly, we also discuss rarely addressed limitations of RTRL in real-world applications, such as its complexity in the multi-layer case.", "pdf": "https://openreview.net/pdf/9107e97b85399a8a37e9379bb2cdb2ef3e226b56.pdf"} {"title": "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting", "url": "https://openreview.net/forum?id=YH5w12OUuU", "detail_url": "https://openreview.net/forum?id=YH5w12OUuU", "authors": "Defu Cao,Furong Jia,Sercan O Arik,Tomas Pfister,Yixiang Zheng,Wen Ye,Yan Liu", "tags": "ICLR 2024,Poster", "abstract": "The past decade has witnessed significant advances in time series modeling with deep learning. While achieving state-of-the-art results, the best-performing architectures vary highly across applications and domains. Meanwhile, for natural language processing, the Generative Pre-trained Transformer (GPT) has demonstrated impressive performance via training one general-purpose model across various textual datasets. It is intriguing to explore whether GPT-type architectures can be effective for time series, capturing the intrinsic dynamic attributes and leading to significant accuracy improvements. In this paper, we propose a novel framework, TEMPO, that can effectively learn time series representations. We focus on utilizing two essential inductive biases of the time series task for pre-trained models: (i) decomposition of the complex interaction between trend, seasonal and residual components; and (ii) introducing the design of prompts to facilitate distribution adaptation in different types of time series. TEMPO expands the capability for dynamically modeling real-world temporal phenomena from data within diverse domains. Our experiments demonstrate the superior performance of TEMPO over state-of-the-art methods on zero shot setting for a number of time series benchmark datasets. This performance gain is observed not only in scenarios involving previously unseen datasets but also in scenarios with multi-modal inputs. This compelling finding highlights TEMPO's potential to constitute a foundational model-building framework.", "pdf": "https://openreview.net/pdf/aa8beab0b2913d71a83f4ec71411d87d5c2409b4.pdf"} {"title": "Scaling physics-informed hard constraints with mixture-of-experts", "url": "https://openreview.net/forum?id=u3dX2CEIZb", "detail_url": "https://openreview.net/forum?id=u3dX2CEIZb", "authors": "Nithin Chalapathi,Yiheng Du,Aditi S. Krishnapriyan", "tags": "ICLR 2024,Poster", "abstract": "Imposing known physical constraints, such as conservation laws, during neural network training introduces an inductive bias that can improve accuracy, reliability, convergence, and data efficiency for modeling physical dynamics. While such constraints can be softly imposed via loss function penalties, recent advancements in differentiable physics and optimization improve performance by incorporating PDE-constrained optimization as individual layers in neural networks. This enables a stricter adherence to physical constraints. However, imposing hard constraints significantly increases computational and memory costs, especially for complex dynamical systems. This is because it requires solving an optimization problem over a large number of points in a mesh, representing spatial and temporal discretizations, which greatly increases the complexity of the constraint. To address this challenge, we develop a scalable approach to enforce hard physical constraints using Mixture-of-Experts (MoE), which can be used with any neural network architecture. Our approach imposes the constraint over smaller decomposed domains, each of which is solved by an ``expert'' through differentiable optimization. During training, each expert independently performs a localized backpropagation step by leveraging the implicit function theorem; the independence of each expert allows for parallelization across multiple GPUs. Compared to standard differentiable optimization, our scalable approach achieves greater accuracy in the neural PDE solver setting for predicting the dynamics of challenging non-linear systems. We also improve training stability and require significantly less computation time during both training and inference stages.", "pdf": "https://openreview.net/pdf/0dbb1a4e1eb20fc5d0e7c94834773579d30e5b4b.pdf"} {"title": "Structural Fairness-aware Active Learning for Graph Neural Networks", "url": "https://openreview.net/forum?id=bvjcMvMn7B", "detail_url": "https://openreview.net/forum?id=bvjcMvMn7B", "authors": "Haoyu Han,Xiaorui Liu,Li Ma,MohamadAli Torkamani,Hui Liu,Jiliang Tang,Makoto Yamada", "tags": "ICLR 2024,Poster", "abstract": "Graph Neural Networks (GNNs) have seen significant achievements in semi-supervised node classification. Yet, their efficacy often hinges on access to high-quality labeled node samples, which may not always be available in real-world scenarios. While active learning is commonly employed across various domains to pinpoint and label high-quality samples based on data features, graph data present unique challenges due to their intrinsic structures that render nodes non-i.i.d. Furthermore, biases emerge from the positioning of labeled nodes; for instance, nodes closer to the labeled counterparts often yield better performance. To better leverage graph structure and mitigate structural bias in active learning, we present a unified optimization framework (SCARCE), which is also easily incorporated with node features. Extensive experiments demonstrate that the proposed method not only improves the GNNs performance but also paves the way for more fair results.", "pdf": "https://openreview.net/pdf/1e818b0f851160468dacdfb012e8de08fabf1511.pdf"} {"title": "Neural-Symbolic Recursive Machine for Systematic Generalization", "url": "https://openreview.net/forum?id=FWJAmwE0xH", "detail_url": "https://openreview.net/forum?id=FWJAmwE0xH", "authors": "Qing Li,Yixin Zhu,Yitao Liang,Ying Nian Wu,Song-Chun Zhu,Siyuan Huang", "tags": "ICLR 2024,Poster", "abstract": "Current learning models often struggle with human-like systematic generalization, particularly in learning compositional rules from limited data and extrapolating them to novel combinations. We introduce the Neural-Symbolic Recursive Ma- chine ( NSR), whose core is a Grounded Symbol System ( GSS), allowing for the emergence of combinatorial syntax and semantics directly from training data. The NSR employs a modular design that integrates neural perception, syntactic parsing, and semantic reasoning. These components are synergistically trained through a novel deduction-abduction algorithm. Our findings demonstrate that NSR\u2019s design, imbued with the inductive biases of equivariance and compositionality, grants it the expressiveness to adeptly handle diverse sequence-to-sequence tasks and achieve unparalleled systematic generalization. We evaluate NSR\u2019s efficacy across four challenging benchmarks designed to probe systematic generalization capabilities: SCAN for semantic parsing, PCFG for string manipulation, HINT for arithmetic reasoning, and a compositional machine translation task. The results affirm NSR \u2019s superiority over contemporary neural and hybrid models in terms of generalization and transferability.", "pdf": "https://openreview.net/pdf/440caa44d54237dd2cb02bb7f536e9fdfceadfba.pdf"} {"title": "Davidsonian Scene Graph: Improving Reliability in Fine-grained Evaluation for Text-to-Image Generation", "url": "https://openreview.net/forum?id=ITq4ZRUT4a", "detail_url": "https://openreview.net/forum?id=ITq4ZRUT4a", "authors": "Jaemin Cho,Yushi Hu,Jason Michael Baldridge,Roopal Garg,Peter Anderson,Ranjay Krishna,Mohit Bansal,Jordi Pont-Tuset,Su Wang", "tags": "ICLR 2024,Poster", "abstract": "Evaluating text-to-image models is notoriously difficult. A strong recent approach for assessing text-image faithfulness is based on QG/A (question generation and answering), which uses pre-trained foundational models to automatically generate a set of questions and answers from the prompt, and output images are scored based on whether these answers extracted with a visual question answering model are consistent with the prompt-based answers. This kind of evaluation is naturally dependent on the quality of the underlying QG and VQA models. We identify and address several reliability challenges in existing QG/A work: (a) QG questions should respect the prompt (avoiding hallucinations, duplications, and omissions) and (b) VQA answers should be consistent (not asserting that there is no motorcycle in an image while also claiming the motorcycle is blue). We address these issues with Davidsonian Scene Graph (DSG), an empirically grounded evaluation framework inspired by formal semantics, which is adaptable to any QG/A frameworks. DSG produces atomic and unique questions organized in dependency graphs, which (i) ensure appropriate semantic coverage and (ii) sidestep inconsistent answers. With extensive experimentation and human evaluation on a range of model configurations (LLM, VQA, and T2I), we empirically demonstrate that DSG addresses the challenges noted above. Finally, we present DSG-1k, an open-sourced evaluation benchmark that includes 1,060 prompts, covering a wide range of fine-grained semantic categories with a balanced distribution. We release the DSG-1k prompts and the corresponding DSG questions.", "pdf": "https://openreview.net/pdf/713250bdf458c9b54444d2bf78bfd594a06adead.pdf"} {"title": "Chain of Thought Empowers Transformers to Solve Inherently Serial Problems", "url": "https://openreview.net/forum?id=3EWTEy9MTM", "detail_url": "https://openreview.net/forum?id=3EWTEy9MTM", "authors": "Zhiyuan Li,Hong Liu,Denny Zhou,Tengyu Ma", "tags": "ICLR 2024,Poster", "abstract": "Generating a sequence of intermediate steps, \\emph{a.k.a.}, a chain of thought (CoT), is a highly effective method to improve the accuracy of large language models (LLMs) on arithmetics and symbolic reasoning tasks. However, the mechanism behind CoT remains unclear. \nThis work provides a theoretical understanding of the power of CoT for decoder-only transformers through the lens of expressiveness. Conceptually, CoT empowers the model with the ability to perform inherently serial computation, which is otherwise lacking in transformers, especially when depth is low. Given input length $n$, previous works have constant-depth transformers with finite precision $\\mathsf{poly}(n)$ embedding size can only solve problems in $\\mathsf{TC}^0$ without CoT. We first show an even tighter expressiveness upper bound for constant-depth transformers with constant-bit precision, which can only solve problems in $\\mathsf{AC}^0$, a proper subset of $ \\mathsf{TC}^0$. However, with $T$ steps of CoT, constant-depth transformers using constant-bit precision and $O(\\log n)$ embedding size can solve any problem solvable by boolean circuits of size $T$. Empirically, enabling CoT dramatically improves the accuracy for tasks that are hard for parallel computation, including the composition of permutation groups, iterated squaring, and circuit value problems, especially for low-depth transformers.", "pdf": "https://openreview.net/pdf/2c3e913d1014164603f487f70dace6570bb0a1d0.pdf"} {"title": "Tractable MCMC for Private Learning with Pure and Gaussian Differential Privacy", "url": "https://openreview.net/forum?id=pmweVpJ229", "detail_url": "https://openreview.net/forum?id=pmweVpJ229", "authors": "Yingyu Lin,Yian Ma,Yu-Xiang Wang,Rachel Emily Redberg,Zhiqi Bu", "tags": "ICLR 2024,Poster", "abstract": "Posterior sampling, i.e., exponential mechanism to sample from the posterior distribution, provides $\\varepsilon$-pure differential privacy (DP) guarantees and does not suffer from potentially unbounded privacy breach introduced by $(\\varepsilon,\\delta)$-approximate DP. In practice, however, one needs to apply approximate sampling methods such as Markov chain Monte Carlo (MCMC), thus re-introducing the unappealing $\\delta$-approximation error into the privacy guarantees. To bridge this gap, we propose the Approximate SAample Perturbation (abbr. ASAP) algorithm which perturbs an MCMC sample with noise proportional to its Wasserstein-infinity ($W_\\infty$) distance from a reference distribution that satisfies pure DP or pure Gaussian DP (i.e., $\\delta=0$). We then leverage a Metropolis-Hastings algorithm to generate the sample and prove that the algorithm converges in W$_\\infty$ distance. We show that by combining our new techniques with a localization step, we obtain the first nearly linear-time algorithm that achieves the optimal rates in the DP-ERM problem with strongly convex and smooth losses.", "pdf": "https://openreview.net/pdf/98377763ad27f9f0dab3a807c831a7d2b1e123ef.pdf"} {"title": "Optimal Sketching for Residual Error Estimation for Matrix and Vector Norms", "url": "https://openreview.net/forum?id=RsJwmWvE6Q", "detail_url": "https://openreview.net/forum?id=RsJwmWvE6Q", "authors": "Yi Li,Honghao Lin,David Woodruff", "tags": "ICLR 2024,Poster", "abstract": "We study the problem of residual error estimation for matrix and vector norms using a linear sketch. Such estimates can be used, for example, to quickly assess how useful a more expensive low-rank approximation computation will be. The matrix case concerns the Frobenius norm and the task is to approximate the $k$-residual $\\|A - A_k\\|_F$ of the input matrix $A$ within a $(1+\\epsilon)$-factor, where $A_k$ is the optimal rank-$k$ approximation. We provide a tight bound of $\\Theta(k^2/\\epsilon^4)$ on the size of bilinear sketches, which have the form of a matrix product $SAT$. This improves the previous $O(k^2/\\epsilon^6)$ upper bound in (Andoni et al. SODA 2013) and gives the first non-trivial lower bound, to the best of our knowledge. \nIn our algorithm, our sketching matrices $S$ and $T$ can both be sparse matrices, allowing for a very fast update time. \nWe demonstrate that this gives a substantial advantage empirically, for roughly the same sketch size and accuracy as in previous work. \n\nFor the vector case, we consider the $\\ell_p$-norm for $p>2$, where the task is to approximate the $k$-residual $\\|x - x_k\\|_p$ up to a constant factor, where $x_k$ is the optimal $k$-sparse approximation to $x$. Such vector norms are frequently studied in the data stream literature and are useful for finding frequent items or so-called heavy hitters. We establish an upper bound of $O(k^{2/p}n^{1-2/p}\\operatorname{poly}(\\log n))$ for constant $\\epsilon$ on the dimension of a linear sketch for this problem. Our algorithm can be extended to the $\\ell_p$ sparse recovery problem with the same sketching dimension, which seems to be the first such bound for $p > 2$. We also show an $\\Omega(k^{2/p}n^{1-2/p})$ lower bound for the sparse recovery problem, which is tight up to a $\\mathrm{poly}(\\log n)$ factor.", "pdf": "https://openreview.net/pdf/7f94d35697799a150e50b5657014861583975c50.pdf"} {"title": "Reverse Diffusion Monte Carlo", "url": "https://openreview.net/forum?id=kIPEyMSdFV", "detail_url": "https://openreview.net/forum?id=kIPEyMSdFV", "authors": "Xunpeng Huang,Hanze Dong,Yifan HAO,Yian Ma,Tong Zhang", "tags": "ICLR 2024,Poster", "abstract": "We propose a Monte Carlo sampler from the reverse diffusion process. Unlike the practice of diffusion models, where the intermediary updates---the score functions---are learned with a neural network, we transform the score matching problem into a mean estimation one.\nBy estimating the means of the regularized posterior distributions, we derive a novel Monte Carlo sampling algorithm called reverse diffusion Monte Carlo (rdMC), which is distinct from the Markov chain Monte Carlo (MCMC) methods. We determine the sample size from the error tolerance and the properties of the posterior distribution to yield an algorithm that can approximately sample the target distribution with any desired accuracy. Additionally, we demonstrate and prove under suitable conditions that sampling with rdMC can be significantly faster than that with MCMC. For multi-modal target distributions such as those in Gaussian mixture models, rdMC greatly improves over the Langevin-style MCMC sampling methods both theoretically and in practice. The proposed rdMC method offers a new perspective and solution beyond classical MCMC algorithms for the challenging complex distributions.", "pdf": "https://openreview.net/pdf/5e05dd8867eb805ba66920ee894e0234bbfd718d.pdf"} {"title": "Counting Graph Substructures with Graph Neural Networks", "url": "https://openreview.net/forum?id=qaJxPhkYtD", "detail_url": "https://openreview.net/forum?id=qaJxPhkYtD", "authors": "Charilaos Kanatsoulis,Alejandro Ribeiro", "tags": "ICLR 2024,Poster", "abstract": "Graph Neural Networks (GNNs) are powerful representation learning tools that have achieved remarkable performance in various downstream tasks. However, there are still open questions regarding their ability to count and list substructures, which play a crucial role in biological and social networks. In this work, we fill this gap and characterize the representation {and generalization} power of GNNs in terms of their ability to produce powerful representations that count substructures. In particular, we study the message-passing operations of GNNs with random node input in a novel fashion, and show how they can produce equivariant representations that are associated with high-order statistical moments. Using these representations, we prove that GNNs can learn how to count cycles, {cliques}, quasi-cliques, and the number of connected components in a graph. We also provide new insights into the generalization capacity of GNNs. Our analysis is constructive and enables the design of a generic GNN architecture that shows remarkable performance in four distinct tasks: cycle detection, cycle counting, graph classification, and molecular property prediction.", "pdf": "https://openreview.net/pdf/b832a3a871aa5f5adcfb1053797c2fbd754232a2.pdf"} {"title": "Are Models Biased on Text without Gender-related Language?", "url": "https://openreview.net/forum?id=w1JanwReU6", "detail_url": "https://openreview.net/forum?id=w1JanwReU6", "authors": "Catarina G Bel\u00e9m,Preethi Seshadri,Yasaman Razeghi,Sameer Singh", "tags": "ICLR 2024,Poster", "abstract": "Gender bias research has been pivotal in revealing undesirable behaviors in large language models, exposing serious gender stereotypes associated with occupations, and emotions. A key observation in prior work is that models reinforce stereotypes as a consequence of the gendered correlations that are present in the training data. In this paper, we focus on bias where the effect from training data is unclear, and instead address the question: *Do language models still exhibit gender bias in non-stereotypical settings?* To do so, we introduce **UnStereoEval (USE)**, a novel framework tailored for investigating gender bias in stereotype-free scenarios. USE defines a sentence-level score based on pretraining data statistics to determine if the sentence contain minimal word-gender associations. To systematically benchmark the fairness of popular language models in stereotype-free scenarios, we utilize USE to automatically generate benchmarks without any gender-related language. By leveraging USE's sentence-level score, we also repurpose prior gender bias benchmarks (Winobias and Winogender) for non-stereotypical evaluation. Surprisingly, we find low fairness across all 28 tested models. Concretely, models demonstrate fair behavior in only 9%-41% of stereotype-free sentences, suggesting that bias does not solely stem from the presence of gender-related words. These results raise important questions about where underlying model biases come from and highlight the need for more systematic and comprehensive bias evaluation. We release the full dataset and code at [ucinlp.github.io/unstereo-eval](https://ucinlp.github.io/unstereo-eval).", "pdf": "https://openreview.net/pdf/bd1813ac5b333e7445f4c1a4ac8d3680ace9c572.pdf"} {"title": "PlaSma: Procedural Knowledge Models for Language-based Planning and Re-Planning", "url": "https://openreview.net/forum?id=dFcXJgnrGB", "detail_url": "https://openreview.net/forum?id=dFcXJgnrGB", "authors": "Faeze Brahman,Chandra Bhagavatula,Valentina Pyatkin,Jena D. Hwang,Xiang Lorraine Li,Hirona Jacqueline Arai,Soumya Sanyal,Keisuke Sakaguchi,Xiang Ren,Yejin Choi", "tags": "ICLR 2024,Poster", "abstract": "Procedural planning, which entails decomposing a high-level goal into a sequence of temporally ordered steps, is an important yet intricate task for machines. It involves integrating common-sense knowledge to reason about complex and often contextualized situations, e.g. ``scheduling a doctor's appointment without a phone''. While current approaches show encouraging results using large language models (LLMs), they are hindered by drawbacks such as costly API calls and reproducibility issues. In this paper, we advocate planning using smaller language models. We present PlaSma, a novel two-pronged approach to endow small language models with procedural knowledge and (constrained) language-based planning capabilities. More concretely, we develop *symbolic procedural knowledge distillation* to enhance the commonsense knowledge in small language models and an *inference-time algorithm* to facilitate more structured and accurate reasoning. In addition, we introduce a new related task, *Replanning*, that requires a revision of a plan to cope with a constrained situation. In both the planning and replanning settings, we show that orders-of-magnitude smaller models (770M-11B parameters) can compete and often surpass their larger teacher models' capabilities. Finally, we showcase successful application of PlaSma in an embodied environment, VirtualHome.", "pdf": "https://openreview.net/pdf/b88eaa0cc84120348881ebfc5dd4fe00210337df.pdf"} {"title": "From Molecules to Materials: Pre-training Large Generalizable Models for Atomic Property Prediction", "url": "https://openreview.net/forum?id=PfPnugdxup", "detail_url": "https://openreview.net/forum?id=PfPnugdxup", "authors": "Nima Shoghi,Adeesh Kolluru,John R. Kitchin,Zachary Ward Ulissi,C. Lawrence Zitnick,Brandon M Wood", "tags": "ICLR 2024,Poster", "abstract": "Foundation models have been transformational in machine learning fields such as natural language processing and computer vision. Similar success in atomic property prediction has been limited due to the challenges of training effective models across multiple chemical domains. To address this, we introduce Joint Multi-domain Pre-training (JMP), a supervised pre-training strategy that simultaneously trains on multiple datasets from different chemical domains, treating each dataset as a unique pre-training task within a multi-task framework. Our combined training dataset consists of $\\sim$120M systems from OC20, OC22, ANI-1x, and Transition-1x. We evaluate performance and generalization by fine-tuning over a diverse set of downstream tasks and datasets including: QM9, rMD17, MatBench, QMOF, SPICE, and MD22. JMP demonstrates an average improvement of 59% over training from scratch and matches or sets state-of-the-art on 34 out of 40 tasks. Our work highlights the potential of pre-training strategies that utilize diverse data to advance property prediction across chemical domains, especially for low-data tasks.", "pdf": "https://openreview.net/pdf/7ae9e4f5f396605dfda891057be51e0ec4e42fdc.pdf"} {"title": "Towards Foundational Models for Molecular Learning on Large-Scale Multi-Task Datasets", "url": "https://openreview.net/forum?id=Zc2aIcucwc", "detail_url": "https://openreview.net/forum?id=Zc2aIcucwc", "authors": "Dominique Beaini,Shenyang Huang,Joao Alex Cunha,Zhiyi Li,Gabriela Moisescu-Pareja,Oleksandr Dymov,Samuel Maddrell-Mander,Callum McLean,Frederik Wenkel,Luis M\u00fcller,Jama Hussein Mohamud,Ali Parviz,Michael Craig,Micha\u0142 Koziarski,Jiarui Lu,Zhaocheng Zhu,Cristian Gabellini,Kerstin Klaser,Josef Dean,Cas Wognum,Maciej Sypetkowski,Guillaume Rabusseau,Reihaneh Rabbany,Jian Tang,Christopher Morris,Mirco Ravanelli,Guy Wolf,Prudencio Tossou,Hadrien Mary,Therence Bois,Andrew W Fitzgibbon,Blazej Banaszewski,Chad Martin,Dominic Masters", "tags": "ICLR 2024,Poster", "abstract": "Recently, pre-trained foundation models have enabled significant advancements in multiple fields. In molecular machine learning, however, where datasets are often hand-curated, and hence typically small, the lack of datasets with labeled features, and codebases to manage those datasets, has hindered the development of foundation models. In this work, we present seven novel datasets categorized by size into three distinct categories: ToyMix, LargeMix and UltraLarge. These datasets push the boundaries in both the scale and the diversity of supervised labels for molecular learning. They cover nearly 100 million molecules and over 3000 sparsely defined tasks, totaling more than 13 billion individual labels of both quantum and biological nature. In comparison, our datasets contain 300 times more data points than the widely used OGB-LSC PCQM4Mv2 dataset, and 13 times more than the quantum-only QM1B dataset. In addition, to support the development of foundational models based on our proposed datasets, we present the Graphium graph machine learning library which simplifies the process of building and training molecular machine learning models for multi-task and multi-level molecular datasets. Finally, we present a range of baseline results as a starting point of multi-task and multi-level training on these datasets. Empirically, we observe that performance on low-resource biological datasets show improvement by also training on large amounts of quantum data. This indicates that there may be potential in multi-task and multi-level training of a foundation model and fine-tuning it to resource-constrained downstream tasks. The Graphium library is publicly available on Github and the dataset links are available in Part 1 and Part 2.", "pdf": "https://openreview.net/pdf/070fcd7e5f031fc5d671ef14723f848bdc7a540b.pdf"} {"title": "Independent-Set Design of Experiments for Estimating Treatment and Spillover Effects under Network Interference", "url": "https://openreview.net/forum?id=w50MQ9Vfty", "detail_url": "https://openreview.net/forum?id=w50MQ9Vfty", "authors": "Chencheng Cai,Xu Zhang,Edoardo Airoldi", "tags": "ICLR 2024,Poster", "abstract": "Interference is ubiquitous when conducting causal experiments over networks. Except for certain network structures, causal inference on the network in the presence of interference is difficult due to the entanglement between the treatment assignments and the interference levels. In this article, we conduct causal inference under interference on an observed, sparse, but connected network, and we propose a novel design of experiments based on an independent set. Compared to conventional designs, the independent-set design focuses on an independent subset of data and controls their interference exposures through the assignments to the rest (auxiliary set). We provide a lower bound on the size of the independent set from a greedy algorithm and justify the theoretical performance of estimators under the proposed design. Our approach is capable of estimating both spillover effects and treatment effects. We justify its superiority over conventional methods and illustrate the empirical performance through simulations.", "pdf": "https://openreview.net/pdf/4a5c46e185cf5147131948b3a81949edc248583b.pdf"} {"title": "FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores", "url": "https://openreview.net/forum?id=gPKTTAfYBp", "detail_url": "https://openreview.net/forum?id=gPKTTAfYBp", "authors": "Daniel Y Fu,Hermann Kumbong,Eric Nguyen,Christopher Re", "tags": "ICLR 2024,Poster", "abstract": "Convolution models with long filters have demonstrated state-of-the-art reasoning abilities in many long-sequence tasks but lag behind the most optimized Transformers in wall-clock time.\nA major bottleneck is the Fast Fourier Transform (FFT)---which allows long convolutions to run in $O(N\\log N)$ time in sequence length $N$ but has poor hardware utilization.\nIn this paper, we study how to optimize the FFT convolution.\nWe find two key bottlenecks: the FFT does not effectively use specialized matrix multiply units, and it incurs expensive I/O between layers of the memory hierarchy.\nIn response, we propose FlashFFTConv.\nFlashFFTConv uses a matrix decomposition that computes the FFT using matrix multiply units and enables kernel fusion for long sequences, reducing I/O.\nWe also present two sparse convolution algorithms---1) partial convolutions and 2) frequency-sparse convolutions---which can be implemented simply by skipping blocks in the matrix decomposition, enabling further opportunities for memory and compute savings.\nFlashFFTConv speeds up exact FFT convolutions by up to 8.7$\\times$ over PyTorch and achieves up to 4.4$\\times$ speedup end-to-end.\nGiven the same compute budget, FlashFFTConv allows Hyena-GPT-s to achieve 2.3 points better perplexity and M2-BERT-base to achieve 3.3 points higher GLUE score---matching models with twice the parameter count.\nFlashFFTConv also achieves 96.1% accuracy on Path-512, a high-resolution vision task where no model had previously achieved better than 50%.\nFurthermore, partial convolutions enable longer-sequence models---yielding the first DNA model that can process the longest human genes (2.3M base pairs)---and frequency-sparse convolutions speed up pretrained models while maintaining or improving model quality.", "pdf": "https://openreview.net/pdf/c77f3c8f339aae3682e72c37d33ce5bf90cd1134.pdf"} {"title": "Transformer-VQ: Linear-Time Transformers via Vector Quantization", "url": "https://openreview.net/forum?id=oDdzXQzP2F", "detail_url": "https://openreview.net/forum?id=oDdzXQzP2F", "authors": "Lucas Dax Lingle", "tags": "ICLR 2024,Poster", "abstract": "We introduce Transformer-VQ, a decoder-only transformer computing softmax-based dense self-attention in linear time. Transformer-VQ's efficient attention is enabled by vector-quantized keys and a novel caching mechanism. \nIn our large-scale experiments, Transformer-VQ is shown highly competitive in quality, obtaining 0.99 bpb on Enwik8, 26.6 ppl on PG-19, and 3.16 bpb on ImageNet64. In addition, the optimized implementation of Transformer-VQ is over 3x faster than a comparable quadratic-time transformer at sequence length 8k, is over 12x faster at 32k, and can scale to 131k with similar throughput. Code available: \\url{https://github.com/transformer-vq/transformer_vq}", "pdf": "https://openreview.net/pdf/9dfab016ded80c8754b4868d9dc2a054a8f347b6.pdf"} {"title": "The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry", "url": "https://openreview.net/forum?id=4g02l2N2Nx", "detail_url": "https://openreview.net/forum?id=4g02l2N2Nx", "authors": "Michael Zhang,Kush Bhatia,Hermann Kumbong,Christopher Re", "tags": "ICLR 2024,Poster", "abstract": "Linear attentions have shown promise for improving Transformer efficiency, reducing attention's quadratic complexity to linear in sequence length. This holds exciting promise for (1) training linear Transformers from scratch, (2) `inetuned-conversion of task-specific Transformers into linear versions that recover task performance, and (3) pretrained-conversion of Transformers, such as language models, into linear versions readily finetunable on downstream tasks. However, linear attentions often underperform compared to standard softmax attention. To close this performance gap, we study the behaviors of softmax and linear attentions in various train-from-scratch and finetuned-conversion settings. We find prior linear attentions lack key properties of softmax attention tied to good performance: low-entropy (or spiky) weights and dot-product monotonicity. We further observe surprisingly simple feature maps that retain these properties match softmax performance, but are inefficient to compute in linear attention. We thus propose Hedgehog, a learnable linear attention that retains the spiky and monotonic properties of softmax attention while maintaining linear complexity. Hedgehog uses simple, trainable MLPs to produce attention weights mimicking softmax attention. Experiments show Hedgehog recovers over 99\\% of standard Transformer performance in train-from-scratch and finetuned-conversion settings, outperforming prior linear attentions by up to 6 perplexity points on WikiText-103 when training causal GPT models from scratch, and up to 8.7 GLUE score points when converting finetuned bidirectional BERT models. Hedgehog also enables pretrained-conversion. Converting a pretrained GPT-2 into a linear attention variant achieves state-of-the-art 16.7 perplexity on WikiText-103 for 125M subquadratic decoder models. We finally turn a pretrained Llama-2 7B into a viable linear attention Llama. With low-rank adaptation, Hedgehog-Llama-2 7B achieves 28.1 higher ROUGE-1 points over the base standard attention model, where prior linear attentions lead to 16.5 point drops.", "pdf": "https://openreview.net/pdf/253a4c0c2132cbb269cad956934997223cc2c5c0.pdf"} {"title": "Abstractors and relational cross-attention: An inductive bias for explicit relational reasoning in Transformers", "url": "https://openreview.net/forum?id=XNa6r6ZjoB", "detail_url": "https://openreview.net/forum?id=XNa6r6ZjoB", "authors": "Awni Altabaa,Taylor Whittington Webb,Jonathan D. Cohen,John Lafferty", "tags": "ICLR 2024,Poster", "abstract": "An extension of Transformers is proposed that enables explicit relational reasoning through a novel module called the *Abstractor*. At the core of the Abstractor is a variant of attention called *relational cross-attention*. The approach is motivated by an architectural inductive bias for relational learning that disentangles relational information from object-level features. This enables explicit relational reasoning, supporting abstraction and generalization from limited data. The Abstractor is first evaluated on simple discriminative relational tasks and compared to existing relational architectures. Next, the Abstractor is evaluated on purely relational sequence-to-sequence tasks, where dramatic improvements are seen in sample efficiency compared to standard Transformers. Finally, Abstractors are evaluated on a collection of tasks based on mathematical problem solving, where consistent improvements in performance and sample efficiency are observed.", "pdf": "https://openreview.net/pdf/24529d393a107ced3db4542a44c29da2edba8d83.pdf"} {"title": "Doubly Robust Instance-Reweighted Adversarial Training", "url": "https://openreview.net/forum?id=OF5x1dzWSS", "detail_url": "https://openreview.net/forum?id=OF5x1dzWSS", "authors": "Daouda Sow,Sen Lin,Zhangyang Wang,Yingbin Liang", "tags": "ICLR 2024,Poster", "abstract": "Assigning importance weights to adversarial data has achieved great success in training adversarially robust networks under limited model capacity. However, existing instance-reweighted adversarial training (AT) methods heavily depend on heuristics and/or geometric interpretations to determine those importance weights, making these algorithms lack rigorous theoretical justification/guarantee. Moreover, recent research has shown that adversarial training suffers from a severe non-uniform robust performance across the training distribution, e.g., data points belonging to some classes can be much more vulnerable to adversarial attacks than others. To address both issues, in this paper, we propose a novel doubly-robust instance reweighted AT framework, which allows to obtain the importance weights via exploring distributionally robust optimization (DRO) techniques, and at the same time boosts the robustness on the most vulnerable examples. In particular, our importance weights are obtained by optimizing the KL-divergence regularized loss function, which allows us to devise new algorithms with a theoretical convergence guarantee. \nExperiments on standard classification datasets demonstrate that our proposed approach outperforms related state-of-the-art baseline methods in terms of average robust performance, and at the same time improves the robustness against attacks on the weakest data points. Codes can be found in the Supplement.", "pdf": "https://openreview.net/pdf/c8cb1e0f5f66335caae7b2b8cebef4a72602b4ad.pdf"} {"title": "Training Diffusion Models with Reinforcement Learning", "url": "https://openreview.net/forum?id=YCWjhGrJFD", "detail_url": "https://openreview.net/forum?id=YCWjhGrJFD", "authors": "Kevin Black,Michael Janner,Yilun Du,Ilya Kostrikov,Sergey Levine", "tags": "ICLR 2024,Poster", "abstract": "Diffusion models are a class of flexible generative models trained with an approximation to the log-likelihood objective. However, most use cases of diffusion models are not concerned with likelihoods, but instead with downstream objectives such as human-perceived image quality or drug effectiveness. In this paper, we investigate reinforcement learning methods for directly optimizing diffusion models for such objectives. We describe how posing denoising as a multi-step decision-making problem enables a class of policy gradient algorithms, which we refer to as denoising diffusion policy optimization ( DDPO), that are more effective than alternative reward-weighted likelihood approaches. Empirically, DDPO can adapt text-to-image diffusion models to objectives that are difficult to express via prompting, such as image compressibility, and those derived from human feedback, such as aesthetic quality. Finally, we show that DDPO can improve prompt-image alignment using feedback from a vision-language model without the need for additional data collection or human annotation. The project\u2019s website can be found at http://rl-diffusion.github.io.", "pdf": "https://openreview.net/pdf/39611b93653580f659f3d4d491f00250c4874376.pdf"} {"title": "Finite-Time Analysis of On-Policy Heterogeneous Federated Reinforcement Learning", "url": "https://openreview.net/forum?id=D2eOVqPX9g", "detail_url": "https://openreview.net/forum?id=D2eOVqPX9g", "authors": "Chenyu Zhang,Han Wang,Aritra Mitra,James Anderson", "tags": "ICLR 2024,Poster", "abstract": "Federated reinforcement learning (FRL) has emerged as a promising paradigm for reducing the sample complexity of reinforcement learning tasks by exploiting information from different agents. However, when each agent interacts with a potentially different environment, little to nothing is known theoretically about the non-asymptotic performance of FRL algorithms. The lack of such results can be attributed to various technical challenges and their intricate interplay: Markovian sampling, linear function approximation, multiple local updates to save communication, heterogeneity in the reward functions and transition kernels of the agents' MDPs, and continuous state-action spaces. Moreover, in the on-policy setting, the behavior policies vary with time, further complicating the analysis. In response, we introduce FedSARSA, a novel federated on-policy reinforcement learning scheme, equipped with linear function approximation, to address these challenges and provide a comprehensive finite-time error analysis. Notably, we establish that FedSARSA converges to a policy that is near-optimal for all agents, with the extent of near-optimality proportional to the level of heterogeneity. Furthermore, we prove that FedSARSA leverages agent collaboration to enable linear speedups as the number of agents increases, which holds for both fixed and adaptive step-size configurations.", "pdf": "https://openreview.net/pdf/f97ca95cc9cf273902a1eba91646203707a42223.pdf"} {"title": "Federated Q-Learning: Linear Regret Speedup with Low Communication Cost", "url": "https://openreview.net/forum?id=fe6ANBxcKM", "detail_url": "https://openreview.net/forum?id=fe6ANBxcKM", "authors": "Zhong Zheng,Fengyu Gao,Lingzhou Xue,Jing Yang", "tags": "ICLR 2024,Poster", "abstract": "In this paper, we consider federated reinforcement learning for tabular episodic Markov Decision Processes (MDP) where, under the coordination of a central server, multiple agents collaboratively explore the environment and learn an optimal policy without sharing their raw data. While linear speedup in the number of agents has been achieved for some metrics, such as convergence rate and sample complexity, in similar settings, it is unclear whether it is possible to design a *model-free* algorithm to achieve linear *regret* speedup with low communication cost. We propose two federated Q-Learning algorithms termed as FedQ-Hoeffding and FedQ-Bernstein, respectively, and show that the corresponding total regrets achieve a linear speedup compared with their single-agent counterparts, while the communication cost scales logarithmically in the total number of time steps $T$. Those results rely on an event-triggered synchronization mechanism between the agents and the server, a novel step size selection when the server aggregates the local estimates of the state-action values to form the global estimates, and a set of new concentration inequalities to bound the sum of non-martingale differences. This is the first work showing that linear regret speedup and logarithmic communication cost can be achieved by model-free algorithms in federated reinforcement learning.", "pdf": "https://openreview.net/pdf/6ae0807140d8835f40da63717a9baa1749faec87.pdf"} {"title": "The Trickle-down Impact of Reward Inconsistency on RLHF", "url": "https://openreview.net/forum?id=MeHmwCDifc", "detail_url": "https://openreview.net/forum?id=MeHmwCDifc", "authors": "Lingfeng Shen,Sihao Chen,Linfeng Song,Lifeng Jin,Baolin Peng,Haitao Mi,Daniel Khashabi,Dong Yu", "tags": "ICLR 2024,Poster", "abstract": "Standard practice within Reinforcement Learning from Human Feedback (RLHF) involves optimizing against a Reward Model (RM), which itself is trained to reflect human preferences for desirable generations. A notable subject that is understudied is the (in-)consistency of RMs --- whether they can recognize the semantic changes to different prompts and \nappropriately adapt their reward assignments\n\n--- and their impact on the downstream RLHF model.\n\nIn this paper, we visit a series of research questions relevant to RM inconsistency:\n(1) How can we measure the consistency of reward models? \n(2) How consistent are the existing RMs and how can we improve them? \n(3) In what ways does reward inconsistency influence the chatbots resulting from the RLHF model training?\n\n\nWe propose **Contrast Instruction** -- a benchmarking strategy for the consistency of RM. \nEach example in **Contrast Instruction** features a pair of lexically similar instructions with different ground truth responses. A consistent RM is expected to rank the corresponding instruction and response higher than other combinations. We observe that current RMs trained with the standard ranking objective fail miserably on \\contrast{} compared to average humans. To show that RM consistency can be improved efficiently without using extra training budget, we propose two techniques **ConvexDA** and **RewardFusion**, which enhance reward consistency \nthrough extrapolation during the RM training and inference stage, respectively.\nWe show that RLHF models trained with a more consistent RM yield more useful responses, suggesting that reward inconsistency exhibits a trickle-down effect on the downstream RLHF process.", "pdf": "https://openreview.net/pdf/a8dad7978440d43316bc0727f7c324cbffe5e4c0.pdf"} {"title": "Efficient Modulation for Vision Networks", "url": "https://openreview.net/forum?id=ip5LHJs6QX", "detail_url": "https://openreview.net/forum?id=ip5LHJs6QX", "authors": "Xu Ma,Xiyang Dai,Jianwei Yang,Bin Xiao,Yinpeng Chen,Yun Fu,Lu Yuan", "tags": "ICLR 2024,Poster", "abstract": "In this work, we present efficient modulation, a novel design for efficient vision networks. We revisit the modulation mechanism, which operates input through convolutional context modeling and feature projection layers, and fuses features via element-wise multiplication and an MLP block. We demonstrate that the abstracted modulation mechanism is particularly well suited for efficient networks and further tailor the modulation design by proposing the efficient modulation (EfficientMod) block, which is considered the essential building block for our networks. Bene- fiting from the prominent representational ability of modulation mechanism and the efficiency of efficient modulation design, our network can accomplish better accuracy-efficiency trade-offs and set new state-of-the-art performance for efficient networks. When integrating EfficientMod block with the vanilla self-attention block, we obtain the hybrid architecture and further improve the performance without sacrificing the efficiency. We carry out comprehensive experiments to verify EfficientMod\u2019s performance. With fewer parameters, our EfficientMod-s performs 0.6 top-1 accuracy better than the prior state-of-the-art approach EfficientFormerV2-s2 without any training tricks and is 25% faster on GPU. Additionally, our method presents a notable improvement in downstream tasks, outperforming EfficientFormerV2-s by 3.6 mIoU on the ADE20K benchmark. Code and checkpoints are available at https://github.com/ma-xu/EfficientMod.", "pdf": "https://openreview.net/pdf/60b84aa807789b9bf4b5e8f2ff637e481c3cd2c8.pdf"} {"title": "Pre-training LiDAR-based 3D Object Detectors through Colorization", "url": "https://openreview.net/forum?id=fB1iiH9xo7", "detail_url": "https://openreview.net/forum?id=fB1iiH9xo7", "authors": "Tai-Yu Pan,Chenyang Ma,Tianle Chen,Cheng Perng Phoo,Katie Z Luo,Yurong You,Mark Campbell,Kilian Q Weinberger,Bharath Hariharan,Wei-Lun Chao", "tags": "ICLR 2024,Poster", "abstract": "Accurate 3D object detection and understanding for self-driving cars heavily relies on LiDAR point clouds, necessitating large amounts of labeled data to train. In this work, we introduce an innovative pre-training approach, Grounded Point Colorization (GPC), to bridge the gap between data and labels by teaching the model to colorize LiDAR point clouds, equipping it with valuable semantic cues. To tackle challenges arising from color variations and selection bias, we incorporate color as \"context\" by providing ground-truth colors as hints during colorization.\nExperimental results on the KITTI and Waymo datasets demonstrate GPC's remarkable effectiveness. Even with limited labeled data, GPC significantly improves fine-tuning performance; notably, on just 20% of the KITTI dataset, GPC outperforms training from scratch with the entire dataset. \nIn sum, we introduce a fresh perspective on pre-training for 3D object detection, aligning the objective with the model's intended role and ultimately advancing the accuracy and efficiency of 3D object detection for autonomous vehicles.", "pdf": "https://openreview.net/pdf/bcbf057d60c81f10beb0dc381109a0616f19d030.pdf"} {"title": "An Emulator for Fine-tuning Large Language Models using Small Language Models", "url": "https://openreview.net/forum?id=Eo7kv0sllr", "detail_url": "https://openreview.net/forum?id=Eo7kv0sllr", "authors": "Eric Mitchell,Rafael Rafailov,Archit Sharma,Chelsea Finn,Christopher D Manning", "tags": "ICLR 2024,Poster", "abstract": "Widely used language models (LMs) are typically built by scaling up a two-stage training pipeline: a pre-training stage that uses a very large, diverse dataset of text and a fine-tuning (sometimes, 'alignment') stage that uses targeted examples or other specifications of desired behaviors. While it has been hypothesized that knowledge and skills come from pre-training, and fine-tuning mostly filters this knowledge and skillset, this intuition has not been extensively tested. To aid in doing so, we introduce a novel technique for decoupling the knowledge and skills gained in these two stages, enabling a direct answer to the question, *What would happen if we combined the knowledge learned by a large model during pre-training with the knowledge learned by a small model during fine-tuning (or vice versa)?* Using an RL-based framework derived from recent developments in learning from human preferences, we introduce *emulated fine-tuning (EFT)*, a principled and practical method for sampling from a distribution that approximates (or 'emulates') the result of pre-training and fine-tuning at different scales. Our experiments with EFT show that scaling up fine-tuning tends to improve helpfulness, while scaling up pre-training tends to improve factuality. Beyond decoupling scale, we show that EFT enables test-time adjustment of competing behavioral traits like helpfulness and harmlessness without additional training. Finally, a special case of emulated fine-tuning, which we call LM *up-scaling*, avoids resource-intensive fine-tuning of large pre-trained models by ensembling them with small fine-tuned models, essentially emulating the result of fine-tuning the large pre-trained model. Up-scaling consistently improves helpfulness and factuality of instruction-following models in the Llama, Llama-2, and Falcon families, without additional hyperparameters or training. For reference implementation, see [https://github.com/eric-mitchell/emulated-fine-tuning](https://github.com/eric-mitchell/emulated-fine-tuning).", "pdf": "https://openreview.net/pdf/b0342e1462535a7e1a40bac079cefdfd493a9912.pdf"} {"title": "Toward Student-oriented Teacher Network Training for Knowledge Distillation", "url": "https://openreview.net/forum?id=wsWGcw6qKD", "detail_url": "https://openreview.net/forum?id=wsWGcw6qKD", "authors": "Chengyu Dong,Liyuan Liu,Jingbo Shang", "tags": "ICLR 2024,Poster", "abstract": "How to conduct teacher training for knowledge distillation is still an open problem. It has been widely observed that a best-performing teacher does not necessarily yield the best-performing student, suggesting a fundamental discrepancy between the current teacher training practice and the ideal teacher training strategy. To fill this gap, we explore the feasibility of training a teacher that is oriented toward student performance with empirical risk minimization (ERM). Our analyses are inspired by the recent findings that the effectiveness of knowledge distillation hinges on the teacher\u2019s capability to approximate the true label distribution of training inputs. We theoretically establish that ERM minimizer can approximate the true label distribution of training data as long as the feature extractor of the learner network is Lipschitz continuous and is robust to feature transformations. In light of our theory, we propose a teacher training method SoTeacher which incorporates Lipschitz regularization and consistency regularization into ERM. Experiments on benchmark datasets using various knowledge distillation algorithms and teacher-student pairs confirm that SoTeacher can improve student accuracy consistently.", "pdf": "https://openreview.net/pdf/ce44a21c493bb4865d221811484ede3d170750a4.pdf"} {"title": "Language Models Represent Space and Time", "url": "https://openreview.net/forum?id=jE8xbmvFin", "detail_url": "https://openreview.net/forum?id=jE8xbmvFin", "authors": "Wes Gurnee,Max Tegmark", "tags": "ICLR 2024,Poster", "abstract": "The capabilities of large language models (LLMs) have sparked debate over whether such systems just learn an enormous collection of superficial statistics or a set of more coherent and grounded representations that reflect the real world. We find evidence for the latter by analyzing the learned representations of three spatial datasets (world, US, NYC places) and three temporal datasets (historical figures, artworks, news headlines) in the Llama-2 family of models. We discover that LLMs learn linear representations of space and time across multiple scales. These representations are robust to prompting variations and unified across different entity types (e.g. cities and landmarks). In addition, we identify individual \"space neurons\" and \"time neurons\" that reliably encode spatial and temporal coordinates. While further investigation is needed, our results suggest modern LLMs learn rich spatiotemporal representations of the real world and possess basic ingredients of a world model.", "pdf": "https://openreview.net/pdf/c50ebbefe4d77434016a143c31d247f30948dd6c.pdf"} {"title": "Causal Modelling Agents: Causal Graph Discovery through Synergising Metadata- and Data-driven Reasoning", "url": "https://openreview.net/forum?id=pAoqRlTBtY", "detail_url": "https://openreview.net/forum?id=pAoqRlTBtY", "authors": "Ahmed Abdulaal,adamos hadjivasiliou,Nina Montana-Brown,Tiantian He,Ayodeji Ijishakin,Ivana Drobnjak,Daniel C. Castro,Daniel C. Alexander", "tags": "ICLR 2024,Poster", "abstract": "Scientific discovery hinges on the effective integration of metadata, which refers to a set of 'cognitive' operations such as determining what information is relevant for inquiry, and data, which encompasses physical operations such as observation and experimentation. This paper introduces the Causal Modelling Agent (CMA), a novel framework that synergizes the metadata-based reasoning capabilities of Large Language Models (LLMs) with the data-driven modelling of Deep Structural Causal Models (DSCMs) for the task of causal discovery. We evaluate the CMA's performance on a number of benchmarks, as well as on the real-world task of modelling the clinical and radiological phenotype of Alzheimer's Disease (AD). Our experimental results indicate that the CMA can outperform previous data-driven or metadata-driven approaches to causal discovery. In our real-world application, we use the CMA to derive new insights into the causal relationships among biomarkers of AD.", "pdf": "https://openreview.net/pdf/62fc3766e10c6f5fa2f2a9b44b46098519f89596.pdf"} {"title": "Fast-ELECTRA for Efficient Pre-training", "url": "https://openreview.net/forum?id=8OBuqbLb8h", "detail_url": "https://openreview.net/forum?id=8OBuqbLb8h", "authors": "Chengyu Dong,Liyuan Liu,Hao Cheng,Jingbo Shang,Jianfeng Gao,Xiaodong Liu", "tags": "ICLR 2024,Poster", "abstract": "ELECTRA pre-trains language models by detecting tokens in a sequence that have been replaced by an auxiliary model. Although ELECTRA offers a significant boost in efficiency, its potential is constrained by the training cost brought by the auxiliary model. Notably, this model, which is jointly trained with the main model, only serves to assist the training of the main model and is discarded post-training. This results in a substantial amount of training cost being expended in vain. To mitigate this issue, we propose Fast-ELECTRA, which leverages an existing language model as the auxiliary model. To construct a learning curriculum for the main model, we smooth its output distribution via temperature scaling following a descending schedule. Our approach rivals the performance of state-of-the-art ELECTRA-style pre-training methods, while significantly eliminating the computation and memory cost brought by the joint training of the auxiliary model. Our method also reduces the sensitivity to hyper-parameters and enhances the pre-training stability.", "pdf": "https://openreview.net/pdf/1d4b13edd818d04501c0e1edf4751b54a5858f09.pdf"} {"title": "Maximum Entropy Model Correction in Reinforcement Learning", "url": "https://openreview.net/forum?id=kNpSUN0uCc", "detail_url": "https://openreview.net/forum?id=kNpSUN0uCc", "authors": "Amin Rakhsha,Mete Kemertas,Mohammad Ghavamzadeh,Amir-massoud Farahmand", "tags": "ICLR 2024,Poster", "abstract": "We propose and theoretically analyze an approach for planning with an approximate model in reinforcement learning that can reduce the adverse impact of model error. If the model is accurate enough, it accelerates the convergence to the true value function too. One of its key components is the MaxEnt Model Correction (MoCo) procedure that corrects the model\u2019s next-state distributions based on a Maximum Entropy density estimation formulation. Based on MoCo, we introduce the Model Correcting Value Iteration (MoCoVI) algorithm, and its sampled-based variant MoCoDyna. We show that MoCoVI and MoCoDyna\u2019s convergence can be much faster than the conventional model-free algorithms. Unlike traditional model-based algorithms, MoCoVI and MoCoDyna effectively utilize an approximate model and still converge to the correct value function.", "pdf": "https://openreview.net/pdf/1f91adc5c8d10f07321994671b62ab5b8ced10cb.pdf"} {"title": "SpaCE: The Spatial Confounding Environment", "url": "https://openreview.net/forum?id=D9rJdtmIG6", "detail_url": "https://openreview.net/forum?id=D9rJdtmIG6", "authors": "Mauricio Tec,Ana Trisovic,Michelle Audirac,Sophie Mirabai Woodward,Jie Kate Hu,Naeem Khoshnevis,Francesca Dominici", "tags": "ICLR 2024,Poster", "abstract": "Spatial confounding poses a significant challenge in scientific studies involving spatial data, where unobserved spatial variables can influence both treatment and outcome, possibly leading to spurious associations. To address this problem, we introduce SpaCE: The Spatial Confounding Environment, the first toolkit to provide realistic benchmark datasets and tools for systematically evaluating causal inference methods designed to alleviate spatial confounding. Each dataset includes training data, true counterfactuals, a spatial graph with coordinates, and smoothness and confounding scores characterizing the effect of a missing spatial confounder. It also includes realistic semi-synthetic outcomes and counterfactuals, generated using state-of-the-art machine learning ensembles, following best practices for causal inference benchmarks. The datasets cover real treatment and covariates from diverse domains, including climate, health and social sciences. SpaCE facilitates an automated end-to-end pipeline, simplifying data loading, experimental setup, and evaluating machine learning and causal inference models. The SpaCE project provides several dozens of datasets of diverse sizes and spatial complexity. It is publicly available as a Python package, encouraging community feedback and contributions.", "pdf": "https://openreview.net/pdf/eeea8a3c3f7c7a89a04d83b01ce69520da74f097.pdf"} {"title": "Language Model Detectors Are Easily Optimized Against", "url": "https://openreview.net/forum?id=4eJDMjYZZG", "detail_url": "https://openreview.net/forum?id=4eJDMjYZZG", "authors": "Charlotte Nicks,Eric Mitchell,Rafael Rafailov,Archit Sharma,Christopher D Manning,Chelsea Finn,Stefano Ermon", "tags": "ICLR 2024,Poster", "abstract": "The fluency and general applicability of large language models (LLMs) has motivated significant interest in detecting whether a piece of text was written by a language model. While both academic and commercial detectors have been deployed in some settings, particularly education, other research has highlighted the fragility of these systems. In this paper, we demonstrate a data-efficient attack that fine-tunes language models to confuse existing detectors, leveraging recent developments in reinforcement learning of language models. We use the `human-ness' score (often just a log probability) of various open-source and commercial detectors as a reward function for reinforcement learning, subject to a KL-divergence constraint that the resulting model does not differ significantly from the original. For a 7B parameter Llama-2 model, fine-tuning for under a day reduces the AUROC of the OpenAI RoBERTa-Large detector from 0.84 to 0.63, while perplexity on OpenWebText increases from 8.7 to only 9.0; with a larger perplexity budget, we can drive AUROC to 0.30 (worse than random). Similar to traditional adversarial attacks, we find that this increase in 'detector evasion' generalizes to other detectors not used during training. In light of our empirical results, we advise against continued reliance on LLM-generated text detectors. Models, datasets, and selected experiment code will be released at https://github.com/charlottttee/llm-detector-evasion.", "pdf": "https://openreview.net/pdf/9a04cc3f1effc953fdd1e29092804ea28ce0eb7f.pdf"} {"title": "Zero-Shot Robotic Manipulation with Pre-Trained Image-Editing Diffusion Models", "url": "https://openreview.net/forum?id=c0chJTSbci", "detail_url": "https://openreview.net/forum?id=c0chJTSbci", "authors": "Kevin Black,Mitsuhiko Nakamoto,Pranav Atreya,Homer Rich Walke,Chelsea Finn,Aviral Kumar,Sergey Levine", "tags": "ICLR 2024,Poster", "abstract": "If generalist robots are to operate in truly unstructured environments, they need to be able to recognize and reason about novel objects and scenarios. Such objects and scenarios might not be present in the robot\u2019s own training data. We propose SuSIE, a method that leverages an image-editing diffusion model to act as a high-level planner by proposing intermediate subgoals that a low-level controller can accomplish. Specifically, we finetune InstructPix2Pix on video data, consisting of both human videos and robot rollouts, such that it outputs hypothetical future \u201csubgoal\u201d observations given the robot\u2019s current observation and a language command. We also use the robot data to train a low-level goal-conditioned policy to act as the aforementioned low-level controller. We find that the high-level subgoal predictions can utilize Internet scale pretraining and visual understanding to guide the low-level goal-conditioned policy, achieving significantly better generalization and precision than conventional language-conditioned policies. We achieve state-of-the-art results on the CALVIN benchmark, and also demonstrate robust generalization on real-world manipulation tasks, beating strong baselines that have access to privileged information or that utilize orders of magnitude more compute and training data. The project website can be found at http://rail-berkeley.github.io/susie.", "pdf": "https://openreview.net/pdf/c0ae16a0a57aa4ec4f933f90e44a5e9f250f076e.pdf"} {"title": "Simple Hierarchical Planning with Diffusion", "url": "https://openreview.net/forum?id=kXHEBK9uAY", "detail_url": "https://openreview.net/forum?id=kXHEBK9uAY", "authors": "Chang Chen,Fei Deng,Kenji Kawaguchi,Caglar Gulcehre,Sungjin Ahn", "tags": "ICLR 2024,Poster", "abstract": "Diffusion-based generative methods have proven effective in modeling trajectories with offline datasets. However, they often face computational challenges and can falter in generalization, especially in capturing temporal abstractions for long-horizon tasks. To overcome this, we introduce the Hierarchical Diffuser, a simple, fast, yet effective planning method combining the advantages of hierarchical and diffusion-based planning. Our model adopts a \u201cjumpy\u201d planning strategy at the high level, which allows it to have a larger receptive field but at a lower computational cost\u2014a crucial factor for diffusion-based planning methods, as we have empirically verified. Additionally, the jumpy sub-goals guide our low-level planner, facilitating a fine-tuning stage and further improving our approach\u2019s effectiveness. We conducted empirical evaluations on standard offline reinforcement learning benchmarks, demonstrating our method\u2019s superior performance and efficiency in terms of training and planning speed compared to the non-hierarchical Diffuser as well as other hierarchical planning methods. Moreover, we explore our model\u2019s generalization capability, particularly on how our method improves generalization capabilities on compositional out-of-distribution tasks.", "pdf": "https://openreview.net/pdf/cd376c92489e21ca9086764bc0ac0d95877b8ad5.pdf"} {"title": "Stochastic Gradient Descent for Gaussian Processes Done Right", "url": "https://openreview.net/forum?id=fj2E5OcLFn", "detail_url": "https://openreview.net/forum?id=fj2E5OcLFn", "authors": "Jihao Andreas Lin,Shreyas Padhy,Javier Antoran,Austin Tripp,Alexander Terenin,Csaba Szepesvari,Jos\u00e9 Miguel Hern\u00e1ndez-Lobato,David Janz", "tags": "ICLR 2024,Poster", "abstract": "As is well known, both sampling from the posterior and computing the mean of the posterior in Gaussian process regression reduces to solving a large linear system of equations. We study the use of stochastic gradient descent for solving this linear system, and show that when done right---by which we mean using specific insights from the optimisation and kernel communities---stochastic gradient descent is highly effective. To that end, we introduce a particularly simple stochastic dual descent algorithm, explain its design in an intuitive manner and illustrate the design choices through a series of ablation studies. Further experiments demonstrate that our new method is highly competitive. In particular, our evaluations on the UCI regression tasks and on Bayesian optimisation set our approach apart from preconditioned conjugate gradients and variational Gaussian process approximations. Moreover, our method places Gaussian process regression on par with state-of-the-art graph neural networks for molecular binding affinity prediction.", "pdf": "https://openreview.net/pdf/1a24ae39cc44caaeb65f4c46067a7b5c53a0ed95.pdf"} {"title": "GAFormer: Enhancing Timeseries Transformers Through Group-Aware Embeddings", "url": "https://openreview.net/forum?id=c56TWtYp0W", "detail_url": "https://openreview.net/forum?id=c56TWtYp0W", "authors": "Jingyun Xiao,Ran Liu,Eva L Dyer", "tags": "ICLR 2024,Poster", "abstract": "Analyzing multivariate time series is important in many domains. However, it has been difficult to learn robust and generalizable representations within multivariate datasets due to complex inter-channel relationships and dynamic shifts. In this paper, we introduce a novel approach for learning spatiotemporal structure and using it to improve the application of transformers to timeseries datasets. Our framework learns a set of group tokens, and builds an instance-specific group embedding (GE) layer that assigns input tokens to a small number of group tokens to incorporate structure into learning. We then introduce a novel architecture, Group-Aware transFormer (GAFormer), which incorporates both spatial and temporal group embeddings to achieve state-of-the-art performance on a number of time-series classification and regression tasks. In evaluations on a number of diverse timeseries datasets, we show that GE on its own can provide a nice enhancement to a number of backbones, and that by coupling spatial and temporal group embeddings, the GAFormer can outperform the existing baselines. Finally, we show how our approach discerns latent structures in data even without information about the spatial ordering of channels, and yields a more interpretable decomposition of spatial and temporal structure underlying complex multivariate datasets.", "pdf": "https://openreview.net/pdf/c6fe3e477e52b832a4b26eaa7a2d211c301b44b7.pdf"} {"title": "Why is SAM Robust to Label Noise?", "url": "https://openreview.net/forum?id=3aZCPl3ZvR", "detail_url": "https://openreview.net/forum?id=3aZCPl3ZvR", "authors": "Christina Baek,J Zico Kolter,Aditi Raghunathan", "tags": "ICLR 2024,Poster", "abstract": "Sharpness-Aware Minimization (SAM) is most known for achieving state-of the-art performances on natural image and language tasks. However, its most pronounced improvements (of tens of percent) is rather in the presence of label noise. Understanding SAM's label noise robustness requires a departure from characterizing the robustness of minimas lying in ``flatter'' regions of the loss landscape. In particular, the peak performance under label noise occurs with early stopping, far before the loss converges. We decompose SAM's robustness into two effects: one induced by changes to the logit term and the other induced by changes to the network Jacobian. The first can be observed in linear logistic regression where SAM provably up-weights the gradient contribution from clean examples. Although this explicit up-weighting is also observable in neural networks, when we intervene and modify SAM to remove this effect, surprisingly, we see no visible degradation in performance. We infer that SAM's effect in deeper networks is instead explained entirely by the effect SAM has on the network Jacobian. We theoretically derive the implicit regularization induced by this Jacobian effect in two layer linear networks. Motivated by our analysis, we see that cheaper alternatives to SAM that explicitly induce these regularization effects largely recover the benefits in deep networks trained on real-world datasets.", "pdf": "https://openreview.net/pdf/71206b659a568dfd25c11ddc958dfbf262274392.pdf"} {"title": "Revisiting the Last-Iterate Convergence of Stochastic Gradient Methods", "url": "https://openreview.net/forum?id=xxaEhwC1I4", "detail_url": "https://openreview.net/forum?id=xxaEhwC1I4", "authors": "Zijian Liu,Zhengyuan Zhou", "tags": "ICLR 2024,Poster", "abstract": "In the past several years, the last-iterate convergence of the Stochastic Gradient Descent (SGD) algorithm has triggered people's interest due to its good performance in practice but lack of theoretical understanding. For Lipschitz convex functions, different works have established the optimal $O(\\log(1/\\delta)\\log T/\\sqrt{T})$ or $O(\\sqrt{\\log(1/\\delta)/T})$ high-probability convergence rates for the final iterate, where $T$ is the time horizon and $\\delta$ is the failure probability. However, to prove these bounds, all the existing works are either limited to compact domains or require almost surely bounded noises. It is natural to ask whether the last iterate of SGD can still guarantee the optimal convergence rate but without these two restrictive assumptions. Besides this important question, there are still lots of theoretical problems lacking an answer. For example, compared with the last-iterate convergence of SGD for non-smooth problems, only few results for smooth optimization have yet been developed. Additionally, the existing results are all limited to a non-composite objective and the standard Euclidean norm. It still remains unclear whether the last-iterate convergence can be provably extended to wider composite optimization and non-Euclidean norms. In this work, to address the issues mentioned above, we revisit the last-iterate convergence of stochastic gradient methods and provide the first unified way to prove the convergence rates both in expectation and in high probability to accommodate general domains, composite objectives, non-Euclidean norms, Lipschitz conditions, smoothness, and (strong) convexity simultaneously.", "pdf": "https://openreview.net/pdf/49e36604c5405004e38defe39ca3ff6ecf070ca6.pdf"} {"title": "CNN Kernels Can Be the Best Shapelets", "url": "https://openreview.net/forum?id=O8ouVV8PjF", "detail_url": "https://openreview.net/forum?id=O8ouVV8PjF", "authors": "Eric Qu,Yansen Wang,Xufang Luo,Wenqiang He,Kan Ren,Dongsheng Li", "tags": "ICLR 2024,Poster", "abstract": "Shapelets and CNN are two typical approaches to model time series. Shapelets aim at finding a set of sub-sequences that extract feature-based interpretable shapes, but may suffer from accuracy and efficiency issues. CNN performs well by encoding sequences with a series of hidden representations, but lacks interpretability. In this paper, we demonstrate that shapelets are essentially equivalent to a specific type of CNN kernel with a squared norm and pooling. Based on this finding, we propose ShapeConv, an interpretable CNN layer with its kernel serving as shapelets to conduct time-series modeling tasks in both supervised and unsupervised settings. By incorporating shaping regularization, we enforce the similarity for maximum interpretability. We also find human knowledge can be easily injected to ShapeConv by adjusting its initialization and model performance is boosted with it. Experiments show that ShapeConv can achieve state-of-the-art performance on time-series benchmarks without sacrificing interpretability and controllability.", "pdf": "https://openreview.net/pdf/69b308cc2c2320f9051d94361939bb8848074ab0.pdf"} {"title": "Fine-Tuning Language Models for Factuality", "url": "https://openreview.net/forum?id=WPZ2yPag4K", "detail_url": "https://openreview.net/forum?id=WPZ2yPag4K", "authors": "Katherine Tian,Eric Mitchell,Huaxiu Yao,Christopher D Manning,Chelsea Finn", "tags": "ICLR 2024,Poster", "abstract": "The fluency and creativity of large pre-trained language models (LLMs) have led to their widespread use, sometimes even as a replacement for traditional search engines. Yet language models are prone to making convincing but factually inaccurate claims, often referred to as `hallucinations.' These errors can inadvertently spread misinformation or harmfully perpetuate misconceptions. Further, manual fact-checking of model responses is a time-consuming process, making human factuality labels expensive to acquire. In this work, we fine-tune language models to be more factual, without human labeling and targeting more open-ended generation settings than past work. We leverage two key recent innovations in NLP to do so. First, several recent works have proposed methods for judging the factuality of open-ended text by measuring consistency with an external knowledge base or simply a large model's confidence scores. Second, the Direct Preference Optimization algorithm enables straightforward fine-tuning of language models on objectives other than supervised imitation, using a preference ranking over possible model responses. We show that learning from automatically generated factuality preference rankings, generated either through existing retrieval systems or our novel retrieval-free approach, significantly improves the factuality (percent of generated claims that are correct) of Llama-2 on held-out topics compared with RLHF or decoding strategies targeted at factuality. At 7B scale, compared to Llama-2-Chat, we observe 53% and 50% reduction in factual error rate when generating biographies and answering medical questions, respectively. A reference implementation can be found at https://github.com/kttian/llm_factuality_tuning.", "pdf": "https://openreview.net/pdf/f90a225c9859565a8e1ed01840ea046b406c7d4f.pdf"} {"title": "Soft Robust MDPs and Risk-Sensitive MDPs: Equivalence, Policy Gradient, and Sample Complexity", "url": "https://openreview.net/forum?id=dEz3ge8QSo", "detail_url": "https://openreview.net/forum?id=dEz3ge8QSo", "authors": "Runyu Zhang,Yang Hu,Na Li", "tags": "ICLR 2024,Poster", "abstract": "Robust Markov Decision Processes (MDPs) and risk-sensitive MDPs are both powerful tools for making decisions in the presence of uncertainties. Previous efforts have aimed to establish their connections, revealing equivalences in specific formulations. This paper introduces a new formulation for risk-sensitive MDPs, which assesses risk in a slightly different manner compared to the classical Markov risk measure [Ruszczy \u0301nski 2010], and establishes its equivalence with a class of soft robust MDP (RMDP) problems, including the standard RMDP as a special case. Leveraging this equivalence, we further derive the policy gradient theorem for both problems, proving gradient domination and global convergence of the exact policy gradient method under the tabular setting with direct parameterization. This forms a sharp contrast to the Markov risk measure, known to be potentially non-gradient-dominant [Huang et al. 2021]. We also propose a sample-based offline learning algorithm, namely the robust fitted-Z iteration (RFZI), for a specific soft RMDP problem with a KL-divergence regularization term (or equivalently the risk-sensitive MDP with an entropy risk measure). We showcase its streamlined\ndesign and less stringent assumptions due to the equivalence and analyze its sample complexity.", "pdf": "https://openreview.net/pdf/317b11f5d5ce4a86be220bbd6715b66f4a55103a.pdf"} {"title": "Tensor Programs VI: Feature Learning in Infinite Depth Neural Networks", "url": "https://openreview.net/forum?id=17pVDnpwwl", "detail_url": "https://openreview.net/forum?id=17pVDnpwwl", "authors": "Greg Yang,Dingli Yu,Chen Zhu,Soufiane Hayou", "tags": "ICLR 2024,Poster", "abstract": "Empirical studies have consistently demonstrated that increasing the size of neural networks often yields superior performance in practical applications. However, there is a lack of consensus regarding the appropriate scaling strategy, particularly when it comes to increasing the depth of neural networks. In practice, excessively large depths can lead to model performance degradation. In this paper, we introduce Depth-$\\mu$P, a principled approach for depth scaling, allowing for the training of arbitrarily deep architectures while maximizing feature learning and diversity among nearby layers. Our method involves dividing the contribution of each residual block and the parameter update by the square root of the depth. Through the use of Tensor Programs, we rigorously establish the existence of a limit for infinitely deep neural networks under the proposed scaling scheme. This scaling strategy ensures more stable training for deep neural networks and guarantees the transferability of hyperparameters from shallow to deep models. To substantiate the efficacy of our scaling method, we conduct empirical validation on neural networks with depths up to $2^{10}$.", "pdf": "https://openreview.net/pdf/a2d69b77708b87f741baac0303581cfb7924d0b7.pdf"} {"title": "Demystifying Poisoning Backdoor Attacks from a Statistical Perspective", "url": "https://openreview.net/forum?id=BPHcEpGvF8", "detail_url": "https://openreview.net/forum?id=BPHcEpGvF8", "authors": "Ganghua Wang,Xun Xian,Ashish Kundu,Jayanth Srinivasa,Xuan Bi,Mingyi Hong,Jie Ding", "tags": "ICLR 2024,Poster", "abstract": "Backdoor attacks pose a significant security risk to machine learning applications due to their stealthy nature and potentially serious consequences. Such attacks involve embedding triggers within a learning model with the intention of causing malicious behavior when an active trigger is present while maintaining regular functionality without it. This paper derives a fundamental understanding of backdoor attacks that applies to both discriminative and generative models, including diffusion models and large language models. We evaluate the effectiveness of any backdoor attack incorporating a constant trigger, by establishing tight lower and upper boundaries for the performance of the compromised model on both clean and backdoor test data. The developed theory answers a series of fundamental but previously underexplored problems, including (1) what are the determining factors for a backdoor attack's success, (2) what is the direction of the most effective backdoor attack, and (3) when will a human-imperceptible trigger succeed. We demonstrate the theory by conducting experiments using benchmark datasets and state-of-the-art backdoor attack scenarios. Our code is available \\href{https://github.com/KeyWgh/DemystifyBackdoor}{here}.", "pdf": "https://openreview.net/pdf/2b97a4885cd767d5e6fad5ceeb1c8e5da20147c4.pdf"} {"title": "Learning to Make Adherence-aware Advice", "url": "https://openreview.net/forum?id=RgELE1dQXx", "detail_url": "https://openreview.net/forum?id=RgELE1dQXx", "authors": "Guanting Chen,Xiaocheng Li,Chunlin Sun,Hanzhao Wang", "tags": "ICLR 2024,Poster", "abstract": "As artificial intelligence (AI) systems play an increasingly prominent role in human decision-making, challenges surface in the realm of human-AI interactions. One challenge arises from the suboptimal AI policies due to the inadequate consideration of humans disregarding AI recommendations, as well as the need for AI to provide advice selectively when it is most pertinent. This paper presents a sequential decision-making model that (i) takes into account the human's adherence level (the probability that the human follows/rejects machine advice) and (ii) incorporates a defer option so that the machine can temporarily refrain from making advice. We provide learning algorithms that learn the optimal advice policy and make advice only at critical time stamps. Compared to problem-agnostic reinforcement learning algorithms, our specialized learning algorithms not only enjoy better theoretical convergence properties but also show strong empirical performance.", "pdf": "https://openreview.net/pdf/23fc1fd51c338383a74e3c5989a6dcd7a273a1c0.pdf"} {"title": "Beyond Spatio-Temporal Representations: Evolving Fourier Transform for Temporal Graphs", "url": "https://openreview.net/forum?id=uvFhCUPjtI", "detail_url": "https://openreview.net/forum?id=uvFhCUPjtI", "authors": "Anson Bastos,Kuldeep Singh,Abhishek Nadgeri,Manish Singh,Toyotaro Suzumura", "tags": "ICLR 2024,Poster", "abstract": "We present the Evolving Graph Fourier Transform (EFT), the first invertible spectral transform that captures evolving representations on temporal graphs. We motivate our work by the inadequacy of existing methods for capturing the evolving graph spectra, which are also computationally expensive due to the temporal aspect along with the graph vertex domain. We view the problem as an optimization over the Laplacian of the continuous time dynamic graph. Additionally, we propose pseudo-spectrum relaxations that decompose the transformation process, making it highly computationally efficient. The EFT method adeptly captures the evolving graph's structural and positional properties, making it effective for downstream tasks on evolving graphs. Hence, as a reference implementation, we develop a simple neural model induced with \\eft for capturing evolving graph spectra. We empirically validate our theoretical findings on a number of large-scale and standard temporal graph benchmarks and demonstrate that our model achieves state-of-the-art performance.", "pdf": "https://openreview.net/pdf/5386623c7dfc54fe602556b341b906eb0ec58d06.pdf"} {"title": "Cycle Consistency Driven Object Discovery", "url": "https://openreview.net/forum?id=f1xnBr4WD6", "detail_url": "https://openreview.net/forum?id=f1xnBr4WD6", "authors": "Aniket Rajiv Didolkar,Anirudh Goyal,Yoshua Bengio", "tags": "ICLR 2024,Poster", "abstract": "Developing deep learning models that effectively learn object-centric representations, akin to human cognition, remains a challenging task. Existing approaches facilitate object discovery by representing objects as fixed-size vectors, called ``slots'' or ``object files''. While these approaches have shown promise in certain scenarios, they still exhibit certain limitations. First, they rely on architectural priors which can be unreliable and usually require meticulous engineering to identify the correct objects. Second, there has been a notable gap in investigating the practical utility of these representations in downstream tasks. To address the first limitation, we introduce a method that explicitly optimizes the constraint that each object in a scene should be associated with a distinct slot. We formalize this constraint by introducing consistency objectives which are cyclic in nature. By integrating these consistency objectives into various existing slot-based object-centric methods, we showcase substantial improvements in object-discovery performance. These enhancements consistently hold true across both synthetic and real-world scenes, underscoring the effectiveness and adaptability of the proposed approach. To tackle the second limitation, we apply the learned object-centric representations from the proposed method to two downstream reinforcement learning tasks, demonstrating considerable performance enhancements compared to conventional slot-based and monolithic representation learning methods. Our results suggest that the proposed approach not only improves object discovery, but also provides richer features for downstream tasks.", "pdf": "https://openreview.net/pdf/18ace982ecbf580ad919f876edd9b3a6e1652550.pdf"} {"title": "Sufficient conditions for offline reactivation in recurrent neural networks", "url": "https://openreview.net/forum?id=RVrINT6MT7", "detail_url": "https://openreview.net/forum?id=RVrINT6MT7", "authors": "Nanda H Krishna,Colin Bredenberg,Daniel Levenstein,Blake Aaron Richards,Guillaume Lajoie", "tags": "ICLR 2024,Poster", "abstract": "During periods of quiescence, such as sleep, neural activity in many brain circuits resembles that observed during periods of task engagement. However, the precise conditions under which task-optimized networks can autonomously reactivate the same network states responsible for online behavior is poorly understood. In this study, we develop a mathematical framework that outlines sufficient conditions for the emergence of neural reactivation in circuits that encode features of smoothly varying stimuli. We demonstrate mathematically that noisy recurrent networks optimized to track environmental state variables using change-based sensory information naturally develop denoising dynamics, which, in the absence of input, cause the network to revisit state configurations observed during periods of online activity. We validate our findings using numerical experiments on two canonical neuroscience tasks: spatial position estimation based on self-motion cues, and head direction estimation based on angular velocity cues. Overall, our work provides theoretical support for modeling offline reactivation as an emergent consequence of task optimization in noisy neural circuits.", "pdf": "https://openreview.net/pdf/70a1d8b98f869d61b787855ba26e719054884ab3.pdf"} {"title": "Forward Learning of Graph Neural Networks", "url": "https://openreview.net/forum?id=Abr7dU98ME", "detail_url": "https://openreview.net/forum?id=Abr7dU98ME", "authors": "Namyong Park,Xing Wang,Antoine Simoulin,Shuai Yang,Grey Yang,Ryan A. Rossi,Puja Trivedi,Nesreen K. Ahmed", "tags": "ICLR 2024,Poster", "abstract": "Graph neural networks (GNNs) have achieved remarkable success across a wide range of applications, such as recommendation, drug discovery, and question answering. Behind the success of GNNs lies the backpropagation (BP) algorithm, which is the de facto standard for training deep neural networks (NNs). However, despite its effectiveness, BP imposes several constraints, which are not only biologically implausible, but also limit the scalability, parallelism, and flexibility in learning NNs. Examples of such constraints include storage of neural activities computed in the forward pass for use in the subsequent backward pass, and the dependence of parameter updates on non-local signals. To address these limitations, the forward-forward algorithm (FF) was recently proposed as an alternative to BP in the image classification domain, which trains NNs by performing two forward passes over positive and negative data. Inspired by this advance, we propose ForwardGNN in this work, a new forward learning procedure for GNNs, which avoids the constraints imposed by BP via an effective layer-wise local forward training. ForwardGNN extends the original FF to deal with graph data and GNNs, and makes it possible to operate without generating negative inputs (hence no longer forward-forward). Further, ForwardGNN enables each layer to learn from both the bottom-up and top-down signals without relying on the backpropagation of errors. Extensive experiments on real-world datasets show the effectiveness and generality of the proposed forward graph learning framework. We release our code at https://github.com/facebookresearch/forwardgnn.", "pdf": "https://openreview.net/pdf/78ce77aec3cb18418df9c216e801999677415163.pdf"} {"title": "Curriculum reinforcement learning for quantum architecture search under hardware errors", "url": "https://openreview.net/forum?id=rINBD8jPoP", "detail_url": "https://openreview.net/forum?id=rINBD8jPoP", "authors": "Yash J. Patel,Akash Kundu,Mateusz Ostaszewski,Xavier Bonet-Monroig,Vedran Dunjko,Onur Danaci", "tags": "ICLR 2024,Poster", "abstract": "The key challenge in the noisy intermediate-scale quantum era is finding useful circuits compatible with current device limitations.\nVariational quantum algorithms (VQAs) offer a potential solution by fixing the circuit architecture and optimizing individual gate parameters in an external loop. However, parameter optimization can become intractable, and the overall performance of the algorithm depends heavily on the initially chosen circuit architecture. Several quantum architecture search (QAS) algorithms have been developed to design useful circuit architectures automatically. In the case of parameter optimization alone, noise effects have been observed to dramatically influence the performance of the optimizer and final outcomes, which is a key line of study. However, the effects of noise on the architecture search, which could be just as critical, are poorly understood. This work addresses this gap by introducing a curriculum-based reinforcement learning QAS (CRLQAS) algorithm designed to tackle challenges in realistic VQA deployment. The algorithm incorporates (i) a 3D architecture encoding and restrictions on environment dynamics to explore the search space of possible circuits efficiently, (ii) an episode halting scheme to steer the agent to find shorter circuits, and (iii) a novel variant of simultaneous perturbation stochastic approximation as an optimizer for faster convergence. To facilitate studies, we developed an optimized simulator for our algorithm, significantly improving computational efficiency in simulating noisy quantum circuits by employing the Pauli-transfer matrix formalism in the Pauli-Liouville basis. Numerical experiments focusing on quantum chemistry tasks demonstrate that CRLQAS outperforms existing QAS algorithms across several metrics in both noiseless and noisy environments.", "pdf": "https://openreview.net/pdf/e6b75e0b94d3c31e7f0e5d1c7d11b4ccb1aca361.pdf"} {"title": "Does CLIP\u2019s generalization performance mainly stem from high train-test similarity?", "url": "https://openreview.net/forum?id=tnBaiidobu", "detail_url": "https://openreview.net/forum?id=tnBaiidobu", "authors": "Prasanna Mayilvahanan,Thadd\u00e4us Wiedemer,Evgenia Rusak,Matthias Bethge,Wieland Brendel", "tags": "ICLR 2024,Poster", "abstract": "Foundation models like CLIP are trained on hundreds of millions of samples and effortlessly generalize to new tasks and inputs. Out of the box, CLIP shows stellar zero-shot and few-shot capabilities on a wide range of out-of-distribution (OOD) benchmarks, which prior works attribute mainly to today's large and comprehensive training dataset (like LAION). However, it is questionable how meaningful terms like out-of-distribution generalization are for CLIP as it seems likely that web-scale datasets like LAION simply contain many samples that are similar to common OOD benchmarks originally designed for ImageNet. To test this hypothesis, we retrain CLIP on pruned LAION splits that replicate ImageNet\u2019s train-test similarity with respect to common OOD benchmarks. While we observe a performance drop on some benchmarks, surprisingly, CLIP\u2019s overall performance remains high. This shows that high train-test similarity is insufficient to explain CLIP\u2019s OOD performance, and other properties of the training data must drive CLIP to learn more generalizable representations. Additionally, by pruning data points that are dissimilar to the OOD benchmarks, we uncover a 100M split of LAION (\u00bc of its original size) on which CLIP can be trained to match its original OOD performance.", "pdf": "https://openreview.net/pdf/914ba616ab5450c89e489fa002bc6f6587152c84.pdf"} {"title": "Unified Projection-Free Algorithms for Adversarial DR-Submodular Optimization", "url": "https://openreview.net/forum?id=H4A9e8HvIn", "detail_url": "https://openreview.net/forum?id=H4A9e8HvIn", "authors": "Mohammad Pedramfar,Yididiya Y. Nadew,Christopher John Quinn,Vaneet Aggarwal", "tags": "ICLR 2024,Poster", "abstract": "This paper introduces unified projection-free Frank-Wolfe type algorithms for adversarial continuous DR-submodular optimization, spanning scenarios such as full information and (semi-)bandit feedback, monotone and non-monotone functions, different constraints, and types of stochastic queries. For every problem considered in the non-monotone setting, the proposed algorithms are either the first with proven sub-linear $\\alpha$-regret bounds or have better $\\alpha$-regret bounds than the state of the art, where $\\alpha$ is a corresponding approximation bound in the offline setting. In the monotone setting, the proposed approach gives state-of-the-art sub-linear $\\alpha$-regret bounds among projection-free algorithms in 7 of the 8 considered cases while matching the result of the remaining case. Additionally, this paper addresses semi-bandit and bandit feedback for adversarial DR-submodular optimization, advancing the understanding of this optimization area.", "pdf": "https://openreview.net/pdf/c8701ec65ea9e2a9ef05a94e26ebcad8d0d3e58f.pdf"} {"title": "When Scaling Meets LLM Finetuning: The Effect of Data, Model and Finetuning Method", "url": "https://openreview.net/forum?id=5HCnKDeTws", "detail_url": "https://openreview.net/forum?id=5HCnKDeTws", "authors": "Biao Zhang,Zhongtao Liu,Colin Cherry,Orhan Firat", "tags": "ICLR 2024,Poster", "abstract": "While large language models (LLMs) often adopt finetuning to unlock their capabilities for downstream applications, our understanding on the inductive biases (especially the scaling properties) of different finetuning methods is still limited. To fill this gap, we conduct systematic experiments studying whether and how different scaling factors, including LLM model size, pretraining data size, new finetuning parameter size and finetuning data size, affect the finetuning performance. We consider two types of finetuning \u2013 full-model tuning (FMT) and parameter efficient tuning (PET, including prompt tuning and LoRA), and explore their scaling behaviors in the data-limited regime where the LLM model size substantially outweighs the finetuning data size. Based on two sets of pretrained bilingual LLMs from 1B to 16B and experiments on bilingual machine translation and multilingual summarization benchmarks, we find that 1) LLM finetuning follows a powerbased multiplicative joint scaling law between finetuning data size and each other scaling factor; 2) LLM finetuning benefits more from LLM model scaling than pretraining data scaling, and PET parameter scaling is generally ineffective; and 3) the optimal finetuning method is highly task- and finetuning data-dependent. We hope our findings could shed light on understanding, selecting and developing LLM finetuning methods.", "pdf": "https://openreview.net/pdf/c50285d47fae2fec5f5aa87de6d9e6a921de02b9.pdf"} {"title": "Learning to design protein-protein interactions with enhanced generalization", "url": "https://openreview.net/forum?id=xcMmebCT7s", "detail_url": "https://openreview.net/forum?id=xcMmebCT7s", "authors": "Anton Bushuiev,Roman Bushuiev,Petr Kouba,Anatolii Filkin,Marketa Gabrielova,Michal Gabriel,Jiri Sedlar,Tomas Pluskal,Jiri Damborsky,Stanislav Mazurenko,Josef Sivic", "tags": "ICLR 2024,Poster", "abstract": "Discovering mutations enhancing protein-protein interactions (PPIs) is critical for advancing biomedical research and developing improved therapeutics. While machine learning approaches have substantially advanced the field, they often struggle to generalize beyond training data in practical scenarios. The contributions of this work are three-fold. First, we construct PPIRef, the largest and non-redundant dataset of 3D protein-protein interactions, enabling effective large-scale learning. Second, we leverage the PPIRef dataset to pre-train PPIformer, a new SE(3)-equivariant model generalizing across diverse protein-binder variants. We fine-tune PPIformer to predict effects of mutations on protein-protein interactions via a thermodynamically motivated adjustment of the pre-training loss function. Finally, we demonstrate the enhanced generalization of our new PPIformer approach by outperforming other state-of-the-art methods on new, non-leaking splits of standard labeled PPI mutational data and independent case studies optimizing a human antibody against SARS-CoV-2 and increasing the thrombolytic activity of staphylokinase.", "pdf": "https://openreview.net/pdf/03bb8ce604b1ff6fb6f9d6504d13322265401a20.pdf"} {"title": "L2MAC: Large Language Model Automatic Computer for Extensive Code Generation", "url": "https://openreview.net/forum?id=EhrzQwsV4K", "detail_url": "https://openreview.net/forum?id=EhrzQwsV4K", "authors": "Samuel Holt,Max Ruiz Luyten,Mihaela van der Schaar", "tags": "ICLR 2024,Poster", "abstract": "Transformer-based large language models (LLMs) are constrained by the fixed context window of the underlying transformer architecture, hindering their ability to produce long and coherent outputs. Memory-augmented LLMs are a promising solution, but current approaches cannot handle long output generation tasks since they (1) only focus on reading memory and reduce its evolution to the concatenation of new memories or (2) use very specialized memories that cannot adapt to other domains. This paper presents L2MAC, the first practical LLM-based general-purpose stored-program automatic computer (von Neumann architecture) framework, an LLM-based multi-agent system, for long and consistent output generation. Its memory has two components: the instruction registry, which is populated with a prompt program to solve the user-given task, and a file store, which will contain the final and intermediate outputs. Each instruction in turn is executed by a separate LLM agent, whose context is managed by a control unit capable of precise memory reading and writing to ensure effective interaction with the entire file store. These components enable L2MAC to generate extensive outputs, bypassing the constraints of the finite context window while producing outputs that fulfill a complex user-specified task. We empirically demonstrate that L2MAC achieves state-of-the-art performance in generating large codebases for system design tasks, significantly outperforming other coding methods in implementing the detailed user-specified task; we show that L2MAC works for general-purpose extensive text-based tasks, such as writing an entire book; and we provide valuable insights into L2MAC's performance improvement over existing methods.", "pdf": "https://openreview.net/pdf/85b60ddd90be827351cfe094c4e8fbcd35267ffe.pdf"} {"title": "BadChain: Backdoor Chain-of-Thought Prompting for Large Language Models", "url": "https://openreview.net/forum?id=c93SBwz1Ma", "detail_url": "https://openreview.net/forum?id=c93SBwz1Ma", "authors": "Zhen Xiang,Fengqing Jiang,Zidi Xiong,Bhaskar Ramasubramanian,Radha Poovendran,Bo Li", "tags": "ICLR 2024,Poster", "abstract": "Large language models (LLMs) are shown to benefit from chain-of-thought (COT) prompting, particularly when tackling tasks that require systematic reasoning processes. On the other hand, COT prompting also poses new vulnerabilities in the form of backdoor attacks, wherein the model will output unintended malicious content under specific backdoor-triggered conditions during inference. Traditional methods for launching backdoor attacks involve either contaminating the training dataset with backdoored instances or directly manipulating the model parameters during deployment. However, these approaches are not practical for commercial LLMs that typically operate via API access. In this paper, we propose BadChain, the first backdoor attack against LLMs employing COT prompting, which does not require access to the training dataset or model parameters and imposes low computational overhead. BadChain leverages the inherent reasoning capabilities of LLMs by inserting a backdoor reasoning step into the sequence of reasoning steps of the model output, thereby altering the final response when a backdoor trigger is embedded in the query prompt. In particular, a subset of demonstrations will be manipulated to incorporate a backdoor reasoning step in COT prompting. Consequently, given any query prompt containing the backdoor trigger, the LLM will be misled to output unintended content. Empirically, we show the effectiveness of BadChain for two COT strategies across four LLMs (Llama2, GPT-3.5, PaLM2, and GPT-4) and six complex benchmark tasks encompassing arithmetic, commonsense, and symbolic reasoning. We show that the baseline backdoor attacks designed for simpler tasks such as semantic classification will fail on these complicated tasks. In addition, our findings reveal that LLMs endowed with stronger reasoning capabilities exhibit higher susceptibility to BadChain, exemplified by a high average attack success rate of 97.0\\% across the six benchmark tasks on GPT-4. We also demonstrate the interpretability of BadChain by showing that the relationship between the trigger and the backdoor reasoning step can be well-explained based on the output of the backdoored model. Finally, we propose two defenses based on shuffling and demonstrate their overall ineffectiveness against BadChain. Therefore, BadChain remains a severe threat to LLMs, underscoring the urgency for the development of robust and effective future defenses.", "pdf": "https://openreview.net/pdf/f55f665827c60d9ab1815886945cb4b0fcd9b12b.pdf"} {"title": "NeuroBack: Improving CDCL SAT Solving using Graph Neural Networks", "url": "https://openreview.net/forum?id=samyfu6G93", "detail_url": "https://openreview.net/forum?id=samyfu6G93", "authors": "Wenxi Wang,Yang Hu,Mohit Tiwari,Sarfraz Khurshid,Kenneth McMillan,Risto Miikkulainen", "tags": "ICLR 2024,Poster", "abstract": "Propositional satisfiability (SAT) is an NP-complete problem that impacts many\nresearch fields, such as planning, verification, and security. Mainstream modern\nSAT solvers are based on the Conflict-Driven Clause Learning (CDCL) algorithm.\nRecent work aimed to enhance CDCL SAT solvers using Graph Neural Networks\n(GNNs). However, so far this approach either has not made solving more effective,\nor required substantial GPU resources for frequent online model inferences. Aiming\nto make GNN improvements practical, this paper proposes an approach called\nNeuroBack, which builds on two insights: (1) predicting phases (i.e., values) of\nvariables appearing in the majority (or even all) of the satisfying assignments are\nessential for CDCL SAT solving, and (2) it is sufficient to query the neural model\nonly once for the predictions before the SAT solving starts. Once trained, the\noffline model inference allows NeuroBack to execute exclusively on the CPU,\nremoving its reliance on GPU resources. To train NeuroBack, a new dataset called\nDataBack containing 120,286 data samples is created. Finally, NeuroBack is implemented\nas an enhancement to a state-of-the-art SAT solver called Kissat. As a result,\nit allowed Kissat to solve 5.2% more problems on the recent SAT competition\nproblem set, SATCOMP-2022. NeuroBack therefore shows how machine learning\ncan be harnessed to improve SAT solving in an effective and practical manner.", "pdf": "https://openreview.net/pdf/68668727b395677105cd2bfd38dc554dc5b67212.pdf"} {"title": "Group Preference Optimization: Few-Shot Alignment of Large Language Models", "url": "https://openreview.net/forum?id=DpFeMH4l8Q", "detail_url": "https://openreview.net/forum?id=DpFeMH4l8Q", "authors": "Siyan Zhao,John Dang,Aditya Grover", "tags": "ICLR 2024,Poster", "abstract": "Many applications of large language models (LLMs), ranging from chatbots to\ncreative writing, require nuanced subjective judgments that can differ significantly\nacross different groups. Existing alignment algorithms can be expensive to align\nfor each group, requiring prohibitive amounts of group-specific preference data\nand computation for real-world use cases. We introduce Group Preference Optimization (GPO), an alignment framework that steers language models to preferences of individual groups in a few-shot manner. In GPO, we augment the base\nLLM with an independent transformer module trained to predict the preferences\nof a group for the LLM generations. For few-shot learning, we parameterize this\nmodule as an in-context autoregressive transformer and train it via meta-learning\non several groups. We empirically validate the efficacy of GPO through rigorous evaluations using LLMs with varied sizes on three human opinion adaptation tasks. These tasks involve adapting to the preferences of US demographic\ngroups, global countries, and individual users. Our results demonstrate that GPO\nnot only aligns models more accurately but also requires fewer group-specific\npreferences and less training and inference computing resources, outperforming\nexisting strategies such as in-context steering and fine-tuning methods.", "pdf": "https://openreview.net/pdf/e743356473984605880070de9402840ffe780599.pdf"} {"title": "MERT: Acoustic Music Understanding Model with Large-Scale Self-supervised Training", "url": "https://openreview.net/forum?id=w3YZ9MSlBu", "detail_url": "https://openreview.net/forum?id=w3YZ9MSlBu", "authors": "Yizhi LI,Ruibin Yuan,Ge Zhang,Yinghao Ma,Xingran Chen,Hanzhi Yin,Chenghao Xiao,Chenghua Lin,Anton Ragni,Emmanouil Benetos,Norbert Gyenge,Roger Dannenberg,Ruibo Liu,Wenhu Chen,Gus Xia,Yemin Shi,Wenhao Huang,Zili Wang,Yike Guo,Jie Fu", "tags": "ICLR 2024,Poster", "abstract": "Self-supervised learning (SSL) has recently emerged as a promising paradigm for training generalisable models on large-scale data in the fields of vision, text, and speech. \nAlthough SSL has been proven effective in speech and audio, its application to music audio has yet to be thoroughly explored. This is partially due to the distinctive challenges associated with modelling musical knowledge, particularly tonal and pitched characteristics of music.\nTo address this research gap, we propose an acoustic **M**usic und**ER**standing model with large-scale self-supervised **T**raining (**MERT**), which incorporates teacher models to provide pseudo labels in the masked language modelling (MLM) style acoustic pre-training.\nIn our exploration, we identified an effective combination of teacher models, which outperforms conventional speech and audio approaches in terms of performance. \nThis combination includes an acoustic teacher based on Residual Vector Quantization - Variational AutoEncoder (RVQ-VAE) and a musical teacher based on the Constant-Q Transform (CQT). \nFurthermore, we explore a wide range of settings to overcome the instability in acoustic language model pre-training, which allows our designed paradigm to scale from 95M to 330M parameters.\nExperimental results indicate that our model can generalise and perform well on 14 music understanding tasks and attain state-of-the-art (SOTA) overall scores.", "pdf": "https://openreview.net/pdf/b0691ff4b8bef0f41861ce3bd2ea50491707b93c.pdf"} {"title": "Variance-aware Regret Bounds for Stochastic Contextual Dueling Bandits", "url": "https://openreview.net/forum?id=rDH7dIFn20", "detail_url": "https://openreview.net/forum?id=rDH7dIFn20", "authors": "Qiwei Di,Tao Jin,Yue Wu,Heyang Zhao,Farzad Farnoud,Quanquan Gu", "tags": "ICLR 2024,Poster", "abstract": "Dueling bandits is a prominent framework for decision-making involving preferential feedback, a valuable feature that fits various applications involving human interaction, such as ranking, information retrieval, and recommendation systems. While substantial efforts have been made to minimize the cumulative regret in dueling bandits, a notable gap in the current research is the absence of regret bounds that account for the inherent uncertainty in pairwise comparisons between the dueling arms. Intuitively, greater uncertainty suggests a higher level of difficulty in the problem. To bridge this gap, this paper studies the problem of contextual dueling bandits, where the binary comparison of dueling arms is generated from a generalized linear model (GLM). We propose a new SupLinUCB-type algorithm that enjoys computational efficiency and a variance-aware regret bound $\\tilde O\\big(d\\sqrt{\\sum_{t=1}^T\\sigma_t^2} + d\\big)$, where $\\sigma_t$ is the variance of the pairwise comparison at round $t$, $d$ is the dimension of the context vectors, and $T$ is the time horizon. Our regret bound naturally aligns with the intuitive expectation \u2014 in scenarios where the comparison is deterministic, the algorithm only suffers from an $\\tilde O(d)$ regret. We perform empirical experiments on synthetic data to confirm the advantage of our method over previous variance-agnostic algorithms.", "pdf": "https://openreview.net/pdf/4c1a08aad3975e9de1adbba054eea8e3f1287418.pdf"} {"title": "A Discretization Framework for Robust Contextual Stochastic Optimization", "url": "https://openreview.net/forum?id=ueTdErd5Ib", "detail_url": "https://openreview.net/forum?id=ueTdErd5Ib", "authors": "Rares C Cristian,Georgia Perakis", "tags": "ICLR 2024,Poster", "abstract": "We study contextual stochastic optimization problems. Optimization problems have uncertain parameters stemming from unknown, context-dependent, distributions. Due to the inherent uncertainty in these problems, one is often interested not only in minimizing expected cost, but also to be robust and protect against worst case scenarios. We propose a novel method that combines the learning stage with knowledge of the downstream optimization task. The method prescribes decisions which aim to maximize the likelihood that the cost is below a (user-controlled) threshold. The key idea is (1) to discretize the feasible region into subsets so that the uncertain objective function can be well approximated deterministically within each subset, and (2) devise a secondary optimization problem to prescribe decisions by integrating the individual approximations determined in step (1). We provide theoretical guarantees bounding the underlying regret of decisions proposed by our method. In addition, experimental results demonstrate that our approach is competitive in terms of average regret and yields more robust solutions than other methods proposed in the literature, including up to 20 times lower worst-case cost on a real-world electricity generation problem.", "pdf": "https://openreview.net/pdf/2097641b7d16f85a6d284604d4e690fc32f2e79e.pdf"} {"title": "Risk Bounds of Accelerated SGD for Overparameterized Linear Regression", "url": "https://openreview.net/forum?id=AcoXPIPh4A", "detail_url": "https://openreview.net/forum?id=AcoXPIPh4A", "authors": "Xuheng Li,Yihe Deng,Jingfeng Wu,Dongruo Zhou,Quanquan Gu", "tags": "ICLR 2024,Poster", "abstract": "Accelerated stochastic gradient descent (ASGD) is a workhorse in deep learning and often achieves better generalization performance than SGD. However, existing optimization theory can only explain the faster convergence of ASGD, but cannot explain its better generalization. In this paper, we study the generalization of ASGD for overparameterized linear regression, which is possibly the simplest setting of learning with overparameterization. We establish an instance-dependent excess risk bound for ASGD within each eigen-subspace of the data covariance matrix. Our analysis shows that (i) ASGD outperforms SGD in the subspace of small eigenvalues, exhibiting a faster rate of exponential decay for bias error, while in the subspace of large eigenvalues, its bias error decays slower than SGD; and (ii) the variance error of ASGD is always larger than that of SGD. Our result suggests that ASGD can outperform SGD when the difference between the initialization and the true weight vector is mostly confined to the subspace of small eigenvalues. Additionally, when our analysis is specialized to linear regression in the strongly convex setting, it yields a tighter bound for bias error than the best-known result.", "pdf": "https://openreview.net/pdf/c07f79889c3181ba8d25abe3732708eda102cceb.pdf"} {"title": "Task structure and nonlinearity jointly determine learned representational geometry", "url": "https://openreview.net/forum?id=k9t8dQ30kU", "detail_url": "https://openreview.net/forum?id=k9t8dQ30kU", "authors": "Matteo Alleman,Jack Lindsey,Stefano Fusi", "tags": "ICLR 2024,Poster", "abstract": "The utility of a learned neural representation depends on how well its geometry supports performance in downstream tasks. This geometry depends on the structure of the inputs, the structure of the target outputs, and on the architecture of the network. By studying the learning dynamics of networks with one hidden layer, we discovered that the network's activation function has an unexpectedly strong impact on the representational geometry: Tanh networks tend to learn representations that reflect the structure of the target outputs, while ReLU networks retain more information about the structure of the raw inputs. This difference is consistently observed across a broad class of parameterized tasks in which we modulated the degree of alignment between the geometry of the task inputs and that of the task labels. We analyzed the learning dynamics in weight space and show how the differences between the networks with Tanh and ReLU nonlinearities arise from the asymmetric saturation of ReLU, which leads feature neurons to specialize for different regions of input space. Feature neurons in Tanh networks, by contrast, tend to inherit the task label structure. Consequently, when the target outputs are low dimensional, Tanh networks generate neural representations that are more disentangled than those obtained with a ReLU nonlinearity. Our findings shed light on the interplay between input-output geometry, nonlinearity, and learned representations in neural networks.", "pdf": "https://openreview.net/pdf/b8a9d014f343b3669b6457839a9e6aa1801d1174.pdf"} {"title": "Llemma: An Open Language Model for Mathematics", "url": "https://openreview.net/forum?id=4WnqRR915j", "detail_url": "https://openreview.net/forum?id=4WnqRR915j", "authors": "Zhangir Azerbayev,Hailey Schoelkopf,Keiran Paster,Marco Dos Santos,Stephen Marcus McAleer,Albert Q. Jiang,Jia Deng,Stella Biderman,Sean Welleck", "tags": "ICLR 2024,Poster", "abstract": "We present Llemma, a large language model for mathematics. We continue pretraining Code Llama on the Proof-Pile-2, a mixture of scientific papers, web data containing mathematics, and mathematical code, yielding Llemma. On the MATH benchmark Llemma outperforms all known openly released models, as well as the unreleased Minerva model suite on an equi-parameter basis. Moreover, Llemma is capable of tool use and formal theorem proving without any finetuning. We openly release all artifacts, including 7 billion and 34 billion parameter models, the Proof-Pile-2, and code to replicate our experiments.", "pdf": "https://openreview.net/pdf/be75740b06066f002fed0867925b737ec1f8757f.pdf"} {"title": "Directly Fine-Tuning Diffusion Models on Differentiable Rewards", "url": "https://openreview.net/forum?id=1vmSEVL19f", "detail_url": "https://openreview.net/forum?id=1vmSEVL19f", "authors": "Kevin Clark,Paul Vicol,Kevin Swersky,David J. Fleet", "tags": "ICLR 2024,Poster", "abstract": "We present Direct Reward Fine-Tuning (DRaFT), a simple and effective method for fine-tuning diffusion models to maximize differentiable reward functions, such as scores from human preference models. We first show that it is possible to backpropagate the reward function gradient through the full sampling procedure, and that doing so achieves strong performance on a variety of rewards, outperforming reinforcement learning-based approaches. We then propose more efficient variants of DRaFT: DRaFT-K, which truncates backpropagation to only the last K steps of sampling, and DRaFT-LV, which obtains lower-variance gradient estimates for the case when K=1. We show that our methods work well for a variety of reward functions and can be used to substantially improve the aesthetic quality of images generated by Stable Diffusion 1.4. Finally, we draw connections between our approach and prior work, providing a unifying perspective on the design space of gradient-based fine-tuning algorithms.", "pdf": "https://openreview.net/pdf/4f6a4d187763cd647549b92a33f6f1c84f23bdec.pdf"} {"title": "Maximum Likelihood Estimation is All You Need for Well-Specified Covariate Shift", "url": "https://openreview.net/forum?id=eoTCKKOgIs", "detail_url": "https://openreview.net/forum?id=eoTCKKOgIs", "authors": "Jiawei Ge,Shange Tang,Jianqing Fan,Cong Ma,Chi Jin", "tags": "ICLR 2024,Poster", "abstract": "A key challenge of modern machine learning systems is to achieve Out-of-Distribution (OOD) generalization---generalizing to target data whose distribution differs from that of source data. Despite its significant importance, the fundamental question of ``what are the most effective algorithms for OOD generalization'' remains open even under the standard setting of covariate shift.\nThis paper addresses this fundamental question by proving that, surprisingly, classical Maximum Likelihood Estimation (MLE) purely using source data (without any modification) achieves the *minimax* optimality for covariate shift under the *well-specified* setting. That is, *no* algorithm performs better than MLE in this setting (up to a constant factor), justifying MLE is all you need.\nOur result holds for a very rich class of parametric models, and does not require any boundedness condition on the density ratio. We illustrate the wide applicability of our framework by instantiating it to three concrete examples---linear regression, logistic regression, and phase retrieval. This paper further complement the study by proving that, under the *misspecified setting*, MLE is no longer the optimal choice, whereas Maximum Weighted Likelihood Estimator (MWLE) emerges as minimax optimal in certain scenarios.", "pdf": "https://openreview.net/pdf/a7c9e2c8b894e55b7cdff72359975a63bd0b405a.pdf"} {"title": "Differentiable Learning of Generalized Structured Matrices for Efficient Deep Neural Networks", "url": "https://openreview.net/forum?id=pAVJKp3Dvn", "detail_url": "https://openreview.net/forum?id=pAVJKp3Dvn", "authors": "Changwoo Lee,Hun-Seok Kim", "tags": "ICLR 2024,Poster", "abstract": "This paper investigates efficient deep neural networks (DNNs) to replace dense unstructured weight matrices with structured ones that possess desired properties. The challenge arises because the optimal weight matrix structure in popular neural network models is obscure in most cases and may vary from layer to layer even in the same network. Prior structured matrices proposed for efficient DNNs were mostly hand-crafted without a generalized framework to systematically learn them. To address this issue, we propose a generalized and differentiable framework to learn efficient structures of weight matrices by gradient descent. We first define a new class of structured matrices that covers a wide range of structured matrices in the literature by adjusting the structural parameters. Then, the frequency-domain differentiable parameterization scheme based on the Gaussian-Dirichlet kernel is adopted to learn the structural parameters by proximal gradient descent. On the image and language tasks, our method learns efficient DNNs with structured matrices, achieving lower complexity and/or higher performance than prior approaches that employ low-rank, block-sparse, or block-low-rank matrices.", "pdf": "https://openreview.net/pdf/f49f086736d3cb7e71497a7fe09351bb020f7175.pdf"} {"title": "A Flexible Generative Model for Heterogeneous Tabular EHR with Missing Modality", "url": "https://openreview.net/forum?id=W2tCmRrj7H", "detail_url": "https://openreview.net/forum?id=W2tCmRrj7H", "authors": "Huan He,William hao,Yuanzhe Xi,Yong Chen,Bradley Malin,Joyce Ho", "tags": "ICLR 2024,Poster", "abstract": "Realistic synthetic electronic health records (EHRs) can be leveraged to acceler- ate methodological developments for research purposes while mitigating privacy concerns associated with data sharing. However, the training of Generative Ad- versarial Networks remains challenging, often resulting in issues like mode col- lapse. While diffusion models have demonstrated progress in generating qual- ity synthetic samples for tabular EHRs given ample denoising steps, their perfor- mance wanes when confronted with missing modalities in heterogeneous tabular EHRs data. For example, some EHRs contain solely static measurements, and some contain only contain temporal measurements, or a blend of both data types. To bridge this gap, we introduce FLEXGEN-EHR\u2013 a versatile diffusion model tai- lored for heterogeneous tabular EHRs, equipped with the capability of handling missing modalities in an integrative learning framework. We define an optimal transport module to align and accentuate the common feature space of hetero- geneity of EHRs. We empirically show that our model consistently outperforms existing state-of-the-art synthetic EHR generation methods both in fidelity by up to 3.10% and utility by up to 7.16%. Additionally, we show that our method can be successfully used in privacy-sensitive settings, where the original patient-level data cannot be shared.", "pdf": "https://openreview.net/pdf/4a599e95342021d11914b38bfcbd9f85362f685b.pdf"} {"title": "Designing Skill-Compatible AI: Methodologies and Frameworks in Chess", "url": "https://openreview.net/forum?id=79rfgv3jw4", "detail_url": "https://openreview.net/forum?id=79rfgv3jw4", "authors": "Karim Hamade,Reid McIlroy-Young,Siddhartha Sen,Jon Kleinberg,Ashton Anderson", "tags": "ICLR 2024,Poster", "abstract": "Powerful artificial intelligence systems are often used in settings where they must interact with agents that are computationally much weaker, for example when they work alongside humans or operate in complex environments where some tasks are handled by algorithms, heuristics, or other entities of varying computational power. For AI agents to successfully interact in these settings, however, achieving superhuman performance alone is not sufficient; they also need to account for suboptimal actions or idiosyncratic style from their less-skilled counterparts. We propose a formal evaluation framework for assessing the compatibility of near-optimal AI with interaction partners who may have much lower levels of skill; we use popular collaborative chess variants as model systems to study and develop AI agents that can successfully interact with lower-skill entities. Traditional chess engines designed to output near-optimal moves prove to be inadequate partners when paired with engines of various lower skill levels in this domain, as they are not designed to consider the presence of other agents. We contribute three methodologies to explicitly create skill-compatible AI agents in complex decision-making settings, and two chess game frameworks designed to foster collaboration between powerful AI agents and less-skilled partners. On these frameworks, our agents outperform state-of-the-art chess AI (based on AlphaZero) despite being weaker in conventional chess, demonstrating that skill-compatibility is a tangible trait that is qualitatively and measurably distinct from raw performance. Our evaluations further explore and clarify the mechanisms by which our agents achieve skill-compatibility.", "pdf": "https://openreview.net/pdf/8b21e7de733143ca1fd7e2af2fc6d9be736bd73b.pdf"} {"title": "Tree Search-Based Policy Optimization under Stochastic Execution Delay", "url": "https://openreview.net/forum?id=RaqZX9LSGA", "detail_url": "https://openreview.net/forum?id=RaqZX9LSGA", "authors": "David Valensi,Esther Derman,Shie Mannor,Gal Dalal", "tags": "ICLR 2024,Poster", "abstract": "The standard formulation of Markov decision processes (MDPs) assumes that the agent's decisions are executed immediately.\nHowever, in numerous realistic applications such as robotics or healthcare, actions are performed with a delay whose value can even be stochastic. In this work, we introduce stochastic delayed execution MDPs, a new formalism addressing random delays without resorting to state augmentation. We show that given observed delay values, it is sufficient to perform a policy search in the class of Markov policies in order to reach optimal performance, thus extending the deterministic fixed delay case. Armed with this insight, we devise DEZ, a model-based algorithm that optimizes over the class of Markov policies. DEZ leverages Monte-Carlo tree search similar to its non-delayed variant EfficientZero to accurately infer future states from the action queue. Thus, it handles delayed execution while preserving the sample efficiency of EfficientZero. Through empirical analysis, we stress that none of the prior benchmarks consistently outperforms others across different delays. We demonstrate that our algorithm surpasses all benchmark methods in Atari games when dealing with constant or stochastic delays. The code is available at \\url{https://github.com/davidva1/Delayed-EZ}.", "pdf": "https://openreview.net/pdf/8830f6b3cafc9913288b14d81260c6f589144619.pdf"} {"title": "Context-Aware Meta-Learning", "url": "https://openreview.net/forum?id=lJYAkDVnRU", "detail_url": "https://openreview.net/forum?id=lJYAkDVnRU", "authors": "Christopher Fifty,Dennis Duan,Ronald Guenther Junkins,Ehsan Amid,Jure Leskovec,Christopher Re,Sebastian Thrun", "tags": "ICLR 2024,Poster", "abstract": "Large Language Models like ChatGPT demonstrate a remarkable capacity to learn new concepts during inference without any fine-tuning. However, visual models trained to detect new objects during inference have been unable to replicate this ability, and instead either perform poorly or require meta-training and/or fine-tuning on similar objects. In this work, we propose a meta-learning algorithm that emulates Large Language Models by learning new visual concepts during inference without fine-tuning. Our approach leverages a frozen pre-trained feature extractor, and analogous to in-context learning, recasts meta-learning as sequence modeling over datapoints with known labels and a test datapoint with an unknown label. On 8 out of 11 meta-learning benchmarks, our approach---without meta-training or fine-tuning---exceeds or matches the state-of-the-art algorithm, P>M>F, which is meta-trained on these benchmarks.", "pdf": "https://openreview.net/pdf/2b890b5667cfd1af349d0022584e53a3221e408c.pdf"} {"title": "Beyond Accuracy: Evaluating Self-Consistency of Code Large Language Models with IdentityChain", "url": "https://openreview.net/forum?id=caW7LdAALh", "detail_url": "https://openreview.net/forum?id=caW7LdAALh", "authors": "Marcus J. Min,Yangruibo Ding,Luca Buratti,Saurabh Pujar,Gail Kaiser,Suman Jana,Baishakhi Ray", "tags": "ICLR 2024,Poster", "abstract": "Code Large Language Models (Code LLMs) are being increasingly employed in real-life applications, so evaluating them is critical. While the conventional accuracy evaluates the performance of Code LLMs on a set of individual tasks, their self-consistency across different tasks is overlooked. Intuitively, a trustworthy model should be self-consistent when generating natural language specifications for its own code and generating code for its own specifications. Failure to preserve self-consistency reveals a lack of understanding of the shared semantics underlying natural language and programming language, and therefore undermines the trustworthiness of a model. In this paper, we first formally define the self-consistency of Code LLMs and then design a framework, IdentityChain, which effectively and efficiently evaluates the self-consistency and conventional accuracy of a model at the same time. We study eleven Code LLMs and show that they fail to preserve self-consistency, which is indeed a distinct aspect from conventional accuracy. Furthermore, we show that IdentityChain can be used as a model debugging tool to expose weaknesses of Code LLMs by demonstrating three major weaknesses that we identify in current models using IdentityChain. Our code is available at https://github.com/marcusm117/IdentityChain.", "pdf": "https://openreview.net/pdf/cc6882135f5b0d5720cfc7705e5db5b102b72cbf.pdf"} {"title": "Ito Diffusion Approximation of Universal Ito Chains for Sampling, Optimization and Boosting", "url": "https://openreview.net/forum?id=fjpfCOV4ru", "detail_url": "https://openreview.net/forum?id=fjpfCOV4ru", "authors": "Aleksei Ustimenko,Aleksandr Beznosikov", "tags": "ICLR 2024,Poster", "abstract": "In this work, we consider rather general and broad class of Markov chains, Ito chains, that look like Euler-Maryama discretization of some Stochastic Differential Equation. The chain we study is a unified framework for theoretical analysis. It comes with almost arbitrary isotropic and state-dependent noise instead of normal and state-independent one as in most related papers. Moreover, in our chain the drift and diffusion coefficient can be inexact in order to cover wide range of applications as Stochastic Gradient Langevin Dynamics, sampling, Stochastic Gradient Descent or Stochastic Gradient Boosting. We prove the bound in $\\mathcal{W}_{2}$-distance between the laws of our Ito chain and corresponding differential equation. These results improve or cover most of the known estimates. And for some particular cases, our analysis is the first.", "pdf": "https://openreview.net/pdf/6096e5c08f636d6144a7e20b80523e0168fe0a4a.pdf"} {"title": "Modeling Boundedly Rational Agents with Latent Inference Budgets", "url": "https://openreview.net/forum?id=W3VsHuga3j", "detail_url": "https://openreview.net/forum?id=W3VsHuga3j", "authors": "Athul Paul Jacob,Abhishek Gupta,Jacob Andreas", "tags": "ICLR 2024,Poster", "abstract": "We study the problem of modeling a population of agents pursuing unknown goals subject to unknown computational constraints. In standard models of bounded rationality, sub-optimal decision-making is simulated by adding homoscedastic noise to optimal decisions rather than actually simulating constrained inference. In this work, we introduce a latent inference budget model (L-IBM) that models these constraints explicitly, via a latent variable (inferred jointly with a model of agents\u2019 goals) that controls the runtime of an iterative inference algorithm. L-IBMs make it possible to learn agent models using data from diverse populations of suboptimal actors. In three modeling tasks\u2014inferring navigation goals from routes, inferring communicative intents from human utterances, and predicting next moves in human chess games\u2014we show that L-IBMs match or outperforms Boltzmann models of decision-making under uncertainty. Moreover, the inferred inference budgets are themselves meaningful, efficient to compute, and correlated with measures of player skill, partner skill and task difficulty.", "pdf": "https://openreview.net/pdf/b6fcabc7a4521fa9b8ff72f6d09fbafc1e118951.pdf"} {"title": "The Effectiveness of Random Forgetting for Robust Generalization", "url": "https://openreview.net/forum?id=MEGQGNUfPx", "detail_url": "https://openreview.net/forum?id=MEGQGNUfPx", "authors": "Vijaya Raghavan T Ramkumar,Bahram Zonooz,Elahe Arani", "tags": "ICLR 2024,Poster", "abstract": "Deep neural networks are susceptible to adversarial attacks, which can compromise their performance and accuracy. Adversarial Training (AT) has emerged as a popular approach for protecting neural networks against such attacks. However, a key challenge of AT is robust overfitting, where the network's robust performance on test data deteriorates with further training, thus hindering generalization. Motivated by the concept of active forgetting in the brain, we introduce a novel learning paradigm called \"Forget to Mitigate Overfitting (FOMO)\". FOMO alternates between the forgetting phase, which randomly forgets a subset of weights and regulates the model's information through weight reinitialization, and the relearning phase, which emphasizes learning generalizable features. Our experiments on benchmark datasets and adversarial attacks show that FOMO alleviates robust overfitting by significantly reducing the gap between the best and last robust test accuracy while improving the state-of-the-art robustness. Furthermore, FOMO provides a better trade-off between the standard and robust accuracy outperforming baseline adversarial methods. Finally, our framework is robust to AutoAttacks and increases generalization in many real-world scenarios.", "pdf": "https://openreview.net/pdf/23ffd81fb8a0b8a1c30e255410b712c66490ed39.pdf"} {"title": "Diffeomorphic Mesh Deformation via Efficient Optimal Transport for Cortical Surface Reconstruction", "url": "https://openreview.net/forum?id=gxhRR8vUQb", "detail_url": "https://openreview.net/forum?id=gxhRR8vUQb", "authors": "Thanh Tung Le,Khai Nguyen,shanlin sun,Kun Han,Nhat Ho,Xiaohui Xie", "tags": "ICLR 2024,Poster", "abstract": "Mesh deformation plays a pivotal role in many 3D vision tasks including dynamic simulations, rendering, and reconstruction. However, defining an efficient discrepancy between predicted and target meshes remains an open problem. A prevalent approach in current deep learning is the set-based approach which measures the discrepancy between two surfaces by comparing two randomly sampled point-clouds from the two meshes with Chamfer pseudo-distance. Nevertheless, the set-based approach still has limitations such as lacking a theoretical guarantee for choosing the number of points in sampled point-clouds, and the pseudo-metricity and the quadratic complexity of the Chamfer divergence. To address these issues, we propose a novel metric for learning mesh deformation. The metric is defined by sliced Wasserstein distance on meshes represented as probability measures that generalize the set-based approach. By leveraging probability measure space, we gain flexibility in encoding meshes using diverse forms of probability measures, such as continuous, empirical, and discrete measures via \\textit{varifold} representation. After having encoded probability measures, we can compare meshes by using the sliced Wasserstein distance which is an effective optimal transport distance with linear computational complexity and can provide a fast statistical rate for approximating the surface of meshes. To the end, we employ a neural ordinary differential equation (ODE) to deform the input surface into the target shape by modeling the trajectories of the points on the surface. Our experiments on cortical surface reconstruction demonstrate that our approach surpasses other competing methods in multiple datasets and metrics.", "pdf": "https://openreview.net/pdf/57b2a95817608783ca78a3f2e620841dc4facc1e.pdf"} {"title": "Lie Group Decompositions for Equivariant Neural Networks", "url": "https://openreview.net/forum?id=p34fRKp8qA", "detail_url": "https://openreview.net/forum?id=p34fRKp8qA", "authors": "Mircea Mironenco,Patrick Forr\u00e9", "tags": "ICLR 2024,Poster", "abstract": "Invariance and equivariance to geometrical transformations have proven to be very useful inductive biases when training (convolutional) neural network models, especially in the low-data regime.\nMuch work has focused on the case where the symmetry group employed is compact or abelian, or both.\nRecent work has explored enlarging the class of transformations used to the case of Lie groups, principally through the use of their Lie algebra, as well as the group exponential and logarithm maps.\nThe applicability of such methods to larger transformation groups is limited by the fact that depending on the group of interest $G$, the exponential map may not be surjective.\nFurther limitations are encountered when $G$ is neither compact nor abelian.\nUsing the structure and geometry of Lie groups and their homogeneous spaces, we present a framework by which it is possible to work with such groups primarily focusing on the Lie groups $G = \\textnormal{GL}^{+}(n, \\mathbb{R})$ and $G = \\textnormal{SL}(n, \\mathbb{R})$, as well as their representation as affine transformations $\\mathbb{R}^{n} \\rtimes G$.\nInvariant integration as well as a global parametrization is realized by decomposing the \"larger\" groups into subgroups and submanifolds which can be handled individually.\nUnder this framework, we show how convolution kernels can be parametrized to build models equivariant with respect to affine transformations.\nWe evaluate the robustness and out-of-distribution generalisation capability of our model on the standard affine-invariant benchmark classification task, where we outperform all previous equivariant models as well as all Capsule Network proposals.", "pdf": "https://openreview.net/pdf/a8637c53d3fd6b8302d16ec775feadb732c3d649.pdf"} {"title": "Efficient Heterogeneous Meta-Learning via Channel Shuffling Modulation", "url": "https://openreview.net/forum?id=QiJuMJl0QS", "detail_url": "https://openreview.net/forum?id=QiJuMJl0QS", "authors": "Minh Hoang,Carl Kingsford", "tags": "ICLR 2024,Poster", "abstract": "We tackle the problem of meta-learning across heterogenous tasks. This problem seeks to extract and generalize transferable meta-knowledge through streaming task sets from a multi-modal task distribution. The extracted meta-knowledge can be used to create predictors for new tasks using a small number of labeled samples. Most meta-learning methods assume a homogeneous task distribution, thus limiting their generalization capacity when handling multi-modal task distributions. Recent work has shown that the generalization of meta-learning depends on the similarity of tasks in the training distribution, and this has led to many clustering approaches that aim to detect homogeneous clusters of tasks. However, these methods suffer from a significant increase in parameter complexity. To overcome this weakness, we propose a new heterogeneous meta-learning strategy that efficiently captures the multi-modality of the task distribution via modulating the routing between convolution channels in the network, instead of directly modulating the network weights. This new mechanism can be cast as a permutation learning problem. We further introduce a novel neural permutation layer based on the classical Benes routing network, which has sub-quadratic parameter complexity in the total number of channels, as compared to the quadratic complexity of the state-of-the-art Gumbel-Sinkhorn layer. We demonstrate our approach on various multi-modal meta-learning benchmarks, showing that our framework outperforms previous methods in both generalization accuracy and convergence speed.", "pdf": "https://openreview.net/pdf/2372b26cc6c55bb1b97f26c7f0b9cc2e897ede18.pdf"} {"title": "To Grok or not to Grok: Disentangling Generalization and Memorization on Corrupted Algorithmic Datasets", "url": "https://openreview.net/forum?id=UHjE5v5MB7", "detail_url": "https://openreview.net/forum?id=UHjE5v5MB7", "authors": "Darshil Doshi,Aritra Das,Tianyu He,Andrey Gromov", "tags": "ICLR 2024,Poster", "abstract": "Robust generalization is a major challenge in deep learning, particularly when the number of trainable parameters is very large. In general, it is very difficult to know if the network has memorized a particular set of examples or understood the underlying rule (or both). Motivated by this challenge, we study an interpretable model where generalizing representations are understood analytically, and are easily distinguishable from the memorizing ones. Namely, we consider multi-layer perceptron (MLP) and Transformer architectures trained on modular arithmetic tasks, where ($\\xi \\cdot 100\\\\%$) of labels are corrupted (*i.e.* some results of the modular operations in the training set are incorrect). We show that (i) it is possible for the network to memorize the corrupted labels *and* achieve $100\\\\%$ generalization at the same time; (ii) the memorizing neurons can be identified and pruned, lowering the accuracy on corrupted data and improving the accuracy on uncorrupted data; (iii) regularization methods such as weight decay, dropout and BatchNorm force the network to ignore the corrupted data during optimization, and achieve $100\\\\%$ accuracy on the uncorrupted dataset; and (iv) the effect of these regularization methods is (\"mechanistically\") interpretable: weight decay and dropout force all the neurons to learn generalizing representations, while BatchNorm de-amplifies the output of memorizing neurons and amplifies the output of the generalizing ones. Finally, we show that in the presence of regularization, the training dynamics involves two consecutive stages: first, the network undergoes *grokking* dynamics reaching high train *and* test accuracy; second, it unlearns the memorizing representations, where the train accuracy suddenly jumps from $100\\\\%$ to $100 (1-\\xi)\\\\%$.", "pdf": "https://openreview.net/pdf/b5a7d2b34d54912009b544ef9c33ac0e3d8050a2.pdf"} {"title": "VCR-Graphormer: A Mini-batch Graph Transformer via Virtual Connections", "url": "https://openreview.net/forum?id=SUUrkC3STJ", "detail_url": "https://openreview.net/forum?id=SUUrkC3STJ", "authors": "Dongqi Fu,Zhigang Hua,Yan Xie,Jin Fang,Si Zhang,Kaan Sancak,Hao Wu,Andrey Malevich,Jingrui He,Bo Long", "tags": "ICLR 2024,Poster", "abstract": "Graph transformer has been proven as an effective graph learning method for its adoption of attention mechanism that is capable of capturing expressive representations from complex topological and feature information of graphs. Graph transformer conventionally performs dense attention (or global attention) for every pair of nodes to learn node representation vectors, resulting in quadratic computational costs that are unaffordable for large-scale graph data. Therefore, mini-batch training for graph transformers is a promising direction, but limited samples in each mini-batch can not support effective dense attention to encode informative representations. Facing this bottleneck, (1) we start by assigning each node a token list that is sampled by personalized PageRank (PPR) and then apply standard multi-head self-attention only on this list to compute its node representations. This PPR tokenization method decouples model training from complex graph topological information and makes heavy feature engineering offline and independent, such that mini-batch training of graph transformers is possible by loading each node's token list in batches. We further prove this PPR tokenization is viable as a graph convolution network with a fixed polynomial filter and jumping knowledge. However, only using personalized PageRank may limit information carried by a token list, which could not support different graph inductive biases for model training. To this end, (2) we rewire graphs by introducing multiple types of virtual connections through structure- and content-based super nodes that enable PPR tokenization to encode local and global contexts, long-range interaction, and heterophilous information into each node's token list, and then formalize our $\\underline{\\textbf{V}}$irtual $\\underline{\\textbf{C}}$onnection $\\underline{\\textbf{R}}$anking based $\\underline{\\textbf{Graph}}$ Trans$\\underline{\\textbf{former}}$ (VCR-Graphormer). Overall, VCR-Graphormer needs $O(m+klogk)$ complexity for graph tokenization as compared to $O(n^{3})$ of previous works. The [code](https://github.com/DongqiFu/VCR-Graphormer) is provided.", "pdf": "https://openreview.net/pdf/5914f5217eeed4b267eb84117a931ad9f7dbd355.pdf"} {"title": "Optimistic Bayesian Optimization with Unknown Constraints", "url": "https://openreview.net/forum?id=D4NJFfrqoq", "detail_url": "https://openreview.net/forum?id=D4NJFfrqoq", "authors": "Quoc Phong Nguyen,Wan Theng Ruth Chew,Le Song,Bryan Kian Hsiang Low,Patrick Jaillet", "tags": "ICLR 2024,Poster", "abstract": "Though some research efforts have been dedicated to constrained Bayesian optimization (BO), there remains a notable absence of a principled approach with a theoretical performance guarantee in the decoupled setting. Such a setting involves independent evaluations of the objective function and constraints at different inputs, and is hence a relaxation of the commonly-studied coupled setting where functions must be evaluated together. As a result, the decoupled setting requires an adaptive selection between evaluating either the objective function or a constraint, in addition to selecting an input (in the coupled setting). This paper presents a novel constrained BO algorithm with a provable performance guarantee that can address the above relaxed setting. Specifically, it considers the fundamental trade-off between exploration and exploitation in constrained BO, and, interestingly, affords a noteworthy connection to active learning. The performance of our proposed algorithms is also empirically evaluated using several synthetic and real-world optimization problems.", "pdf": "https://openreview.net/pdf/b9f8e9fc8f8a6b5a79bf0426ca34210e1c6552ff.pdf"} {"title": "DRSM: De-Randomized Smoothing on Malware Classifier Providing Certified Robustness", "url": "https://openreview.net/forum?id=m7aPLHwsLr", "detail_url": "https://openreview.net/forum?id=m7aPLHwsLr", "authors": "Shoumik Saha,Wenxiao Wang,Yigitcan Kaya,Soheil Feizi,Tudor Dumitras", "tags": "ICLR 2024,Poster", "abstract": "Machine Learning (ML) models have been utilized for malware detection for over two decades. Consequently, this ignited an ongoing arms race between malware authors and antivirus systems, compelling researchers to propose defenses for malware-detection models against evasion attacks. However, most if not all existing defenses against evasion attacks suffer from sizable performance degradation and/or can defend against only specific attacks, which makes them less practical in real-world settings. In this work, we develop a certified defense, DRSM (De-Randomized Smoothed MalConv), by redesigning the *de-randomized smoothing* technique for the domain of malware detection. Specifically, we propose a *window ablation* scheme to provably limit the impact of adversarial bytes while maximally preserving local structures of the executables. After showing how DRSM is theoretically robust against attacks with contiguous adversarial bytes, we verify its performance and certified robustness experimentally, where we observe only marginal accuracy drops as the cost of robustness. To our knowledge, we are the first to offer certified robustness in the realm of static detection of malware executables. More surprisingly, through evaluating DRSM against $9$ empirical attacks of different types, we observe that the proposed defense is empirically robust to some extent against a diverse set of attacks, some of which even fall out of the scope of its original threat model. In addition, we collected $15.5K$ recent benign raw executables from diverse sources, which will be made public as a dataset called PACE (Publicly Accessible Collection(s) of Executables) to alleviate the scarcity of publicly available benign datasets for studying malware detection and provide future research with more representative data of the time. Our code and dataset are available at - https://github.com/ShoumikSaha/DRSM", "pdf": "https://openreview.net/pdf/f16a6785a007c48f8e9c460b978613135d30e8c3.pdf"} {"title": "On the Variance of Neural Network Training with respect to Test Sets and Distributions", "url": "https://openreview.net/forum?id=pEGSdJu52I", "detail_url": "https://openreview.net/forum?id=pEGSdJu52I", "authors": "Keller Jordan", "tags": "ICLR 2024,Poster", "abstract": "Neural network trainings are stochastic, causing the performance of trained networks to vary across repeated runs of training.\nWe contribute the following results towards understanding this variation.\n(1) Despite having significant variance on their test-sets, we demonstrate that standard CIFAR-10 and ImageNet trainings have little variance in their performance on the test-distributions from which their test-sets are sampled.\n(2) We introduce the independent errors assumption and show that it suffices to recover the structure and variance of the empirical accuracy distribution across repeated runs of training.\n(3) We prove that test-set variance is unavoidable given the observation that ensembles of identically trained networks are calibrated (Jiang et al., 2021), and demonstrate that the variance of binary classification trainings closely follows a simple formula based on the error rate and number of test examples.\n(4) We conduct preliminary studies of data augmentation, learning rate, finetuning instability and distribution-shift through the lens of variance between runs.", "pdf": "https://openreview.net/pdf/2de59bd3955e067d8a7c6a4eee59896af773422d.pdf"} {"title": "Large Language Models to Enhance Bayesian Optimization", "url": "https://openreview.net/forum?id=OOxotBmGol", "detail_url": "https://openreview.net/forum?id=OOxotBmGol", "authors": "Tennison Liu,Nicol\u00e1s Astorga,Nabeel Seedat,Mihaela van der Schaar", "tags": "ICLR 2024,Poster", "abstract": "Bayesian optimization (BO) is a powerful approach for optimizing complex and expensive-to-evaluate black-box functions. Its importance is underscored in many applications, notably including hyperparameter tuning, but its efficacy depends on efficiently balancing exploration and exploitation. While there has been substantial progress in BO methods, striking this balance remains a delicate process. In this light, we present \\texttt{LLAMBO}, a novel approach that integrates the capabilities of Large Language Models (LLM) within BO. At a high level, we frame the BO problem in natural language, enabling LLMs to iteratively \\emph{propose} and \\emph{evaluate} promising solutions conditioned on historical evaluations. More specifically, we explore how combining contextual understanding, few-shot learning proficiency, and domain knowledge of LLMs can improve model-based BO. Our findings illustrate that \\texttt{LLAMBO} is effective at zero-shot warmstarting, and enhances surrogate modeling and candidate sampling, especially in the early stages of search when observations are sparse. Our approach is performed in context and does not require LLM finetuning. Additionally, it is modular by design, allowing individual components to be integrated into existing BO frameworks, or function cohesively as an end-to-end method. We empirically validate \\texttt{LLAMBO}'s efficacy on the problem of hyperparameter tuning, highlighting strong empirical performance across a range of diverse benchmarks, proprietary, and synthetic tasks.", "pdf": "https://openreview.net/pdf/b22b6ea9b84dd474cf4871acc78a784c45bac294.pdf"} {"title": "Towards Identifiable Unsupervised Domain Translation: A Diversified Distribution Matching Approach", "url": "https://openreview.net/forum?id=55uj7mU7Cv", "detail_url": "https://openreview.net/forum?id=55uj7mU7Cv", "authors": "Sagar Shrestha,Xiao Fu", "tags": "ICLR 2024,Poster", "abstract": "Unsupervised domain translation (UDT) aims to find functions that convert samples from one domain (e.g., sketches) to another domain (e.g., photos) without changing the high-level semantic meaning (also referred to as \"content\"). The translation functions are often sought by probability distribution matching of the transformed source domain and target domain. CycleGAN stands as arguably the most representative approach among this line of work. However, it was noticed in the literature that CycleGAN and variants could fail to identify the desired translation functions and produce content-misaligned translations.\nThis limitation arises due to the presence of multiple translation functions---referred to as ``measure-preserving automorphism\" (MPA)---in the solution space of the learning criteria. Despite awareness of such identifiability issues, solutions have remained elusive. This study delves into the core identifiability inquiry and introduces an MPA elimination theory. Our analysis shows that MPA is unlikely to exist, if multiple pairs of diverse cross-domain conditional distributions are matched by the learning function.\nOur theory leads to a UDT learner using distribution matching over auxiliary variable-induced subsets of the domains---other than over the entire data domains as in the classical approaches. The proposed framework is the first to rigorously establish translation identifiability under reasonable UDT settings, to our best knowledge.\nExperiments corroborate with our theoretical claims.", "pdf": "https://openreview.net/pdf/310f0697e921696dbb723b2f938345c38befe7d6.pdf"} {"title": "SineNet: Learning Temporal Dynamics in Time-Dependent Partial Differential Equations", "url": "https://openreview.net/forum?id=LSYhE2hLWG", "detail_url": "https://openreview.net/forum?id=LSYhE2hLWG", "authors": "Xuan Zhang,Jacob Helwig,Yuchao Lin,Yaochen Xie,Cong Fu,Stephan Wojtowytsch,Shuiwang Ji", "tags": "ICLR 2024,Poster", "abstract": "We consider using deep neural networks to solve time-dependent partial differential equations (PDEs), where multi-scale processing is crucial for modeling complex, time-evolving dynamics. While the U-Net architecture with skip connections is commonly used by prior studies to enable multi-scale processing, our analysis shows that the need for features to evolve across layers results in temporally misaligned features in skip connections, which limits the model\u2019s performance. To address this limitation, we propose SineNet, consisting of multiple sequentially connected U-shaped network blocks, referred to as waves. In SineNet, high-resolution features are evolved progressively through multiple stages, thereby reducing the amount of misalignment within each stage. We furthermore analyze the role of skip connections in enabling both parallel and sequential processing of multi-scale information. Our method is rigorously tested on multiple PDE datasets, including the Navier-Stokes equations and shallow water equations, showcasing the advantages of our proposed approach over conventional U-Nets with a comparable parameter budget. We further demonstrate that increasing the number of waves in SineNet while maintaining the same number of parameters leads to a monotonically improved performance. The results highlight the effectiveness of SineNet and the potential of our approach in advancing the state-of-the-art in neural PDE solver design. Our code is available as part of AIRS (https://github.com/divelab/AIRS).", "pdf": "https://openreview.net/pdf/e09653bff41011adca550b8b7e3e27bd1e59e9f1.pdf"} {"title": "GNNBoundary: Towards Explaining Graph Neural Networks through the Lens of Decision Boundaries", "url": "https://openreview.net/forum?id=WIzzXCVYiH", "detail_url": "https://openreview.net/forum?id=WIzzXCVYiH", "authors": "Xiaoqi Wang,Han Wei Shen", "tags": "ICLR 2024,Poster", "abstract": "While Graph Neural Networks (GNNs) have achieved remarkable performance on various machine learning tasks on graph data, they also raised questions regarding their transparency and interpretability. Recently, there have been extensive research efforts to explain the decision-making process of GNNs. These efforts often focus on explaining why a certain prediction is made for a particular instance, or what discriminative features the GNNs try to detect for each class. However, to the best of our knowledge, there is no existing study on understanding the decision boundaries of GNNs, even though the decision-making process of GNNs is directly determined by the decision boundaries. To bridge this research gap, we propose a model-level explainability method called GNNBoundary, which attempts to gain deeper insights into the decision boundaries of graph classifiers. Specifically, we first develop an algorithm to identify the pairs of classes whose decision regions are adjacent. For an adjacent class pair, the near-boundary graphs between them are effectively generated by optimizing a novel objective function specifically designed for boundary graph generation. Thus, by analyzing the nearboundary graphs, the important characteristics of decision boundaries can be uncovered. To evaluate the efficacy of GNNBoundary, we conduct experiments on both synthetic and public real-world datasets. The results demonstrate that, via the analysis of faithful near-boundary graphs generated by GNNBoundary, we can thoroughly assess the robustness and generalizability of the explained GNNs. The official implementation can be found at https://github.com/yolandalalala/GNNBoundary.", "pdf": "https://openreview.net/pdf/4c88af58e7fea802a541ca1a850768e448700110.pdf"} {"title": "Sparse Spiking Neural Network: Exploiting Heterogeneity in Timescales for Pruning Recurrent SNN", "url": "https://openreview.net/forum?id=0jsfesDZDq", "detail_url": "https://openreview.net/forum?id=0jsfesDZDq", "authors": "Biswadeep Chakraborty,Beomseok Kang,Harshit Kumar,Saibal Mukhopadhyay", "tags": "ICLR 2024,Poster", "abstract": "Recurrent Spiking Neural Networks (RSNNs) have emerged as a computationally efficient and brain-inspired machine learning model. The design of sparse RSNNs with fewer neurons and synapses helps reduce the computational complexity of RSNNs. Traditionally, sparse SNNs are obtained by first training a dense and complex SNN for a target task and, next, eliminating neurons with low activity (activity-based pruning) while maintaining task performance. In contrast, this paper presents a task-agnostic methodology for designing sparse RSNNs by pruning an untrained (arbitrarily initialized) large model. \nWe introduce a novel Lyapunov Noise Pruning (LNP) algorithm that uses graph sparsification methods and utilizes Lyapunov exponents to design a stable sparse RSNN from an untrained RSNN. We show that the LNP can leverage diversity in neuronal timescales to design a sparse Heterogeneous RSNN (HRSNN). Further, we show that the same sparse HRSNN model can be trained for different tasks, such as image classification and time-series prediction. The experimental results show that, in spite of being task-agnostic, LNP increases computational efficiency (fewer neurons and synapses) and prediction performance of RSNNs compared to traditional activity-based pruning of trained dense models.", "pdf": "https://openreview.net/pdf/7e55c6eb14311819313269f9b7f133c8cee91597.pdf"} {"title": "Investigating the Benefits of Projection Head for Representation Learning", "url": "https://openreview.net/forum?id=GgEAdqYPNA", "detail_url": "https://openreview.net/forum?id=GgEAdqYPNA", "authors": "Yihao Xue,Eric Gan,Jiayi Ni,Siddharth Joshi,Baharan Mirzasoleiman", "tags": "ICLR 2024,Poster", "abstract": "An effective technique for obtaining high-quality representations is adding a projection head on top of the encoder during training, then discarding it and using the pre-projection representations. Despite its proven practical effectiveness, the reason behind the success of this technique is poorly understood. The pre-projection representations are not directly optimized by the loss function, raising the question: what makes them better? In this work, we provide a rigorous theoretical answer to this question. We start by examining linear models trained with self-supervised contrastive loss. We reveal that the implicit bias of training algorithms leads to layer-wise progressive feature weighting, where features become increasingly unequal as we go deeper into the layers. Consequently, lower layers tend to have more normalized and less specialized representations. We theoretically characterize scenarios where such representations are more beneficial, highlighting the intricate interplay between data augmentation and input features. Additionally, we demonstrate that introducing non-linearity into the network allows lower layers to learn features that are completely absent in higher layers. Finally, we show how this mechanism improves the robustness in supervised contrastive learning and supervised learning. We empirically validate our results through various experiments on CIFAR-10/100, UrbanCars and shifted versions of ImageNet. We also introduce a potential alternative to projection head, which offers a more interpretable and controllable design.", "pdf": "https://openreview.net/pdf/274d4b0de6e23f1e7b5f267ab4aa084835b1eec5.pdf"} {"title": "A Variational Perspective on Solving Inverse Problems with Diffusion Models", "url": "https://openreview.net/forum?id=1YO4EE3SPB", "detail_url": "https://openreview.net/forum?id=1YO4EE3SPB", "authors": "Morteza Mardani,Jiaming Song,Jan Kautz,Arash Vahdat", "tags": "ICLR 2024,Poster", "abstract": "Diffusion models have emerged as a key pillar of foundation models in visual domains. One of their critical applications is to universally solve different downstream inverse tasks via a single diffusion prior without re-training for each task. Most inverse tasks can be formulated as inferring a posterior distribution over data (e.g., a full image) given a measurement (e.g., a masked image). This is however challenging in diffusion models since the nonlinear and iterative nature of the diffusion process renders the posterior intractable. To cope with this challenge, we propose a variational approach that by design seeks to approximate the true posterior distribution. We show that our approach naturally leads to regularization by denoising diffusion process (RED-diff) where denoisers at different timesteps concurrently impose different structural constraints over the image. To gauge the contribution of denoisers from different timesteps, we propose a weighting mechanism based on signal-to-noise-ratio (SNR). Our approach provides a new variational perspective for solving inverse problems with diffusion models, allowing us to formulate sampling as stochastic optimization, where one can simply apply off-the-shelf solvers with lightweight iterates. Our experiments for various linear and nonlinear image restoration tasks demonstrate the strengths of our method compared with state-of-the-art sampling-based diffusion models. The code is available online \\footnote{\\url{https://github.com/NVlabs/RED-diff}}.", "pdf": "https://openreview.net/pdf/98565461a56247bb80f25ac785745223d4ec97f7.pdf"} {"title": "Can Large Language Models Infer Causation from Correlation?", "url": "https://openreview.net/forum?id=vqIH0ObdqL", "detail_url": "https://openreview.net/forum?id=vqIH0ObdqL", "authors": "Zhijing Jin,Jiarui Liu,Zhiheng LYU,Spencer Poff,Mrinmaya Sachan,Rada Mihalcea,Mona T. Diab,Bernhard Sch\u00f6lkopf", "tags": "ICLR 2024,Poster", "abstract": "Causal inference is one of the hallmarks of human intelligence. While the field of CausalNLP has attracted much interest in the recent years, existing causal inference datasets in NLP primarily rely on discovering causality from empirical knowledge (e.g., commonsense knowledge). In this work, we propose the first benchmark dataset to test the pure causal inference skills of large language models (LLMs). Specifically, we formulate a novel task Corr2Cause, which takes a set of correlational statements and determines the causal relationship between the variables. We curate a large-scale dataset of more than 200K samples, on which we evaluate seventeen existing LLMs. Through our experiments, we identify a key shortcoming of LLMs in terms of their causal inference skills, and show that these models achieve almost close to random performance on the task. This shortcoming is somewhat mitigated when we try to re-purpose LLMs for this skill via finetuning, but we find that these models still fail to generalize \u2013 they can only perform causal inference in in-distribution settings when variable names and textual expressions used in the queries are similar to those in the training set, but fail in out-of-distribution settings generated by perturbing these queries. Corr2Cause is a challenging task for LLMs, and can be helpful in guiding future research on improving LLMs\u2019 pure reasoning skills and generalizability. Our data is at https://huggingface.co/datasets/causalnlp/corr2cause. Our code is at https://github.com/causalNLP/corr2cause.", "pdf": "https://openreview.net/pdf/c32089ff7c53b6e59610da92bace7b326b0a622c.pdf"} {"title": "Improved statistical and computational complexity of the mean-field Langevin dynamics under structured data", "url": "https://openreview.net/forum?id=Of2nEDc4s7", "detail_url": "https://openreview.net/forum?id=Of2nEDc4s7", "authors": "Atsushi Nitanda,Kazusato Oko,Taiji Suzuki,Denny Wu", "tags": "ICLR 2024,Poster", "abstract": "Recent works have shown that neural networks optimized by gradient-based methods can adapt to sparse or low-dimensional target functions through feature learning; an often studied target is the sparse parity function on the unit hypercube. However, such isotropic data setting does not capture the anisotropy and low intrinsic dimensionality exhibited in realistic datasets. In this work, we address this shortcoming by studying how gradient-based feature learning interacts with structured (anisotropic) input data: we consider the classification of $k$-sparse parity on high-dimensional orthotope where the feature coordinates have varying magnitudes, and analyze the learning complexity of the mean-field Langevin dynamics (MFLD), which describes the noisy gradient descent update on two-layer neural network. We show that the statistical complexity (i.e. sample size) and computational complexity (i.e. network width) of MFLD can both be improved when prominent directions of the anisotropic input data align with the support of the target function. Moreover, by employing a coordinate transform determined by the gradient covariance, the width can be made independent of the target degree $k$. Lastly, we demonstrate the benefit of feature learning by establishing a kernel lower bound on the classification error, which applies to neural networks in the lazy regime.", "pdf": "https://openreview.net/pdf/5513243838f990d8d7323b2afa870a627e9c9484.pdf"} {"title": "Jointly-Learned Exit and Inference for a Dynamic Neural Network", "url": "https://openreview.net/forum?id=jX2DT7qDam", "detail_url": "https://openreview.net/forum?id=jX2DT7qDam", "authors": "florence regol,Joud Chataoui,Mark Coates", "tags": "ICLR 2024,Poster", "abstract": "Large pretrained models, coupled with fine-tuning, are slowly becoming established as the dominant architecture in machine learning. Even though these models offer impressive performance, their practical application is often limited by the prohibitive amount of resources required for $\\textit{every}$ inference. Early-exiting dynamic neural networks (EDNN) circumvent this issue by allowing a model to make some of its predictions from intermediate layers (i.e., early-exit). Training an EDNN architecture is challenging as it consists of two intertwined components: the gating mechanism (GM) that controls early-exiting decisions and the intermediate inference modules (IMs) that perform inference from intermediate representations. As a result, most existing approaches rely on thresholding confidence metrics for the gating mechanism and strive to improve the underlying backbone network and the inference modules. Although successful, this approach has two fundamental shortcomings: 1) the GMs and the IMs are decoupled during training, leading to a train-test mismatch; and 2) the thresholding gating mechanism introduces a positive bias into the predictive probabilities, making it difficult to readily extract uncertainty information. We propose a novel architecture that connects these two modules. This leads to significant performance improvements on classification datasets and enables better uncertainty characterization capabilities.", "pdf": "https://openreview.net/pdf/530fc74c8de9fa89f925873092154bf72fa2e92c.pdf"} {"title": "Understanding the Robustness of Multi-modal Contrastive Learning to Distribution Shift", "url": "https://openreview.net/forum?id=rtl4XnJYBh", "detail_url": "https://openreview.net/forum?id=rtl4XnJYBh", "authors": "Yihao Xue,Siddharth Joshi,Dang Nguyen,Baharan Mirzasoleiman", "tags": "ICLR 2024,Poster", "abstract": "Recently, multimodal contrastive learning (MMCL) approaches, such as CLIP, have achieved a remarkable success in learning representations that are robust against distribution shift and generalize to new domains. Despite the empirical success, the mechanism behind learning such generalizable representations is not understood. In this work, we rigorously analyze this problem and \nuncover two mechanisms behind MMCL's robustness: \\emph{intra-class contrasting}, which allows the model to learn features with a high variance, and \\emph{inter-class feature sharing}, where annotated details in one class help learning other classes better. Both mechanisms prevent spurious features that are over-represented in the training data to overshadow the generalizable core features. This yields superior zero-shot classification accuracy under distribution shift. Furthermore, we theoretically demonstrate the benefits of using rich captions on robustness and explore the effect of annotating different types of details in the captions. We validate our theoretical findings through experiments, including a well-designed synthetic experiment and an experiment involving training CLIP models on MSCOCO/Conceptual Captions and evaluating them on shifted ImageNets.", "pdf": "https://openreview.net/pdf/6a626c46d27f126eacb827909e73e287e464009d.pdf"} {"title": "SequenceMatch: Imitation Learning for Autoregressive Sequence Modelling with Backtracking", "url": "https://openreview.net/forum?id=FJWT0692hw", "detail_url": "https://openreview.net/forum?id=FJWT0692hw", "authors": "Chris Cundy,Stefano Ermon", "tags": "ICLR 2024,Poster", "abstract": "In many domains, autoregressive models can attain high likelihood on the task of predicting the next observation. However, this maximum-likelihood (MLE) objective does not necessarily match a downstream use-case of autoregressively generating high-quality sequences. The MLE objective weights sequences proportionally to their frequency under the data distribution, with no guidance for the model's behaviour out of distribution (OOD): leading to compounding error during autoregressive generation. In order to address this compounding error problem, we formulate sequence generation as an imitation learning (IL) problem. This allows us to minimize a variety of divergences between the distribution of sequences generated by an autoregressive model and sequences from a dataset, including divergences with weight on OOD generated sequences. The IL framework also allows us to incorporate backtracking by introducing a backspace action into the generation process. This further mitigates the compounding error problem by allowing the model to revert a sampled token if it takes the sequence OOD. Our resulting method, SequenceMatch, can be implemented without adversarial training or major architectural changes. We identify the SequenceMatch-\u03c72 divergence as a more suitable training objective for autoregressive models which are used for generation. We show that empirically, SequenceMatch training leads to improvements over MLE on text generation with language models and arithmetic", "pdf": "https://openreview.net/pdf/c034cb58427202bff4ad7164db6d76d3e815b9d6.pdf"} {"title": "Layer-wise linear mode connectivity", "url": "https://openreview.net/forum?id=LfmZh91tDI", "detail_url": "https://openreview.net/forum?id=LfmZh91tDI", "authors": "Linara Adilova,Maksym Andriushchenko,Michael Kamp,Asja Fischer,Martin Jaggi", "tags": "ICLR 2024,Poster", "abstract": "Averaging neural network parameters is an intuitive method for fusing the knowledge of two independent models. It is most prominently used in federated learning. If models are averaged at the end of training, this can only lead to a good performing model if the loss surface of interest is very particular, i.e., the loss in the midpoint between the two models needs to be sufficiently low. This is impossible to guarantee for the non-convex losses of state-of-the-art networks. For averaging models trained on vastly different datasets, it was proposed to average only the parameters of particular layers or combinations of layers, resulting in better performing models. To get a better understanding of the effect of layer-wise averaging, we analyse the performance of the models that result from averaging single layers, or groups of layers. Based on our empirical and theoretical investigation, we introduce a novel notion of the layer-wise linear connectivity, and show that deep networks do not have layer-wise barriers between them.", "pdf": "https://openreview.net/pdf/039a426405636e0c6ee6fd6ea0035e1d20b6bc28.pdf"} {"title": "Understanding Certified Training with Interval Bound Propagation", "url": "https://openreview.net/forum?id=h05eQniJsQ", "detail_url": "https://openreview.net/forum?id=h05eQniJsQ", "authors": "Yuhao Mao,Mark Niklas Mueller,Marc Fischer,Martin Vechev", "tags": "ICLR 2024,Poster", "abstract": "As robustness verification methods are becoming more precise, training certifiably robust neural networks is becoming ever more relevant. To this end, certified training methods compute and then optimize an upper bound on the worst-case loss over a robustness specification. Curiously, training methods based on the imprecise interval bound propagation (IBP) consistently outperform those leveraging more precise bounds. Still, we lack a theoretical understanding of the mechanisms making IBP so successful. In this work, we investigate these mechanisms by leveraging a novel metric measuring the tightness of IBP bounds. We first show theoretically that, for deep linear models (DLNs), tightness decreases with width and depth at initialization, but improves with IBP training. We, then, derive sufficient and necessary conditions on weight matrices for IBP bounds to become exact and demonstrate that these impose strong regularization, providing an explanation for the observed robustness-accuracy trade-off. Finally, we show how these results on DLNs transfer to ReLU networks, before conducting an extensive empirical study, (i) confirming this transferability and yielding state-of-the-art certified accuracy, (ii) finding that while all IBP-based training methods lead to high tightness, this increase is dominated by the size of the propagated input regions rather than the robustness specification, and finally (iii) observing that non-IBP-based methods do not increase tightness. Together, these results help explain the success of recent certified training methods and may guide the development of new ones.", "pdf": "https://openreview.net/pdf/3368ca8a5f0ec99ee88301ab1de8634647f5ce72.pdf"} {"title": "Offline RL with Observation Histories: Analyzing and Improving Sample Complexity", "url": "https://openreview.net/forum?id=GnOLWS4Llt", "detail_url": "https://openreview.net/forum?id=GnOLWS4Llt", "authors": "Joey Hong,Anca Dragan,Sergey Levine", "tags": "ICLR 2024,Poster", "abstract": "Offline reinforcement learning (RL) can in principle synthesize more optimal behavior from a dataset consisting only of suboptimal trials. One way that this can happen is by \"stitching\" together the best parts of otherwise suboptimal trajectories that overlap on similar states, to create new behaviors where each individual state is in-distribution, but the overall returns are higher. However, in many interesting and complex applications, such as autonomous navigation and dialogue systems, the state is partially observed. Even worse, the state representation is unknown or not easy to define. In such cases, policies and value functions are often conditioned on observation histories instead of states. In these cases, it is not clear if the same kind of \"stitching\" is feasible at the level of observation histories, since two different trajectories would always have different histories, and thus \"similar states\" that might lead to effective stitching cannot be leveraged. Theoretically, we show that standard offline RL algorithms conditioned on observation histories suffer from poor sample complexity, in accordance with the above intuition. We then identify sufficient conditions under which offline RL can still be efficient -- intuitively, it needs to learn a compact representation of history comprising only features relevant for action selection. We introduce a bisimulation loss that captures the extent to which this happens, and propose that offline RL can explicitly optimize this loss to aid worst-case sample complexity. Empirically, we show that across a variety of tasks either our proposed loss improves performance, or the value of this loss is already minimized as a consequence of standard offline RL, indicating that it correlates well with good performance.", "pdf": "https://openreview.net/pdf/efd5f3cf56edfff0bb4ff14fb974d994bfaa976f.pdf"} {"title": "Graph-based Virtual Sensing from Sparse and Partial Multivariate Observations", "url": "https://openreview.net/forum?id=CAqdG2dy5s", "detail_url": "https://openreview.net/forum?id=CAqdG2dy5s", "authors": "Giovanni De Felice,Andrea Cini,Daniele Zambon,Vladimir Gusev,Cesare Alippi", "tags": "ICLR 2024,Poster", "abstract": "Virtual sensing techniques allow for inferring signals at new unmonitored locations by exploiting spatio-temporal measurements coming from physical sensors at different locations. However, as the sensor coverage becomes sparse due to costs or other constraints, physical proximity cannot be used to support interpolation. In this paper, we overcome this challenge by leveraging dependencies between the target variable and a set of correlated variables (covariates) that can frequently be associated with each location of interest. From this viewpoint, covariates provide partial observability, and the problem consists of inferring values for unobserved channels by exploiting observations at other locations to learn how such variables can correlate. We introduce a novel graph-based methodology to exploit such relationships and design a graph deep learning architecture, named GgNet, implementing the framework. The proposed approach relies on propagating information over a nested graph structure that is used to learn dependencies between variables as well as locations. GgNet is extensively evaluated under different virtual sensing scenarios, demonstrating higher reconstruction accuracy compared to the state-of-the-art.", "pdf": "https://openreview.net/pdf/1f134ca075964391b30baa22a0bf6a338d01598c.pdf"} {"title": "NEFTune: Noisy Embeddings Improve Instruction Finetuning", "url": "https://openreview.net/forum?id=0bMmZ3fkCk", "detail_url": "https://openreview.net/forum?id=0bMmZ3fkCk", "authors": "Neel Jain,Ping-yeh Chiang,Yuxin Wen,John Kirchenbauer,Hong-Min Chu,Gowthami Somepalli,Brian R. Bartoldson,Bhavya Kailkhura,Avi Schwarzschild,Aniruddha Saha,Micah Goldblum,Jonas Geiping,Tom Goldstein", "tags": "ICLR 2024,Poster", "abstract": "We show that language model finetuning can be improved, sometimes dramatically, with a simple augmentation. \nNEFTune adds noise to the embedding vectors during training.\nStandard finetuning of LLaMA-2-7B using Alpaca achieves $29.79$\\% on AlpacaEval, which rises to $64.69$\\% using noisy embeddings. NEFTune also improves over strong baselines on modern instruction datasets.\nModels trained with Evol-Instruct see a $10$\\% improvement, with ShareGPT an $8$\\% improvement, and with OpenPlatypus an $8$\\% improvement. \nEven powerful models further refined with RLHF such as LLaMA-2-Chat benefit from additional training with NEFTune. Particularly, we see these improvements on the conversational abilities of the instruction model and not on traditional tasks like those on the OpenLLM Leaderboard, where performance is the same.", "pdf": "https://openreview.net/pdf/ab62341bb5a8f427d4042033bd3b7c9f59652b51.pdf"} {"title": "An operator preconditioning perspective on training in physics-informed machine learning", "url": "https://openreview.net/forum?id=WWlxFtR5sV", "detail_url": "https://openreview.net/forum?id=WWlxFtR5sV", "authors": "Tim De Ryck,Florent Bonnet,Siddhartha Mishra,Emmanuel de Bezenac", "tags": "ICLR 2024,Poster", "abstract": "In this paper, we investigate the behavior of gradient descent algorithms in physics-informed machine learning methods like PINNs, which minimize residuals connected to partial differential equations (PDEs). Our key result is that the difficulty in training these models is closely related to the conditioning of a specific differential operator. This operator, in turn, is associated to the Hermitian square of the differential operator of the underlying PDE. If this operator is ill-conditioned, it results in slow or infeasible training. Therefore, preconditioning this operator is crucial. We employ both rigorous mathematical analysis and empirical evaluations to investigate various strategies, explaining how they better condition this critical operator, and consequently improve training.", "pdf": "https://openreview.net/pdf/bb2db7fa354d99feb2b3b02f70bf881aab7de2f6.pdf"} {"title": "Two-stage LLM Fine-tuning with Less Specialization and More Generalization", "url": "https://openreview.net/forum?id=pCEgna6Qco", "detail_url": "https://openreview.net/forum?id=pCEgna6Qco", "authors": "Yihan Wang,Si Si,Daliang Li,Michal Lukasik,Felix Yu,Cho-Jui Hsieh,Inderjit S Dhillon,Sanjiv Kumar", "tags": "ICLR 2024,Poster", "abstract": "Pretrained large language models (LLMs) are general purpose problem solvers applicable to a diverse set of tasks with prompts. They can be further improved towards a specific task by fine-tuning on a specialized dataset. However, fine-tuning usually makes the model narrowly specialized on this dataset with reduced general in-context learning performances, which is undesirable whenever the fine-tuned model needs to handle additional tasks where no fine-tuning data is available. \nIn this work, we first demonstrate that fine-tuning on a single task indeed decreases LLMs' general in-context learning performance. We discover one important cause of such forgetting, format specialization, where the model overfits to the format of the fine-tuned task.We further show that format specialization happens at the very beginning of fine-tuning. To solve this problem, we propose Prompt Tuning with MOdel Tuning (ProMoT), a simple yet effective two-stage fine-tuning framework that reduces format specialization and improves generalization.ProMoT offloads task-specific format learning into additional and removable parameters by first doing prompt tuning and then fine-tuning the model itself with this soft prompt attached. \nWith experiments on several fine-tuning tasks and 8 in-context evaluation tasks, we show that ProMoT achieves comparable performance on fine-tuned tasks to standard fine-tuning, but with much less loss of in-context learning performances across a board range of out-of-domain evaluation tasks. More importantly, ProMoT can even enhance generalization on in-context learning tasks that are semantically related to the fine-tuned task, e.g. ProMoT on En-Fr translation significantly improves performance on other language pairs, and ProMoT on NLI improves performance on summarization.\nExperiments also show that ProMoT can improve the generalization performance of multi-task training.", "pdf": "https://openreview.net/pdf/967d482dc9ddfbf84e0116bc826d1f2205123e77.pdf"} {"title": "Expressive Losses for Verified Robustness via Convex Combinations", "url": "https://openreview.net/forum?id=mzyZ4wzKlM", "detail_url": "https://openreview.net/forum?id=mzyZ4wzKlM", "authors": "Alessandro De Palma,Rudy R Bunel,Krishnamurthy Dj Dvijotham,M. Pawan Kumar,Robert Stanforth,Alessio Lomuscio", "tags": "ICLR 2024,Poster", "abstract": "In order to train networks for verified adversarial robustness, it is common to over-approximate the worst-case loss over perturbation regions, resulting in networks that attain verifiability at the expense of standard performance.\nAs shown in recent work, better trade-offs between accuracy and robustness can be obtained by carefully coupling adversarial training with over-approximations. \nWe hypothesize that the expressivity of a loss function, which we formalize as the ability to span a range of trade-offs between lower and upper bounds to the worst-case loss through a single parameter (the over-approximation coefficient), is key to attaining state-of-the-art performance. \nTo support our hypothesis, we show that trivial expressive losses, obtained via convex combinations between adversarial attacks and IBP bounds, yield state-of-the-art results across a variety of settings in spite of their conceptual simplicity.\nWe provide a detailed analysis of the relationship between the over-approximation coefficient and performance profiles across different expressive losses, showing that, while expressivity is essential, better approximations of the worst-case loss are not necessarily linked to superior robustness-accuracy trade-offs.", "pdf": "https://openreview.net/pdf/40ce9bb1de770f815cc5905ce78d2b4957712ed0.pdf"} {"title": "Learning Adaptive Multiresolution Transforms via Meta-Framelet-based Graph Convolutional Network", "url": "https://openreview.net/forum?id=5RielfrDkP", "detail_url": "https://openreview.net/forum?id=5RielfrDkP", "authors": "Tianze Luo,Zhanfeng Mo,Sinno Jialin Pan", "tags": "ICLR 2024,Poster", "abstract": "Graph Neural Networks are popular tools in graph representation learning that capture the graph structural properties. However, most GNNs employ single-resolution graph feature extraction, thereby failing to capture micro-level local patterns (high resolution) and macro-level graph cluster and community patterns (low resolution) simultaneously. Many multiresolution methods have been developed to capture graph patterns at multiple scales, but most of them depend on predefined and handcrafted multiresolution transforms that remain fixed throughout the training process once formulated. Due to variations in graph instances and distributions, fixed handcrafted transforms can not effectively tailor multiresolution representations to each graph instance. To acquire multiresolution representation suited to different graph instances and distributions, we introduce the Multiresolution Meta-Framelet-based Graph Convolutional Network (MM-FGCN), facilitating comprehensive and adaptive multiresolution analysis across diverse graphs. Extensive experiments demonstrate that our MM-FGCN achieves SOTA performance on various graph learning tasks.", "pdf": "https://openreview.net/pdf/78d38cf91a8f58ca68193e324a079c22002fffa8.pdf"} {"title": "REFACTOR: Learning to Extract Theorems from Proofs", "url": "https://openreview.net/forum?id=fgKjiVrm6u", "detail_url": "https://openreview.net/forum?id=fgKjiVrm6u", "authors": "Jin Peng Zhou,Yuhuai Wu,Qiyang Li,Roger Baker Grosse", "tags": "ICLR 2024,Poster", "abstract": "Human mathematicians are often good at recognizing modular and reusable theorems that make complex mathematical results within reach. In this paper, we propose a novel method called theoREm-from-prooF extrACTOR (REFACTOR) for training neural networks to mimic this ability in formal mathematical theorem proving. We show on a set of unseen proofs, REFACTOR is able to extract 19.6\\% of the theorems that humans would use to write the proofs. When applying the model to the existing Metamath library, REFACTOR extracted 16 new theorems. With newly extracted theorems, we show that the existing proofs in the MetaMath database can be refactored. The new theorems are used very frequently after refactoring, with an average usage of 733.5 times, and help shorten the proof lengths. Lastly, we demonstrate that the prover trained on the new-theorem refactored dataset proves more test theorems and outperforms state-of-the-art baselines by frequently leveraging a diverse set of newly extracted theorems. Code can be found at https://github.com/jinpz/refactor.", "pdf": "https://openreview.net/pdf/e4944ce26133003e30b0fe056311fca837baa9a1.pdf"} {"title": "Let's do the time-warp-attend: Learning topological invariants of dynamical systems", "url": "https://openreview.net/forum?id=Fj7Fzm5lWL", "detail_url": "https://openreview.net/forum?id=Fj7Fzm5lWL", "authors": "Noa Moriel,Matt Ricci,Mor Nitzan", "tags": "ICLR 2024,Poster", "abstract": "Dynamical systems across the sciences, from electrical circuits to ecological networks, undergo qualitative and often catastrophic changes in behavior, called bifurcations, when their underlying parameters cross a threshold. Existing methods predict oncoming catastrophes in individual systems but are primarily time-series-based and struggle both to categorize qualitative dynamical regimes across diverse systems and to generalize to real data. To address this challenge, we propose a data-driven, physically-informed deep-learning framework for classifying dynamical regimes and characterizing bifurcation boundaries based on the extraction of topologically invariant features. We focus on the paradigmatic case of the supercritical Hopf bifurcation, which is used to model periodic dynamics across a wide range of applications. Our convolutional attention method is trained with data augmentations that encourage the learning of topological invariants which can be used to detect bifurcation boundaries in unseen systems and to design models of biological systems like oscillatory gene regulatory networks. We further demonstrate our method's use in analyzing real data by recovering distinct proliferation and differentiation dynamics along pancreatic endocrinogenesis trajectory in gene expression space based on single-cell data. Our method provides valuable insights into the qualitative, long-term behavior of a wide range of dynamical systems, and can detect bifurcations or catastrophic transitions in large-scale physical and biological systems.", "pdf": "https://openreview.net/pdf/8dc9a0a901c8cc5a3348f7f9bb9b466a836c04e9.pdf"} {"title": "Sparse MoE with Language Guided Routing for Multilingual Machine Translation", "url": "https://openreview.net/forum?id=ySS7hH1smL", "detail_url": "https://openreview.net/forum?id=ySS7hH1smL", "authors": "Xinyu Zhao,Xuxi Chen,Yu Cheng,Tianlong Chen", "tags": "ICLR 2024,Poster", "abstract": "Sparse Mixture-of-Experts (SMoE) has gained increasing popularity as a promising framework for scaling up multilingual machine translation (MMT) models with negligible extra computational overheads. However, current SMoE solutions neglect the intrinsic structures of the MMT problem: ($a$) $\\textit{Linguistics Hierarchy.}$ Languages are naturally grouped according to their lingual properties like genetic families, phonological characteristics, etc; ($b$) $\\textit{Language Complexity.}$ The learning difficulties are varied for diverse languages due to their grammar complexity, available resources, etc. Therefore, routing a fixed number of experts (e.g., $1$ or $2$ experts in usual) only at the word level leads to inferior performance. To fill in the missing puzzle, we propose $\\textbf{\\texttt{Lingual-SMoE}}$ by equipping the SMoE with adaptive and linguistic-guided routing policies. Specifically, it ($1$) extracts language representations to incorporate linguistic knowledge and uses them to allocate experts into different groups; ($2$) determines the number of activated experts for each target language in an adaptive and automatic manner, according to their translation difficulties, which aims to mitigate the potential over-/under-fitting issues of learning simple/challenges translations. Sufficient experimental studies on MMT benchmarks with {$16$, $50$, $100$} language pairs and various network architectures, consistently validate the superior performance of our proposals. For instance, $\\texttt{Lingual-SMoE}$ outperforms its dense counterpart by over $5\\%$ BLEU scores on $\\texttt{OPUS-100}$ dataset.", "pdf": "https://openreview.net/pdf/2f14a8bdcce65ee1907499fef8fcdf05b7ce4f91.pdf"} {"title": "Detecting Pretraining Data from Large Language Models", "url": "https://openreview.net/forum?id=zWqr3MQuNs", "detail_url": "https://openreview.net/forum?id=zWqr3MQuNs", "authors": "Weijia Shi,Anirudh Ajith,Mengzhou Xia,Yangsibo Huang,Daogao Liu,Terra Blevins,Danqi Chen,Luke Zettlemoyer", "tags": "ICLR 2024,Poster", "abstract": "Although large language models (LLMs) are widely deployed, the data used to train them is rarely disclosed. Given the incredible scale of this data, up to trillions of tokens, it is all but certain that it includes potentially problematic text such as copyrighted materials, personally identifiable information, and test data for widely reported reference benchmarks. However, we currently have no way to know which data of these types is included or in what proportions. In this paper, we study the pretraining data detection problem: given a piece of text and black-box access to an LLM without knowing the pretraining data, can we determine if the model was trained on the provided text? To facilitate this study, we introduce a dynamic benchmark WIKIMIA that uses data created before and after model training to support gold truth detection. We also introduce a new detection method MIN-K PROB based on a simple hypothesis: an unseen example is likely to contain a few outlier words with low probabilities under the LLM, while a seen example is less likely to have words with such low probabilities. MIN-K PROB can be applied without any knowledge about the pretrainig corpus or any additional training, departing from previous detection methods that require training a reference model on data that is similar to the pretraining data. Moreover, our experiments demonstrate that MIN-K PROB achieves a 7.4% improvement on WIKIMIA over these previous methods. We apply MIN-K PROB to two real-world scenarios, copyrighted book detection and contaminated downstream example detection, and find that it to be a consistently effective solution.", "pdf": "https://openreview.net/pdf/0d224da3c14b884532028169f92fd03868dcd86e.pdf"} {"title": "Don't Trust: Verify -- Grounding LLM Quantitative Reasoning with Autoformalization", "url": "https://openreview.net/forum?id=V5tdi14ple", "detail_url": "https://openreview.net/forum?id=V5tdi14ple", "authors": "Jin Peng Zhou,Charles E Staats,Wenda Li,Christian Szegedy,Kilian Q Weinberger,Yuhuai Wu", "tags": "ICLR 2024,Poster", "abstract": "Large language models (LLM), such as Google's Minerva and OpenAI's GPT families, are becoming increasingly capable of solving mathematical quantitative reasoning problems. However, they still make unjustified logical and computational errors in their reasoning steps and answers. In this paper, we leverage the fact that if the training corpus of LLMs contained sufficiently many examples of formal mathematics (e.g. in Isabelle, a formal theorem proving environment), they can be prompted to translate i.e. autoformalize informal mathematical statements into formal Isabelle code --- which can be verified automatically for internal consistency. This provides a mechanism to automatically reject solutions whose formalized versions are inconsistent within themselves or with the formalized problem statement. We evaluate our method on GSM8K, MATH and MultiArith datasets and demonstrate that our approach provides a consistently better heuristic than vanilla majority voting --- the previously best method to identify correct answers, by more than 12\\% on GSM8K. In our experiments it improves results consistently across all datasets and LLM model sizes. The code can be found at https://github.com/jinpz/dtv.", "pdf": "https://openreview.net/pdf/f2feee3c9af794ca03a85d1ab69dd06c8330981e.pdf"} {"title": "PubDef: Defending Against Transfer Attacks From Public Models", "url": "https://openreview.net/forum?id=Tvwf4Vsi5F", "detail_url": "https://openreview.net/forum?id=Tvwf4Vsi5F", "authors": "Chawin Sitawarin,Jaewon Chang,David Huang,Wesson Altoyan,David Wagner", "tags": "ICLR 2024,Poster", "abstract": "Adversarial attacks have been a looming and unaddressed threat in the industry. However, through a decade-long history of the robustness evaluation literature, we have learned that mounting a strong or optimal attack is challenging. It requires both machine learning and domain expertise. In other words, the white-box threat model, religiously assumed by a large majority of the past literature, is unrealistic. In this paper, we propose a new practical threat model where the adversary relies on transfer attacks through publicly available surrogate models. We argue that this setting will become the most prevalent for security-sensitive applications in the future. We evaluate the transfer attacks in this setting and propose a specialized defense method based on a game-theoretic perspective. The defenses are evaluated under 24 public models and 11 attack algorithms across three datasets (CIFAR-10, CIFAR-100, and ImageNet). Under this threat model, our defense, PubDef, outperforms the state-of-the-art white-box adversarial training by a large margin with almost no loss in the normal accuracy. For instance, on ImageNet, our defense achieves 62% accuracy under the strongest transfer attack vs only 36% of the best adversarially trained model. Its accuracy when not under attack is only 2% lower than that of an undefended model (78% vs 80%). We release our code at https://github.com/wagner-group/pubdef.", "pdf": "https://openreview.net/pdf/1c48a415047176b3fbdb5c594a119648fea8e3d8.pdf"} {"title": "AutomaTikZ: Text-Guided Synthesis of Scientific Vector Graphics with TikZ", "url": "https://openreview.net/forum?id=v3K5TVP8kZ", "detail_url": "https://openreview.net/forum?id=v3K5TVP8kZ", "authors": "Jonas Belouadi,Anne Lauscher,Steffen Eger", "tags": "ICLR 2024,Poster", "abstract": "Generating bitmap graphics from text has gained considerable attention, yet for scientific figures, vector graphics are often preferred. Given that vector graphics are typically encoded using low-level graphics primitives, generating them directly is difficult. To address this, we propose the use of TikZ, a well-known abstract graphics language that can be compiled to vector graphics, as an intermediate representation of scientific figures. TikZ offers human-oriented, high-level commands, thereby facilitating conditional language modeling with any large language model. To this end, we introduce DaTikZ the first large-scale TikZ dataset, consisting of 120k TikZ drawings aligned with captions. We fine-tune LLaMA on DaTikZ, as well as our new model CLiMA, which augments LLaMA with multimodal CLIP embeddings. In both human and automatic evaluation, CLiMA and LLaMA outperform commercial GPT-4 and Claude 2 in terms of similarity to human-created figures, with CLiMA additionally improving text-image alignment. Our detailed analysis shows that all models generalize well and are not susceptible to memorization. GPT-4 and Claude 2, however, tend to generate more simplistic figures compared to both humans and our models. We make our framework, AutomaTikZ, along with model weights and datasets, publicly available.", "pdf": "https://openreview.net/pdf/eb9ccf8fa1a97129a5a38ac09ddf4f1257daa864.pdf"} {"title": "Sharpness-Aware Minimization Enhances Feature Quality via Balanced Learning", "url": "https://openreview.net/forum?id=3xDaj4pRna", "detail_url": "https://openreview.net/forum?id=3xDaj4pRna", "authors": "Jacob Mitchell Springer,Vaishnavh Nagarajan,Aditi Raghunathan", "tags": "ICLR 2024,Poster", "abstract": "Sharpness-Aware Minimization (SAM) has emerged as a promising alternative optimizer to stochastic gradient descent (SGD). The originally-proposed motivation behind SAM was to bias neural networks towards flatter minima that are believed to generalize better. However, recent studies have shown conflicting evidence on the relationship between flatness and generalization, suggesting that flatness does fully explain SAM's success. Sidestepping this debate, we identify an orthogonal effect of SAM that is beneficial out-of-distribution: we argue that SAM implicitly balances the quality of diverse features. SAM achieves this effect by adaptively suppressing well-learned features which gives remaining features opportunity to be learned. We show that this mechanism is beneficial in datasets that contain redundant or spurious features where SGD falls for the simplicity bias and would not otherwise learn all available features. Our insights are supported by experiments on real data: we demonstrate that SAM improves the quality of features in datasets containing redundant or spurious features, including CelebA, Waterbirds, CIFAR-MNIST, and DomainBed.", "pdf": "https://openreview.net/pdf/b581b037f272dae403bd0933d1d9c3f7c144a1a9.pdf"} {"title": "Can LLM-Generated Misinformation Be Detected?", "url": "https://openreview.net/forum?id=ccxD4mtkTU", "detail_url": "https://openreview.net/forum?id=ccxD4mtkTU", "authors": "Canyu Chen,Kai Shu", "tags": "ICLR 2024,Poster", "abstract": "The advent of Large Language Models (LLMs) has made a transformative impact. However, the potential that LLMs such as ChatGPT can be exploited to generate misinformation has posed a serious concern to online safety and public trust. A fundamental research question is: will LLM-generated misinformation cause more harm than human-written misinformation? We propose to tackle this question from the perspective of detection difficulty. We first build a taxonomy of LLM-generated misinformation. Then we categorize and validate the potential real-world methods for generating misinformation with LLMs. Then, through extensive empirical investigation, we discover that LLM-generated misinformation can be harder to detect for humans and detectors compared to human-written misinformation with the same semantics, which suggests it can have more deceptive styles and potentially cause more harm. We also discuss the implications of our discovery on combating misinformation in the age of LLMs and the countermeasures.", "pdf": "https://openreview.net/pdf/4070fd2a7564a91bda1c7908c2650e500a1d18d8.pdf"} {"title": "A Simple Interpretable Transformer for Fine-Grained Image Classification and Analysis", "url": "https://openreview.net/forum?id=bkdWThqE6q", "detail_url": "https://openreview.net/forum?id=bkdWThqE6q", "authors": "DIPANJYOTI PAUL,Arpita Chowdhury,Xinqi Xiong,Feng-Ju Chang,David Edward Carlyn,Samuel Stevens,Kaiya L Provost,Anuj Karpatne,Bryan Carstens,Daniel Rubenstein,Charles Stewart,Tanya Berger-Wolf,Yu Su,Wei-Lun Chao", "tags": "ICLR 2024,Poster", "abstract": "We present a novel usage of Transformers to make image classification interpretable. Unlike mainstream classifiers that wait until the last fully connected layer to incorporate class information to make predictions, we investigate a proactive approach, asking each class to search for itself in an image. We realize this idea via a Transformer encoder-decoder inspired by DEtection TRansformer (DETR). We learn ''class-specific'' queries (one for each class) as input to the decoder, enabling each class to localize its patterns in an image via cross-attention. We name our approach INterpretable TRansformer (INTR), which is fairly easy to implement and exhibits several compelling properties. We show that INTR intrinsically encourages each class to attend distinctively; the cross-attention weights thus provide a faithful interpretation of the prediction. Interestingly, via ''multi-head'' cross-attention, INTR could identify different ''attributes'' of a class, making it particularly suitable for fine-grained classification and analysis, which we demonstrate on eight datasets. Our code and pre-trained models are publicly accessible at the Imageomics Institute GitHub site: https://github.com/Imageomics/INTR.", "pdf": "https://openreview.net/pdf/edc88ec0b85f5a82a7aed747683ae161b423d1ab.pdf"} {"title": "One-shot Active Learning Based on Lewis Weight Sampling for Multiple Deep Models", "url": "https://openreview.net/forum?id=EDXkkUAIFW", "detail_url": "https://openreview.net/forum?id=EDXkkUAIFW", "authors": "Sheng-Jun Huang,Yi Li,Yiming Sun,Ying-Peng Tang", "tags": "ICLR 2024,Poster", "abstract": "Active learning (AL) for multiple target models aims to reduce labeled data querying while effectively training multiple models concurrently. Existing AL algorithms often rely on iterative model training, which can be computationally expensive, particularly for deep models. In this paper, we propose a one-shot AL method to address this challenge, which performs all label queries without repeated model training. Specifically, we extract different representations of the same dataset using distinct network backbones, and actively learn the linear prediction layer on each representation via an $\\ell_p$-regression formulation. The regression problems are solved approximately by \nsampling and reweighting the unlabeled instances based on their maximum Lewis weights across the representations. An upper bound on the number of samples needed is provided with a rigorous analysis for $p\\in [1, +\\infty)$. Experimental results on 11 benchmarks show that our one-shot approach achieves competitive performances with the state-of-the-art AL methods for multiple target models.", "pdf": "https://openreview.net/pdf/151a8c3cbd0951af8b2abf96dd2fee8beaee4ae7.pdf"} {"title": "Disentangling Time Series Representations via Contrastive Independence-of-Support on l-Variational Inference", "url": "https://openreview.net/forum?id=iI7hZSczxE", "detail_url": "https://openreview.net/forum?id=iI7hZSczxE", "authors": "Khalid Oublal,Said Ladjal,David Benhaiem,Emmanuel LE BORGNE,Fran\u00e7ois Roueff", "tags": "ICLR 2024,Poster", "abstract": "Learning disentangled representations for time series is a promising path to facilitate reliable generalization to in- and out-of distribution (OOD), offering benefits like feature derivation and improved interpretability and fairness, thereby enhancing downstream tasks. We focus on disentangled representation learning for home appliance electricity usage, enabling users to understand and optimize their consumption for a reduced carbon footprint. Our approach frames the problem as disentangling each attribute's role in total consumption. Unlike existing methods assuming attribute independence which leads to non-identiability, we acknowledge real-world time series attribute correlations, learned up to a smooth bijection using contrastive learning and a single autoencoder. To address this, we propose a Disentanglement under Independence-Of-Support via Contrastive Learning (DIOSC), facilitating representation generalization across diverse correlated scenarios. Our method utilizes innovative \\textit{l}-variational inference layers with self-attention, effectively addressing temporal dependencies across bottom-up and top-down networks. We find that DIOSC can enhance the task of representation of time series electricity consumption. We introduce TDS (Time Disentangling Score) to gauge disentanglement quality. TDS reliably reflects disentanglement performance, making it a valuable metric for evaluating time series representations disentanglement. Code available at https://institut-polytechnique-de-paris.github.io/time-disentanglement-lib.", "pdf": "https://openreview.net/pdf/ef5c6cb813437c7c5c77405a079228ac7c8b50bd.pdf"} {"title": "Improved algorithm and bounds for successive projection", "url": "https://openreview.net/forum?id=GlpawHh80l", "detail_url": "https://openreview.net/forum?id=GlpawHh80l", "authors": "Jiashun Jin,Tracy Ke,Gabriel Moryoussef,Jiajun Tang,Jingming Wang", "tags": "ICLR 2024,Poster", "abstract": "Consider a $K$-vertex simplex in a $d$-dimensional space. We measure $n$ points on the simplex, but due to the measurement noise, \nsome of the observed points fall outside the simplex. The interest is vertex hunting (i.e., estimating the vertices of the simplex). The successive projection algorithm (SPA) is one of the most popular approaches to vertex hunting, but it is vulnerable to noise and outliers, and may perform unsatisfactorily. We propose pseudo-point SPA (pp-SPA) as a new approach to vertex hunting. The approach contains \ntwo novel ideas (a projection step and a denoise step) and generates roughly $n$ pseudo-points, which can be fed in to SPA for vertex hunting. For theory, we first derive an improved non-asymptotic bound for the orthodox SPA, and then use the result to derive the bounds for pp-SPA. Compared with the orthodox SPA, pp-SPA has a faster rate and more satisfactory numerical performance in a broad setting. The analysis is quite delicate: the non-asymptotic bound is hard to derive, and we need precise results on the extreme values of (possibly) high-dimensional random vectors.", "pdf": "https://openreview.net/pdf/7ebdcea695b8aa3a70d605879db5781a831c5f96.pdf"} {"title": "Distributional Preference Learning: Understanding and Accounting for Hidden Context in RLHF", "url": "https://openreview.net/forum?id=0tWTxYYPnW", "detail_url": "https://openreview.net/forum?id=0tWTxYYPnW", "authors": "Anand Siththaranjan,Cassidy Laidlaw,Dylan Hadfield-Menell", "tags": "ICLR 2024,Poster", "abstract": "In practice, preference learning from human feedback depends on incomplete data with hidden context. Hidden context refers to data that affects the feedback received, but which is not represented in the data used to train a preference model. This captures common issues of data collection, such as having human annotators with varied preferences, cognitive processes that result in seemingly irrational behavior, and combining data labeled according to different criteria. We prove that standard applications of preference learning, including reinforcement learning from human feedback (RLHF), implicitly aggregate over hidden contexts according to a well-known voting rule called *Borda count*. We show this can produce counter-intuitive results that are very different from other methods which implicitly aggregate via expected utility. Furthermore, our analysis formalizes the way that preference learning from users with diverse values tacitly implements a social choice function. A key implication of this result is that annotators have an incentive to misreport their preferences in order to influence the learned model, leading to vulnerabilities in the deployment of RLHF. As a step towards mitigating these problems, we introduce a class of methods called *distributional preference learning* (DPL). DPL methods estimate a distribution of possible score values for each alternative in order to better account for hidden context. Experimental results indicate that applying DPL to RLHF for LLM chatbots identifies hidden context in the data and significantly reduces subsequent jailbreak vulnerability.", "pdf": "https://openreview.net/pdf/6a63d0436aca59c5082cae0bf2ea3a6c0aea6257.pdf"} {"title": "Estimating Shape Distances on Neural Representations with Limited Samples", "url": "https://openreview.net/forum?id=kvByNnMERu", "detail_url": "https://openreview.net/forum?id=kvByNnMERu", "authors": "Dean A Pospisil,Brett W. Larsen,Sarah E Harvey,Alex H Williams", "tags": "ICLR 2024,Poster", "abstract": "Measuring geometric similarity between high-dimensional network representations is a topic of longstanding interest to neuroscience and deep learning. Although many methods have been proposed, only a few works have rigorously analyzed their statistical efficiency or quantified estimator uncertainty in data-limited regimes. Here, we derive upper and lower bounds on the worst-case convergence\nof standard estimators of shape distance\u2014a measure of representational dissimilarity proposed by Williams et al. (2021). These bounds reveal the challenging nature of the problem in high-dimensional feature spaces. To overcome these challenges, we introduce a novel method-of-moments estimator with a tunable bias-variance tradeoff parameterized by an upper bound on bias. We show that this estimator achieves superior performance to standard estimators in simulation and on neural data, particularly in high-dimensional settings. Our theoretical work and estimator thus respectively define and dramatically expand the scope of neural data for which geometric similarity can be accurately measured.", "pdf": "https://openreview.net/pdf/cc1e959c8a6eec0004bd40509131279fa05e6610.pdf"} {"title": "Learning semilinear neural operators: A unified recursive framework for prediction and data assimilation.", "url": "https://openreview.net/forum?id=ZMv6zKYYUs", "detail_url": "https://openreview.net/forum?id=ZMv6zKYYUs", "authors": "Ashutosh Singh,Ricardo Augusto Borsoi,Deniz Erdogmus,Tales Imbiriba", "tags": "ICLR 2024,Poster", "abstract": "Recent advances in the theory of Neural Operators (NOs) have enabled fast and accurate computation of the solutions to complex systems described by partial differential equations (PDEs). Despite their great success, current NO-based solutions face important challenges when dealing with spatio-temporal PDEs over long time scales. Specifically, the current theory of NOs does not present a systematic framework to perform data assimilation and efficiently correct the evolution of PDE solutions over time based on sparsely sampled noisy measurements. In this paper, we propose a learning-based state-space approach to compute the solution operators to infinite-dimensional semilinear PDEs. Exploiting the structure of semilinear PDEs and the theory of nonlinear observers in function spaces, we develop a flexible recursive method that allows for both prediction and data assimilation by combining prediction and correction operations. The proposed framework is capable of producing fast and accurate predictions over long time horizons, dealing with irregularly sampled noisy measurements to correct the solution, and benefits from the decoupling between the spatial and temporal dynamics of this class of PDEs. We show through experiments on the Kuramoto-Sivashinsky, Navier-Stokes and Korteweg-de Vries equations that the proposed model is robust to noise and can leverage arbitrary amounts of measurements to correct its prediction over a long time horizon with little computational overhead.", "pdf": "https://openreview.net/pdf/1560d1f0bd3603f9f3ab40125342ea9969adfe14.pdf"} {"title": "Eureka: Human-Level Reward Design via Coding Large Language Models", "url": "https://openreview.net/forum?id=IEduRUO55F", "detail_url": "https://openreview.net/forum?id=IEduRUO55F", "authors": "Yecheng Jason Ma,William Liang,Guanzhi Wang,De-An Huang,Osbert Bastani,Dinesh Jayaraman,Yuke Zhu,Linxi Fan,Anima Anandkumar", "tags": "ICLR 2024,Poster", "abstract": "Large Language Models (LLMs) have excelled as high-level semantic planners for sequential decision-making tasks. However, harnessing them to learn complex low-level manipulation tasks, such as dexterous pen spinning, remains an open problem. We bridge this fundamental gap and present Eureka, a human-level reward design algorithm powered by LLMs. Eureka exploits the remarkable zero-shot generation, code-writing, and in-context improvement capabilities of state-of-the-art LLMs, such as GPT-4, to perform evolutionary optimization over reward code. The resulting rewards can then be used to acquire complex skills via reinforcement learning. Without any task-specific prompting or pre-defined reward templates, Eureka generates reward functions that outperform expert human-engineered rewards. In a diverse suite of 29 open-source RL environments that include 10 distinct robot morphologies, Eureka outperforms human experts on 83% of the tasks, leading to an average normalized improvement of 52%. The generality of Eureka also enables a new gradient-free in-context learning approach to reinforcement learning from human feedback (RLHF), readily incorporating human inputs to improve the quality and the safety of the generated rewards without model updating. Finally, using Eureka rewards in a curriculum learning setting, we demonstrate for the first time, a simulated Shadow Hand capable of performing pen spinning tricks, adeptly manipulating a pen in circles at rapid speed.", "pdf": "https://openreview.net/pdf/6c6607629a103a06c7c1b52817845f25aa866b8b.pdf"} {"title": "f-FERM: A Scalable Framework for Robust Fair Empirical Risk Minimization", "url": "https://openreview.net/forum?id=s90VIdza2K", "detail_url": "https://openreview.net/forum?id=s90VIdza2K", "authors": "Sina Baharlouei,Shivam Patel,Meisam Razaviyayn", "tags": "ICLR 2024,Poster", "abstract": "Training and deploying machine learning models that meet fairness criteria for protected groups are fundamental in modern artificial intelligence. \nWhile numerous constraints and regularization terms have been proposed in the literature to promote fairness in machine learning tasks, most of these approaches are not amenable to stochastic optimization due to the complex and nonlinear structure of constraints and regularizers. Here, the term ``stochastic'' refers to the ability of the algorithm to work with small mini-batches of data. Motivated by the limitation of existing literature, this paper presents a unified stochastic optimization framework for fair empirical risk minimization based on $f$-divergence measures ($f$-FERM). The proposed stochastic algorithm enjoys theoretical convergence guarantees. In addition, our experiments demonstrate the superiority of fairness-accuracy tradeoffs offered by $f$-FERM for almost all batch sizes (ranging from full-batch to batch size of one). Moreover, we show that our framework can be extended to the case where there is a distribution shift from training to the test data. \nOur extension is based on a distributionally robust optimization reformulation of $f$-FERM objective under $\\ell_p$ norms as uncertainty sets. Again, in this distributionally robust setting, $f$-FERM not only enjoys theoretical convergence guarantees but also outperforms other baselines in the literature in the tasks involving distribution shifts. \n An efficient stochastic implementation of $f$-FERM is publicly available.", "pdf": "https://openreview.net/pdf/1d2b7d920918fe102d32b04476b29c2c607925bd.pdf"} {"title": "Source-Free and Image-Only Unsupervised Domain Adaptation for Category Level Object Pose Estimation", "url": "https://openreview.net/forum?id=UPvufoBAIs", "detail_url": "https://openreview.net/forum?id=UPvufoBAIs", "authors": "Prakhar Kaushik,Aayush Mishra,Adam Kortylewski,Alan Yuille", "tags": "ICLR 2024,Poster", "abstract": "We consider the problem of source-free unsupervised category-level 3D pose estimation from only RGB images to an non-annotated and unlabelled target domain without any access to source domain data or annotations during adaptation. Collecting and annotating real world 3D data and corresponding images is laborious, expensive yet unavoidable process since even 3D pose domain adaptation methods require 3D data in the target domain. We introduce a method which is capable of adapting to a nuisance ridden target domain without any 3D data or annotations. We represent object categories as simple cuboid meshes, and harness a generative model of neural feature activations modeled as a von Mises Fisher distribution at each mesh vertex learnt using differential rendering. We focus on individual mesh vertex features and iteratively update them based on their proximity to corresponding features in the target domain. Our key insight stems from the observation that specific object subparts remain stable across out-of-domain (OOD) scenarios, enabling strategic utilization of these invariant subcomponents for effective model updates. Our model is then trained in an EM fashion alternating between updating the vertex features and feature extractor. We show that our method simulates fine-tuning on a global-pseudo labelled dataset under mild assumptions which converges to the target domain asymptotically. Through extensive empirical validation, we demonstrate the potency of our simple approach in addressing the domain shift challenge and significantly enhancing pose estimation accuracy. By accentuating robust and less changed object subcomponents, our framework contributes to the evolution of UDA techniques in the context of 3D pose estimation using only images from the target domain.", "pdf": "https://openreview.net/pdf/2713fa61c35aea3d714ed1c74104eb90194b1f28.pdf"} {"title": "Closing the Curious Case of Neural Text Degeneration", "url": "https://openreview.net/forum?id=dONpC9GL1o", "detail_url": "https://openreview.net/forum?id=dONpC9GL1o", "authors": "Matthew Finlayson,John Hewitt,Alexander Koller,Swabha Swayamdipta,Ashish Sabharwal", "tags": "ICLR 2024,Poster", "abstract": "Despite their ubiquity in language generation, it remains unknown why truncation sampling heuristics like nucleus sampling are so effective. We provide a theoretical explanation for the effectiveness of the truncation sampling by proving that truncation methods that discard tokens below some probability threshold (the most common type of truncation) can guarantee that all sampled tokens have nonzero true probability. However, thresholds are a coarse heuristic, and necessarily discard some tokens with nonzero true probability as well. In pursuit of a more precise sampling strategy, we show that we can leverage a known source of model errors, the softmax bottleneck, to prove that certain tokens have nonzero true probability, without relying on a threshold. Based on our findings, we develop an experimental truncation strategy and the present pilot studies demonstrating the promise of this type of algorithm. Our evaluations show that our method outperforms its threshold-based counterparts under automatic and human evaluation metrics for low-entropy (i.e., close to greedy) open-ended text generation. Our theoretical findings and pilot experiments provide both insight into why truncation sampling works, and make progress toward more expressive sampling algorithms that better surface the generative capabilities of large language models.", "pdf": "https://openreview.net/pdf/3208517640b1a7c17b6a44cc2d44f67bdc006e33.pdf"} {"title": "Mediator Interpretation and Faster Learning Algorithms for Linear Correlated Equilibria in General Sequential Games", "url": "https://openreview.net/forum?id=bsKMPAFHO7", "detail_url": "https://openreview.net/forum?id=bsKMPAFHO7", "authors": "Brian Hu Zhang,Gabriele Farina,Tuomas Sandholm", "tags": "ICLR 2024,Poster", "abstract": "A recent paper by Farina and Pipis (2023) established the existence of uncoupled no-linear-swap regret dynamics with polynomial-time iterations in extensive-form games. The equilibrium points reached by these dynamics, known as linear correlated equilibria, are currently the tightest known relaxation of correlated equilibrium that can be learned in polynomial time in any finite extensive-form game. However, their properties remain vastly unexplored, and their computation is onerous. In this paper, we provide several contributions shedding light on the fundamental nature of linear-swap regret. First, we show a connection between linear deviations and a generalization of communication deviations in which the player can make queries to a ``mediator'' who replies with action recommendations, and, critically, the player is not constrained to match the timing of the game as would be the case for communication deviations. We coin this latter set the untimed communication (UTC) deviations. We show that the UTC deviations coincide precisely with the linear deviations, and therefore that any player minimizing UTC regret also minimizes linear-swap regret. We then leverage this connection to develop state-of-the-art no-regret algorithms for computing linear correlated equilibria, both in theory and in practice. In theory, our algorithms achieve polynomially better per-iteration runtimes; in practice, our algorithms represent the state of the art by several orders of magnitude.", "pdf": "https://openreview.net/pdf/cc579eed7f8cd682cf460a081609c0c02ed9a3b9.pdf"} {"title": "3D Feature Prediction for Masked-AutoEncoder-Based Point Cloud Pretraining", "url": "https://openreview.net/forum?id=LokR2TTFMs", "detail_url": "https://openreview.net/forum?id=LokR2TTFMs", "authors": "Siming Yan,Yuqi Yang,Yu-Xiao Guo,Hao Pan,Peng-Shuai Wang,Xin Tong,Yang Liu,Qixing Huang", "tags": "ICLR 2024,Poster", "abstract": "Masked autoencoders (MAE) have recently been introduced to 3D self-supervised pretraining for point clouds due to their great success in NLP and computer vision. Unlike MAEs used in the image domain, where the pretext task is to restore features at the masked pixels, such as colors, the existing 3D MAE works reconstruct the missing geometry only, i.e, the location of the masked points. In contrast to previous studies, we advocate that point location recovery is inessential and restoring intrinsic point features is much superior. To this end, we propose to ignore point position reconstruction and recover high-order features at masked points including surface normals and surface variations, through a novel attention-based decoder which is independent of the encoder design. We validate the effectiveness of our pretext task and decoder design using different encoder structures for 3D training and demonstrate the advantages of our pretrained networks on various point cloud analysis tasks.", "pdf": "https://openreview.net/pdf/2ecf6db8536bd51a51f39e4d291975d6fdf06840.pdf"} {"title": "Understanding Catastrophic Forgetting in Language Models via Implicit Inference", "url": "https://openreview.net/forum?id=VrHiF2hsrm", "detail_url": "https://openreview.net/forum?id=VrHiF2hsrm", "authors": "Suhas Kotha,Jacob Mitchell Springer,Aditi Raghunathan", "tags": "ICLR 2024,Poster", "abstract": "We lack a systematic understanding of the effects of fine-tuning (via methods such as instruction-tuning or reinforcement learning from human feedback), particularly on tasks outside the narrow fine-tuning distribution. In a simplified scenario, we demonstrate that improving performance on tasks within the fine-tuning data distribution comes at the expense of capabilities on other tasks. We hypothesize that language models implicitly infer the task of the prompt and that fine-tuning skews this inference towards tasks in the fine-tuning distribution. To test this, we propose Conjugate Prompting, which artificially makes the task look farther from the fine-tuning distribution while requiring the same capability, and we find that this recovers some of the pretraining capabilities in our synthetic setup. Since real-world fine-tuning distributions are predominantly English, we apply conjugate prompting to recover pretrained capabilities in LLMs by simply translating the prompts to different languages. This allows us to recover in-context learning abilities lost via instruction tuning, natural reasoning capability lost during code fine-tuning, and, more concerningly, harmful content generation suppressed by safety fine-tuning in chatbots like ChatGPT.", "pdf": "https://openreview.net/pdf/fba453c1032d524bc23b1106ab3a598979c31704.pdf"} {"title": "Efficient Subgraph GNNs by Learning Effective Selection Policies", "url": "https://openreview.net/forum?id=gppLqZLQeY", "detail_url": "https://openreview.net/forum?id=gppLqZLQeY", "authors": "Beatrice Bevilacqua,Moshe Eliasof,Eli Meirom,Bruno Ribeiro,Haggai Maron", "tags": "ICLR 2024,Poster", "abstract": "Subgraph GNNs are provably expressive neural architectures that learn graph representations from sets of subgraphs. Unfortunately, their applicability is hampered by the computational complexity associated with performing message passing on many subgraphs. In this paper, we consider the problem of learning to select a small subset of the large set of possible subgraphs in a data-driven fashion. We first motivate the problem by proving that there are families of WL-indistinguishable graphs for which there exist efficient subgraph selection policies: small subsets of subgraphs that can already identify all the graphs within the family. We then propose a new approach, called _Policy-Learn_, that learns how to select subgraphs in an iterative manner. We prove that, unlike popular random policies and prior work addressing the same problem, our architecture is able to learn the efficient policies mentioned above. Our experimental results demonstrate that _Policy-Learn_ outperforms existing baselines across a wide range of datasets.", "pdf": "https://openreview.net/pdf/bb667e3793f646c96519efa8bde1df46581bd0cc.pdf"} {"title": "Parsing neural dynamics with infinite recurrent switching linear dynamical systems", "url": "https://openreview.net/forum?id=YIls9HEa52", "detail_url": "https://openreview.net/forum?id=YIls9HEa52", "authors": "Victor Geadah,International Brain Laboratory,Jonathan W. Pillow", "tags": "ICLR 2024,Poster", "abstract": "Unsupervised methods for dimensionality reduction of neural activity and behavior have provided unprecedented insights into the underpinnings of neural information processing. One popular approach involves the recurrent switching linear dynamical system (rSLDS) model, which describes the latent dynamics of neural spike train data using discrete switches between a finite number of low-dimensional linear dynamical systems. However, a few properties of rSLDS model limit its deployability on trial-varying data, such as a fixed number of states over trials, and no latent structure or organization of states. Here we overcome these limitations by endowing the rSLDS model with a semi-Markov discrete state process, with latent geometry, that captures key properties of stochastic processes over partitions with flexible state cardinality. We leverage partial differential equations (PDE) theory to derive an efficient, semi-parametric formulation for dynamical sufficient statistics to the discrete states. This process, combined with switching dynamics, defines our infinite recurrent switching linear dynamical system (irSLDS) model class. We first validate and demonstrate the capabilities of our model on synthetic data. Next, we turn to the analysis of mice electrophysiological data during decision-making, and uncover strong non-stationary processes underlying both within-trial and trial-averaged neural activity.", "pdf": "https://openreview.net/pdf/69049199784ba1584b80a5be627c8e848d397879.pdf"} {"title": "Active Retrosynthetic Planning Aware of Route Quality", "url": "https://openreview.net/forum?id=h7DGnWGeos", "detail_url": "https://openreview.net/forum?id=h7DGnWGeos", "authors": "Luotian Yuan,Yemin Yu,Ying Wei,Yongwei Wang,Zhihua Wang,Fei Wu", "tags": "ICLR 2024,Poster", "abstract": "Retrosynthetic planning is a sequential decision-making process of identifying synthetic routes from the available building block materials to reach a desired target molecule.\nThough existing planning approaches show promisingly high solving rates and low costs, the trivial route cost evaluation via pre-trained forward reaction prediction models certainly falls short of real-world chemical practice.\nAn alternative option is to annotate the actual cost of a route, such as yield, through chemical experiments or input from chemists, while \nthis often leads to substantial query costs.\nIn order to strike the balance between query costs and route quality evaluation, we propose an Active Retrosynthetic Planning (ARP) framework that remains compatible with the established retrosynthetic planners.\nOn one hand, the proposed ARP trains an actor that decides whether to query the cost of a reaction; on the other hand, it resorts to a critic to estimate the value of a molecule with its preceding reaction cost as input. \nThose molecules with low reaction costs are preferred to expand first.\nWe apply our framework to different existing approaches on both the benchmark and an expert dataset and demonstrate that it outperforms the existing state-of-the-art approach by 6.2\\% in route quality while reducing the query cost by 12.8\\%.\nIn addition, \nARP consistently plans \nhigh-quality routes with either abundant or sparse annotations.", "pdf": "https://openreview.net/pdf/0eba4d55c6b2d267c53e120ccbcbe1c0a441de5a.pdf"} {"title": "How do Language Models Bind Entities in Context?", "url": "https://openreview.net/forum?id=zb3b6oKO77", "detail_url": "https://openreview.net/forum?id=zb3b6oKO77", "authors": "Jiahai Feng,Jacob Steinhardt", "tags": "ICLR 2024,Poster", "abstract": "Language models (LMs) can recall facts mentioned in context, as shown by their performance on reading comprehension tasks. When the context describes facts about more than one entity, the LM has to correctly bind attributes to their corresponding entity. We show, via causal experiments, that LMs' internal activations represent binding information by exhibiting appropriate binding ID vectors at the entity and attribute positions. We further show that binding ID vectors form a subspace and often transfer across tasks. Our results demonstrate that LMs learn interpretable strategies for representing symbolic knowledge in context, and that studying context activations is a fruitful direction for understanding LM cognition.", "pdf": "https://openreview.net/pdf/28b9cffab0e6b66ae449e4a74e7e3c4798f5be18.pdf"} {"title": "Novel Quadratic Constraints for Extending LipSDP beyond Slope-Restricted Activations", "url": "https://openreview.net/forum?id=HfXDrAzFvG", "detail_url": "https://openreview.net/forum?id=HfXDrAzFvG", "authors": "Patricia Pauli,Aaron J Havens,Alexandre Araujo,Siddharth Garg,Farshad Khorrami,Frank Allg\u00f6wer,Bin Hu", "tags": "ICLR 2024,Poster", "abstract": "Recently, semidefinite programming (SDP) techniques have shown great promise in providing accurate Lipschitz bounds for neural networks. Specifically, the LipSDP approach (Fazlyab et al., 2019) has received much attention and provides the least conservative Lipschitz upper bounds that can be computed with polynomial time guarantees. However, one main restriction of LipSDP is that its formulation requires the activation functions to be slope-restricted on $[0,1]$, preventing its further use for more general activation functions such as GroupSort, MaxMin, and Householder. One can rewrite MaxMin activations for example as residual ReLU networks. However, a direct application of LipSDP to the resultant residual ReLU networks is conservative and even fails in recovering the well-known fact that the MaxMin activation is 1-Lipschitz. Our paper bridges this gap and extends LipSDP beyond slope-restricted activation functions. To this end, we provide novel quadratic constraints for GroupSort, MaxMin, and Householder activations via leveraging their underlying properties such as sum preservation. Our proposed analysis is general and provides a unified approach for estimating $\\ell_2$ and $\\ell_\\infty$ Lipschitz bounds for a rich class of neural network architectures, including non-residual and residual neural networks and implicit models, with GroupSort, MaxMin, and HouseHolder activations. Finally, we illustrate the utility of our approach with a variety of experiments and show that our proposed SDPs generate less conservative Lipschitz bounds in comparison to existing approaches.", "pdf": "https://openreview.net/pdf/77c8f16bf84c684bfa1e6014821209b562511cc5.pdf"} {"title": "Unbalancedness in Neural Monge Maps Improves Unpaired Domain Translation", "url": "https://openreview.net/forum?id=2UnCj3jeao", "detail_url": "https://openreview.net/forum?id=2UnCj3jeao", "authors": "Luca Eyring,Dominik Klein,Th\u00e9o Uscidda,Giovanni Palla,Niki Kilbertus,Zeynep Akata,Fabian J Theis", "tags": "ICLR 2024,Poster", "abstract": "In optimal transport (OT), a Monge map is known as a mapping that transports a source distribution to a target distribution in the most cost-efficient way. Recently, multiple neural estimators for Monge maps have been developed and applied in diverse unpaired domain translation tasks, e.g. in single-cell biology and computer vision. However, the classic OT framework enforces mass conservation, which\nmakes it prone to outliers and limits its applicability in real-world scenarios. The latter can be particularly harmful in OT domain translation tasks, where the relative position of a sample within a distribution is explicitly taken into account. While unbalanced OT tackles this challenge in the discrete setting, its integration into neural Monge map estimators has received limited attention. We propose a theoretically\ngrounded method to incorporate unbalancedness into any Monge map estimator. We improve existing estimators to model cell trajectories over time and to predict cellular responses to perturbations. Moreover, our approach seamlessly integrates with the OT flow matching (OT-FM) framework. While we show that OT-FM performs competitively in image translation, we further improve performance by\nincorporating unbalancedness (UOT-FM), which better preserves relevant features. We hence establish UOT-FM as a principled method for unpaired image translation.", "pdf": "https://openreview.net/pdf/4e007c37023074d72d5fdc49366594785523cb68.pdf"} {"title": "Implicit Neural Representations and the Algebra of Complex Wavelets", "url": "https://openreview.net/forum?id=uZfjFyPAvn", "detail_url": "https://openreview.net/forum?id=uZfjFyPAvn", "authors": "T Mitchell Roddenberry,Vishwanath Saragadam,Maarten V. de Hoop,Richard Baraniuk", "tags": "ICLR 2024,Poster", "abstract": "Implicit neural representations (INRs) have arisen as useful methods for representing signals on Euclidean domains. By parameterizing an image as a multilayer perceptron (MLP) on Euclidean space, INRs effectively couple spatial and spectral features of the represented signal in a way that is not obvious in the usual discrete representation. Although INRs using sinusoidal activation functions have been studied in terms of Fourier theory, recent works have shown the advantage of using wavelets instead of sinusoids as activation functions, due to their ability to simultaneously localize in both frequency and space. In this work, we approach such INRs and demonstrate how they resolve high-frequency features of signals from coarse approximations performed in the first layer of the MLP. This leads to multiple prescriptions for the design of INR architectures, including the use of progressive wavelets, decoupling of low and high-pass approximations, and initialization schemes based on the singularities of the target signal.", "pdf": "https://openreview.net/pdf/4204c82f9ab502c4460ea36edf09dabf851c3781.pdf"} {"title": "Fiber Monte Carlo", "url": "https://openreview.net/forum?id=sP1tCl2QBk", "detail_url": "https://openreview.net/forum?id=sP1tCl2QBk", "authors": "Nick Richardson,Deniz Oktay,Yaniv Ovadia,James C Bowden,Ryan P Adams", "tags": "ICLR 2024,Poster", "abstract": "Integrals with discontinuous integrands are ubiquitous, arising from discrete structure in applications like topology optimization, graphics, and computational geometry.\n These integrals are often part of a forward model in an inverse problem where it is necessary to reason backwards about the parameters, ideally using gradient-based optimization. \n Monte Carlo methods are widely used to estimate the value of integrals, but this results in a non-differentiable approximation that is amenable to neither conventional automatic differentiation nor reparameterization-based gradient methods. \n This significantly disrupts efforts to integrate machine learning methods in areas that exhibit these discontinuities: physical simulation and robotics, design, graphics, and computational geometry. \n Although bespoke domain-specific techniques can handle special cases, a general methodology to wield automatic differentiation in these discrete contexts is wanting. \n We introduce a differentiable variant of the simple Monte Carlo estimator which samples line segments rather than points from the domain. \n We justify our estimator analytically as conditional Monte Carlo and demonstrate the diverse functionality of the method as applied to image stylization, topology optimization, and computational geometry.", "pdf": "https://openreview.net/pdf/7d5a83d7a94183d1d9225cc666d97a8d77188d5b.pdf"} {"title": "Learning from Label Proportions: Bootstrapping Supervised Learners via Belief Propagation", "url": "https://openreview.net/forum?id=KQe9tHd0k8", "detail_url": "https://openreview.net/forum?id=KQe9tHd0k8", "authors": "Shreyas Havaldar,Navodita Sharma,Shubhi Sareen,Karthikeyan Shanmugam,Aravindan Raghuveer", "tags": "ICLR 2024,Poster", "abstract": "Learning from Label Proportions (LLP) is a learning problem where only aggregate level labels are available for groups of instances, called bags, during training, and the aim is to get the best performance at the instance-level on the test data. This setting arises in domains like advertising and medicine due to privacy considerations. We propose a novel algorithmic framework for this problem that iteratively performs two main steps. For the first step (Pseudo Labeling) in every iteration, we define a Gibbs distribution over binary instance labels that incorporates a) covariate information through the constraint that instances with similar covariates should have similar labels and b) the bag level aggregated label. We then use Belief Propagation (BP) to marginalize the Gibbs distribution to obtain pseudo labels. In the second step (Embedding Refinement), we use the pseudo labels to provide supervision for a learner that yields a better embedding. Further, we iterate on the two steps again by using the second step's embeddings as new covariates for the next iteration. In the final iteration, a classifier is trained using the pseudo labels. Our algorithm displays strong gains against several SOTA baselines (upto **15%**) for the LLP Binary Classification problem on various dataset types - tabular and Image. We achieve these improvements with minimal computational overhead above standard supervised learning due to Belief Propagation, for large bag sizes, even for a million samples.", "pdf": "https://openreview.net/pdf/fe1295b607d0d088fa07653fbda2b9c5a57e521f.pdf"} {"title": "Modeling state-dependent communication between brain regions with switching nonlinear dynamical systems", "url": "https://openreview.net/forum?id=WQwV7Y8qwa", "detail_url": "https://openreview.net/forum?id=WQwV7Y8qwa", "authors": "Orren Karniol-Tambour,David M. Zoltowski,E. Mika Diamanti,Lucas Pinto,Carlos D Brody,David W. Tank,Jonathan W. Pillow", "tags": "ICLR 2024,Poster", "abstract": "Understanding how multiple brain regions interact to produce behavior is a major challenge in systems neuroscience, with many regions causally implicated in common tasks such as sensory processing and decision making. A precise description of interactions between regions remains an open problem. Moreover, neural dynamics are nonlinear and non-stationary. Here, we propose MR-SDS, a multiregion, switching nonlinear state space model that decomposes global dynamics into local and cross-communication components in the latent space. MR-SDS includes directed interactions between brain regions, allowing for estimation of state-dependent communication signals, and accounts for sensory inputs effects. We show that our model accurately recovers latent trajectories, vector fields underlying switching nonlinear dynamics, and cross-region communication profiles in three simulations. We then apply our method to two large-scale, multi-region neural datasets involving mouse decision making. The first includes hundreds of neurons per region, recorded simultaneously at single-cell-resolution across 3 distant cortical regions. The second is a mesoscale widefield dataset of 8 adjacent cortical regions imaged across both hemispheres. On these multi-region datasets, our model outperforms existing piece-wise linear multi-region models and reveals multiple distinct dynamical states and a rich set of cross-region communication profiles.", "pdf": "https://openreview.net/pdf/e972a0d03b94f15cfd893473ba7abbb0f7ec0e16.pdf"} {"title": "Neural Language of Thought Models", "url": "https://openreview.net/forum?id=HYyRwm367m", "detail_url": "https://openreview.net/forum?id=HYyRwm367m", "authors": "Yi-Fu Wu,Minseung Lee,Sungjin Ahn", "tags": "ICLR 2024,Poster", "abstract": "The Language of Thought Hypothesis suggests that human cognition operates on a structured, language-like system of mental representations. While neural language models can naturally benefit from the compositional structure inherently and explicitly expressed in language data, learning such representations from non-linguistic general observations, like images, remains a challenge. In this work, we introduce the Neural Language of Thought Model (NLoTM), a novel approach for unsupervised learning of LoTH-inspired representation and generation. NLoTM comprises two key components: (1) the Semantic Vector-Quantized Variational Autoencoder, which learns hierarchical, composable discrete representations aligned with objects and their properties, and (2) the Autoregressive LoT Prior, an autoregressive transformer that learns to generate semantic concept tokens compositionally, capturing the underlying data distribution. We evaluate NLoTM on several 2D and 3D image datasets, demonstrating superior performance in downstream tasks, out-of-distribution generalization, and image generation quality compared to patch-based VQ-VAE and continuous object-centric representations. Our work presents a significant step towards creating neural networks exhibiting more human-like understanding by developing LoT-like representations and offers insights into the intersection of cognitive science and machine learning.", "pdf": "https://openreview.net/pdf/49107366def27f8426cf313dc870595042c85c07.pdf"} {"title": "Retrieval-Guided Reinforcement Learning for Boolean Circuit Minimization", "url": "https://openreview.net/forum?id=0t1O8ziRZp", "detail_url": "https://openreview.net/forum?id=0t1O8ziRZp", "authors": "Animesh Basak Chowdhury,Marco Romanelli,Benjamin Tan,Ramesh Karri,Siddharth Garg", "tags": "ICLR 2024,Poster", "abstract": "Logic synthesis, a pivotal stage in chip design, entails optimizing chip specifications encoded in hardware description languages like Verilog into highly efficient implementations using Boolean logic gates. The process involves a sequential application of logic minimization heuristics (``synthesis recipe\"), with their arrangement significantly impacting crucial metrics such as area and delay. Addressing the challenge posed by the broad spectrum of hardware design complexities \u2014 from variations of past designs (e.g., adders and multipliers) to entirely novel configurations (e.g., innovative processor instructions) \u2014 requires a nuanced 'synthesis recipe' guided by human expertise and intuition. This study conducts a thorough examination of learning and search techniques for logic synthesis, unearthing a surprising revelation: pre-trained agents, when confronted with entirely novel designs, may veer off course, detrimentally affecting the search trajectory. We present ABC-RL, a meticulously tuned $\\alpha$ parameter that adeptly adjusts recommendations from pre-trained agents during the search process. Computed based on similarity scores through nearest neighbor retrieval from the training dataset, ABC-RL yields superior synthesis recipes tailored for a wide array of hardware designs. Our findings showcase substantial enhancements in the Quality of Result (QoR) of synthesized circuits, boasting improvements of up to 24.8\\% compared to state-of-the-art techniques. Furthermore, ABC-RL achieves an impressive up to 9x reduction in runtime (iso-QoR) when compared to current state-of-the-art methodologies.", "pdf": "https://openreview.net/pdf/45352e5a0436249e936152d04e6550dff1475ac5.pdf"} {"title": "Enhancing Small Medical Learners with Privacy-preserving Contextual Prompting", "url": "https://openreview.net/forum?id=ztpy1gsUpT", "detail_url": "https://openreview.net/forum?id=ztpy1gsUpT", "authors": "Xinlu Zhang,Shiyang Li,Xianjun Yang,Chenxin Tian,Yao Qin,Linda Ruth Petzold", "tags": "ICLR 2024,Poster", "abstract": "Large language models (LLMs) demonstrate remarkable medical expertise, but data privacy concerns impede their direct use in healthcare environments. Although offering improved data privacy protection, domain-specific small language models (SLMs) often underperform LLMs, emphasizing the need for methods that reduce this performance gap while alleviating privacy concerns. In this paper, we present a simple yet effective method that harnesses LLMs' medical proficiency to boost SLM performance in medical tasks under $privacy-restricted$ scenarios. Specifically, we mitigate patient privacy issues by extracting keywords from medical data and prompting the LLM to generate a medical knowledge-intensive context by simulating clinicians' thought processes. This context serves as additional input for SLMs, augmenting their decision-making capabilities. Our method significantly enhances performance in both few-shot and full training settings across three medical knowledge-intensive tasks, achieving up to a 22.57% increase in absolute accuracy compared to SLM fine-tuning without context, and sets new state-of-the-art results in two medical tasks within privacy-restricted scenarios. Further out-of-domain testing and experiments in two general domain datasets showcase its generalizability and broad applicability.", "pdf": "https://openreview.net/pdf/870c2d2b9c207c1295cc917d1fe533ee69e399a9.pdf"} {"title": "What Makes a Good Prune? Maximal Unstructured Pruning for Maximal Cosine Similarity", "url": "https://openreview.net/forum?id=jsvvPVVzwf", "detail_url": "https://openreview.net/forum?id=jsvvPVVzwf", "authors": "Gabryel Mason-Williams,Fredrik Dahlqvist", "tags": "ICLR 2024,Poster", "abstract": "Pruning is an effective method to reduce the size of deep neural network models, maintain accuracy, and, in some cases, improve the network's overall performance. However, the mechanisms underpinning pruning remain unclear. Why can different methods prune by different percentages yet achieve similar performance? Why can we not prune at the start of training? Why are some models more amenable to being pruned than others? Given a model, what is the maximum amount it can be pruned before significantly affecting the performance? This paper explores and answers these questions from the global unstructured magnitude pruning perspective with one epoch of fine-tuning. We develop the idea that cosine similarity is an effective proxy measure for functional similarity between the parent and the pruned network. We prove that the L1 pruning method is optimal when pruning by cosine similarity. We show that the higher the kurtosis of a model's parameter distribution, the more it can be pruned while maintaining performance. Finally, we present a simple method to determine the optimal amount by which a network can be L1-pruned based on its parameter distribution. The code demonstrating the method is available at https://github.com/gmw99/what_makes_a_good_prune", "pdf": "https://openreview.net/pdf/d36cb7ab7a56f325f5bb6a9e1bbccce752db0563.pdf"} {"title": "Enhancing Human Experience in Human-Agent Collaboration: A Human-Centered Modeling Approach Based on Positive Human Gain", "url": "https://openreview.net/forum?id=BqEvdOS1Hs", "detail_url": "https://openreview.net/forum?id=BqEvdOS1Hs", "authors": "Yiming Gao,Feiyu Liu,Liang Wang,Dehua Zheng,Zhenjie Lian,Weixuan Wang,Wenjin Yang,Siqin Li,Xianliang Wang,Wenhui Chen,Jing Dai,QIANG FU,Yang Wei,Lanxiao Huang,Wei Liu", "tags": "ICLR 2024,Poster", "abstract": "Existing game AI research mainly focuses on enhancing agents' abilities to win games, but this does not inherently make humans have a better experience when collaborating with these agents. For example, agents may dominate the collaboration and exhibit unintended or detrimental behaviors, leading to poor experiences for their human partners. In other words, most game AI agents are modeled in a \"self-centered\" manner. In this paper, we propose a \"human-centered\" modeling scheme for collaborative agents that aims to enhance the experience of humans. Specifically, we model the experience of humans as the goals they expect to achieve during the task. We expect that agents should learn to enhance the extent to which humans achieve these goals while maintaining agents' original abilities (e.g., winning games). To achieve this, we propose the Reinforcement Learning from Human Gain (RLHG) approach. The RLHG approach introduces a \"baseline\", which corresponds to the extent to which humans primitively achieve their goals, and encourages agents to learn behaviors that can effectively enhance humans in achieving their goals better. We evaluate the RLHG agent in the popular Multi-player Online Battle Arena (MOBA) game, Honor of Kings, by conducting real-world human-agent tests. Both objective performance and subjective preference results show that the RLHG agent provides participants better gaming experience.", "pdf": "https://openreview.net/pdf/6ce678ca6108e4a85517e2526c83d7d968f7108b.pdf"} {"title": "Vocos: Closing the gap between time-domain and Fourier-based neural vocoders for high-quality audio synthesis", "url": "https://openreview.net/forum?id=vY9nzQmQBw", "detail_url": "https://openreview.net/forum?id=vY9nzQmQBw", "authors": "Hubert Siuzdak", "tags": "ICLR 2024,Poster", "abstract": "Recent advancements in neural vocoding are predominantly driven by Generative Adversarial Networks (GANs) operating in the time-domain. While effective, this approach neglects the inductive bias offered by time-frequency representations, resulting in reduntant and computionally-intensive upsampling operations. Fourier-based time-frequency representation is an appealing alternative, aligning more accurately with human auditory perception, and benefitting from well-established fast algorithms for its computation. Nevertheless, direct reconstruction of complex-valued spectrograms has been historically problematic, primarily due to phase recovery issues. This study seeks to close this gap by presenting Vocos, a new model that directly generates Fourier spectral coefficients. Vocos not only matches the state-of-the-art in audio quality, as demonstrated in our evaluations, but it also substantially improves computational efficiency, achieving an order of magnitude increase in speed compared to prevailing time-domain neural vocoding approaches. The source code and model weights have been open-sourced.", "pdf": "https://openreview.net/pdf/d6b6fd2b9464f306e29d42b554de0a493bb52ade.pdf"} {"title": "An Agnostic View on the Cost of Overfitting in (Kernel) Ridge Regression", "url": "https://openreview.net/forum?id=YrTI2Zu0dd", "detail_url": "https://openreview.net/forum?id=YrTI2Zu0dd", "authors": "Lijia Zhou,James B Simon,Gal Vardi,Nathan Srebro", "tags": "ICLR 2024,Poster", "abstract": "We study the cost of overfitting in noisy kernel ridge regression (KRR), which we define as the ratio between the test error of the interpolating ridgeless model and the test error of the optimally-tuned model. We take an ``agnostic'' view in the following sense: we consider the cost as a function of sample size for any target function, even if the sample size is not large enough for consistency or the target is outside the RKHS. We analyze the cost of overfitting under a Gaussian universality ansatz using recently derived (non-rigorous) risk estimates in terms of the task eigenstructure. Our analysis provides a more refined characterization of benign, tempered and catastrophic overfitting (cf. Mallinar et al. 2022).", "pdf": "https://openreview.net/pdf/d8be4e90683a3f7e1634819a4732c562ef29b8a9.pdf"} {"title": "Explaining Kernel Clustering via Decision Trees", "url": "https://openreview.net/forum?id=FAGtjl7HOw", "detail_url": "https://openreview.net/forum?id=FAGtjl7HOw", "authors": "Maximilian Fleissner,Leena Chennuru Vankadara,Debarghya Ghoshdastidar", "tags": "ICLR 2024,Poster", "abstract": "Despite the growing popularity of explainable and interpretable machine learning, there is still surprisingly limited work on inherently interpretable clustering methods. Recently, there has been a surge of interest in explaining the classic k-means algorithm, leading to efficient algorithms that approximate k-means clusters using axis-aligned decision trees. However, interpretable variants of k-means have limited applicability in practice, where more flexible clustering methods are often needed to obtain useful partitions of the data. In this work, we investigate interpretable kernel clustering, and propose algorithms that construct decision trees to approximate the partitions induced by kernel k-means, a nonlinear extension of k-means. We further build on previous work on explainable k-means and demonstrate how a suitable choice of features allows preserving interpretability without sacrificing approximation guarantees on the interpretable model.", "pdf": "https://openreview.net/pdf/92e83ec19298dc4bbddbca30bcb2e45b1a0e9a39.pdf"} {"title": "Provable Benefits of Multi-task RL under Non-Markovian Decision Making Processes", "url": "https://openreview.net/forum?id=U6Qulbv2qT", "detail_url": "https://openreview.net/forum?id=U6Qulbv2qT", "authors": "Ruiquan Huang,Yuan Cheng,Jing Yang,Vincent Tan,Yingbin Liang", "tags": "ICLR 2024,Poster", "abstract": "In multi-task reinforcement learning (RL) under Markov decision processes (MDPs), the presence of shared latent structures among multiple MDPs has been shown to yield significant benefits to the sample efficiency compared to single-task RL. In this paper, we investigate whether such a benefit can extend to more general sequential decision making problems such as predictive state representations (PSRs). The main challenge here is that the large and complex model space makes it hard to identify what types of common latent structure of multi-task PSRs can reduce the model complexity and improve sample efficiency.\nTo this end, we posit a joint model class for tasks and use the notion of $\\eta$-bracketing number to quantify its complexity; this number also serves as a general metric to capture the similarity of tasks and thus determines the benefit of multi-task over single-task RL. We first study upstream multi-task learning over PSRs, in which all tasks share the same observation and action spaces. We propose a provably efficient algorithm UMT-PSR for finding near-optimal policies for all PSRs, and demonstrate that the advantage of multi-task learning manifests if the joint model class of PSRs has a smaller $\\eta$-bracketing number compared to that of individual single-task learning. We further investigate downstream learning, in which the agent needs to learn a new target task that shares some commonalities with the upstream tasks via a similarity constraint. By exploiting the learned PSRs from the upstream, we develop a sample-efficient algorithm that provably finds a near-optimal policy. \nUpon specialization to some examples with small $\\eta$-bracketing numbers, our results further highlight the benefit compared to directly learning a single-task PSR.", "pdf": "https://openreview.net/pdf/e04e45de352848ad2466d4c254699426119e2e33.pdf"} {"title": "Learning Delays in Spiking Neural Networks using Dilated Convolutions with Learnable Spacings", "url": "https://openreview.net/forum?id=4r2ybzJnmN", "detail_url": "https://openreview.net/forum?id=4r2ybzJnmN", "authors": "Ilyass Hammouamri,Ismail Khalfaoui-Hassani,Timoth\u00e9e Masquelier", "tags": "ICLR 2024,Poster", "abstract": "Spiking Neural Networks (SNNs) are a promising research direction for building power-efficient information processing systems, especially for temporal tasks such as speech recognition. In SNNs, delays refer to the time needed for one spike to travel from one neuron to another. These delays matter because they influence the spike arrival times, and it is well-known that spiking neurons respond more strongly to coincident input spikes. More formally, it has been shown theoretically that plastic delays greatly increase the expressivity in SNNs. Yet, efficient algorithms to learn these delays have been lacking. Here, we propose a new discrete-time algorithm that addresses this issue in deep feedforward SNNs using backpropagation, in an offline manner. To simulate delays between consecutive layers, we use 1D convolutions across time. The kernels contain only a few non-zero weights \u2013 one per synapse \u2013 whose positions correspond to the delays. These positions are learned together with the weights using the recently proposed Dilated Convolution with Learnable Spacings (DCLS). We evaluated our method on three datasets: the Spiking Heidelberg Dataset (SHD), the Spiking Speech Commands (SSC) and its non spiking version Google Speech Commands v0.02 (GSC) benchmarks, which require detecting temporal patterns. We used feedforward SNNs with two or three hidden fully connected layers, and vanilla leaky integrate-and-fire neurons. We showed that fixed random delays help and that learning them helps even more. Furthermore, our method outperformed the state-of-the-art in the three datasets without using recurrent connections and with substantially fewer parameters. Our work demonstrates the potential of delay learning in developing accurate and precise models for temporal data processing. Our code is based on PyTorch / SpikingJelly and available at: https://github.com/Thvnvtos/SNN-delays", "pdf": "https://openreview.net/pdf/f1fd6afaca66ead7463f3e4cd5a92340f8e9e81e.pdf"} {"title": "The Reversal Curse: LLMs trained on \u201cA is B\u201d fail to learn \u201cB is A\u201d", "url": "https://openreview.net/forum?id=GPKTIktA0k", "detail_url": "https://openreview.net/forum?id=GPKTIktA0k", "authors": "Lukas Berglund,Meg Tong,Maximilian Kaufmann,Mikita Balesni,Asa Cooper Stickland,Tomasz Korbak,Owain Evans", "tags": "ICLR 2024,Poster", "abstract": "We expose a surprising failure of generalization in auto-regressive large language models (LLMs). If a model is trained on a sentence of the form ''_A_ is _B_'', it will not automatically generalize to the reverse direction ''_B_ is _A_''. This is the **Reversal Curse**. For instance, if a model is trained on ''Valentina Tereshkova was the first woman to travel to space'', it will not automatically be able to answer the question, ''Who was the first woman to travel to space?''. Moreover, the likelihood of the correct answer (''Valentina Tershkova'') will not be higher than for a random name. Thus, models do not generalize a prevalent pattern in their training set: if ''_A_ is _B_'' occurs, ''_B_ is _A_'' is more likely to occur. It is worth noting, however, that if ''_A_ is _B_'' appears _in-context_, models can deduce the reverse relationship. \n\nWe provide evidence for the Reversal Curse by finetuning GPT-3 and Llama-1 on fictitious statements such as ''Uriah Hawthorne is the composer of _Abyssal Melodies_'' and showing that they fail to correctly answer ''Who composed _Abyssal Melodies?_''. The Reversal Curse is robust across model sizes and model families and is not alleviated by data augmentation.\n\nWe also evaluate ChatGPT (GPT-3.5 and GPT-4) on questions about real-world celebrities, such as ''Who is Tom Cruise's mother? [A: Mary Lee Pfeiffer]'' and the reverse ''Who is Mary Lee Pfeiffer's son?''. GPT-4 correctly answers questions like the former 79\\% of the time, compared to 33\\% for the latter.\n\n Code available at: https://github.com/lukasberglund/reversal_curse.", "pdf": "https://openreview.net/pdf/8f0b327c2fe27ca52094fdbede8101325e1e7800.pdf"} {"title": "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models", "url": "https://openreview.net/forum?id=7Jwpw4qKkb", "detail_url": "https://openreview.net/forum?id=7Jwpw4qKkb", "authors": "Xiaogeng Liu,Nan Xu,Muhao Chen,Chaowei Xiao", "tags": "ICLR 2024,Poster", "abstract": "The aligned Large Language Models (LLMs) are powerful language understanding and decision-making tools that are created through extensive alignment with human feedback. However, these large models remain susceptible to jailbreak attacks, where adversaries manipulate prompts to elicit malicious outputs that should not be given by aligned LLMs. Investigating jailbreak prompts can lead us to delve into the limitations of LLMs and further guide us to secure them. Unfortunately, existing jailbreak techniques suffer from either (1) scalability issues, where attacks heavily rely on manual crafting of prompts, or (2) stealthiness problems, as attacks depend on token-based algorithms to generate prompts that are often semantically meaningless, making them susceptible to detection through basic perplexity testing. In light of these challenges, we intend to answer this question: Can we develop an approach that can automatically generate stealthy jailbreak prompts? In this paper, we introduce AutoDAN, a novel jailbreak attack against aligned LLMs. AutoDAN can automatically generate stealthy jailbreak prompts by the carefully designed hierarchical genetic algorithm. Extensive evaluations demonstrate that AutoDAN not only automates the process while preserving semantic meaningfulness, but also demonstrates superior attack strength in cross-model transferability, and cross-sample universality compared with the baseline. Moreover, we also compare AutoDAN with perplexity-based defense methods and show that AutoDAN can bypass them effectively. Code is available at https://github.com/SheltonLiu-N/AutoDAN.", "pdf": "https://openreview.net/pdf/7f7805e88bbe936e341887640330e3f74da3b650.pdf"} {"title": "MixSATGEN: Learning Graph Mixing for SAT Instance Generation", "url": "https://openreview.net/forum?id=PXXuLvIH5r", "detail_url": "https://openreview.net/forum?id=PXXuLvIH5r", "authors": "Xinyan Chen,Yang Li,Runzhong Wang,Junchi Yan", "tags": "ICLR 2024,Poster", "abstract": "The Boolean satisfiability problem (SAT) stands as a canonical NP-complete task. In particular, the scarcity of real-world SAT instances and their usefulness for tuning SAT solvers underscore the necessity for effective and efficient ways of hard instance generation, whereas existing methods either struggle to maintain plausible hardness or suffer from limited applicability. Different from the typical construction-based methods, this paper introduces an adaptive and efficient graph interpolation approach that in place modifies the raw structure of graph-represented SAT instance by replacing it with a counterpart from another instance. Specifically, it involves a two-stage matching and mixing pipeline. The matching aims to find a correspondence map of literal nodes from two instance graphs via learned features from a matching network; while the mixing stage involves iteratively exchanging clause pairs with the highest correspondence scores until a specified replacement ratio is achieved. We further show that under our matching-mixing framework, moderate randomness can avoid hardness degradation of instances by introducing Gumbel noise. Experimental results show the superiority of our method with both resemblance in structure and hardness, and general applicability.", "pdf": "https://openreview.net/pdf/deb8544197d35dd8d0d5d3613853b75f5351bb71.pdf"} {"title": "A Stable, Fast, and Fully Automatic Learning Algorithm for Predictive Coding Networks", "url": "https://openreview.net/forum?id=RyUvzda8GH", "detail_url": "https://openreview.net/forum?id=RyUvzda8GH", "authors": "Tommaso Salvatori,Yuhang Song,Yordan Yordanov,Beren Millidge,Lei Sha,Cornelius Emde,Zhenghua Xu,Rafal Bogacz,Thomas Lukasiewicz", "tags": "ICLR 2024,Poster", "abstract": "Predictive coding networks are neuroscience-inspired models with roots in both Bayesian statistics and neuroscience. Training such models, however, is quite inefficient and unstable. In this work, we show how by simply changing the temporal scheduling of the update rule for the synaptic weights leads to an algorithm that is much more efficient and stable than the original one, and has theoretical guarantees in terms of convergence. The proposed algorithm, that we call incremental predictive coding (iPC) is also more biologically plausible than the original one, as it it fully automatic. In an extensive set of experiments, we show that iPC constantly performs better than the original formulation on a large number of benchmarks for image classification, as well as for the training of both conditional and masked language models, in terms of test accuracy, efficiency, and convergence with respect to a large set of hyperparameters.", "pdf": "https://openreview.net/pdf/44d053d08f335ed3b0c3e59491bc957e0c2cc2eb.pdf"} {"title": "Scalable Monotonic Neural Networks", "url": "https://openreview.net/forum?id=DjIsNDEOYX", "detail_url": "https://openreview.net/forum?id=DjIsNDEOYX", "authors": "Hyunho Kim,Jong-Seok Lee", "tags": "ICLR 2024,Poster", "abstract": "In this research, we focus on the problem of learning monotonic neural networks, as preserving the monotonicity of a model with respect to a subset of inputs is crucial for practical applications across various domains. Although several methods have recently been proposed to address this problem, they have limitations such as not guaranteeing monotonicity in certain cases, requiring additional inference time, lacking scalability with increasing network size and number of monotonic inputs, and manipulating network weights during training. To overcome these limitations, we introduce a simple but novel architecture of the partially connected network which incorporates a 'scalable monotonic hidden layer' comprising three units: the exponentiated unit, ReLU unit, and confluence unit. This allows for the repetitive integration of the scalable monotonic hidden layers without other structural constraints. Consequently, our method offers ease of implementation and rapid training through the conventional error-backpropagation algorithm. We accordingly term this method as Scalable Monotonic Neural Networks (SMNN). Numerical experiments demonstrated that our method achieved comparable prediction accuracy to the state-of-the-art approaches while effectively addressing the aforementioned weaknesses.", "pdf": "https://openreview.net/pdf/4659f41f940610cc73b4142d71235fc0bbdff339.pdf"} {"title": "Accurate Retraining-free Pruning for Pretrained Encoder-based Language Models", "url": "https://openreview.net/forum?id=s2NjWfaYdZ", "detail_url": "https://openreview.net/forum?id=s2NjWfaYdZ", "authors": "Seungcheol Park,Hojun Choi,U Kang", "tags": "ICLR 2024,Poster", "abstract": "Given a pretrained encoder-based language model, how can we accurately compress it without retraining? Retraining-free structured pruning algorithms are crucial in pretrained language model compression due to their significantly reduced pruning cost and capability to prune large language models. However, existing retraining-free algorithms encounter severe accuracy degradation, as they fail to handle pruning errors, especially at high compression rates. In this paper, we propose KPrune (Knowledge-preserving pruning), an accurate retraining-free structured pruning algorithm for pretrained encoder-based language models.\nKPrune focuses on preserving the useful knowledge of the pretrained model to minimize pruning errors through a carefully designed iterative pruning process composed of knowledge measurement, knowledge-preserving mask search, and knowledge-preserving weight-tuning. As a result, KPrune shows significant accuracy improvements up to 58.02%p higher F1 score compared to existing retraining-free pruning algorithms under a high compression rate of 80% on the SQuAD benchmark without any retraining process.", "pdf": "https://openreview.net/pdf/7be529579f3a50dbe8a628d518b7986357e95d8c.pdf"} {"title": "PROGRAM: PROtotype GRAph Model based Pseudo-Label Learning for Test-Time Adaptation", "url": "https://openreview.net/forum?id=x5LvBK43wg", "detail_url": "https://openreview.net/forum?id=x5LvBK43wg", "authors": "Haopeng Sun,Lumin Xu,Sheng Jin,Ping Luo,Chen Qian,Wentao Liu", "tags": "ICLR 2024,Poster", "abstract": "Test-time adaptation (TTA) aims to adapt a pre-trained model from a source domain to a target domain only using online unlabeled target data during testing, without accessing to the source data or modifying the original training process. Among the various TTA methods, pseudo-labeling has gained popularity. However, the presence of incorrect pseudo-labels can hinder the effectiveness of target domain adaptation. To overcome this challenge, we propose a novel TTA method, called PROtotype GRAph Model based pseudo-label learning (PROGRAM). PROGRAM consists of two key components: (1) Prototype Graph Model (PGM) for reliable pseudo-label generation; (2) Robust Self-Training (RST) for test-time adaptation with noisy pseudo-labels. PGM constructs the graph using prototypes and test samples, facilitating effective message passing among them to generate more reliable pseudo-labels. RST combines the advantages of consistency regularization and pseudo-labeling to achieve robust target domain adaptation in the presence of noisy pseudo-labels. Our proposed PROGRAM can be easily integrated into existing baselines, resulting in consistent improvement. Extensive experiments show that our PROGRAM outperforms the existing TTA methods on multiple domain generalization and image corruption benchmarks.", "pdf": "https://openreview.net/pdf/abf81b71b6719508d36fb748baf090a1daf94937.pdf"} {"title": "Navigating Dataset Documentations in AI: A Large-Scale Analysis of Dataset Cards on HuggingFace", "url": "https://openreview.net/forum?id=xC8xh2RSs2", "detail_url": "https://openreview.net/forum?id=xC8xh2RSs2", "authors": "Xinyu Yang,Weixin Liang,James Zou", "tags": "ICLR 2024,Poster", "abstract": "Advances in machine learning are closely tied to the creation of datasets. While data documentation is widely recognized as essential to the reliability, reproducibility, and transparency of ML, we lack a systematic empirical understanding of current dataset documentation practices. To shed light on this question, here we take Hugging Face - one of the largest platforms for sharing and collaborating on ML models and datasets - as a prominent case study. By analyzing all 7,433 dataset documentation on Hugging Face, our investigation provides an overview of the Hugging Face dataset ecosystem and insights into dataset documentation practices, yielding 5 main findings: (1) The dataset card completion rate shows marked heterogeneity correlated with dataset popularity: While 86.0\\% of the top 100 downloaded dataset cards fill out all sections suggested by Hugging Face community, only 7.9\\% of dataset cards with no downloads complete all these sections. (2) A granular examination of each section within the dataset card reveals that the practitioners seem to prioritize Dataset Description and Dataset Structure sections, accounting for 36.2\\% and 33.6\\% of the total card length, respectively, for the most downloaded datasets. In contrast, the Considerations for Using the Data section receives the lowest proportion of content, accounting for just 2.1\\% of the text. (3) By analyzing the subsections within each section and utilizing topic modeling to identify key topics, we uncover what is discussed in each section, and underscore significant themes encompassing both technical and social impacts, as well as limitations within the Considerations for Using the Data section. (4) Our findings also highlight the need for improved accessibility and reproducibility of datasets in the Usage sections. (5) In addition, our human annotation evaluation emphasizes the pivotal role of comprehensive dataset content in shaping individuals' perceptions of a dataset card's overall quality. Overall, our study offers a unique perspective on analyzing dataset documentation through large-scale data science analysis and underlines the need for more thorough dataset documentation in machine learning research.", "pdf": "https://openreview.net/pdf/57002f23415c9a3640dd438f61e8487b5eb59bc2.pdf"} {"title": "A Differentially Private Clustering Algorithm for Well-Clustered Graphs", "url": "https://openreview.net/forum?id=hkSjjs4o5d", "detail_url": "https://openreview.net/forum?id=hkSjjs4o5d", "authors": "Weiqiang He,Hendrik Fichtenberger,Pan Peng", "tags": "ICLR 2024,Poster", "abstract": "We study differentially private (DP) algorithms for recovering clusters in well-clustered graphs, which are graphs whose vertex set can be partitioned into a small number of sets, each inducing a subgraph of high inner conductance and small outer conductance. Such graphs have widespread application as a benchmark in the theoretical analysis of spectral clustering.\nWe provide an efficient ($\\epsilon$,$\\delta$)-DP algorithm tailored specifically for such graphs. Our algorithm draws inspiration from the recent work of Chen et al., who developed DP algorithms for recovery of stochastic block models in cases where the graph comprises exactly two nearly-balanced clusters. Our algorithm works for well-clustered graphs with $k$ nearly-balanced clusters, and the misclassification ratio almost matches the one of the best-known non-private algorithms. We conduct experimental evaluations on datasets with known ground truth clusters to substantiate the prowess of our algorithm. We also show that any (pure) $\\epsilon$-DP algorithm would result in substantial error.", "pdf": "https://openreview.net/pdf/932c357f99c7bafe3b57c4d6ba322cd4ad35c5c6.pdf"} {"title": "Minimax optimality of convolutional neural networks for infinite dimensional input-output problems and separation from kernel methods", "url": "https://openreview.net/forum?id=EW8ZExRZkJ", "detail_url": "https://openreview.net/forum?id=EW8ZExRZkJ", "authors": "Yuto Nishimura,Taiji Suzuki", "tags": "ICLR 2024,Poster", "abstract": "Recent deep learning applications, exemplified by text-to-image tasks, often involve high-dimensional inputs and outputs. While several studies have investigated the function estimation capabilities of deep learning, research on dilated convolutional neural networks (CNNs) has mainly focused on cases where input dimensions are infinite but output dimensions are one-dimensional, similar to many other studies. However, many practical deep learning tasks involve high-dimensional (or even infinite dimensional) inputs and outputs.\nIn this paper, we investigate the optimality of dilated CNNs for estimating a map between infinite-dimensional input and output spaces \nby analyzing their approximation and estimation abilities. \nFor that purpose, we first show that approximation and estimation errors depend only on the smoothness and decay rate with respect to the infinity norm of the output, and their estimation accuracy actually achieve the {\\it minimax optimal} rate of convergence.\nSecond, we demonstrate that the dilated CNNs outperform {\\it any} linear estimators including kernel ridge regression and $k$-NN estimators in a minimax error sense, highlighting the usefulness of feature learning realized by deep neural networks.\nOur theoretical analysis particularly explains the success of deep learning in recent high-dimensional input-output tasks.", "pdf": "https://openreview.net/pdf/4b7946387cd61ff2ae4a8ba6793aed54e90f4b5d.pdf"} {"title": "Consistent algorithms for multi-label classification with macro-at-$k$ metrics", "url": "https://openreview.net/forum?id=XOnya9gSdF", "detail_url": "https://openreview.net/forum?id=XOnya9gSdF", "authors": "Erik Schultheis,Wojciech Kotlowski,Marek Wydmuch,Rohit Babbar,Strom Borman,Krzysztof Dembczynski", "tags": "ICLR 2024,Poster", "abstract": "We consider the optimization of complex performance metrics in multi-label classification under the population utility framework. We mainly focus on metrics linearly decomposable into a sum of binary classification utilities applied separately to each label with an additional requirement of exactly $k$ labels predicted for each instance. These \"macro-at-$k$\" metrics possess desired properties for extreme classification problems with long tail labels. Unfortunately, the at-$k$ constraint couples the otherwise independent binary classification tasks, leading to a much more challenging optimization problem than standard macro-averages. We provide a statistical framework to study this problem, prove the existence and the form of the optimal classifier, and propose a statistically consistent and practical learning algorithm based on the Frank-Wolfe method. Interestingly, our main results concern even more general metrics being non-linear functions of label-wise confusion matrices. Empirical results provide evidence for the competitive performance of the proposed approach.", "pdf": "https://openreview.net/pdf/a672c09faabd424ec3161e19cf789efb721061c5.pdf"} {"title": "Dynamic Layer Tying for Parameter-Efficient Transformers", "url": "https://openreview.net/forum?id=d4uL2MSe0z", "detail_url": "https://openreview.net/forum?id=d4uL2MSe0z", "authors": "Tamir David Hay,Lior Wolf", "tags": "ICLR 2024,Poster", "abstract": "In the pursuit of reducing the number of trainable parameters in deep transformer networks, we employ Reinforcement Learning to dynamically select layers during training and tie them together. Every few iterations, the RL agent is asked whether to train each layer $i$ independently or to copy the weights of a previous layer $j