Timezone: »
Generative modeling has recently gained massive attention given high-profile successes in natural language processing and computer vision. However, there remain major challenges in deploying generative models for real-world impact in domains like healthcare and biology. This is a challenging agenda that requires collaboration across multiple research fields and industry stakeholders. This workshop aims to advance such interdisciplinary conversations around challenges in deploying generative models – the lessons learned by deploying large language models could be impactful for high stakes domains like medicine and biology. Specifically, we will solicit contributions that prioritize (1) Multimodal capabilities in generative modeling, (2) Deployment-critical features in generative models such as Safety, Interpretability, Robustness, Ethics, Fairness and Privacy, and (3) Human facing evaluation of generative models. The topic of generative modeling is extremely relevant to the core audience of ICML. Modern generative models impact several fields outside machine learning and hence responsible deployment of such powerful algorithms has become a major concern of researchers in academia and industry alike. ICML, being the flagship conference of Machine learning, is the perfect place to facilitate this cross disciplinary sharing of knowledge.
Fri 12:00 p.m. - 12:05 p.m.
|
Opening Remarks
(
Introductory Remarks
)
SlidesLive Video » |
🔗 |
Fri 12:05 p.m. - 12:30 p.m.
|
Talk: Tim Salimans, Google Research
(
Virtual Talk
)
SlidesLive Video » |
🔗 |
Fri 12:30 p.m. - 12:55 p.m.
|
Talk: Olga Russakovsky, Princeton University
(
Virtual Talk
)
SlidesLive Video » |
🔗 |
Fri 12:55 p.m. - 1:20 p.m.
|
Talk: Alan Aspuru-Guzik, University of Toronto
(
Virtual Talk
)
SlidesLive Video » |
🔗 |
Fri 1:20 p.m. - 1:45 p.m.
|
Talk: Pamela Mishkin, OpenAI
(
Virtual Talk
)
SlidesLive Video » |
🔗 |
Fri 1:45 p.m. - 2:00 p.m.
|
Coffee Break
|
🔗 |
Fri 2:00 p.m. - 3:00 p.m.
|
Poster Session I
(
Poster presentations
)
|
🔗 |
Fri 3:00 p.m. - 4:00 p.m.
|
Lunch
|
🔗 |
Fri 4:00 p.m. - 4:30 p.m.
|
Talk: Deep Ganguli, Anthropic
(
In-person talk
)
SlidesLive Video » |
🔗 |
Fri 4:30 p.m. - 5:00 p.m.
|
Talk: Finale Doshi-Velez, Harvard University
(
In-person talk
)
SlidesLive Video » |
🔗 |
Fri 5:00 p.m. - 5:30 p.m.
|
Talk: Daphne Ippolito, Carnegie Mellon University
(
In-person talk
)
SlidesLive Video » |
🔗 |
Fri 5:30 p.m. - 6:00 p.m.
|
Talk: Kyunghyun Cho, NYU / Genentech
(
In-person talk
)
SlidesLive Video » |
🔗 |
Fri 6:10 p.m. - 7:00 p.m.
|
Poster Session II
(
Poster presentations
)
|
🔗 |
Fri 7:00 p.m. - 7:55 p.m.
|
Panel Discussion: “Challenges and lessons learned in deploying Generative AI”
(
Discussion Panel
)
SlidesLive Video » Panelists:
|
🔗 |
Fri 7:55 p.m. - 8:00 p.m.
|
Closing Remarks
(
Concluding Remarks
)
SlidesLive Video » |
🔗 |
-
|
A comparison of diffusion models and CycleGANs for virtual staining of slide-free microscopy images
(
Poster
)
link »
Slide-free microscopy (SFM) methods can serve as a faster alternative to the standard histological examination of tissue specimens. However, SFM methods often provide images that differ from the hematoxylin- and eosin-stained (H\&E) images commonly obtained in standard histology. Unpaired image-to-image translation has been explored for transforming SFM images into H\&E images, a process known as virtual staining. Here, we compare a standard CycleGAN approach to a diffusion model-based approach for virtual staining of SFM images. We observe that the diffusion model approach, which relies on the inherent semantic preservation of the latent encodings, fails to outperform the standard CycleGAN approach, when tested on two different SFM datasets. This indicates that the semantic preservation of diffusion models is lacking for virtual staining tasks and additional regularization is needed. |
Tanishq Abraham · Richard Levenson 🔗 |
-
|
Temporal Attention Bottleneck is informative? Interpretability through Disentangled Generative Representations for Energy Time Series Disaggregation
(
Poster
)
link »
Generative models have garnered significant attention for their ability to address the challenge of source separation in disaggregation tasks. This approach holds promise for promoting energy conservation by enabling homeowners to obtain detailed information on their energy consumption solely through the analysis of aggregated load curves. Nevertheless, the model's ability to generalize and its interpretability remain two major challenges. To tackle these challenges, we deploy a generative model called TAB-VAE (Temporal Attention Bottleneck for Variational Auto-encoder), based on hierarchical architecture, addresses signature variability, and provides a robust, interpretable separation through the design of its informative representation of latent space. Our implementation and evaluation guidelines are available at https://anonymous.4open.science/r/TAB-VAE-3934/. |
khalid OUBLAL · Said Ladjal · David Benhaiem · Emmanuel le-borgne · François Roueff 🔗 |
-
|
Conditional Diffusion Replay for Continual Learning in Medical Settings
(
Poster
)
link »
Episodic replay methods, which store and replay past data, have proven effective for handling distribution shifts in continual learning. However, due to regulatory and privacy concerns for data sharing, their applicability can be limited, especially in healthcare.In this work, we advance the state of art,focusing our inquiry on two novel benchmarks for domain incremental continual learning: diabetic retinopathy severity classification and dermoscopy skin lesion detection.First, we demonstrate the poor forward and backward transferability of simple baselines. Then, to overcome these challenges, we propose a novel method called conditional diffusion replay. By leveraging a text-to-image diffusion model for synthetic data generation, our approach effectively preservesperformance on previously encountered domainswhile adapting to new ones.We observe that compared to standard sequential fine-tuning, our conditional diffusion replay method improves average AUC by up to 6.2 points and 3.3 points for the skin lesions and diabetic retinopathy benchmarks, respectively. |
Yewon Byun · Saurabh Garg · Sanket Vaibhav Mehta · Praveer Singh · Jayashree Kalpathy-cramer · Bryan Wilder · Zachary Lipton 🔗 |
-
|
Deep Generative Clustering with Multimodal Variational Autoencoders
(
Poster
)
link »
Multimodal VAEs have recently received significant attention as generative models for weakly-supervised learning with multiple heterogeneous modalities. In parallel, VAE-based methods have been explored as probabilistic approaches for clustering tasks. Our work lies at the intersection of these two research directions. We propose a novel multimodal VAE model, in which the latent space is extended to learn data clusters, leveraging shared information across modalities. Our experiments show that our proposed model improves generative performance over existing multimodal VAEs, particularly for unconditional generation. Furthermore, our method favourably compares to alternative clustering approaches, in weakly-supervised settings. Notably, we propose a post-hoc procedure that avoids the need for our method to have a priori knowledge of the true number of clusters, mitigating a critical limitation of previous clustering frameworks. |
Emanuele Palumbo · Sonia Laguna · Daphné Chopard · Julia Vogt 🔗 |
-
|
Do Users Write More Insecure Code with AI Assistants?
(
Poster
)
link »
We conduct the first large-scale user study examining how users interact with an AI Code assistant to solve avariety of security related tasks across different programminglanguages. Overall, we find that participants who had accessto an AI assistant based on OpenAI’s \texttt{codex-davinci-002}model wrote less secure code than those withoutaccess. Additionally, participants with access to an AI assistantwere more likely to believe they wrote secure code than thosewithout access to the AI assistant. Furthermore, we find thatparticipants who trusted the AI less and engaged more withthe language and format of their prompts (e.g. re-phrasing,adjusting temperature) provided code with fewer securityvulnerabilities. Finally, in order to better inform the designof future AI Assistants, we provide an in-depthanalysis of participants’ language and interaction behavior, aswell as release our user interface as an instrument to conductsimilar studies in the future. |
Neil Perry · Megha Srivastava · Deepak Kumar · Dan Boneh 🔗 |
-
|
Risk-Aware Image Generation by Estimating and Propagating Uncertainty
(
Poster
)
link »
While generative AI models have revolutionized content creation across various modalities, they have yet to be deployed in safety-critical scenarios. This is in part due to limited understanding of their underlying uncertainty, as general-purpose frameworks for estimating uncertainty in large-scale generative models are lacking. Here we analyze the effects of uncertainty and risk estimation methods on generative AI systems and their applications to two critical domains of deployment -- identification of failures, and fast optimization of input prompts. As a case study, we apply our approach to create an uncertainty-aware variant of the Stable Diffusion text-to-image model, allowing us to estimate and propagate uncertainty over inputs, latent representations, and outputs. We demonstrate that our method enables the identification of uncertain output regions and the optimization of input prompts to minimize output uncertainty. We envision that our framework will enable the deployment of more robust and auditable generative AI systems. |
Alejandro Perez · Iaroslav Elistratov · Fynn Schmitt-Ulms · Ege Demir · Sadhana Lolla · Elaheh Ahmadi · Daniela Rus · Alexander Amini 🔗 |
-
|
Answering Causal Questions with Augmented LLMs
(
Poster
)
link »
Large Language Models (LLMs) are revolutionising the way we interact with machines and enable never before seen applications. A common use-case of LLMs is as a chat interface for more complicated underlying systems to enable natural interaction without the need for learning system specifics. However, LLMs in their current form alone are not sufficient to do causal reasoning. In this paper, we explore different ways to augment the LLM with existing large scale end-to-end causal models to enable causal question answering abilities. Specifically, we compare the effectiveness of answering causal questions using two different approaches that both rely on the output of a causal expert model: 1) providing the predicted causal graph and related treatment effects in the context; 2) access to an API to derive insights from the output of the causal model. Our experiments show that context-augmented LLMs make significantly more mistakes than the data-access API-augmented LLMs, which are invariant to the size of the causal problem. |
Nick Pawlowski · Joel Jennings · Cheng Zhang 🔗 |
-
|
Circuit Breaking: Removing Model Behaviors with Targeted Ablation
(
Poster
)
link »
Language models often exhibit behaviors that improve performance on a pre-training objective but harm performance on downstream tasks. We propose a novel approach to removing undesirable behaviors by ablating a small number of causal pathways between model components, with the intention of disabling the computational circuit responsible for the bad behavior. Given a small dataset of inputs where the model behaves poorly, we learn to ablate a small number of important causal pathways. In the setting of reducing GPT-2 toxic language generation, we find ablating just 12 of the 11.6K causal edges mitigates toxic generation with minimal degradation of performance on other inputs. |
Maximilian Li · Xander Davies · Max Nadeau 🔗 |
-
|
LLM-Based Aspect Augmentations for Recommendation Systems
(
Poster
)
link »
Large language models (LLMs) have shown to be effective in different task settings, including recommendation-related tasks. In this study, we aim at measuring the effectiveness of using item aspects (justifications for users' intentions when buying the item) generated by LLMs in improving the results for ranking tasks. For this purpose, we carefully design prompts for LLMs to derive aspects for items using their textual data in an eCommerce setting. The extracted aspects are used as augmentations for Learning-to-Rank models. Specifically, we input the generated aspects as summarized embeddings using three approaches: (i) augmenting using feature concatenation, (ii) adding a wide aspect component beside a deep component of features, and (iii) adding an aspect embedding tower to create a two-tower model. We conduct extensive experiments on real-world eCommerce dataset and show the effectiveness of including LLM-based aspects in improving ranking metrics such as MRR and NDCG, even when they are compared to models augmented by pre-trained language models (PLM). |
Reza Yousefi Maragheh · LALITESH MORISHETTI · Ramin Giahi · Kaushiki Nag · Jianpeng Xu · Jason Cho · Evren Korpeoglu · Sushant Kumar · Kannan Achan 🔗 |
-
|
Can Public Large Language Models Help Private Cross-device Federated Learning?
(
Poster
)
link »
We study (differentially) private federated learning (FL) of language models. The language models in cross-device FL are relatively small, which can be trained with meaningful formal user-level differential privacy (DP) guarantees when massive parallelism in training is enabled by the participation of a moderate size of users. Recently, public data has been used to improve privacy-utility trade-offs for both large and small language models. In this work, we provide a systematic study of using large-scale public data and LLMs to help differentially private training of on-device FL models, and further improve the privacy-utility tradeoff by techniques of distillation. Moreover, we propose a novel distribution matching algorithm with theoretical grounding to sample public data close to private data distribution, which significantly improves the sample efficiency of (pre)training on public data. The proposed method is efficient and effective for training private models by taking advantage of public data, especially for customized on-device architectures that do not have ready-to-use pre-trained models. |
Boxin Wang · Yibo J. Zhang · Yuan Cao · Bo Li · Hugh B McMahan · Sewoong Oh · Zheng Xu · Manzil Zaheer 🔗 |
-
|
Soft prompting might be a bug, not a feature
(
Poster
)
link »
Prompt tuning, or "soft prompting," replaces text prompts to generative models with learned embeddings (i.e. vectors) and is used as an alternative to parameter-efficient fine-tuning. Prior work suggests analyzing soft prompts by interpreting them as natural language prompts. However, we find that soft prompts occupy regions in the embedding space that are distinct from those containing natural language, meaning that direct comparisons may be misleading. We argue that because soft prompts are currently uninterpretable, they could potentially be a source of vulnerability of LLMs to malicious manipulations during deployment. |
Luke Bailey · Gustaf Ahdritz · Anat Kleiman · Siddharth Swaroop · Finale Doshi-Velez · Weiwei Pan 🔗 |
-
|
(Un)interpretability of Transformers: a case study with Dyck grammars
(
Poster
)
link »
Understanding the algorithm implemented by a model is important for trustworthiness when deploying large-scale models, which has been a topic of great interest for interpretability. In this work, we take a critical view of methods that exclusively focus on individual parts of the model, rather than consider the network as a whole. We consider a simple synthetic setup of learning a Dyck language. Theoretically, we show that the set of models that can solve this task satisfies a structural characterization derived from ideas in formal languages (the pumping lemma). We use this characterization to show that the set of optima is qualitatively rich: in particular, the attention pattern of a single layer can be ``nearly randomized'', while preserving the functionality of the network. We also show via extensive experiments that these constructions are not merely a theoretical artifact: even with severe constraints to the architecture of the model, vastly different solutions can be reached via standard training. Thus, interpretability claims based on individual heads or weight matrices in the Transformer can be misleading. |
Kaiyue Wen · Yuchen Li · Bingbin Liu · Andrej Risteski 🔗 |
-
|
Discovering Variable Binding Circuitry with Desiderata
(
Poster
)
link »
Recent work has shown that computation in language models may be human-understandable, with successful efforts to localize and intervene on both single-unit features and input-output circuits. Here, we introduce an approach which extends causal mediation experiments to automatically identify model components responsible for performing a specific subtask by solely specifying a set of $\textit{desiderata}$, or causal attributes of the model components executing that subtask. As a proof of concept, we apply our method to automatically discover shared variable binding circuitry in LLaMA-13B, which retrieves variable values for multiple arithmetic tasks. Our method successfully localizes variable binding to only 9 attention heads (of the 1.6k) and one MLP in the final token's residual stream.
|
Xander Davies · Max Nadeau · Nikhil Prakash · Tamar Shaham · David Bau 🔗 |
-
|
Decision Stacks: Flexible Reinforcement Learning via Modular Generative Models
(
Poster
)
link »
Deployment of reinforcement learning algorithms in real-world scenarios often presents numerous challenges such as dealing with complex goals, planning future observations and actions, and critiquing their utilities, demanding a balance between expressivity and flexible modeling for efficient learning and inference.We present Decision Stacks, a generative framework that decomposes goal-conditioned policy agents into 3 generative modules which simulate the temporal evolution of observations, rewards, and actions. Our framework guarantees both expressivity and flexibility in designing individual modules to account for key factors such as architectural bias, optimization objective and dynamics, transferrability across domains, and inference speed. Our empirical results demonstrate the effectiveness of Decision Stacks for offline policy optimization for several MDP and POMDP environments. |
Siyan Zhao · Aditya Grover 🔗 |
-
|
Generative Autoencoders as Watermark Attackers: Analyses of Vulnerabilities and Threats
(
Poster
)
link »
Invisible watermarks safeguard images' copyrights by embedding hidden messages detectable by owners. It also prevents people from misusing images, especially those generated by AI models. Malicious adversaries can violate these rights by removing the watermarks. In order to remove watermarks without damaging the visual quality, the adversary needs to erase them while retaining the essential information in the image. This is analogous to the encoding and decoding process of generative autoencoders, especially variational autoencoders (VAEs) and diffusion models. We propose a framework using generative autoencoders to remove invisible watermarks and test it using VAEs and diffusions. Our results reveal that, even without specific training, off-the-shelf Stable Diffusion effectively removes most watermarks, surpassing all current attackers. The result underscores the vulnerabilities in existing watermarking schemes and calls for more robust methods for copyright protection. |
Xuandong Zhao · Kexun Zhang · Yu-Xiang Wang · Lei Li 🔗 |
-
|
The Journey, Not the Destination: How Data Guides Diffusion Models
(
Poster
)
link »
Diffusion-based generative models can synthesize photo-realistic images of unprecedented quality and diversity. However, attributing these images back to the training data---that is, identifying specific training examples which caused the images to be generated---remains challenging. In this paper, we propose a framework that: i) formalizes data attribution in the context of diffusion models, and ii) provides a method for computing attributions efficiently. By applying our framework to CIFAR-10 and MS COCO, we uncover visually compelling attributions, which we validate through counterfactual analysis. |
Kristian Georgiev · Joshua Vendrow · Hadi Salman · Sung Min (Sam) Park · Aleksander Madry 🔗 |
-
|
One-Step Diffusion Distillation via Deep Equilibrium Models
(
Poster
)
link »
Diffusion models excel at producing high-quality samples but naively require hundreds of iterations, prompting multiple attempts to distill this process into a faster network. Existing approaches, however, often require complex multi-stage distillation and perform sub-optimally in single-step image generation. In response, we introduce a simple yet effective means of diffusion distillation---*directly* mapping initial noise to the resulting image. Of particular importance to our approach is to leverage a new Deep Equilibrium (DEQ) model for distillation: the Generative Equilibrium Transformer (GET). Our method enables fully offline training with just noise/image pairs from the diffusion model while achieving superior performance compared to existing one-step methods on comparable training budgets. The DEQ architecture proves crucial, as GET matches a $5\times$ larger ViT in terms of FID scores while striking a critical balance of computational cost and image quality. Code, checkpoints, and datasets will be released.
|
Zhengyang Geng · Ashwini Pokle · Zico Kolter 🔗 |
-
|
Seeing Through the Facade: Understanding the Realism, Expressivity, and Limitations of Diffusion Models
(
Poster
)
link »
While text-to-image generation models such as DALLE-2 and Stable Diffusion 2.0 have captured the public psyche with the ability to create photorealistic images, just how "fake" are their outputs? To better understand this question, we present a three-prong process for extracting insights from diffusion models. First, we show strong results in classifying real vs. fake images by using transfer learning with a nearly decade-old model, setting an initial benchmark of realism not yet achieved. After visualizing the classifier's inference decisions, we conclude that concrete, singular subject objects -- like buildings and hands -- helped distinguish real from fake images. However, we found no consensus on which features were distinct to each of DALLE-2 and Stable Diffusion. Finally, after dissecting the prompts used to generate fake images, we found that prompts that failed to trick our classifier contained similar types of nouns while prompts that succeeded in this task differed for each model. We believe our work can serve as the first step in an iterative process that continuously establishes increasingly difficult benchmarks of realism for diffusion models to overcome. The code for our project is open source: https://github.com/cpondoc/diffusion-model-analysis. |
Christopher Pondoc · Joseph O'Brien · Joseph Guman 🔗 |
-
|
Local Differential Privacy with Entropic Wasserstein Distance
(
Poster
)
link »
Local differential privacy (LDP) is a powerful method for privacy-preserving data collection. In this paper, we develop a framework for training Generative Adversarial Networks (GAN) on differentially privatized data. We show that entropic regularization of the Wasserstein distance - a popular regularization method in the literature that has been often leveraged for its computational benefits - can be used to denoise the data distribution when data is privatized by popular additive noise mechanisms, such as Laplace and Gaussian. This combination uniquely enables the mitigation of both the regularization bias and the effects of privatization noise, thereby enhancing the overall efficacy of the model. We analyze the proposed method, provide sample complexity results and experimental evidence to support its efficacy. |
Daria Reshetova · Wei-Ning Chen · Ayfer Ozgur 🔗 |
-
|
Provable Robust Watermarking for AI-Generated Text
(
Poster
)
link »
As AI-generated text increasingly resembles human-written content, the ability to detect machine-generated text becomes crucial. To address this challenge, we present GPTWatermark, a robust and high-quality solution designed to ascertain whether a piece of text originates from a specific model. Our approach extends existing watermarking strategies and employs a fixed group design to enhance robustness against editing and paraphrasing attacks. We show that our watermarked language model enjoys strong provable guarantees on generation quality, correctness in detection, and security against evasion attacks. Experimental results on various large language models (LLMs) and diverse datasets demonstrate that our method achieves superior detection accuracy and comparable generation quality in perplexity, thus promoting the responsible use of LLMs. |
Xuandong Zhao · Prabhanjan Ananth · Lei Li · Yu-Xiang Wang 🔗 |
-
|
Exposing Attention Glitches with Flip-Flop Language Modeling
(
Poster
)
link »
Why do large language models hallucinate? This work identifies and analyzes the phenomenon of \emph{attention glitches},in which the Transformer architecture's inductive biases intermittently fail to capture robust reasoning.To isolate the issue, we introduce \emph{flip-flop language modeling} (FFLM), a parametric family of synthetic benchmarks designed to probe the extrapolation of language models. This simple generative task requires a model to copy binary symbols over long-range dependencies, ignoring the tokens in between. We find that Transformer FFLMs suffer from a long tail of sporadic reasoning errors, some of which we can eliminate using various regularization techniques. Our preliminary mechanistic analyses show why the remaining errors may be very difficult to diagnose and resolve. We hypothesize that attention glitches account for (some of) the closed-domain hallucinations in natural LLMs. |
Bingbin Liu · Jordan Ash · Surbhi Goel · Akshay Krishnamurthy · Cyril Zhang 🔗 |
-
|
Calibrating Language Models via Augmented Prompt Ensembles
(
Poster
)
link »
Large Language Models (LLMs) have achieved remarkable success, but often exhibit overconfidence and poor calibration, particularly after instruction-finetuning, which limits their reliability and applicability. To address this, we investigate ensembles, a technique known to enhance neural network calibration but underexplored in LLMs, possibly due to the computational cost of training and evaluating multiple LLMs. We introduce Calibration via Augmented Prompt Ensembles (CAPE), a practical approach to LLM ensembles that leverages the inherent prompt sensitivity of LLMs by augmenting prompts, e.g., by template paraphrasing or option permutation. Our method requires no additional training and can be efficiently evaluated in batch mode, yielding significant calibration improvements for instruction-tuned LLMs. |
Mingjian Jiang · Yangjun Ruan · Sicong Huang · Saifei Liao · Silviu Pitis · Roger Grosse · Jimmy Ba 🔗 |
-
|
Adapting Blackbox Generative Models via Inversion
(
Poster
)
link »
Adapting large-scale generative AI tools to differ-ent end uses continues to be challenging, as manyindustry grade image generator models are notpublicly available. Thus, to finetune an industrygrade image generator is not currently feasiblein the classical sense of finetuning certain layersof a given deep-network. Instead, we present analternative perspective for the problem of adapt-ing large-scale generative models that does notrequire access to the full model. Recognizingthe expense of storing and fine-tuning generativemodels, as well as the restricted access to weightsand gradients (often limited to API calls only), weintroduce AdvIN (Adapting via Inversion). Thisapproach advocates the use of inversion methods,followed by training a latent generative model asbeing equivalent to adaptation. We evaluate thefeasibility of such a framework on StyleGANswith real distribution shifts, and outline someopen research questions. Even with simple in-version and latent generation strategies, AdvINis surprisingly competitive to fine-tuning basedmethods, making it a promising alternative forend-to-end fine-tuning |
Sinjini Mitra · Rakshith Subramanyam · Rushil Anirudh · Jayaraman J. Thiagarajan · Ankita Shukla · Pavan Turaga 🔗 |
-
|
Leaving Reality to Imagination: Robust Classification via Generated Datasets
(
Poster
)
link »
Recent research on robustness has revealed significant performance gaps between neural image classifiers trained on datasets that are similar to the test set, and those that are from a naturally shifted distribution, such as sketches, paintings, and animations of the object categories observed during training. Prior work focuses on reducing this gap by designing engineered augmentations of training data or through unsupervised pretraining of a single large model on massive in-the-wild training datasets scraped from the Internet. However, the notion of a dataset is also undergoing a paradigm shift in recent years. With drastic improvements in the quality, ease-of-use, and access to modern generative models, generated data is pervading the web. In this light, we study the question: How do these generated datasets influence the natural robustness of image classifiers? We find that Imagenet classifiers trained on real data augmented with generated data achieve higher accuracy and effective robustness than standard training and popular augmentation strategies in the presence of natural distribution shifts. We analyze various factors influencing these results, including the choice of conditioning strategies and the amount of generated data. Lastly, we find that the standard ImageNet classifiers suffer a performance degradation of upto 20% on the generated data, indicating their fragility at accurately classifying the objects under novel variations. |
Hritik Bansal · Aditya Grover 🔗 |
-
|
Beyond Scale: the Diversity Coefficient as a Data Quality Metric Demonstrates LLMs are Pre-trained on Formally Diverse Data
(
Poster
)
link »
Current trends to pre-train capable Large Language Models (LLMs) mostly focus on scaling of model and dataset size.However, the $\textit{quality}$ of pre-training data is an important factor for training powerful LLMs, yet it is a nebulous concept that has not been fully characterized.Therefore, we use the recently proposed Task2Vec diversity coefficient to understand formal aspects of data quality that go beyond scale alone.Specifically, we measure the diversity coefficient of publicly available pre-training datasets to demonstrate that their formal diversity is high when compared to theoretical lower and upper bounds.In addition, to build confidence in the diversity coefficient, we conduct interpretability experiments and find that the coefficient aligns with intuitive properties of diversity,e.g., it increases as the number of latent concepts increases. We conclude the diversity coefficient is reliable and conjecture it can be used to build useful diverse datasets for LLMs.
|
Alycia Lee · Brando Miranda · Sanmi Koyejo 🔗 |
-
|
Differentially Private Generation of High Fidelity Samples From Diffusion Models
(
Poster
)
link »
Diffusion based generative models achieve unprecedented image quality but are known to leak private information about the training data. Our goal is to provide provable guarantees on privacy leakage of training data while simultaneously enabling generation of high-fidelity samples. Our proposed approach first non-privately trains an ensemble of diffusion models and then aggregates their prediction to provide privacy guarantees for generated samples. We demonstrate the success of our approach on the MNIST and CIFAR-10. |
Vikash Sehwag · Ashwinee Panda · Ashwini Pokle · Xinyu Tang · Saeed Mahloujifar · Mung Chiang · Zico Kolter · Prateek Mittal 🔗 |
-
|
Towards Safe Self-Distillation of Internet-Scale Text-to-Image Diffusion Models
(
Poster
)
link »
Large-scale image generation models, with impressive quality made possible by the vast amount of data available on the Internet, raise social concerns that these models may generate harmful or copyrighted content. The biases and harmfulness arise throughout the entire training process and are hard to completely remove, which have become significant hurdles to the safe deployment of these models. In this paper, we propose a method called SDD to prevent problematic content generation in text-to-image diffusion models. We self-distill the diffusion model to guide the noise estimate conditioned on the target removal concept to match the unconditional one. Compared to the previous methods, our method eliminates a much greater proportion of harmful content from the generated images without degrading the overall image quality. Furthermore, our method allows the removal of multiple concepts at once, whereas previous works are limited to removing a single concept at a time. |
Sanghyun Kim · Seohyeon Jung · Balhae Kim · Moonseok Choi · Jinwoo Shin · Juho Lee 🔗 |
-
|
DisasterResponseGPT: Large Language Models for Accelerated Plan of Action Development in Disaster Response Scenarios
(
Poster
)
link »
The development of plans of action in disaster response scenarios is a time-consuming process. Large Language Models (LLMs) offer a powerful solution to expedite this process through in-context learning. This study presents DisasterResponseGPT, an algorithm that leverages LLMs to generate valid plans of action quickly by incorporating disaster response and planning guidelines in the initial prompt. In DisasterResponseGPT, users input the scenario description and receive a plan of action as output. The proposed method generates multiple plans within seconds, which can be further refined following the user's feedback. Preliminary results indicate that the plans of action developed by DisasterResponseGPT are comparable to human-generated ones while offering greater ease of modification in real-time. This approach has the potential to revolutionize disaster response operations by enabling rapid updates and adjustments during the plan's execution. |
Vinicius G. Goecks · Nicholas Waytowich 🔗 |
-
|
Identifying Implicit Social Biases in Vision-Language Models
(
Poster
)
link »
Vision-language models like CLIP are widely used for multimodal retrieval tasks. However, they can learn historical biases from their training data, resulting in the perpetuation of stereotypes and potential harm. In this study, we analyze the social biases present in CLIP, particularly in the interaction between image and text. We introduce a taxonomy of social biases called So-B-IT, consisting of 374 words categorized into ten types of bias. These biases can have negative societal effects when associated with specific demographic groups. Using this taxonomy, we investigate the images retrieved by CLIP from a facial image dataset using each word as a prompt. We observe that CLIP often exhibits undesirable associations between harmful words and particular demographic groups. Furthermore, we explore the source of these biases by demonstrating their presence in a large image-text dataset used to train CLIP models. Our findings emphasize the significance of evaluating and mitigating bias in vision-language models, underscoring the necessity for transparent and fair curation of extensive pre-training datasets. |
Kimia Hamidieh · Haoran Zhang · Thomas Hartvigsen · Marzyeh Ghassemi 🔗 |
-
|
Analyzing Chain-of-Thought Prompting in Large Language Models via Gradient-based Feature Attributions
(
Poster
)
link »
Chain-of-thought (CoT) prompting has been shown to empirically improve the accuracy of large language models (LLMs) on various question answering tasks. While understanding $\textit{why}$ CoT prompting is effective is crucial to ensuring that this phenomenon is a consequence of desired model behavior, little work has addressed this; nonetheless, such an understanding is a critical prerequisite for responsible model deployment. We address this question by leveraging gradient-based feature attribution methods which produce saliency scores that capture the influence of input tokens on model output. Specifically, we probe several open-source LLMs to investigate whether CoT prompting affects the relative importances they assign to particular input tokens. Our results indicate that while CoT prompting does not increase the magnitude of saliency scores attributed to semantically relevant tokens in the prompt compared to standard few-shot prompting, it increases the robustness of saliency scores to question perturbations and variations in model output.
|
Skyler Wu · Eric Shen · Charumathi Badrinath · Jiaqi Ma · Himabindu Lakkaraju 🔗 |
-
|
A Demand-Driven Perspective on Generative Audio AI
(
Poster
)
link »
To achieve successful deployment of AI research, it is crucial to understand the demands of the industry. In this paper, we present the results of a survey conducted with professional audio engineers. The survey was conducted to determine research priorities and define various research tasks. Additionally, we summarize the current challenges in audio quality and controllability, based on the survey results. Our analysis reveals that the availability of datasets is currently the main bottleneck for achieving high-quality audio generation. Lastly, drawing on our experience, we suggest potential solutions and provide supporting empirical evidence. |
Sangshin Oh · Minsung Kang · Hyeongi Moon · Keunwoo Choi · Ben Sangbae Chon 🔗 |
-
|
Predicting Task Forgetting in Large Language Models
(
Poster
)
link »
In this paper, we offer a comprehensive evaluation of forgetting in large language models (LLMs) during sequential learning of finetuning tasks in a pretrained model. We empirically track the degradation of performance across diverse tasks and find that the validation perplexity can be predicted using a linear function, regardless of the specific task, model architecture, or task order. This knowledge sheds light on the dynamics of knowledge acquisition and retention, offering practical implications for managing and mitigating task forgetting in LLM-based systems. |
Anat Kleiman · Jonathan Frankle · Sham Kakade · Mansheej Paul 🔗 |
-
|
Understanding Data Replication in Diffusion Models
(
Poster
)
link »
Images generated by diffusion models like Stable Diffusion are increasingly widespread. Recent works and even lawsuits have shown that these models are prone to replicating their training data, unbeknownst to the user. In this paper, we first analyze this memorization problem in text-to-image diffusion models. Contrary to the prevailing belief attributing content replication solely to duplicated images in the training set, our findings highlight the equally significant role of text conditioning in this phenomenon. Specifically, we observe that the combination of image and caption duplication contributes to the memorization of training data, while the sole duplication of images either fails to contribute or even diminishes the occurrence of memorization in the examined cases. |
Gowthami Somepalli · Vasu Singla · Micah Goldblum · Jonas Geiping · Tom Goldstein 🔗 |
-
|
AutoBiasTest: Controllable Test Sentence Generation for Open-Ended Social Bias Testing in Language Models at Scale
(
Poster
)
link »
Social bias in Pretrained Language Models (PLMs) affects text generation and other downstream NLP tasks. Existing bias testing methods rely predominantly on manual templates or on expensive crowd-sourced data. We propose a novel AutoBiasTest method that automatically generates controlled sentences for testing bias in PLMs, hence providing a flexible and low-cost alternative. Our approach uses another PLM for generation controlled by conditioning on social group and attribute terms. We show that generated sentences are natural and similar to human-produced content in terms of word length and diversity. We find that our bias scores are well correlated with manual templates, but AutoBiasTest highlights biases not captured by these templates due to more diverse and realistic contexts. By automating large-scale test sentence generation, we enable better estimation of underlying bias distributions. |
Rafal Kocielnik · Shrimai Prabhumoye · Vivian Zhang · R. Alvarez · Anima Anandkumar 🔗 |
-
|
Word-Level Explanations for Analyzing Bias in Text-to-Image Models
(
Poster
)
link »
Text-to-image models take a sentence (i.e. prompt) and generate images associated with this input prompt. These models have created award wining-art, videos, and even synthetic datasets. However, text-to-image (T2I) models can generate images that underrepresent minorities based on race and sex. This paper investigates which word in the input prompt is responsible for bias in generated images. We introduce a method for computing scores for each word in the prompt; these scores represent its influence on biases in the model’s output. Our method follows the principle of explaining by removing, leveraging masked language models to calculate the influence scores. We perform experiments on Stable Diffusion to demonstrate that our method identifies the replication of societal stereotypes in generated images. |
Alexander Lin · Lucas Monteiro Paes · Sree Harsha Tanneru · Suraj Srinivas · Himabindu Lakkaraju 🔗 |
-
|
Squeezing Large-Scale Diffusion Models for Mobile
(
Poster
)
link »
The emergence of diffusion models has greatly broadened the scope of high-fidelity image synthesis, resulting in notable advancements in both practical implementation and academic research.With the active adoption of the model in various real-world applications, the need for on-device deployment has grown considerably. However, deploying large diffusion models such as Stable Diffusion with more than one billion parameters to mobile devices poses distinctive challenges due to the limited computational and memory resources, which may vary according to the device. In this paper, we present the challenges and solutions for deploying Stable Diffusion on mobile devices with TensorFlow Lite framework, which supports both iOS and Android devices. The resulting Mobile Stable Diffusion achieves the inference latency of smaller than 7 seconds for a 512 $\times$ 512 image generation on Android devices with mobile GPUs.
|
Jiwoong Choi · Minkyu Kim · Daehyun Ahn · Taesu Kim · Yulhwa Kim · Dongwon Jo · Hyesung Jeon · jae-joon kim · Hyungjun Kim 🔗 |
-
|
MargCTGAN: A ``Marginally'' Better CTGAN for the Low Sample Regime
(
Poster
)
link »
The potential of realistic and useful synthetic data is significant. However, current evaluation methods for synthetic tabular data generation predominantly focus on downstream task usefulness, often neglecting the importance of statistical properties. This oversight becomes particularly prominent in low sample scenarios, accompanied by a swift deterioration of these statistical measures. In this paper, we address this issue by conducting an evaluation of three state-of-the-art synthetic tabular data generators based on their marginal distribution, column-pair correlation, joint distribution and downstream task utility performance across high to low sample regimes. The popular CTGAN models shows strong utility, but underperforms in low sample settings in terms of utility. To overcome this limitation, we propose MargCTGAN that adds feature matching of de-correlated marginals, which results in a consistent improvement in downstream utility as well as statistical properties of the synthetic data. |
Tejumade Afonja · Dingfan Chen · Mario Fritz 🔗 |
-
|
On the Challenges of Deploying Privacy-Preserving Synthetic Data in the Enterprise
(
Poster
)
link »
Generative AI technologies are gaining unprecedented popularity, causing a mix of excitement and apprehension through their remarkable capabilities.In this paper, we study the challenges associated with deploying synthetic data, a subfield of Generative AI.Our focus centers on enterprise deployment, with an emphasis on privacy concerns caused by the vast amount of personal and highly sensitive data.We identify 40+ challenges and systematize them into five main groups -- i) generation, ii) infrastructure \& architecture, iii) governance, iv) compliance \& regulation, and v) adoption.Additionally, we discuss a strategic and systematic approach that enterprises can employ to effectively address the challenges and achieve their goals by establishing trust in the implemented solutions. |
Lauren Arthur · Jason Costello · Jonathan Hardy · Will O’Brien · James Rea · Gareth Rees · Georgi Ganev 🔗 |
-
|
Neuro-Symbolic Models of Human Moral Judgment: LLMs as Automatic Feature Extractors
(
Poster
)
link »
As AI systems gain prominence in society, concerns about their safety become crucial to address. There have been repeated calls to align powerful AI systems with human morality. However, attempts to do this have used black-box systems that cannot be interpreted or explained. In response, we introduce a methodology leveraging the natural language processing abilities of large language models (LLMs) and the interpretability of symbolic models to form competitive neuro-symbolic models for predicting human moral judgment. Our method involves using LLMs to extract morally-relevant features from a stimulus and then passing those features through a cognitive model that predicts human moral judgment. This approach achieves state-of-the-art performance on the MoralExceptQA benchmark, improving on the previous F1 score by 20 points and accuracy by 18 points, while also enhancing model interpretability by baring all key features in the model's computation. |
joseph kwon · Sydney Levine · Josh Tenenbaum 🔗 |
-
|
Tree Variational Autoencoders
(
Poster
)
link »
We propose a new generative hierarchical clustering model that learns a flexible tree-based posterior distribution over latent variables. The proposed Tree Variational Autoencoder (TreeVAE) hierarchically divides samples according to their intrinsic characteristics, shedding light on hidden structures in the data. It adapts its architecture to discover the optimal tree for encoding dependencies between latent variables, improving generative performance. We show that TreeVAE uncovers underlying clusters in the data and finds meaningful hierarchical relations between the different groups on several datasets. Due to its generative nature, TreeVAE can generate new samples from the discovered clusters via conditional sampling. |
Laura Manduchi · Moritz Vandenhirtz · Alain Ryser · Julia Vogt 🔗 |
-
|
Mitigating Inappropriateness in Image Generation: Can there be Value in Reflecting the Worlds Ugliness?
(
Poster
)
link »
Text-conditioned image generation models have recently achieved astonishing results in image quality and text alignment and are consequently employed in a fast-growing number of applications. Since they are highly data-driven, relying on billion-sized datasets randomly scraped from the web, they also reproduce inappropriate human behavior. Specifically, we demonstrate inappropriate degeneration on a large-scale for various generative text-to-image models, thus motivating the need for monitoring and moderating them at deployment. To this end, we evaluate mitigation strategies at inference to suppress the generation of inappropriate content. Our findings show that we can use models’ representations of the world’s ugliness to align them with human preferences. |
Manuel Brack · Felix Friedrich · Patrick Schramowski · Kristian Kersting 🔗 |
-
|
RustGen: An Augmentation Approach for Generating Compilable Rust Code with Large Language Models
(
Poster
)
link »
Foundation models show an impressive ability to write code snippets. However, there are still challenges when generating code for resource-poor programming languages. In this work, using Rust as an example, we tackle these challenges through in-context learning, with additional components that feed back compile errors to the LLM until it converges on a runnable code that is free of several common programming errors. We describe the specific techniques that allow us to do this -- history-based search, prompt engineering, and syntax-based skeletonization -- and evaluate their benefits on a set of code generation tasks in Rust. |
Xingbo Wu · Nathanaël Cheriere · Cheng Zhang · Dushyanth Narayanan 🔗 |
-
|
E3-VITS: Emotional End-to-End TTS with Cross-speaker Style Transfer
(
Poster
)
link »
Since previous emotional TTS models are based on a two-stage pipeline or additional labels, their training process is complex and requires a high labeling cost. To deal with this problem, this paper presents E3-VITS, an end-to-end emotional TTS model that addresses the limitations of existing models. E3-VITS synthesizes high-quality speeches for multi-speaker conditions, supports both reference speech and textual description-based emotional speech synthesis, and enables cross-speaker emotion transfer with a disjoint dataset. To implement E3-VITS, we propose batch-permuted style perturbation, which generates audio samples with unpaired emotion to increase the quality of cross-speaker emotion transfer. Results show that E3-VITS outperforms the baseline model in terms of naturalness, speaker and emotion similarity, and inference speed. |
Wonbin Jung · Junhyeok Lee 🔗 |
-
|
Are Emergent Abilities of Large Language Models a Mirage?
(
Poster
)
link »
Recent work claims that large language models display \textit{emergent abilities}, abilities not present in smaller-scale models that are present in larger-scale models. What makes emergent abilities intriguing is two-fold: their \textit{sharpness}, transitioning seemingly instantaneously from not present to present, and their \textit{unpredictability}, appearing at seemingly unforeseeable model scales.We present an alternative explanation for emergent abilities: that for a particular task and model family, when analyzing fixed model outputs, emergent abilities appear due to the researcher’s choice of metric. Specifically, nonlinear or discontinuous metrics produce apparent emergent abilities, whereas linear or continuous metrics produce smooth, continuous, predictable changes in model performance.We present our alternative explanation in a simple mathematical model, then test it in three ways: we (1) make, test and confirm predictions on the effect of metric choice using the InstructGPT/GPT-3 family; (2) make, test and confirm predictions about metric choices in a meta-analysis on BIG-Bench; and (3) show how to choose metrics to produce never-before-seen seemingly emergent abilities on vision tasks.These analyses provide evidence that alleged emergent abilities disappear with different metrics or better statistics.Our work challenging a popular conception speaks to challenges with accurately evaluating generative AI models. |
Rylan Schaeffer · Brando Miranda · Sanmi Koyejo 🔗 |
-
|
Seeing Seeds Beyond Weeds: Green Teaming Generative AI for Beneficial Uses
(
Poster
)
link »
Large generative AI models (GMs) like GPT and DALL-E are trained to generate content for general, wide-ranging purposes. GM content filters are generalized to filter out content which has a risk of harm in many cases, e.g., hate speech. However, prohibited content is not always harmful -- there are instances where generating prohibited content can be beneficial. So, when GMs filter out content, they preclude beneficial use cases along with harmful ones. Which use cases are precluded reflects the values embedded in GM content filtering. Recent work on red teaming proposes methods to bypass GM content filters to generate harmful content. We coin the term green teaming to describe methods of bypassing GM content filters to design for beneficial use cases. We showcase green teaming by: 1) Using ChatGPT as a virtual patient to simulate a person experiencing suicidal ideation, for suicide support training; 2) Using Codex to intentionally generate buggy solutions to train students on debugging; and 3) Examining an Instagram page using Midjourney to generate images of anti-LGBTQ+ politicians in drag. Finally, we discuss how our use cases demonstrate green teaming as both a practical design method and a mode of critique, which problematizes and subverts current understandings of harms and values in generative AI. |
Logan Stapleton · Jordan Taylor · Sarah Fox · Sherry Tongshuang Wu · Haiyi Zhu 🔗 |
-
|
Continual Learning for Forgetting in Deep Generative Models
(
Poster
)
link »
The recent proliferation of large-scale text-to-image models has led to growing concerns that such models may be misused to generate harmful, misleading, and inappropriate content. Motivated by this issue, we derive a technique inspired by continual learning to selectively forget concepts in pretrained text-to-image generative models. Our method enables controllable forgetting, where a user can specify how a concept should be forgotten. We apply our method to the open-source Stable Diffusion model and focus on tackling the problem of deepfakes, where experiments show that the model effectively forgets the depictions of various celebrities. |
Alvin Heng · Harold Soh 🔗 |
-
|
Compositional Abilities Emerge Multiplicatively: Exploring Diffusion Models on a Synthetic Task
(
Poster
)
link »
Modern generative models exhibit unprecedented capabilities to generate extremely realistic data. However, given the inherent compositionality of real world, reliable use of these models in practical applications mandates they exhibit the ability to compose their capabilities, generating and reasoning over entirely novel samples never seen in the training distribution. Prior work demonstrates recent vision diffusion models exhibit intriguing compositional generalization abilities, but also fail rather unpredictably. What are the reasons underlying this behavior? Which concepts does the model generally find difficult to compose to form novel data? To address these questions, we perform a controlled study of compositional generalization in conditional diffusion models in a synthetic setting, varying different attributes of the training data and measuring the model's ability to generate samples out-of-distribution. Our results show that: (i) the compositional structure of the data-generating process governs the order in which capabilities and an ability to compose them emerges; (ii) learning individual concepts impacts performance on compositional tasks, multiplicatively explaining sudden emergence; and (iii) learning and composing capabilities is difficult under correlations. We hope our study inspires further grounded research on understanding capabilities and compositionality in generative models from a data-centric perspective. |
Maya Okawa · Ekdeep Singh Lubana · Robert Dick · Hidenori Tanaka 🔗 |
-
|
Functional Renyi Differential Privacy for Generative Modeling
(
Poster
)
link »
Recently, R\'enyi differential privacy (RDP) becomes an alternative to the ordinary differential privacy (DP) notion, for its convenient compositional rules and flexibility. However, existing mechanisms with RDP guarantees are based on randomizing a fixed, finite-dimensional vector output. In this work, following Hall et al. (2013) we further extend RDP to functional outputs, where the output space can be infinite-dimensional, and develop all necessary tools, e.g. (subsampled) Gaussian mechanism, composition, and post-processing rules, to facilitate its practical adoption. As an illustration, we apply functional RDP (f-RDP) to functions in the reproducing kernel Hilbert space (RKHS) to develop a differentially private generative model (DPGM), where training can be interpreted as releasing loss functions (in an RKHS) with RDP guarantees. Empirically, the new training paradigm achieves a significant improvement in privacy-utility trade-off compared to existing alternatives when $\epsilon=0.2$.
|
Dihong Jiang · Sun Sun · Yaoliang Yu 🔗 |
-
|
DP-LFlow: Differentially Private Latent Flow for Scalable Sensitive Image Generation
(
Poster
)
link »
Differentially private generative model (DPGM) is designed to generate data that are distributionally similar to the original sensitive data yet with differential privacy (DP) guarantees. While GAN attracts major attention, existing DPGMs based on flow generative models are limited and only developed on low-dimensional tabular datasets. The capability of exact density estimation makes the flow model exceptional especially when density estimation is of interest. In this work, we will first show that it is challenging (or even infeasible) to train a DP-flow via DP-SGD, i.e. the workhorse algorithm for private deep learning, on high-dimensional image sets with acceptable utility, and then we give an effective solution by reducing the generation from the pixel space to a lower dimensional latent space. We show the effectiveness and scalability of the proposed method via extensive experiments. Notably, our method is scalable to high-resolution image sets, which is rarely studied in related works. |
Dihong Jiang · Sun Sun 🔗 |
-
|
Chain-of-Thought Hub: A Continuous Effort to Measure Large Language Models’ Reasoning Performance
(
Poster
)
link »
As large language models (LLMs) are continuously being developed, their evaluation becomes increasingly important yet challenging. This work proposes Chain-of-Thought Hub, an open-source evaluation suite on the multi-step reasoning capabilities of large language models. We are interested in this setting for two reasons:(1) from the behavior of GPT and PaLM model family, we observe that complex reasoning is likely to be a key differentiator between weaker and stronger LLMs;(2) we envisage large language models to become the next-generation computational platform and foster an ecosystem of LLM-based new applications, this naturally requires the foundation models to perform complex tasks that often involve the composition of linguistic and logical operations. Our approach is to compile a suite of challenging reasoning benchmarks to track the progress of LLMs. Our current results show that: (1) model scale clearly correlates with reasoning capabilities;(2) As of May 2023, Claude-v1.3 and PaLM-2 are the only two models that are comparable with GPT-4, while open-sourced models still lag behind;(3) LLaMA-65B performs closely to code-davinci-002, indicating that with successful further development such as reinforcement learning from human feedback (RLHF), it has great potential to be close to GPT-3.5-Turbo. Our results also suggest that for the open-source efforts to catch up, the community may focus more on building better base models and exploring RLHF. |
Yao Fu · Litu Ou · Yuhao Wan · Mingyu Chen · Hao Peng · Tushar Khot 🔗 |
-
|
Surely You’re Lying, Mr. Model: Improving and Analyzing CCS
(
Poster
)
link »
Contrast Consistent Search (Burns et al., 2022) is a method for eliciting latent knowledge without supervision. In this paper, we explore a few directions for improving CCS. We use conjunctive logic to make CCS fully unsupervised. We investigate which factors contribute to CCS’s poor performance on autoregressive models. Replicating (Belrose & Mallen, 2023), we improve CCS’s performance on autoregressive models and study the effect of multi-shot context. And we better characterize where CCS techniques add value by adding early exit baselines to the original CCS experiments, replicating (Halawi et al., 2023). |
Naomi Bashkansky · Chloe Loughridge · Chuyue Tang 🔗 |
-
|
Ensuring Visual Commonsense Morality for Text-to-Image Generation
(
Poster
)
link »
Text-to-image generation methods produce high-resolution and high-quality images, but these methods should not produce immoral images that may contain inappropriate content from the perspective of commonsense morality. In this paper, we aim to automatically judge the immorality of synthesized images and manipulate these images into morally acceptable alternatives. To this end, we build a model that has three main primitives: (1) recognition of the visual commonsense immorality in a given image, (2) localization or highlighting of immoral visual (and textual) attributes that contribute to the immorality of the image, and (3) manipulation of an immoral image to create a morally-qualifying alternative. We conduct experiments and human studies using the state-of-the-art Stable Diffusion text-to-image generation model, demonstrating the effectiveness of our ethical image manipulation approach. |
Seongbeom Park · Suhong Moon · Jinkyu Kim 🔗 |
-
|
Concept Bottleneck Generative Models
(
Poster
)
link »
Despite their increasing prevalence, generative models remain opaque and difficult to steer reliably. To address these challenges, we present concept bottleneck (CB) generative models, a type of generative model where one of its internal layers—a concept bottleneck (CB) layer—is constrained to encode human-understandable features. While concept-botttleneck layers have been used to improved interpretability for supervised learning tasks, here we extend them generative models. The concept bottleneck layer partitions the generative model into three parts: the pre-concept bottleneck portion, the CB layer, and the post-concept bottleneck portion. To train CB generative models, we complement the traditional task-based loss function for training generative models with three additional loss terms: a concept loss, an orthogonality loss, and a concept sensitivity loss. The CB layer and these corresponding loss termsare model agnostic, which we demonstrate by applying them to three different families of generativemodels: generative adversarial networks, variational autoencoders, and diffusion models. Onreal-world datasets, across three types of generative models, steering a generative model with the CB layer outperforms several baselines. |
Aya Ismail · Julius Adebayo · Hector Corrada Bravo · Stephen Ra · Kyunghyun Cho 🔗 |
-
|
Language Model Tokenizers Introduce Unfairness Between Languages
(
Poster
)
link »
Recent language models have shown impressive multilingual performance, even when not explicitly trained for it. Despite this, there are concerns about the quality of their outputs across different languages. In this paper, we show how disparity in the treatment of languages arises at the tokenization stage, well before a model is even invoked. The same text translated into different languages can have drastically different tokenization lengths, with differences up to 15 times in some cases. These disparities persist across the 17 tokenizers we evaluate, even if they are intentionally trained for multilingual support. Character-level and byte-level models also exhibit over 4 times the difference in the encoding length for some language pairs. This induces unfair treatment for some language communities in regard to the cost of accessing commercial language services, the processing time and latency, as well as the amount of content that can be provided as context to the models. Therefore, we make the case that we should train future language models using multilingually fair subword tokenizers |
Aleksandar Petrov · Emanuele La Malfa · Phil Torr · Adel Bibi 🔗 |
-
|
CLAM: Selective Clarification for Ambiguous Questions with Generative Language Models
(
Poster
)
link »
Users often ask dialogue systems ambiguous questions that require clarification. We show that current language models rarely ask users to clarify ambiguous questions and instead provide incorrect answers. To address this, we introduce CLAM: a framework for getting language models to selectively ask for clarification about ambiguous user questions. In particular, we show that we can prompt language models to detect whether a given question is ambiguous, generate an appropriate clarifying question to ask the user, and give a final answer after receiving clarification. We also show that we can simulate users by providing language models with privileged information. This lets us automatically evaluate multi-turn clarification dialogues. Finally, CLAM significantly improves language models' accuracy on mixed ambiguous and unambiguous questions relative to SotA. |
Lorenz Kuhn · Yarin Gal · Sebastian Farquhar 🔗 |
-
|
The Power of Sound (TPoS): Audio Reactive Video Generation with Stable Diffusion
(
Poster
)
link »
In recent years, video generation has drawn significant attention. However, there is little consideration in audio-to-video generation, though audio contains unique qualities like temporal semantics and magnitude. Hence, we propose The Power of Sound (TPoS) model to incorporate audio input that includes both changeable temporal semantics and magnitude. To generate video frames, TPoS utilizes a latent stable diffusion model with textual semantic information, which is then guided by the sequential audio embedding from our pretrained Audio Encoder. As a result, this method produces audio reactive video contents. We demonstrate the effectiveness of TPoS across various tasks and compare its results with current state-of-the-art techniques in the field of audio-to-video generation. |
Yujin Jeong · Wonjeong Ryoo · Seung Hyun Lee · Da Bin Seo · Wonmin Byeon · Sangpil Kim · Jinkyu Kim 🔗 |
-
|
Large Language Models for Code: Security Hardening and Adversarial Testing
(
Poster
)
link »
Large language models (large LMs) are increasingly used to generate code. However, LMs lack awareness of security and are found to frequently produce unsafe code. This work studies the security of LMs along two important axes: (i) security hardening, which enhances LMs' reliability in generating secure code, and (ii) adversarial testing, which evaluates LMs' security at an adversarial standpoint. To address both, we propose a novel method called SVEN, which leverages continuous prompts to control LMs to generate secure or unsafe code. We optimize these continuous vectors by enforcing specialized loss terms on different code regions, using a high-quality dataset carefully curated by us. Our extensive evaluation shows that SVEN achieves strong security control and preserves functional correctness. |
Jingxuan He · Martin Vechev 🔗 |
-
|
TRAC: Trustworthy Retrieval Augmented Chatbot
(
Poster
)
link »
Although neural conversational AIs have demonstrated fantastic performance, they often generate incorrect information, or \textit{hallucinations}. Retrieval augmented generation has emerged as a promising solution to reduce these hallucinations. However, these techniques still cannot guarantee correctness. Focusing on question answering, we propose a framework that can provide statistical guarantees for the retrieval augmented question answering system by combining conformal prediction and global testing. In addition, we use Bayesian optimization to choose hyperparameters of the global test to maximize the performance of the system. Our empirical results on the Natural Questions dataset demonstrate that our method can provide the desired coverage guarantee while minimizing the average prediction set size. |
Shuo Li · Sangdon Park · Insup Lee · Osbert Bastani 🔗 |
-
|
Evaluation Metrics for Protein Structure Generation
(
Poster
)
link »
Generative models have become increasingly popular for sampling novel proteins. To compare and evaluate these models, we need metrics that can assess the quality of the generated structures. We propose a set of standardized metrics for benchmarking protein generation. We experimentally show that these metrics can measure differences between proteins on a distributional level, as well as quantify the novelty, diversity and designability of the generated proteins. |
Josh Southern · Arne Schneuing · Michael Bronstein · Bruno Correia 🔗 |
-
|
De-stereotyping Text-to-image Models through Prompt Tuning
(
Poster
)
link »
Recent text-to-image (TTI) generation models have been reported to generate images demographically stereotyped in various sensitive attributes such as gender or race. This may seriously harm the fairness of the generative model to be deployed. We propose a novel and efficient framework to de-stereotype the existing TTI model through soft prompt tuning. Utilizing a newly designed de-stereotyping loss, we train a small number of parameters consisting of the soft prompt. We demonstrate that our framework effectively balances the generated images with respect to sensitive attributes, which can also generalize to unseen text prompts. |
Eunji Kim · Siwon Kim · Chaehun Shin · Sungroh Yoon 🔗 |
-
|
Using Synthetic Data for Data Augmentation to Improve Classification Accuracy
(
Poster
)
link »
Obtaining high quality data for training classification models is challenging when sufficient data covering the real manifold is difficult to find in the wild. In this paper, we present Diffusion Inversion, a dataset-agnostic augmentation strategy for training classification models. Diffusion Inversion is a simple yet effective method that leverages the powerful pretrained Stable Diffusion model to generate synthetic datasets that ensure coverage of the original data manifold while also generating novel samples that extrapolate the training domain to allow for better generalization. We ensure data coverage by inverting each image in the original set to its condition vector in the latent space of Stable Diffusion. We ensure sample diversity by adding noise to the learned embeddings or performing interpolation in the latentspace, and using the new vector as the conditioning signal. The method produces high-quality and diverse samples, consistently outperforming generic prompt-based steering methods and KNN retrieval baselines across a wide range of common and specialized datasets. Furthermore, we demonstrate the compatibility of our approach with widely-used data augmentation techniques, and assess the reliability of the generated data in both supporting various neural architectures and enhancing few-shot learning performance. |
Yongchao Zhou · Hshmat Sahak · Jimmy Ba 🔗 |
-
|
Goal-conditioned GFlowNets for Controllable Multi-Objective Molecular Design
(
Poster
)
link »
In recent years, in-silico molecular design has received much attention from the machine learning community. When designing a new compound for pharmaceutical applications, there are usually multiple properties of such molecules that need to be optimised: binding energy to the target, synthesizability, toxicity, EC50, and so on. While previous approaches have employed a scalarization scheme to turn the multi-objective problem into a preference-conditioned single objective, it has been established that this kind of reduction may produce solutions that tend to slide towards the extreme points of the objective space when presented with a problem that exhibits a concave Pareto front. In this work we experiment with an alternative formulation of goal-conditioned molecular generation to obtain a more controllable conditional model that can uniformly explore solutions along the entire Pareto front. |
Julien Roy · Pierre-Luc Bacon · Christopher Pal · Emmanuel Bengio 🔗 |
-
|
The Unseen A+ Student: Navigating the Impact of Large Language Models in the Classroom
(
Poster
)
link »
The recent boom of so-called generative artificial intelligence (AI) applications, namely large language models such as ChatGPT, took the public discourse by storm, disrupting many fields and industries. Education, being one of them, was left with little to no information about the impacts of such an unprecedented technology, let alone holistic methodologies on how to reflect and adopt it. Therefore, many educators rush to adopt policies banning the technology but find themselves incapable of reliably detecting the cheaters. All in all, the distribution established chaos in the education community. To address this problem, we propose a methodology assessing relevant information for educational institutions that can guide their decision-making around generative AI practices and policies. We demonstrate this methodology in a Czech high school, conduct a student assessment study, and make the dataset available to the public. |
Matyáš Boháček 🔗 |
-
|
Interpolating between Images with Diffusion Models
(
Poster
)
link »
One little-explored frontier of image generation and editing is the task of interpolating between two input images, a feature missing from all currently deployed image generation pipelines. We argue that such a feature can expand the creative applications of such models, and propose a method for zero-shot controllable interpolation using latent diffusion models. We apply interpolation in latent space at a sequence of decreasing noise levels, then perform denoising conditioned on interpolated text embeddings derived from textual inversion and (optionally) subject poses derived from OpenPose. For greater consistency, or to specify additional criteria, we can generate several candidates and use CLIP to select the highest quality image. We obtain convincing interpolations across diverse subject poses, image styles, and image content, and show that standard quantitative metrics such as FID are insufficient to identify successful interpolations. |
Clinton Wang · Polina Golland 🔗 |
Author Information
Swami Sankaranarayanan (Massachusetts Institute of Technology)
Thomas Hartvigsen (MIT)
Camille Bilodeau (University of Virginia)
Dr. Bilodeau is currently an assistant professor in Chemical Engineering at the University of Virginia. She received her B.S. and M.S. from Northwestern University and her Ph.D. from Rensselaer Polytechnic Institute, both in Chemical and Biological Engineering. During her Ph.D., she received the Lawrence Livermore Advanced Simulations and Computation Graduate Fellowship, through which she carried out research at Lawrence Livermore National Laboratory. Her research explores the intersection between artificial intelligence and molecular simulations with the goal of designing new molecules and materials.
Ryutaro Tanno (DeepMind)
Cheng Zhang (Microsoft Research, Cambridge)
Florian Tramer (ETH Zürich)
Phillip Isola (MIT)
More from the Same Authors
-
2021 : Optimal transport for causal discovery »
Ruibo Tu · Kun Zhang · Hedvig Kjellström · Cheng Zhang -
2021 : VICAUSE: Simultaneous missing value imputation and causal discovery »
Pablo Morales-Alvarez · Angus Lamb · Simon Woodhead · Simon Pyton Jones · Miltiadis Allamanis · Cheng Zhang -
2023 : Counterfactual Memorization in Neural Language Models »
Chiyuan Zhang · Daphne Ippolito · Katherine Lee · Matthew Jagielski · Florian Tramer · Nicholas Carlini -
2023 : Identifying Implicit Social Biases in Vision-Language Models »
Kimia Hamidieh · Haoran Zhang · Thomas Hartvigsen · Marzyeh Ghassemi -
2023 : A Pipeline for Interpretable Clinical Subtyping with Deep Metric Learning »
Haoran Zhang · Qixuan Jin · Thomas Hartvigsen · Miriam Udler · Marzyeh Ghassemi -
2023 : A Pipeline for Interpretable Clinical Subtyping with Deep Metric Learning »
Haoran Zhang · Qixuan Jin · Thomas Hartvigsen · Miriam Udler · Marzyeh Ghassemi -
2023 : Continuous Time Evidential Distributions for Irregular Time Series »
Taylor Killian · Haoran Zhang · Thomas Hartvigsen · Ava Amini -
2023 : BayesDAG: Gradient-Based Posterior Sampling for Causal Discovery »
Yashas Annadani · Nick Pawlowski · Joel Jennings · Stefan Bauer · Cheng Zhang · Wenbo Gong -
2023 : Answering Causal Questions with Augmented LLMs »
Nick Pawlowski · Joel Jennings · Cheng Zhang -
2023 : Identifying Implicit Social Biases in Vision-Language Models »
Kimia Hamidieh · Haoran Zhang · Thomas Hartvigsen · Marzyeh Ghassemi -
2023 : RustGen: An Augmentation Approach for Generating Compilable Rust Code with Large Language Models »
Xingbo Wu · Nathanaël Cheriere · Cheng Zhang · Dushyanth Narayanan -
2023 : Panel Discussion »
Peter Kairouz · Song Han · Kamalika Chaudhuri · Florian Tramer -
2023 Poster: Optimal Goal-Reaching Reinforcement Learning via Quasimetric Learning »
Tongzhou Wang · Antonio Torralba · Phillip Isola · Amy Zhang -
2023 Poster: CO-BED: Information-Theoretic Contextual Optimization via Bayesian Experimental Design »
Desi Ivanova · Joel Jennings · Tom Rainforth · Cheng Zhang · Adam Foster -
2023 Poster: Straightening Out the Straight-Through Estimator: Overcoming Optimization Challenges in Vector Quantized Networks »
Minyoung Huh · Brian Cheung · Pulkit Agrawal · Phillip Isola -
2022 Poster: Denoised MDPs: Learning World Models Better Than the World Itself »
Tongzhou Wang · Simon Du · Antonio Torralba · Phillip Isola · Amy Zhang · Yuandong Tian -
2022 Spotlight: Denoised MDPs: Learning World Models Better Than the World Itself »
Tongzhou Wang · Simon Du · Antonio Torralba · Phillip Isola · Amy Zhang · Yuandong Tian -
2020 : Invited talk 4: Divergence Measures in Variational Inference and How to Choose Them »
Cheng Zhang -
2020 Poster: Understanding Contrastive Representation Learning through Alignment and Uniformity on the Hypersphere »
Tongzhou Wang · Phillip Isola -
2019 : Cheng Zhang: Active Mini-Batch Sampling using Repulsive Point Processes »
Cheng Zhang -
2019 Poster: EDDI: Efficient Dynamic Discovery of High-Value Information with Partial VAE »
Chao Ma · Sebastian Tschiatschek · Konstantina Palla · Jose Miguel Hernandez-Lobato · Sebastian Nowozin · Cheng Zhang -
2019 Oral: EDDI: Efficient Dynamic Discovery of High-Value Information with Partial VAE »
Chao Ma · Sebastian Tschiatschek · Konstantina Palla · Jose Miguel Hernandez-Lobato · Sebastian Nowozin · Cheng Zhang