Paper Reading AI Learner

Large Language Models as In-context AI Generators for Quality-Diversity

2024-04-24 10:35:36
Bryan Lim, Manon Flageat, Antoine Cully

Abstract

Quality-Diversity (QD) approaches are a promising direction to develop open-ended processes as they can discover archives of high-quality solutions across diverse niches. While already successful in many applications, QD approaches usually rely on combining only one or two solutions to generate new candidate solutions. As observed in open-ended processes such as technological evolution, wisely combining large diversity of these solutions could lead to more innovative solutions and potentially boost the productivity of QD search. In this work, we propose to exploit the pattern-matching capabilities of generative models to enable such efficient solution combinations. We introduce In-context QD, a framework of techniques that aim to elicit the in-context capabilities of pre-trained Large Language Models (LLMs) to generate interesting solutions using the QD archive as context. Applied to a series of common QD domains, In-context QD displays promising results compared to both QD baselines and similar strategies developed for single-objective optimization. Additionally, this result holds across multiple values of parameter sizes and archive population sizes, as well as across domains with distinct characteristics from BBO functions to policy search. Finally, we perform an extensive ablation that highlights the key prompt design considerations that encourage the generation of promising solutions for QD.

Abstract (translated)

质量多样性(QD)方法是开发开放性过程的有前途的方向,因为它们可以揭示不同领域的优质解决方案。虽然已经在许多应用中取得成功,但QD方法通常仅结合一个或两个解决方案来生成新的候选解决方案。如开放性过程(如技术进步)中所观察到的,明智地结合这些解决方案的大多样性可能会导致更具创新性的解决方案,并可能提高QD搜索的生产力。在这项工作中,我们提出了利用生成模型的模式匹配能力来实现这种有效的解决方案组合。我们引入了In-Context QD,一个旨在通过将预训练的大语言模型(LLMs)在QD档案中生成有趣解决方案的技术框架。将In-Context QD应用于一系列常见的QD领域,与QD基线和为单目标优化 similar strategies 相比,显示出积极的结果。此外,这一结果在参数大小和档案人口大小不同的领域以及具有不同特性的领域中均成立。最后,我们进行了一项广泛的消融,重点突出了鼓励QD生成有前途的解决方案的关键提示设计考虑。

URL

https://arxiv.org/abs/2404.15794

PDF

https://arxiv.org/pdf/2404.15794.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot