LLaRA: Supercharging Robot Learning Data for Vision-Language Policy (2024.06.28)
Xiang Li, Cristina Mata, Jong Sung Park, Kumara Kahatapitiya, Yoo Sung Jang, etc
Dataset Size Recovery from LoRA Weights (2024.06.27)
Mohammad Salama, Jonathan Kahana, Eliahu Horwitz, Yedid Hoshen
Dual-Phase Accelerated Prompt Optimization (2024.06.19)
Muchen Yang, Moxin Li, Yongle Li, Zijun Chen, Chongming Gao, etc
Hitesh Wadhwa, Rahul Seetharaman, Somyaa Aggarwal, Reshmi Ghosh, Samyadeep Basu, etc
VoCo-LLaMA: Towards Vision Compression with Large Language Models (2024.06.18)
Xubing Ye, Yukang Gan, Xiaoke Huang, Yixiao Ge, Ying Shan, etc
LaMDA: Large Model Fine-Tuning via Spectrally Decomposed Low-Dimensional Adaptation (2024.06.18)
Seyedarmin Azizi, Souvik Kundu, M. Pedram
The Impact of Initialization on LoRA Finetuning Dynamics (2024.06.12)
Soufiane Hayou, Nikhil Ghosh, Bin Yu
An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language Models (2024.06.07)
Xiongtao Zhou, Jie He, Yuhua Ke, Guangyao Zhu, V'ictor Guti'errez-Basulto, etc
Cross-Context Backdoor Attacks against Graph Prompt Learning (2024.05.28)
Xiaoting Lyu, Yufei Han, Wei Wang, Hangwei Qian, Ivor Tsang, etc . - 【arXiv.org】
Yuan 2.0-M32: Mixture of Experts with Attention Router (2024.05.28)
Shaohua Wu, Jiangang Luo, Xi Chen, Lingjun Li, Xudong Zhao, etc . - 【arXiv.org】
Efficient Prompt Tuning by Multi-Space Projection and Prompt Fusion (2024.05.19)
Pengxiang Lan, Enneng Yang, Yuting Liu, Guibing Guo, Linying Jiang, etc . - 【arXiv.org】
Unsupervised Image Prior via Prompt Learning and CLIP Semantic Guidance for Low-Light Image Enhancement (2024.05.19)
Igor Morawski, Kai He, Shusil Dangi, Winston H. Hsu . - 【arXiv.org】
Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision Transformers (2024.05.17)
Shengyuan Yang, Jiawang Bai, Kuofeng Gao, Yong Yang, Yiming Li, etc . - 【arXiv.org】
Revisiting OPRO: The Limitations of Small-Scale LLMs as Optimizers (2024.05.16)
Tuo Zhang, Jinyue Yuan, A. Avestimehr . - 【arXiv.org】
Pseudo-Prompt Generating in Pre-trained Vision-Language Models for Multi-Label Medical Image Classification (2024.05.10)
Yaoqin Ye, Junjie Zhang, Hongwei Shi . - 【arXiv.org】
Memory-Space Visual Prompting for Efficient Vision-Language Fine-Tuning (2024.05.09)
Shibo Jie, Yehui Tang, Ning Ding, Zhi-Hong Deng, Kai Han, etc . - 【arXiv.org】
On the test-time zero-shot generalization of vision-language models: Do we really need prompt learning? (2024.05.03)
Maxime Zanella, Ismail Ben Ayed . - 【arXiv.org】
AdaMoLE: Fine-Tuning Large Language Models with Adaptive Mixture of Low-Rank Adaptation Experts (2024.05.01)
Zefang Liu, Jiahua Luo . - 【arXiv.org】
Self-Refine Instruction-Tuning for Aligning Reasoning in Language Models (2024.05.01)
Leonardo Ranaldi, André Freitas . - 【arXiv.org】
RST-LoRA: A Discourse-Aware Low-Rank Adaptation for Long Document Abstractive Summarization (2024.05.01)
Dongqi Pu, Vera Demberg . - 【North American Chapter of the Association for Computational Linguistics】
Training-Free Unsupervised Prompt for Vision-Language Models (2024.04.25)
Sifan Long, Linbin Wang, Zhen Zhao, Zichang Tan, Yiming Wu, etc . - 【arXiv.org】
AAPL: Adding Attributes to Prompt Learning for Vision-Language Models (2024.04.25)
Gahyeon Kim, Sohee Kim, Seokju Lee . - 【arXiv.org】
V2Xum-LLM: Cross-Modal Video Summarization with Temporal Prompt Instruction Tuning (2024.04.18)
Hang Hua, Yunlong Tang, Chenliang Xu, Jiebo Luo . - 【arXiv.org】
LLMTune: Accelerate Database Knob Tuning with Large Language Models (2024.04.17)
Xinmei Huang, Haoyang Li, Jing Zhang, Xinxin Zhao, Zhiming Yao, etc . - 【arXiv.org】
DLoRA: Distributed Parameter-Efficient Fine-Tuning Solution for Large Language Model (2024.04.08)
Chao Gao, Sai Qian Zhang
3P-LLM: Probabilistic Path Planning using Large Language Model for Autonomous Robot Navigation (2024.03.27)
Ehsan Latif . - 【arXiv.org】
SAMCT: Segment Any CT Allowing Labor-Free Task-Indicator Prompts (2024.03.20)
Xian Lin, Yangyang Xiang, Zhehao Wang, Kwang-Ting Cheng, Zengqiang Yan, etc
AFLoRA: Adaptive Freezing of Low Rank Adaptation in Parameter Efficient Fine-Tuning of Large Models (2024.03.20)
Zeyu Liu, Souvik Kundu, Anni Li, Junrui Wan, Lianghao Jiang, etc
Few-Shot Class Incremental Learning with Attention-Aware Self-Adaptive Prompt (2024.03.14)
Chenxi Liu, Zhenyi Wang, Tianyi Xiong, Ruibo Chen, Yihan Wu, etc
Unveiling the Generalization Power of Fine-Tuned Large Language Models (2024.03.14)
Haoran Yang, Yumeng Zhang, Jiaqi Xu, Hongyuan Lu, P. Heng, etc
Attention Prompt Tuning: Parameter-efficient Adaptation of Pre-trained Models for Spatiotemporal Modeling (2024.03.11)
W. G. C. Bandara, Vishal M. Patel
VidProM: A Million-scale Real Prompt-Gallery Dataset for Text-to-Video Diffusion Models (2024.03.10)
Wenhao Wang, Yi Yang
Localized Zeroth-Order Prompt Optimization (2024.03.05)
Wenyang Hu, Yao Shu, Zongmin Yu, Zhaoxuan Wu, Xiangqiang Lin, etc
RIFF: Learning to Rephrase Inputs for Few-shot Fine-tuning of Language Models (2024.03.04)
Saeed Najafi, Alona Fyshe
Prompt-Driven Dynamic Object-Centric Learning for Single Domain Generalization (2024.02.28)
Deng Li, Aming Wu, Yaowei Wang, Yahong Han
Meta-Task Prompting Elicits Embedding from Large Language Models (2024.02.28)
Yibin Lei, Di Wu, Tianyi Zhou, Tao Shen, Yu Cao, etc
GISTEmbed: Guided In-sample Selection of Training Negatives for Text Embedding Fine-tuning (2024.02.26)
Aivin V. Solatorio . - 【arXiv.org】
Fujian Jia, Xin Liu, Lixi Deng, Jiwen Gu, Chunchao Pu, etc . - 【arXiv.org】
Stepwise Self-Consistent Mathematical Reasoning with Large Language Models (2024.02.24)
Zilong Zhao, Yao Rong, Dongyang Guo, Emek Gözlüklü, Emir Gülboy, etc . - 【arXiv.org】
Double-I Watermark: Protecting Model Copyright for LLM Fine-tuning (2024.02.22)
Shen Li, Liuyi Yao, Jinyang Gao, Lan Zhang, Yaliang Li . - 【arXiv.org】
Semantic Mirror Jailbreak: Genetic Algorithm Based Jailbreak Prompts Against Open-source LLMs (2024.02.21)
Xiaoxia Li, Siyuan Liang, Jiyi Zhang, Hansheng Fang, Aishan Liu, etc . - 【arXiv.org】
Qihuang Zhong, Liang Ding, Juhua Liu, Bo Du, Dacheng Tao . - 【arXiv.org】
LoRETTA: Low-Rank Economic Tensor-Train Adaptation for Ultra-Low-Parameter Fine-Tuning of Large Language Models (2024.02.18)
Yifan Yang, Jiajun Zhou, Ngai Wong, Zheng Zhang . - 【North American Chapter of the Association for Computational Linguistics】
MoRAL: MoE Augmented LoRA for LLMs' Lifelong Learning (2024.02.17)
Shu Yang, Muhammad Asif Ali, Cheng-Long Wang, Lijie Hu, Di Wang . - 【arXiv.org】
E. Buehler, M. J. Buehler . - 【APL Machine Learning】
Self-Discover: Large Language Models Self-Compose Reasoning Structures (2024.02.06)
Pei Zhou, J. Pujara, Xiang Ren, Xinyun Chen, Heng-Tze Cheng, etc . - 【arXiv.org】
Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding (2024.01.23)
Mirac Suzgun, A. Kalai . - 【arXiv.org】
Chain-of-Table: Evolving Tables in the Reasoning Chain for Table Understanding (2024.01.09)
Zilong Wang, Hao Zhang, Chun-Liang Li, Julian Martin Eisenschlos, Vincent Perot, etc . - 【arXiv.org】
Rephrase and Respond: Let Large Language Models Ask Better Questions for Themselves (2023.11.07)
Yihe Deng, Weitong Zhang, Zixiang Chen, Quanquan Gu . - 【arXiv.org】
Take a Step Back: Evoking Reasoning via Abstraction in Large Language Models (2023.10.09)
Huaixiu Steven Zheng, Swaroop Mishra, Xinyun Chen, Heng-Tze Cheng, E. Chi, etc . - 【arXiv.org】
Walking Down the Memory Maze: Beyond Context Limit through Interactive Reading (2023.10.08)
Howard Chen, Ramakanth Pasunuru, Jason Weston, Asli Celikyilmaz . - 【arXiv.org】
Chain-of-Verification Reduces Hallucination in Large Language Models (2023.09.20)
S. Dhuliawala, M. Komeili, Jing Xu, Roberta Raileanu, Xian Li, etc . - 【arXiv.org】
Giraffe: Adventures in Expanding Context Lengths in LLMs (2023.08.21)
Arka Pal, Deep Karkhanis, Manley Roberts, S. Dooley, Arvind Sundararajan, etc . - 【arXiv.org】
Exploring the Intersection of Large Language Models and Agent-Based Modeling via Prompt Engineering (2023.08.14)
Edward Junprung . - 【arXiv.org】
Cumulative Reasoning with Large Language Models (2023.08.08)
Yifan Zhang, Jingqin Yang, Yang Yuan, A. Yao . - 【arXiv.org】
Scaling Relationship on Learning Mathematical Reasoning with Large Language Models (2023.08.03)
Zheng Yuan, Hongyi Yuan, Cheng Li, Guanting Dong, Chuanqi Tan, etc . - 【arXiv.org】
Skeleton-of-Thought: Large Language Models Can Do Parallel Decoding (2023.07.28)
Xuefei Ning, Zinan Lin, Zixuan Zhou, Huazhong Yang, Yu Wang . - 【arXiv.org】
ChatSpot: Bootstrapping Multimodal LLMs via Precise Referring Instruction Tuning (2023.07.18)
Liang Zhao, En Yu, Zheng Ge, Jinrong Yang, Hao-Ran Wei, etc . - 【arXiv.org】
Zhenhailong Wang, Shaoguang Mao, Wenshan Wu, Tao Ge, Furu Wei, etc
Self-consistency for open-ended generations (2023.07.11)
Brain in a Vat: On Missing Pieces Towards Artificial General Intelligence in Large Language Models (2023.07.07)
Yuxi Ma, Chi Zhang, Song-Chun Zhu . - 【arXiv.org】
Conformer LLMs - Convolution Augmented Large Language Models (2023.07.02)
Prateek Verma . - 【arXiv.org】
Learning Multi-Step Reasoning by Solving Arithmetic Tasks (2023.06.02)
Tianduo Wang, Wei Lu . - 【arXiv.org】
AutoPlan: Automatic Planning of Interactive Decision-Making Tasks With Large Language Models (2023.05.24)
Siqi Ouyang, Lei Li . - 【Conference on Empirical Methods in Natural Language Processing】
OverPrompt: Enhancing ChatGPT Capabilities through an Efficient In-Context Learning Approach (2023.05.24)
Jiazheng Li, Runcong Zhao, Yulan He, Lin Gui
In-Context Impersonation Reveals Large Language Models' Strengths and Biases (2023.05.24)
Leonard Salewski, Stephan Alaniz, Isabel Rio-Torto, Eric Schulz, Zeynep Akata
Frugal Prompting for Dialog Models (2023.05.24)
Bishal Santra, Sakya Basak, Abhinandan De, Manish Gupta, Pawan Goyal
Chain-of-Questions Training with Latent Answers for Robust Multistep Question Answering (2023.05.24)
Wang Zhu, Jesse Thomason, Robin Jia
Improving Probability-based Prompt Selection Through Unified Evaluation and Analysis (2023.05.24)
Sohee Yang, Jonghyeon Kim, Joel Jang, Seonghyeon Ye, Hyunji Lee, etc
MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions (2023.05.24)
Zexuan Zhong, Zhengxuan Wu, Christopher D. Manning, Christopher Potts, Danqi Chen
BeamSearchQA: Large Language Models are Strong Zero-Shot QA Solver (2023.05.24)
Hao Sun, Xiao Liu, Yeyun Gong, Yan Zhang, Nan Duan
TACR: A Table-alignment-based Cell-selection and Reasoning Model for Hybrid Question-Answering (2023.05.24)
Jian Wu, Yicheng Xu, Yan Gao, Jian-Guang Lou, Börje F. Karlsson, etc
Mixture of Prompt Experts for Generalizable and Interpretable Question Answering (2023.05.24)
Chenglei Si, Weijia Shi, Chen Zhao, Luke Zettlemoyer, Jordan L. Boyd-Graber
Chain-of-Knowledge: Grounding Large Language Models via Dynamic Knowledge Adapting over Heterogeneous Sources (2023.05.22)
Xingxuan Li, Ruochen Zhao, Yew Ken Chia, Bosheng Ding, Shafiq R. Joty, etc
Logic-LM: Empowering Large Language Models with Symbolic Solvers for Faithful Logical Reasoning (2023.05.20)
Liangming Pan, Alon Albalak, Xinyi Wang, William Yang Wang
SelfzCoT: a Self-Prompt Zero-shot CoT from Semantic-level to Code-level for a Better Utilization of LLMs (2023.05.19)
IokTong Lei, ZhiDong Deng . - 【arXiv.org】
Giorgi Kokaia, Pratyush Sinha, Yutong Jiang, N. Boujemaa . - 【arXiv.org】
Efficient Prompting via Dynamic In-Context Learning (2023.05.18)
Wangchunshu Zhou, Yuchen Jiang, Ryan Cotterell, Mrinmaya Sachan . - 【arXiv.org】
Distilling Step-by-Step! Outperforming Larger Language Models with Less Training Data and Smaller Model Sizes (2023.05.03)
Cheng-Yu Hsieh, Chun-Liang Li, Chih-Kuan Yeh, Hootan Nakhost, Yasuhisa Fujii, etc . - 【Annual Meeting of the Association for Computational Linguistics】
WizardLM: Empowering Large Language Models to Follow Complex Instructions (2023.04.24)
Can Xu, Qingfeng Sun, Kai Zheng, Xiubo Geng, Pu Zhao, etc
LLM+P: Empowering Large Language Models with Optimal Planning Proficiency (2023.04.22)
B. Liu, Yuqian Jiang, Xiaohan Zhang, Qian Liu, Shiqi Zhang, etc
Why Johnny Can’t Prompt: How Non-AI Experts Try (and Fail) to Design LLM Prompts (2023.04.19)
J. Zamfirescu-Pereira, Richmond Y. Wong, Bjoern Hartmann, Qiang Yang . - 【International Conference on Human Factors in Computing Systems】
Progressive-Hint Prompting Improves Reasoning in Large Language Models (2023.04.19)
Chuanyang Zheng, Zhengying Liu, Enze Xie, Zhenguo Li, Yu Li
Boosted Prompt Ensembles for Large Language Models (2023.04.12)
Silviu Pitis, Michael Ruogu Zhang, Andrew Wang, Jimmy Ba . - 【arXiv.org】
Prompt Pre-Training with Twenty-Thousand Classes for Open-Vocabulary Visual Recognition (2023.04.10)
REFINER: Reasoning Feedback on Intermediate Representations (2023.04.04)
Debjit Paul, Mete Ismayilzada, Maxime Peyrard, Beatriz Borges, Antoine Bosselut, etc
Self-Refine: Iterative Refinement with Self-Feedback (2023.03.30)
Context-faithful Prompting for Large Language Models (2023.03.20)
Wenxuan Zhou, Sheng Zhang, Hoifung Poon, Muhao Chen . - 【arXiv.org】
Reflexion: an autonomous agent with dynamic memory and self-reflection (2023.03.20)
Noah Shinn, Beck Labash, Ashwin Gopinath . - 【arXiv.org】
SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models (2023.03.18)
Vithursan Thangarasa, Abhay Gupta, William Marshall, Tianda Li, Kevin Leong, etc . - 【Conference on Uncertainty in Artificial Intelligence】
A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT (2023.02.21)
Jules White, Quchen Fu, Sam Hays, M. Sandborn, Carlos Olea, etc . - 【ArXiv】
GraphPrompt: Unifying Pre-Training and Downstream Tasks for Graph Neural Networks (2023.02.16)
Zemin Liu, Xingtong Yu, Yuan Fang, Xinming Zhang . - 【ArXiv】
Progressive Prompts: Continual Learning for Language Models (2023.01.29)
Anastasia Razdaibiedina, Yuning Mao, Rui Hou, Madian Khabsa, M. Lewis, etc . - 【ArXiv】
Batch Prompting: Efficient Inference with Large Language Model APIs (2023.01.19)
Zhoujun Cheng, Jungo Kasai, Tao Yu . - 【ArXiv】
Successive Prompting for Decomposing Complex Questions (2022.12.08)
Dheeru Dua, Shivanshu Gupta, Sameer Singh, Matt Gardner . - 【Conference on Empirical Methods in Natural Language Processing】
PAL: Program-aided Language Models (2022.11.18)
Luyu Gao, Aman Madaan, Shuyan Zhou, Uri Alon, Pengfei Liu, etc . - 【ArXiv】
Measuring and Narrowing the Compositionality Gap in Language Models (2022.10.07)
Ofir Press, Muru Zhang, Sewon Min, Ludwig Schmidt, Noah A. Smith, etc . - 【ArXiv】
ReAct: Synergizing Reasoning and Acting in Language Models (2022.10.06)
Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, I. Shafran, etc . - 【arXiv.org】
Interactive and Visual Prompt Engineering for Ad-hoc Task Adaptation with Large Language Models (2022.08.16)
Hendrik Strobelt, Albert Webson, Victor Sanh, Benjamin Hoover, J. Beyer, etc . - 【IEEE Transactions on Visualization and Computer Graphics】
Black-box Prompt Learning for Pre-trained Language Models (2022.01.21)
Shizhe Diao, Xuechun Li, Yong Lin, Zhichao Huang, Xiao Zhou, etc . - 【ArXiv】
Design Guidelines for Prompt Engineering Text-to-Image Generative Models (2021.09.14)
Vivian Liu, Lydia B. Chilton . - 【International Conference on Human Factors in Computing Systems】
Do Prompt-Based Models Really Understand the Meaning of Their Prompts? (2021.09.02)
Albert Webson, Ellie Pavlick . - 【North American Chapter of the Association for Computational Linguistics】
PADA: Example-based Prompt Learning for on-the-fly Adaptation to Unseen Domains (2021.02.24)
Eyal Ben-David, Nadav Oved, Roi Reichart . - 【International Conference on Topology, Algebra and Categories in Logic】
Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm (2021.02.15)
Laria Reynolds, Kyle McDonell . - 【CHI Extended Abstracts】
Prompt Engineering for Text-Based Generative Art
J. Oppenlaender . - 【ArXiv】
L4Q: Parameter Efficient Quantization-Aware Training on Large Language Models via LoRA-wise LSQ
Hyesung Jeon, Yulhwa Kim, Jae-Joon Kim . - 【arXiv.org】