Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

A list of data-efficient and data-centric LLM (Large Language Model) papers. Our Survey Paper: Towards Efficient LLM Post Training: A Data-centric Perspective

NotificationsYou must be signed in to change notification settings

luo-junyu/Awesome-Data-Efficient-LLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 

Repository files navigation

A list of data-efficient and data-centric LLM (Large Language Model) papers

flywheel

❖ Paper List

TilteTLDRCategoryPaper LinkYearPublish
Data-efficient Fine-tuning for LLM-based RecommendationPropose data pruning method for efficient LLM - based recommendation.Data Selectionlink2024ACM
CoachLM: Automatic Instruction Revisions Improve the Data Quality in LLM Instruction TuningCoachLM automatically revises samples to enhance instruction dataset quality.Data Selection, Data Quality Enhancementlink2023IEEE
Alpagasus:Training a Better Alpaca with Fewer DataPropose data selection strategy, filter low - quality data for IFT, ALPAGASUS as example.Data Selectionlink2024NIPS/ICML/ICLR
From Quantity to Quality: Boosting LLM Performance with Self-Guided Data Selection for Instruction TuningIntroduce self - guided method for LLMs to select samples, key innovation IFD metric.Data Selectionlink2024*ACL
Rethinking the Instruction Quality: LIFT is What You NeedLIFT elevates instruction quality by broadening data distribution.Data Selectionlink2023arxiv
Instag:Instruction tagging for analyzing supervised fine-tuning of large language models.pdfPropose INSTAG to tag instructions, find benefits for LLMs, and a data sampling procedure.Data Selectionlink2024NIPS/ICML/ICLR
MoDS: Model-oriented Data Selection for Instruction TuningMoDS selects instruction data by quality, coverage and necessity.Data Selectionlink2023arxiv
SELF-INSTRUCT: Aligning Language Models with Self-Generated InstructionsSELF - INSTRUCT bootstraps from LM for instruction - following, nearly annotation - free.Data Selectionlink2023*ACL
Active Instruction Tuning: Improving Cross-Task Generalization by Training on Prompt Sensitive TasksPropose active IT based on prompt uncertainty to select tasks for LLM tuning.Data Selectionlink2023*ACL
Automated Data Curation for Robust Language Model Fine-TuningIntroduce CLEAR for data curation in LLM fine - tuning without extra computations.Data Selectionlink2024*ACL
CLUES: Collaborative Private-domain High-quality Data Selection for LLMs via Training DynamicsPropose data quality control via training dynamics for collaborative LLM training.Data Selectionlink2024NIPS/ICML/ICLR
Compute-Constrained Data SelectionFormalize data selection problem cost - aware, model trade - offs.Data Selectionlink2025NIPS/ICML/ICLR
DATA ADVISOR: Dynamic Data Curation for Safety Alignment of Large Language ModelsDATA ADVISOR for data generation to enhance LLM safety.Data Selectionlink2024*ACL
Data Curation Alone Can Stabilize In-context LearningTwo methods curate training data subsets to stabilize ICL without algorithm changes.Data Selectionlink2023*ACL
Get more for less: Principled Data Selection for Warming Up Fine-Tuning in LLMsSelect data to nudge pre - training dist. closer to target dist. for cost - effective fine - tuning.Data Selectionlink2024NIPS/ICML/ICLR
Improving Data Efficiency via Curating LLM-Driven Rating SystemsDS2 corrects LLM - based scores for data selection promoting diversity.Data Selectionlink2025NIPS/ICML/ICLR
LLM-Select: Feature Selection with Large Language ModelsLLMs can select predictive features without seeing training data.Data Selectionlink2024Journal
One-Shot Learning as Instruction Data Prospector for Large Language ModelsNUGGETS uses one - shot learning to select high - quality instruction data.Data Selectionlink2024*ACL
SAMPLE-EFFICIENT ALIGNMENT FOR LLMSIntroduce unified algorithm for LLM alignment based on Thompson sampling.Data Selectionlink2024arxiv
LESS: Selecting Influential Data for Targeted Instruction TuningPropose LESS to select data for targeted instruction tuning in LLMs.Data Selectionlink2024NIPS/ICML/ICLR
An Experimental Design Framework for Label-Efficient Supervised Finetuning of Large Language ModelsPropose experimental design for SFT in LLMs to mitigate annotation cost.Data Selectionlink2024*ACL
DELE: Data Efficient LLM EvaluationPropose adaptive sampling for LLM evaluation to reduce cost without losing integrity.Data Selectionlink2024NIPS/ICML/ICLR
Towards a Theoretical Understanding of Synthetic Data in LLM Post-Training: A Reverse-Bottleneck PerspectiveModel synthetic data gen process, relate generalization & info gain.Data Synthesislink2024arxiv
Advancing Theorem Proving in LLMs through Large-Scale Synthetic DataGenerate Lean 4 proof data to enhance LLM theorem - proving, without experimental focus.Data Synthesislink2024NIPS/ICML/ICLR
Are LLMs Naturally Good at Synthetic Tabular Data Generation?LLMs as-is or fine - tuned are bad at tabular data generation; permutation - aware can help.Data Synthesislink2024arxiv
Balancing Cost and Effectiveness of Synthetic Data Generation Strategies for LLMsGroup synthetic data strategies, study LLM training, propose selection framework.Data Synthesislink2024NIPS/ICML/ICLR
Best Practices and Lessons Learned on Synthetic Data for Language ModelsThe paper focuses on synthetic data for LMs, its use, challenges and responsible use.Data Synthesislink2024arxiv
ChatTS: Aligning Time Series with LLMs via Synthetic Data for Enhanced Understanding and ReasoningChatTS, a TS - MLLM, uses synthetic data for time series analysis.Data Synthesislink2024arxiv
Data extraction for evidence synthesis using a large language model: A proof-of-concept studyThe study assesses Claude 2's data extraction in evidence synthesis.Data Synthesislink2024Journal
Illuminating Blind Spots of Language Models with Targeted Agent-in-the-Loop Synthetic DataUse intelligent agents as teachers to generate samples for blind spot mitigation.Data Synthesislink2024arxiv
Generating Faithful Synthetic Data with Large Language Models: A Case Study in Computational Social ScienceThe paper studies strategies to increase synthetic data faithfulness.Data Synthesislink2023arxiv
Generative LLMs for Synthetic Data Generation: Methods, Challenges and the FutureThe paper focuses on using LLMs for synthetic data generation & related aspects.Data Synthesislink2023Journal
HARMONIC: Harnessing LLMs for Tabular Data Synthesis and Privacy ProtectionIntroduce HARMONIC for tabular data synth & privacy, use LLMs w/ fine - tuning.Data Synthesislink2024NIPS/ICML/ICLR
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with NothingMAGPIE self - synthesizes alignment data from aligned LLMs without human prompts.Data Synthesislink2024arxiv
Synthesizing Post-Training Data for LLMs through Multi-Agent SimulationMATRIX multi - agent simulator creates scenarios for data synthesis in LLM post - training.Data Synthesislink2025NIPS/ICML/ICLR
Synthetic Data Generation with Large Language Models for Text Classification: Potential and LimitationsExplore factors moderating LLM - generated data effectiveness in text classification.Data Synthesislink2023*ACL
Synthetic Oversampling: Theory and A Practical Approach Using LLMs to Address Data ImbalanceDevelop theoretical foundations for synthetic oversampling using LLMs.Data Synthesislink2024arxiv
Unveiling the Flaws: Exploring Imperfections in Synthetic Data and Mitigation Strategies for Large Language ModelsThis paper explores synthetic data flaws in LLM & presents a mitigation method.Data Synthesislink2024*ACL
Condor: Enhance LLM Alignment with Knowledge-Driven Data Synthesis and RefinementCondor generates high - quality SFT data with two - stage framework for LLMs.Data Synthesislink2025arxiv
Data Augmentation using LLMs: Data Perspectives, Learning Paradigms and ChallengesThe paper explores LLM - based data augmentation, challenges & learning paradigms.Data Augmentationlink2024*ACL
Data is all you need: Finetuning LLMs for Chip Design via an Automated design-data augmentation frameworkPropose an automated design - data augmentation framework for LLMs in chip design.Data Augmentationlink2024ACM
LLM-powered Data Augmentation for Enhanced Cross-lingual PerformanceUses LLMs for data augmentation in limited multilingual datasets.Data Augmentation, Surveylink2023*ACL
LLM-DA: Data Augmentation via Large Language Models for Few-Shot Named Entity RecognitionLLM - DA augments data at context/entity levels for few - shot NER.Data Augmentationlink2024arxiv
LLM-Generated Natural Language Meets Scaling Laws: New Explorations and Data Augmentation MethodsCalculates LLMNL and HNL by scaling laws, proposes ZGPTDA for data augmentation.Data Augmentationlink2024arxiv
A Survey on Data Augmentation in Large Model EraPaper reviews large - model - driven data aug. methods, applications & future challenges.Data Augmentationlink2024arxiv
ChatGPT Based Data Augmentation for Improved Parameter-Efficient Debiasing of LLMsUse ChatGPT to generate data for LLM debiasing with two strategies.Data Augmentationlink2024COLM
A Guide To Effectively Leveraging LLMs for Low-Resource Text Summarization: Data Augmentation and Semi-supervised ApproachesTwo new methods for low - resource text summarization are proposed.Data Augmentationlink2025*ACL
Empowering Large Language Models for Textual Data AugmentationPropose a solution to auto - generate LLM augmentation instructions for quality data.Data Augmentationlink2024*ACL
LLM-Generated Natural Language Meets Scaling Laws: New Explorations and Data Augmentation MethodsIntroduce scaling laws for LLMNL and HNL, a new data augmentation method ZGPTDA.Data Augmentationlink2024arxiv
LLM-AutoDA: Large Language Model-Driven Automatic Data Augmentation for Long-tailed ProblemsProposes LLM - AutoDA for long - tailed data augmentation by leveraging large - scale models.Data Augmentationlink2024NIPS/ICML/ICLR
Building a Family of Data Augmentation Models for Low-cost LLM Fine-tuning on the CloudPresent data augmentation models for low - cost LLM fine - tuning with key functionalities.Data Augmentationlink2025*ACL
Mini-DA: Improving Your Model Performance through Minimal Data Augmentation using LLMMini - DA selects challenging samples for augmentation, improving resource utilization.Data Augmentationlink2024*ACL
Data Augmentation for Text-based Person Retrieval Using Large Language ModelsPropose LLM - DA for TPR, use TFF & BSS to augment data concisely & efficiently.Data Augmentationlink2024*ACL
Data Augmentation for Cross-domain Parsing via Lightweight LLM Generation and Tree HybridizationPropose data augmentation via LLM & tree hybridization for cross - domain parsing.Data Augmentationlink2025*ACL
AugGPT: Leveraging ChatGPT for Text Data AugmentationPropose AugGPT for text data augmentation, rephrasing training samples.Data Augmentationlink2025IEEE
PGA-SciRE: Harnessing LLM on Data Augmentation for Enhancing Scientific Relation ExtractionPropose PGA framework for RE in scientific domain, two data aug. ways.Data Augmentationlink2024arxiv
Improving Topic Relevance Model by Mix-structured Summarization and LLM-based Data AugmentationUse query/doc summaries & LLM data augmentation for topic relevance modeling.Data Augmentationlink2024arxiv
Retrieval-Augmented Data Augmentation for Low-Resource Domain TasksPropose RADA framework to augment data for low - resource domain tasks.Data Augmentationlink2024arxiv
The Applicability of LLMs in Generating Textual Samples for Analysis of Imbalanced DatasetsThe paper compares approaches for handling text data class imbalance.Data Augmentationlink2024IEEE
Self-Rewarding Language ModelsStudy self - rewarding LMs, use LLM - as - a - Judge for self - rewards during training.Self Evolutionlink2024NIPS/ICML/ICLR
Self-Play Fine-Tuning Converts Weak Language Models to Strong Language ModelsPropose SPIN method for LLM, self - play mechanism refines its own capabilities.Self Evolutionlink2024NIPS/ICML/ICLR
Self-Boosting Large Language Models with Synthetic Preference DataSynPO self - boosts LLMs via synthetic preference data, eliminating large - scale annotation.Self Evolutionlink2024arxiv
MEMORYLLM: Towards Self-Updatable Large Language ModelsMEMORYLLM is self - updatable, can integrate new knowledge and retain long - term info.Self Evolutionlink2024NIPS/ICML/ICLR
Self-Refine: Iterative Refinement with Self-FeedbackSelf - Refine iteratively refines LLM outputs without extra training data or RL.Self Evolutionlink2023NIPS/ICML/ICLR
META-REWARDING LANGUAGE MODELS: Self-Improving Alignment with LLM-as-a-Meta-JudgeIntroduce Meta - Rewarding step for self - improving LLMs' judgment skills.Self Evolutionlink2024arxiv
Automated Proof Generation for Rust Code via Self-EvolutionSAFE framework enables Rust code proof generation via self - evolving cycle.Self Evolutionlink2025NIPS/ICML/ICLR
Arxiv Copilot: A Self-Evolving and Efficient LLM System for Personalized Academic AssistanceArxiv Copilot is a self - evolving LLM system for personalized academic assistance.Self Evolutionlink2024*ACL
Automatic programming via large language models with population self-evolution for dynamic job shop scheduling problemThis paper proposes SeEvo method for HDRs design inspired by experts' strategies.Self Evolutionlink2024arxiv
Benchmark Self-Evolving: A Multi-Agent Framework for Dynamic LLM EvaluationA multi - agent framework for dynamic LLM evaluation through instance reframing.Self Evolutionlink2025*ACL
Bias Amplification in Language Model Evolution: An Iterated Learning PerspectiveDraws parallels between LLM behavior & human culture evolution via Iterated Learning.Self Evolutionlink2024NIPS/ICML/ICLR
Enhanced Fine-Tuning of Lightweight Domain-Specific Q&A Model Based on Large Language ModelsPropose Self - Evolution framework for lightweight LLM fine - tuning.Self Evolutionlink2024IEEE
Interactive Evolution: A Neural-Symbolic Self-Training Framework For Large Language ModelsPropose ENVISIONS to self - train LLMs in neural - symbolic scenarios, overcoming two challenges.Self Evolutionlink2024arxiv
I-SHEEP: Self-Alignment of LLM from Scratch through an Iterative Self-Enhancement ParadigmI - SHEEP paradigm enables LLMs to self - improve iteratively in low - resource scenarios.Self Evolutionlink2024arxiv
Language Models as Continuous Self-Evolving Data EngineersPropose LANCE for LLMs to self - train by auto - data operations, reducing post - training cost.Self Evolutionlink2024arxiv
LLM Guided Evolution - The Automation of Models Advancing ModelsGE uses LLMs to directly modify code for model evolution.Self Evolutionlink2024arxiv
LLM-Evolve: Evaluation for LLM's Evolving Capability on BenchmarksProposes LLM - Evolve framework to evaluate LLMs' evolving ability on benchmarks.Self Evolutionlink2024*ACL
Long Term Memory : The Foundation of AI Self-EvolutionThis paper explores AI self - evolution with LTM, not on experimental performance.Self Evolutionlink2024arxiv
METEOR: Evolutionary Journey of Large Language Models from Guidance to Self-GrowthPropose Meteor method for model evolution with 3 training phases to maximize domain capabilities.Self Evolution, Distillationlink2024arxiv
Promptbreeder: Self-referential self-improvement via prompt evolutionPromptbreeder self - improves prompts via self - referential evolution.Self Evolutionlink2024NIPS/ICML/ICLR
rStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep ThinkingrStar - Math uses deep thinking via MCTS for SLMs to master math reasoning.Self Evolutionlink2025arxiv
Self: Language-driven self-evolution for large language modelSELF enables LLMs to self - evolve without human intervention via language feedback.Self Evolutionlink2024NIPS/ICML/ICLR
Self-Evolution Fine-Tuning for Policy OptimizationSEFT for policy optimization eliminates need for annotated samples.Self Evolutionlink2024*ACL
Self-Evolutionary Group-wise Log Parsing Based on Large Language ModelSelfLog self - evolves by LLM - extracted similar pairs and uses N - Gram - based methods.Self Evolutionlink2024IEEE
Self-Evolutionary Large Language Models through Uncertainty-Enhanced Preference OptimizationUPO framework mitigates noisy pref data for LLM self - evolution via reliable feedback.Self Evolutionlink2024arxiv
Self-Evolved Reward Learning for LLMsSelf - Evolved Reward Learning (SER) iteratively improves RM with self - generated data.Self Evolutionlink2025NIPS/ICML/ICLR
AugmenToxic: Leveraging Reinforcement Learning to Optimize LLM Instruction Fine-Tuning for Data Augmentation to Enhance Toxicity DetectionPropose RL - based method for LLM fine - tuning to augment toxic language data.Toxicity / Trust-worthylink2024ACM
Benchmarking LLMs in Political Content Text-Annotation: Proof-of-Concept with Toxicity and Incivility DataBenchmarked LLMs in political text -annotation, not focusing on exp. performance.Toxicity / Trust-worthylink2024arxiv
Can LLMs Recognize Toxicity? A Structured Investigation Framework and Toxicity MetricIntroduce LLM - based toxicity metric, analyze factors, evaluate its performance.Toxicity / Trust-worthylink2024*ACL
Characterizing Large Language Model Geometry Helps Solve Toxicity Detection and GenerationThe paper uses geometry to understand LLMs and solve toxicity - related issues.Toxicity / Trust-worthylink2024NIPS/ICML/ICLR
Detectors for Safe and Reliable LLMs: Implementations, Uses, and LimitationsPaper presents detectors library for LLM harms, uses & challenges, not exp perf.Toxicity / Trust-worthylink2024arxiv
Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMsThis paper creates an open - source dataset to evaluate LLM safeguards.Toxicity / Trust-worthylink2023arxiv
Effcient Toxic Content Detection by Bootstrapping and Distilling Large Language ModelsBD - LLM bootstraps & distills LLMs for toxic content detection via DToT.Toxicity / Trust-worthylink2024AAAI/IJCAL
Evaluating the Impact of Model Size on Toxicity and Stereotyping in Generative LLMExplore LLM size's relation to toxicity & stereotyping, smallest model performs best.Toxicity / Trust-worthylink2023Journal
How Toxic Can You Get? Search-based Toxicity Testing for Large Language ModelsEvoTox tests LLM toxicity post - alignment via iterative evolution strategy.Toxicity / Trust-worthylink2025arxiv
Improving Covert Toxicity Detection by Retrieving and Generating ReferencesThis paper explores refs' potential for covert toxicity detection.Toxicity / Trust-worthylink2024*ACL
Leak, Cheat, Repeat: Data Contamination and Evaluation Malpractices in Closed-Source LLMsThe paper analyzes data contamination & eval malpractices in closed - source LLMs.Toxicity / Trust-worthylink2024*ACL
LLM-Based Synthetic Datasets: Applications and Limitations in Toxicity DetectionThe paper explores LLM - based synthetic data in toxicity detection, its potential and limits.Toxicity / Trust-worthylink2024*ACL
Mitigating Biases to Embrace Diversity: A Comprehensive Annotation Benchmark for Toxic LanguageNew annotation benchmark reduces bias, shows LLM annotation value.Toxicity / Trust-worthylink2024*ACL
People Make Better Edits: Measuring the Efficacy of LLM-Generated Counterfactually Augmented Data for Harmful Language DetectionAssess if CAD generation for harmful lang. detection can be automated using NLP models.Toxicity / Trust-worthylink2023*ACL
Realistic Evaluation of Toxicity in Large Language ModelsNew TET dataset helps rigorously evaluate toxicity in popular LLMs.Toxicity / Trust-worthylink2024*ACL
TOXICCHAT: Unveiling Hidden Challenges of Toxicity Detection in Real-World User-AI ConversationThis paper isn't about Efficient LLM Post Training, so can't provide relevant summary.Toxicity / Trust-worthylink2023*ACL
Toxicity Detection with Generative Prompt-based InferenceExplore generative zero - shot prompt - based toxicity detection.Toxicity / Trust-worthylink2022arxiv
Toxicity in CHATGPT: Analyzing Persona-assigned Language ModelsThe paper evaluates ChatGPT toxicity based on persona - assigned language models.Toxicity / Trust-worthylink2023*ACL
ToxiCraft:A Novel Framework for Synthetic Generation of Harmful InformationThe paper proposes ToxiCraft to generate harmful info datasets, addressing two issues.Toxicity / Trust-worthylink2024*ACL
TOXIGEN: A Large-Scale Machine-Generated Dataset for Adversarial and Implicit Hate Speech DetectionCreate TOXIGEN dataset, new method for generating text, human evaluation.Toxicity / Trust-worthylink2022arxiv
Dialectal Toxicity Detection: Evaluating LLM-as-a-Judge Consistency Across Language VarietiesThis paper focuses on dialectal toxicity detection in LLMs, not relevant to efficient post - training.Toxicity / Trust-worthy, LLM-as-Judgerlink2024arxiv
Do-Not-Answer: Evaluating Safeguards in LLMsThe paper curates a dataset to evaluate LLM safeguards for safer deployment.Toxicity / Trust-worthylink2024*ACL
An Empirical Study of LLM-as-a-Judge for LLM Evaluation: Fine-tuned Judge Model is not a General Substitute for GPT-4Fine - tuned judge models have limitations, integrated method improves them.LLM-as-Judgerlink2024*ACL
CalibraEval: Calibrating Prediction Distribution to Mitigate Selection Bias in LLMs-as-JudgesCalibraEval mitigates LLM - as - Judges selection bias via NOA.LLM-as-Judgerlink2024arxiv
Can LLMs be Good Graph Judger for Knowledge Graph Construction?The paper proposes GraphJudger to address KG construction challenges.LLM-as-Judgerlink2024arxiv
CodeUltraFeedback: An LLM-as-a-Judge Dataset for Aligning Large Language Models to Coding PreferencesPropose LLM - as - a - Judge methodology for evaluating LLM coding preference alignment.LLM-as-Judgerlink2024arxiv
Crowd score: A method for the evaluation of jokes using large language model AI voters as judgesCrowd Score method assesses joke funniness via LLMs as AI judges.LLM-as-Judgerlink2022arxiv
Foundational Autoraters: Taming Large Language Models for Better Automatic EvaluationIntroduce FLAMe, trained on quality tasks, less biased than other LLM - as - a - Judge models.LLM-as-Judgerlink2024*ACL
Judging LLM-as-a-Judge with MT-Bench and Chatbot ArenaUse LLM - as - a - judge to evaluate chat assistants, verify with two benchmarks.LLM-as-Judgerlink2023NIPS/ICML/ICLR
Judgelm: Fine-tuned large language models are scalable judgesFine - tune LLMs as scalable judges, propose dataset & techniques.LLM-as-Judgerlink2023arxiv
Judging the Judges: Evaluating Alignment and Vulnerabilities in LLMs-as-JudgesThe paper studies LLM - as - judges, judges' performance and vulnerabilities.LLM-as-Judgerlink2024arxiv
Large Language Models are Inconsistent and Biased EvaluatorsLLMs are inconsistent/biased evaluators; recipes to mitigate limitations are shared.LLM-as-Judgerlink2024arxiv
Llm-as-a-judge & reward model- What they can and cannot doAnalysis of automated evaluators: English eval & limitations.LLM-as-Judgerlink2024arxiv
LLMs instead of Human Judges? A Large Scale Empirical Study across 20 NLP Evaluation TasksEvaluated 11 LLMs on 20 datasets; LLMs need human - validation before use as evaluators.LLM-as-Judgerlink2024arxiv
Meta-rewarding language models: Self-improving alignment with llm-as-a-meta-judgeIntroduce Meta - Rewarding step to self - improve LLM's judgment skills.LLM-as-Judgerlink2024arxiv
MLLM-as-a-Judge: Assessing Multimodal LLM-as-a-Judge with Vision-Language BenchmarkThis paper introduces MLLM - as - a - Judge benchmark to assess MLLMs' judging ability.LLM-as-Judgerlink2024NIPS/ICML/ICLR
R-Judge: Benchmarking Safety Risk Awareness for LLM AgentsR - Judge benchmarks LLM agents' safety risk awareness in interactions.LLM-as-Judgerlink2024arxiv
Self-Taught EvaluatorsAn approach improves evaluators using only synthetic training data.LLM-as-Judgerlink2024arxiv
Style Over Substance: Evaluation Biases for Large Language ModelsStudy shows evaluation bias for LLMs, proposes MERS to improve LLM - based evaluations.LLM-as-Judgerlink2025*ACL
Wider and Deeper LLM Networks are Fairer LLM EvaluatorsThe paper uses wider & deeper LLM networks for fairer LLM evaluation.LLM-as-Judgerlink2023arxiv
Internal Consistency and Self-Feedback in Large Language Models: A SurveyThis paper uses internal consistency perspective to explain LLM issues and introduce Self - Feedback.Surveylink2024arxiv
A Survey on Self-Evolution of Large Language ModelsThe paper surveys self - evolution in LLMs, including its process and challenges.Survey, Self Evolutionlink2024arxiv
Automatically Correcting Large Language Models: Surveying the Landscape of Diverse Automated Correction StrategiesReviews advances in auto - correcting LLMs via feedback, categorizes approaches.Surveylink2024Journal
A Survey on Data Selection for LLM Instruction TuningThis paper surveys data selection for LLM instruction tuning.Survey, Data Selectionlink2024arxiv
Large Language Models for Data Annotation and Synthesis: A SurveyThis paper focuses on LLM post - training from a data - centric view.Survey, Data Synthesislink2024*ACL
On LLMs-Driven Synthetic Data Generation, Curation, and Evaluation: A SurveyThe paper organizes LLMs - driven data gen. studies to show research gaps and future ways.Surveylink2024*ACL
Trustworthy LLMs: A survey and guideline for evaluating large language models' alignmentThe paper surveys LLM trustworthiness dimensions for alignment evaluation.Survey, Toxicity / Trust-worthylink2024NIPS/ICML/ICLR
A Survey on Data Selection for Language ModelsComprehensive review of data selection for LMs to accelerate related research.Survey, Data Selectionlink2024Journal
LLMs-as-Judges: A Comprehensive Survey on LLM-based Evaluation MethodsI'm sorry, but the given data is about "LLMs - as - Judges" not "Efficient LLM Post Training: A Data - centric Perspective", so I can't provide a relevant summary.Survey, LLM-as-Judgerlink2024arxiv
A Survey on Data Synthesis and Augmentation for Large Language ModelsReviews LLM data generation techniques, discusses constraints.Survey, Data Synthesis, Data Augmentationlink2024arxiv
A Survey on Knowledge Distillation of Large Language ModelsComprehensive survey on KD in LLMs: mechanisms, skills, verticalization & DA interplay.Survey, Distillationlink2024arxiv
Survey on Knowledge Distillation for Large Language Models: Methods, Evaluation, and ApplicationSurvey on LLM knowledge distillation methods, evaluation & application, not exp perf.Survey, Distillationlink2024ACM
Impossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and ParaphrasingImpossible Distillation: distill high - quality from low - quality for summarization & paraphrasing.Distillationlink2023arxiv
Prompt Distillation for Efficient LLM-based RecommendationPropose prompt distillation to bridge IDs & words & reduce inference time.Distillationlink2023ACM
Performance-Guided LLM Knowledge Distillation for Efficient Text Classification at ScalePGKD for text classification, an LLM distillation method with versatile framework.Distillationlink2024*ACL
Knowledge Distillation in Automated Annotation: Supervised Text Classification with LLM-Generated Training LabelsThe paper tests LLM - generated labels for supervised text classification workflows.Distillationlink2024*ACL
Multistage Collaborative Knowledge Distillation from a Large Language Model for Semi-Supervised Sequence GenerationPropose MCKD for semi - supervised seq. gen., iteratively improve pseudolabels.Distillationlink2024*ACL
Self-Data Distillation for Recovering Quality in Pruned Large Language ModelsSelf - data distillation fine - tuning mitigates quality loss from pruning and SFT.Distillationlink2024NIPS/ICML/ICLR
Distillation Matters: Empowering Sequential Recommenders to Match the Performance of Large Language ModelsProposes DLLM2Rec for LLM-based rec. model distillation to sequential models.Distillationlink2024ACM
Towards Cross-Tokenizer Distillation: the Universal Logit Distillation Loss for LLMsIntroduce ULD loss for cross - tokenizer distillation in LLMs.Distillationlink2025Journal
Self-Evolution Knowledge Distillation for LLM-based Machine TranslationSelf - Evolution KD dynamically integrates prior knowledge for better knowledge transfer.Distillation, Self Evolutionlink2025*ACL
Efficiently Distilling LLMs for Edge ApplicationsPropose MLFS for parameter - efficient supernet training of LLMs.Distillationlink2024*ACL
Xai-driven knowledge distillation of large language models for efficient deployment on low-resource devicesDiXtill uses XAI to distill LLM knowledge into a self - explainable student model.Distillationlink2024Journal
Compact Language Models via Pruning and Knowledge DistillationDevelop compression practices for LLMs via pruning and distillation.Distillationlink2024NIPS/ICML/ICLR
LLM-Enhanced Multi-Teacher Knowledge Distillation for Modality-Incomplete Emotion Recognition in Daily HealthcarePropose LLM - enhanced multi - teacher KD for emotion rec in modality - incomplete cases.Distillationlink2024IEEE
BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-DistillationBitDistiller combines QAT and KD for sub - 4 - bit LLMs with new techniques.Distillationlink2024*ACL
Reducing LLM Hallucination Using Knowledge Distillation: A Case Study with Mistral Large and MMLU BenchmarkKnowledge distillation reduces LLM hallucination via specific methods.Distillationlink2024arxiv
Distilling Large Language Models for Text-Attributed Graph LearningPropose distilling LLMs into local graph model for TAG learning, novel training method.Distillationlink2024ACM
CourseGPT-zh: an Educational Large Language Model Based on Knowledge Distillation Incorporating Prompt OptimizationCourseGPT - zh uses prompt optimization in a distillation framework for educational LLM.Distillationlink2024arxiv
LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt CompressionPropose data distillation for prompt compression, formulate as token classification.Distillationlink2024*ACL
LLM for Patient-Trial Matching: Privacy-Aware Data Augmentation Towards Better Performance and GeneralizabilityPropose LLM - PTM for patient - trial match, ensure data privacy in methodology.Applicationslink2023Others
LLM-Assisted Data Augmentation for Chinese Dialogue-Level Dependency ParsingPresent 3 LLM - based strategies for Chinese dialogue - level dependency parsing.Applicationslink2024Others
Resolving the Imbalance Issue in Hierarchical Disciplinary Topic Inference via LLM-based Data AugmentationUse Llama V1 to augment data for balancing disciplinary topic inference.Applicationslink2023IEEE
LLM-based Privacy Data Augmentation Guided by Knowledge Distillation with a Distribution Tutor for Medical Text ClassificationPropose a DP - based DA method for text classification in private domains.Applicationslink2024Others
Large Language Models for Healthcare Data Augmentation: An Example on Patient-Trial MatchingAn LLM - based patient - trial matching approach with privacy - aware data augmentation.Applicationslink2024Others
Identifying Citizen-Related Issues from Social Media Using LLM-Based Data AugmentationPropose LLM - based method for data augmentation to extract citizen - related data from tweets.Applications, Data Augmentationlink2024Others
Synthetic Data Augmentation Using Large Language Models (LLM): A Case-Study of the Kamyr DigesterIntroduces LLM - based data augmentation technique for data scarcity.Applicationslink2024IEEE
Conditional Label Smoothing For LLM-Based Data Augmentation in Medical Text ClassificationPropose CLS for data augmentation in medical text classification.Applicationslink2024IEEE
Curriculum-style Data Augmentation for LLM-based Metaphor DetectionPropose open - source LLM fine - tuning and CDA for metaphor detection.Applications, Data Augmentationlink2024arxiv
Enhancing Speech De-Identification with LLM-Based Data AugmentationA novel data augmentation method for speech de - id using LLM and end - to - end model.Applicationslink2024IEEE
Enhancing Multilingual Fake News Detection through LLM-Based Data AugmentationUse Llama 3 via LLM - based data augmentation to enrich fake news datasets.Applicationslink2024Others
LLMs Accelerate Annotation for Medical Information ExtractionPropose LLM - human combo for medical text annotation, reducing human burden.Applications, Active Annotationlink2023Others
Crowdsourcing with Enhanced Data Quality Assurance: An Efficient Approach to Mitigate Resource Scarcity Challenges in Training Large Language Models for HealthcarePropose CS framework with quality control for LLM in healthcare, address resource scarcity.Applicationslink2024Others
LLM2LLM: Boosting LLMs with Novel Iterative Data EnhancementLLM2LLM iteratively augments data for LLM fine - tuning in low - data scenarios.Data Quality Enhancement, Data Augmentationlink2024*ACL
Data Quality Enhancement on the Basis of Diversity with Large Language Models for Text Classification: Uncovered, Difficult, and NoisyPropose DQE method for text classification with LLMs, select data by greedy algorithm.Data Quality Enhancementlink2025*ACL
Multi-News+: Cost-efficient Dataset Cleansing via LLM-based Data AnnotationUse LLM for data cleansing in Multi - News dataset, no need for costly human annotators.Data Quality Enhancementlink2024*ACL
LLM-Enhanced Data ManagementLLMDB for data management: avoid hallucination, reduce cost, improve accuracy.Data Quality Enhancementlink2024ACM
Enhancing LLM Fine-tuning for Text-to-SQLs by SQL Quality MeasurementPropose using SQL Quality Measurement to enhance LLM-based Text - to - SQLs performance.Data Quality Enhancementlink2024arxiv
On The Role of Prompt Construction In Enhancing Efficacy and Efficiency of LLM-Based Tabular Data GenerationEnriching prompts with domain insights improves LLM-based tabular data generation.Data Quality Enhancementlink2024arxiv
On LLM-Enhanced Mixed-Type Data Imputation with High-Order Message PassingPropose UnIMP with BiHMP and Xfusion for mixed - type data imputation.Data Quality Enhancementlink2025arxiv
SEMIEVOL: Semi-supervised Fine-tuning for LLM AdaptationSEMIEVOL, a semi - supervised LLM fine - tuning framework, propagates and selects knowledge.Data Curationlink2024arxiv
Curated LLM: Synergy of LLMs and Data Curation for tabular augmentation in low-data regimesIntroduce CLLM for tabular augmentation in low - data, with curation mechanism for data.Data Curationlink2024NIPS/ICML/ICLR
Data to Defense: The Role of Curation in Customizing LLMs Against Jailbreaking AttacksPropose data curation approach & mitigation framework to counter jailbreaking.Data Curationlink2024arxiv
DATA ADVISOR: Dynamic Data Curation for Safety Alignment of Large Language ModelsPropose Data Advisor for data gen. considering dataset char. to enhance quality.Data Curationlink2024*ACL
Data Curation Alone Can Stabilize In-context LearningTwo methods curate data subsets to stabilize ICL without algorithm changes.Data Curationlink2023*ACL
Automated Data Curation for Robust Language Model Fine-TuningIntroduced CLEAR for instruction tuning datasets to curate data without extra computations.Data Curationlink2024*ACL
Improving Data Efficiency via Curating LLM-Driven Rating SystemsDS2, a data selection method, corrects LLM scores and promotes data sample diversity.Data Curation, Data Selectionlink2025NIPS/ICML/ICLR
The RefinedWeb Dataset for Falcon LLM: Outperforming Curated Corpora with Web Data OnlyShow web data alone can lead to powerful models without curated data.Data Curationlink2023NIPS/ICML/ICLR
Use of a Structured Knowledge Base Enhances Metadata Curation by Large Language ModelsLLMs can improve metadata curation with a structured knowledge base.Data Curationlink2024arxiv
Source2Synth: Synthetic Data Generation and Curation Grounded in Real Data SourcesSource2Synth generates synth data from real sources without human annotations.Data Curation, Data Synthesislink2024arxiv
AutoDCWorkflow: LLM-based Data Cleaning Workflow Auto-Generation and BenchmarkInvestigated LLM's data - cleaning workflow auto - gen, proposed a benchmark.Data Curationlink2024arxiv
Dynosaur: A Dynamic Growth Paradigm for Instruction-Tuning Data CurationDynosaur automatically constructs instruction adjustment data and reduces costs by leveraging existing datasets.Data Curationlink2023*ACL
AutoPureData: Automated Filtering of Web Data for LLM Fine-tuningProposes system to auto - filter web data for LLM training with trusted AI models.Data Curationlink2024arxiv
Automatic Dataset Construction (ADC): Sample Collection, Data Curation, and BeyondPropose ADC for efficient dataset construction, offer benchmarks.Data Curationlink2024arxiv
Diversify and Conquer: Diversity-Centric Data Selection with Iterative RefinementProposes k - means & iterative refinement for data selection to finetune LLMs.Data Curationlink2025NIPS/ICML/ICLR
Increasing Diversity While Maintaining Accuracy: Text Data Generation with Large Language Models and Human InterventionsExplore human - AI partnerships for high - quality LLM - based text data generation.Data Curationlink2023*ACL
Balancing performance and cost of LLMs in a multi-agent framework for BIM data retrievalPropose MAS method to match queries with LLMs for balanced BIM data retrieval.Data Curation, Applicationslink2025Others
Optima: Optimizing Effectiveness and Efficiency for LLM-Based Multi-Agent SystemOptima framework in LLM - based MAS improves communication & task effectiveness via LLM training.Data Curationlink2025NIPS/ICML/ICLR
Synergized Data Efficiency and Compression (SEC) Optimization for Large Language ModelsPropose SEC for LLMs to enhance efficiency without sacrificing performance.Data Curationlink2024Others
LLMaAA: Making Large Language Models as Active AnnotatorsLLMaAA uses LLMs as annotators in active learning loop, optimizing annotation & training.Active Annotationlink2023*ACL
Enhancing Review Classification Via Llm-Based Data Annotation and Multi-Perspective Feature Representation LearningPropose MJAR dataset & MPFR approach for review classification.Active Annotationlink2024Others
AutoLabel: Automated Textual Data Annotation Method Based on Active Learning and Large Language ModelAutoLabel uses LLM & active learning to assist text data annotation.Active Annotation, Data Quality Enhancementlink2024Others
Human-LLM Collaborative Annotation Through Effective Verification of LLM LabelsA multi - step human - LLM collaborative approach for accurate annotations.Active Annotationlink2024ACM
PDFChatAnnotator: A Human-LLM Collaborative Multi-Modal Data Annotation Tool for PDF-Format CatalogsPDFChatAnnotator links data & extracts info, user can guide LLM annotations.Active Annotation, Applicationslink2024ACM
Selective Annotation via Data Allocation: These Data Should Be Triaged to Experts for Annotation Rather Than the ModelPropose SANT for selective annotation, allocating data to expert & model effectively.Active Annotationlink2024*ACL
Entity Alignment with Noisy Annotations from Large Language ModelsPropose LLM4EA framework for entity alignment with reduced annotation space and label refiner.Active Annotationlink2024NIPS/ICML/ICLR
CoAnnotating: Uncertainty-Guided Work Allocation between Human and Large Language Models for Data AnnotationThe paper proposes CoAnnotating for human - LLM co - annotation using uncertainty.Active Annotationlink2023*ACL
Code Less, Align More: Efficient LLM Fine-tuning for Code Generation with Data PruningPresent techniques to enhance code LLM training efficiency with data pruning.Data Pruninglink2024*ACL
Data-efficient Fine-tuning for LLM-based RecommendationPropose a data pruning method with two scores for efficient LLM - based recommendation.Data Pruninglink2024ACM
LLM-Pruner: On the Structural Pruning of Large Language ModelsLLM - Pruner compresses LLMs task - agnostically via structural pruning.Data Pruninglink2023NIPS/ICML/ICLR
Pruning as a Domain-specific LLM ExtractorIntroduce D - Pruner for domain - specific LLM compression by dual - pruning.Data Pruninglink2024*ACL
Measuring Sample Importance in Data Pruning for Language Models based on Information EntropyRank training samples by informativeness via entropy for data - pruning of LLMs.Data Pruninglink2024arxiv
P3: A Policy-Driven, Pace-Adaptive, and Diversity-Promoted Framework for data pruning in LLM TrainingP3 optimizes LLM fine - tuning via iterative data pruning with 3 key components.Data Pruninglink2024NIPS/ICML/ICLR
All-in-One Tuning and Structural Pruning for Domain-Specific LLMsATP is a unified approach to pruning & fine - tuning LLMs via a trainable generator.Data Pruninglink2024arxiv
Language Model-Driven Data Pruning Enables Efficient Active LearningActivePrune, a novel pruning strategy for AL, uses LMs to prune unlabeled data.Data Pruninglink2025NIPS/ICML/ICLR
Compresso: Structured Pruning with Collaborative Prompting Learns Compact Large Language ModelsCompresso: Structured Pruning via algo - LLM collaboration, uses LoRA & prompt.Data Pruninglink2024NIPS/ICML/ICLR
Efficient LLM Pruning with Global Token-Dependency Awareness and Hardware-Adapted InferencePropose VIB - based pruning method, post - pruning for LLMs to compress & speed up.Data Pruninglink2024Others
SlimGPT: Layer-wise Structured Pruning for Large Language ModelsSlimGPT, a fast LLM pruning method, uses strategies for near - optimal results.Data Pruninglink2024NIPS/ICML/ICLR
Shortened LLaMA: A Simple Depth Pruning for Large Language ModelsSimple depth pruning can compete with width pruning in zero - shot LLM task.Data Pruninglink2024NIPS/ICML/ICLR
Fewer is More: Boosting LLM Reasoning with Reinforced Context PruningCoT - Influx maximizes concise CoT examples input to boost LLM math reasoning.Data Pruninglink2024*ACL

🤗 Welcome to contribute to this repo! You can create a pull request or email me atluo.junyu@outlook.com.

About

A list of data-efficient and data-centric LLM (Large Language Model) papers. Our Survey Paper: Towards Efficient LLM Post Training: A Data-centric Perspective

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

[8]ページ先頭

©2009-2025 Movatter.jp