Skip to content

Latest commit

 

History

History
778 lines (589 loc) · 45.3 KB

AL_combinations.md

File metadata and controls

778 lines (589 loc) · 45.3 KB

AL with other AI Research Problems

Active learning is also been used with other Learning/Research paradigms. These works could be considered as the applications of AL with other AI research problems. For the information of the real applications of AL in science and industry, please check AL Applications.

(Note that the works list in this pages are works I browsed in background reading. There might be more important works in the corresponding fields not listed here. Besides, I can't ensure the works are representative in the fields. So if you have any comments and recommendations, pls let me know.)

Reducing the labeling cost is a common need in many research fields.

Popular Fields

Computer Vision (CV)

CV is quite a wide conception. Here we only post several subtypes in the fields.

Survey for the whole field:

  • Deep Active Learning for Computer Vision: Past and Future [2022]

Image classification:

Image Semantic Segmentation:

Semantic Segmentation with Domain Adaptation:

  • Multi-Anchor Active Domain Adaptation for Semantic Segmentation [2021, ICCV]: One time selection, not in a conventional loop manner.
  • ADeADA: Adaptive Density-aware Active Domain Adaptation for Semantic Segmentation [2022]: Acquire labels of the samples with high probability density in the target domain yet with low probability density in the source domain
  • Active Domain Adaptation with Multi-level Contrastive Units for Semantic Segmentation [2022]
  • Labeling Where Adapting Fails: Cross-Domain Semantic Segmentation with Point Supervision via Active Selection [2022]: Pixel-level selections.
  • Pixel Exclusion: Uncertainty-aware Boundary Discovery for Active Cross-Domain Semantic Segmentation [2022, MM]
  • MADAv2: Advanced Multi-Anchor Based Active Domain Adaptation Segmentation [2023]
  • Class Balanced Dynamic Acquisition for Domain Adaptive Semantic Segmentation using Active Learning [2023, NeurIPS Workshop ReALML]
  • Prototype Guided Pseudo Labeling and Perturbation-based Active Learning for domain adaptive semantic segmentation [2024, Pattern Recognition]

Object Detection:

Point Cloud Semantic Segmentation:

  • Label-Efficient Point Cloud Semantic Segmentation: An Active Learning Approach [2021, CVPR]
  • ReDAL: Region-based and Diversity-aware Active Learning for Point Cloud Semantic Segmentation [2021, ICCV]
  • Active Learning for Point Cloud Semantic Segmentation via Spatial-Structural Diversity Reasoning [2022]
  • You Never Get a Second Chance To Make a Good First Impression: Seeding Active Learning for 3D Semantic Segmentation [2023]
  • A multi-granularity semisupervised active learning for point cloud semantic segmentation [2023, NCA]
  • Hierarchical Point-based Active Learning for Semi-supervised Point Cloud Semantic Segmentation [2023, ICCV]

Image Captioning:

  • Structural Semantic Adversarial Active Learning for Image Captioning [2020, ACMMM]

Action Recognition:

Video Object Detection:

Video Captioning:

  • MAViC: Multimodal Active Learning for Video Captioning [2022]

Video Action Spotting:

  • Towards Active Learning for Action Spotting in Association Football Videos [2023]

Video Action Recognition:

  • Active Exploration of Multimodal Complementarity for Few-Shot Action Recognition [2023, CVPR]

Video Dataset:

  • FOCAL: A Cost-Aware, Video Dataset for Active Learning [IEEE BigData, 2023]

3D Object Detection:

  • Exploring Diversity-based Active Learning for 3D Object Detection in Autonomous Driving [2022]
  • Exploring Active 3D Object Detection from a Generalization Perspective [2023, ICLR]
  • Bi3D: Bi-domain Active Learning for Cross-domain 3D Object Detection [2023, CVPR]

3D Semantic Segmentation:

  • LiDAL: Inter-frame Uncertainty Based Active Learning for 3D LiDAR Semantic Segmentation [2022, ECCV]
  • Efficient 3D Scene Semantic Segmentation via Active Learning on Rendered 2D Images [2023, TIP]

Visual Tracking:

  • Active Learning for Deep Visual Tracking [2021]
  • Pseudo Loss Active Learning for Deep Visual Tracking [2022, Pattern Recognition]
  • Heterogeneous Diversity Driven Active Learning for Multi-Object Tracking [2023, ICCV]

Visual Question Answering:

  • Mind Your Outliers! Investigating the Negative Impact of Outliers on Active Learning for Visual Question Answering [2021]
  • Single-Modal Entropy based Active Learning for Visual Question Answering [2021]

Vision Language Model (VLM):

  • Active Prompt Learning in Vision Language Models [2023]

Pose Estimation:

  • Deep Active Learning For Human Pose Estimation Via Consistency Weighted Core-Set Approach [2021, ICIP]
  • Active Learning with Pseudo-Labels for Multi-View 3D Pose Estimation [2021]
  • Meta Agent Teaming Active Learning for Pose Estimation [2022, CVPR]
  • VL4Pose: Active Learning Through Out-Of-Distribution Detection For Pose Estimation [2022]
  • Rethinking the Data Annotation Process for Multi-view 3D Pose Estimation with Active Learning and Self-Training [2023, WACV]

Optical Flow Prediction:

  • Optical Flow Training under Limited Label Budget via Active Learning [2022]

Emotion Classification:

In this field, the acquired labels could be highly subjective.

  • An Exploration of Active Learning for Affective Digital Phenotyping [2022]

Micro-Expression Recognition:

  • Tackling Micro-Expression Data Shortage via Dataset Alignment and Active Learning [2022, IEEE Trans Multimedia]

Collision Prediction & Handling:

  • N-Penetrate: Active Learning of Neural Collision Handler for Complex 3D Mesh Deformations [2022, ICML]

Super-resolution:

  • USIM-DAL: Uncertainty-aware Statistical Image Modeling-based Dense Active Learning for Super-resolution [2023, UAI]

Natural Language Processing (NLP)

NLP is also quite a wide conception. Here we only post several subtypes in the fields.

Survey for the whole field:

Sentiment/text classification:

  • Active learning for imbalanced sentiment classification [2012, EMNLP-CoNLL]
  • Deep Active Learning with Simulated Rationales for Text Classification [2020, PRAI]: Utilize auxiliary task to reduce the data scarce affect at the beginning of the AL process.
  • Active Learning via Membership Query Synthesis for Semi-supervised Sentence Classification [2019, CoNLL]
  • Active Learning for BERT: An Empirical Study [2020, EMNLP]
  • Deep Active Learning for Text Classification with Diverse Interpretations [2021, CIKM]
  • Revisiting Uncertainty-based Query Strategies for Active Learning with Transformers [2022, ACL Findings]: Fine-tune the pretrained transformer based models.
  • Active Learning on Pre-trained Language Model with Task-Independent Triplet Loss [2022, AAAI]

Named entity recognition (Information Extraction):

  • Learning How to Actively Learn: A Deep Imitation Learning Approach [2018, ACL]
  • Deep active learning for named entity recognition [ICLR, 2018]: Incremental manner. Uncertainty-based heuristic, select those sentences for which the length-normalized log probability of the current prediction is the lowest.
  • Overcoming Practical Issues of Deep Active Learning and its Applications on Named Entity Recognition [Arxiv, 2019]
  • LTP: A New Active Learning Strategy for CRF-Based Named Entity Recognition [2020, Arxiv]
  • Subsequence Based Deep Active Learning for Named Entity Recognition [2021]
  • Deep Active Learning for Swedish Named Entity Recognition [2021, Master Thesis]
  • FAMIE: A Fast Active Learning Framework for Multilingual Information Extraction [2022]
  • Subsequence Based Deep Active Learning for Named Entity Recognition [2022, ACL/IJCNLP]
  • Active Learning for Name Entity Recognition with External Knowledge [2023, ACM Trans. Asian Low-Resour. Lang. Inf. Process.]
  • EASAL: Entity-Aware Subsequence-Based Active Learning for Named Entity Recognition [2023, AAAI]

Parsing:

Classifier Pruning:

Neural Machine Translation:

Sequence Tagging:

Sequence Generation:

Coreference Resolution:

Learning with Pre-Trained Model:

  • Active Learning for Effectively Fine-Tuning Transfer Learning to Downstream Task
  • Bayesian Active Learning with Pretrained Language Models
  • Multi-class Text Classification using BERT-based Active Learning
  • On the Importance of Effectively Adapting Pretrained Language Models for Active Learning [2022, ACL]
  • ACTUNE: Uncertainty-Based Active Self-Training for Active Fine-Tuning of Pretrained Language Models [2022, NAACL-HLT]: AL + SelfSL.
  • Smooth Sailing: Improving Active Learning for Pre-trained Language Models with Representation Smoothness Analysis [2022]
  • Low-resource Interactive Active Labeling for Fine-tuning Language Models [2022, EMNLP]
  • No Labels? No Problem! Experiments with active learning strategies for multi-class classification in imbalanced low-resource settings [2023, ICAIL]
  • Combining Active Learning and Task Adaptation with BERT for Cost-Effective Annotation of Social Media Datasets [2023, ACL workshop]

Question Answering:

  • Improving Question Answering Performance Using Knowledge Distillation and Active Learning [2021]

Event Extraction (Information Extraction):

  • Active Learning for Event Extraction with Memory-based Loss Prediction Model [2021]

Rational Learning:

  • A Rationale-Centric Framework for Human-in-the-loop Machine Learning [2022]
  • Active Learning on Pre-trained Language Model with Task-Independent Triplet Loss [2022]

Open Relation Extraction:

  • Active Relation Discovery: Towards General and Label-aware Open Relation Extraction [2021]
  • Actively Supervised Clustering for Open Relation Extraction [2023]

Argument Structure Extraction:

  • Efficient Argument Structure Extraction with Transfer Learning and Active Learning [2022]

Claim Verification:

  • Active PETs: Active Data Annotation Prioritisation for Few-Shot Claim Verification with Pattern Exploiting Training [2022]

Abuse Detection:

  • Is More Data Better? Re-thinking the Importance of Efficiency in Abusive Language Detection with Transformers-Based Active Learning [2022, COLING workshop TRAC]

Abstractive Text Summarization:

  • Active Learning for Abstractive Text Summarization [2023]

Natural Language Explanations:

  • Beyond Labels: Empowering Human with Natural Language Explanations through a Novel Active-Learning Architecture [2023]

n-ary relation extraction:

  • Active Learning for Cross-Sentence n-ary Relation Extraction [2023, Information Sciences]

Morphophonological Processing:

  • Deep Active Learning for Morphophonological Processing [2023, ACL]

Domain Adaptation/Transfer Learning

Normally when we use AL in domain adaptation, we can obtain several true labels of the unlabeled instances on source/target domain. And for transfer learning, the concept is even wider. Here we refer active transfer learning to any forms of transfer, eg. data or model.

Domain Adaptation:

  • Active learning for cross-domain sentiment classification [2013, IJCAI]
  • Transfer Learning with Active Queries from Source Domain [2016, IJCAI]
  • Active Sentiment Domain Adaptation [2017, ACL]
  • On Gleaning Knowledge from Multiple Domains for Active Learning [2017, IJCAI]
  • Active Adversarial Domain Adaptation [2020, WACV]
  • Transferable Query Selection for Active Domain Adaptation [2021]: Point out that the strategy of previous works are not transferable (they use the criteria from the source domain to guide the target domain selection).
  • Zero-Round Active Learning [2021, Arxiv]: Cold-start.
  • S3VAADA: Submodular Subset Selection for Virtual Adversarial Active Domain Adaptation [2021, ICCV]
  • Active Domain Adaptation via Clustering Uncertainty-weighted Embeddings [2021, ICCV]
  • Active Universal Domain Adaptation [2021, ICCV]: There are unknown class in the target domain.
  • Active Learning for Domain Adaptation: An Energy-based Approach [2021]
  • Active Learning Over Multiple Domains in Natural Language Tasks [2021]
  • Improving Semi-Supervised Domain Adaptation Using Effective Target Selection and Semantics [2021, CVPR]
  • Discriminative active learning for domain adaptation [2021, KBS]
  • Learning Distinctive Margin toward Active Domain Adaptation [2022]: The feature gradient from both loss and query function share similar update direction and yield high query score
  • Loss-based Sequential Learning for Active Domain Adaptation [2022]: Semi-supervised domain adaptation on target domain. Use predict loss to select.
  • DistillAdapt: Source-Free Active Visual Domain Adaptation [2022]: Access source model instead of source data.
  • Active Model Adaptation Under Unknown Shift [2022]: The distribution is shifted from the source domain.
  • Combating Label Distribution Shift for Active Domain Adaptation [2022]: Consider label distribution mismatch.
  • Source-Free Active Domain Adaptation via Energy-Based Locality Preserving Transfer [2022, MM]
  • Active Multi-Task Representation Learning [2022, ICML]: Select from multiple source domains.
  • TL-ADA: Transferable Loss-based Active Domain Adaptation [2023, Neural Networks]
  • Dirichlet-based Uncertainty Calibration for Active Domain Adaptation [2023, ICLR]
  • MHPL: Minimum Happy Points Learning for Active Source Free Domain Adaptation [2023, CVPR]
  • Divide and Adapt: Active Domain Adaptation via Customized Learning [2023, CVPR]
  • Multi-Discriminator Active Adversarial Network for Multi-Center Brain Disease Diagnosis [2023, TBD]
  • CSAL: Cost sensitive active learning for multi-source drifting stream [2023, KBS]
  • Open set transfer learning through distribution driven active learning [2023, Pattern Recognition]
  • Source-Free Active Domain Adaptation via Augmentation-Based Sample Query and Progressive Model Adaptation [2023, TNNLS]

Transfer Learning:

  • Accelerating active learning with transfer learning [2013, ICDM]
  • Hierarchical Active Transfer Learning [2015, SIAM]
  • Semi-Supervised Active Learning with Cross-Class Sample Transfer [2016, IJCAI]
  • Active learning with cross-class knowledge transfer [2016, AAAI]
  • Active learning with cross-class similarity transfer [2017, AAAI]
  • Rapid Performance Gain through Active Model Reuse [IJCAI, 2019]

Domain Generalization:

  • LOG: Active Model Adaptation for Label-Efficient OOD Generalization [2022, NeuraIPS]
  • Bridging the gap with grad: Integrating active learning into semi-supervised domain generalization [2024, Neural Networks]

One/Few/Zero-Shot Learning or Meta-Learning

One/few-shot learning is to learn a model on a new dataset with one for few available instances. AL could be used to select the instances to build the support set.

For one/few-shot learning:

  • Active one-shot learning [2017, Arxiv]
  • A Meta-Learning Approach to One-Step Active-Learning [2017, Arxiv]
  • Learning Algorithms for Active Learning [2017, ICML]
  • Meta-learning for batch mode active learning [2018, ICLR Workshop]
  • Augmented Memory Networks for Streaming-Based Active One-Shot Learning [2019, Arxiv]
  • Active one-shot learning with Prototypical Networks [2019, ESANN]
  • Active one-shot learning by a deep Q-network strategy [2020, Neurocomputing]
  • On the Utility of Active Instance Selection for Few-Shot Learning [2020]: Show via these “upper bounds” that we do not have a significant room for improving few-shot models through actively selecting instances.
  • Beyond Simple Meta-Learning: Multi-Purpose Models for Multi-Domain, Active and Continual Few-Shot Learning [2022]
  • Active Few-Shot Learning with FASL [2022]: Find that AL methods do not yield strong improvements over a random baseline when applied to datasets with balanced label distributions. However, experiments on modified datasets with a skewed label distributions as well as naturally unbalanced datasets show the value of AL methods.
  • Active Few-Shot Learning for Sound Event Detection [2022]
  • Active Transfer Prototypical Network: An Efficient Labeling Algorithm for Time-Series Data [2022]
  • Active Few-Shot Classification: a New Paradigm for Data-Scarce Learning Settings [2022]
  • MEAL: Stable and Active Learning for Few-Shot Prompting [2022]
  • Few-shot initializing of Active Learner via Meta-Learning [2022, EMNLP]
  • Active Learning for Efficient Few-Shot Classification [2023, ICASSP]
  • Improved prototypical network for active few-shot learning [2023, PRL]

There are also works about zero-shot learning:

  • Graph active learning for GCN-based zero-shot classification [2021, Neurocomputing]

Graph Processing

Graph Embedding/Network representation learning:

  • Active discriminative network representation learning [2018, IJCAI]

Graph node classification:

  • Active learning for streaming networked data [2014, ACM International Conference on Conference on Information and Knowledge Management]
  • Active Semi-Supervised Learning Using Sampling Theory for Graph Signals [2014, KDD]
  • MetAL: Active Semi-Supervised Learning on Graphs via Meta Learning [2020, Arxiv]
  • Active Learning for Node Classification: An Evaluation [2020, Entropy MDPI]
  • Active Learning for Node Classification: The Additional Learning Ability from Unlabelled Nodes [2020]
  • Active Learning for Attributed Graphs [2020, Master Dissertation]
  • ALG: Fast and Accurate Active Learning Framework for Graph Convolutional Networks [2021, SIGMOD]
  • Partition-Based Active Learning for Graph Neural Networks [2022]
  • LSCALE: Latent Space Clustering-Based Active Learning for Node Classification [2022, ECMLPKDD]
  • Active Learning for Node Classification using a Convex Optimization approach [2022, IEEE BigDataService]
  • SmartQuery: An Active Learning Framework for Graph Neural Networks through Hybrid Uncertainty Reduction [2022, CIKM]
  • Open-World Graph Active Learning for Node Classification [2023, ACM TKDD]
  • Improving Graph Neural Networks by combining active learning with self-training [2023, MDKD]
  • Joint European Conference on Machine Learning and Knowledge Discovery in Database [2023, ECML PKDD]
  • Mitigating Semantic Confusion from Hostile Neighborhood for Graph Active Learning [2023, CIKM]

Link Prediction:

  • Complex Query Answering with Neural Link Predictors [2020]

Node response prediction:

Graph transfer/Network Alignment:

Anchor user prediction:

Entity resolution:

Entity Alignment:

  • ActiveEA: Active Learning for Neural Entity Alignment [2021]
  • Impact of the Characteristics of Multi-Source Entity Matching Tasks on the Performance of Active Learning Methods [2022]
  • Active Temporal Knowledge Graph Alignment [2023, IJSWIS]
  • Deep Active Alignment of Knowledge Graph Entities and Schemata [2023, Proc. ACM Manag. Data]

Regression on Graph Data:

  • Batch Mode Deep Active Learning for Regression on Graph Data [2023, IEEE BigData]

Metric learning/Pairwise comparison/Similarity learning

  • Active Ordinal Querying for Tuplewise Similarity Learning [2020 AAAI]: Introduce AL to their new similarity methods (InfoTuple). Active Sampling for Pairwise Comparisons via Approximate Message Passing and Information Gain Maximization.
  • Batch Decorrelation for Active Metric Learning [2020, IJCAI]
  • Maximizing Conditional Entropy for Batch-Mode Active Learning of Perceptual Metrics [2021]
  • A Unified Batch Selection Policy for Active Metric Learning [2021, ECML PKDD]
  • Active metric learning and classification using similarity queries [2022]
  • Active pairwise distance learning for efficient labeling of large datasets by human experts [2023, Applied Intelligence]

Recommendation

Review:

  • A survey of active learning in collaborative filtering recommender systems

Transfer Learning for Recommendation System:

  • Active Transfer Learning for Recommendation System [2020, PhD Dissertation]

Positive Unlabeled learning:

  • PU Active Learning for Recommender Systems [2021, Neural Processing Letters]

Reinforcement Learning

Active Exploration

  • SAMBA: Safe Model-Based & Active Reinforcement Learning [2020, Arxiv]
  • How to Train Your Agent: Active Learning from Human Preferences and Justifications in Safety-Critical Environments [2022, AAMAS]
  • Active Exploration for Inverse Reinforcement Learning [2022]
  • Uncertainty-Guided Active Reinforcement Learning with Bayesian Neural Networks [2023, ICRA]

Save Training Cost in the Measure of Time:

Atari Games:

  • Width-Based Planning and Active Learning for Atari [2021]

Robotics

Review:

Human-Robot Interaction:

Object Detection Learning:

Robot motion planning:

Demonstrate Robots:

  • Active Learning of Bayesian Probabilistic Movement Primitives [2021, IEEE ROBOTICS AND AUTOMATION LETTERS]

Active Exploration:

  • SEAL: Self-supervised Embodied Active Learning using Exploration and 3D Consistency [2021, NeurIPS]

Semantic Mapping:

  • An Informative Path Planning Framework for Active Learning in UAV-based Semantic Mapping [2023]

Multi-Agent:

  • Fleet Active Learning: A Submodular Maximization Approach [2023, CoRL]

Model Interpretability

Clustering

AL could support clustering by provide supervised information.

  • Semi-Supervised Selective Affinity Propagation Ensemble Clustering With Active Constraints [2020, IEEE Access]
  • Active Learning for Constrained Document Clustering with Uncertainty Region [Complexity, 2020]: Must link & cannot link.
  • Cautious Active Clustering [2020]
  • Improving evolutionary constrained clustering using Active Learning [2020, Knowledge-Based Systems]
  • An Active Learning Method Based on Variational Autoencoder and DBSCAN Clustering [2021]
  • Active constrained deep embedded clustering with dual source [2022, Applied Intelligence]
  • Active deep image clustering [2022, KBS]
  • Active Learning with Positive and Negative Pairwise Feedback [2023]
  • Active Clustering Ensemble With Self-Paced Learning [2023, TNNLS]

Review:

  • Interactive clustering: a scoping review [2020, Artificial Intelligence Review]

Other Fields (alphabetical order)

Adversarial Attack

Detection:

Training:

  • Models in the Loop: Aiding Crowdworkers with Generative Annotation Assistants [2021]

Algorithm Fairness

Addressing fairness at the data collection and dataset preparation stages therefore becomes an essential part of training fairer algorithms. For example, the ImageNet was crawled from image databases without considering sensitive attributes such as race or gender. In consequence, models trained (or pre-trained) on this dataset are prone to mimic societal biases.

Anomaly Detection

  • Meta-AAD: Active Anomaly Detection with Deep Reinforcement Learning [2020, Arxiv]
  • A Semi-Supervised VAE Based Active Anomaly Detection Framework in Multivariate Time Series for Online Systems [2022, WWW]
  • Situation-Aware Multivariate Time Series Anomaly Detection Through Active Learning and Contrast VAE-Based Models in Large Distributed Systems [2022]
  • Unsupervised Learning of Distributional Properties can Supplement Human Labeling and Increase Active Learning Efficiency in Anomaly Detection [2023, ICML workshop]
  • Semi-Supervised Learning from Active Noisy Soft Labels for Anomaly Detection [2023, ECML PKDD]

Audio Processing

  • Active Correction for Incremental Speaker Diarization of a Collection with Human in the Loop [2022, Applied Science]
  • Investigating Active-learning-based Training Data Selection for Speech Spoofing Countermeasure [2022]
  • An efficient framework for constructing speech emotion corpus based on integrated active learning strategies [2022, IEEE Trans. Affect. Comput.]

Causal Analysis

Choice Model

  • Active Learning for Non-Parametric Choice Models [2022]

Continual Learning

  • Active Continual Learning: Labelling Queries in a Sequence of Tasks [2023]

Curriculum Learning

Entity Matching

  • Deep entity matching with adversarial active learning [2022, VLDB]

Federated Learning

  • Federated Active Learning (F-AL): an Efficient Annotation Strategy for Federated Learning [2022]
  • LG-FAL : Federated Active Learning Strategy using Local and Global Models [2022, ICML workshop]
  • Knowledge-Aware Federated Active Learning with Non-IID Data [2022]
  • Federated deep active learning for attention-based transaction classification [2022, Applied Intelligence]
  • Re-thinking Federated Active Learning based on Inter-class Diversity [2023]

Fuzzy System

  • Online Active Learning for Evolving Error Feedback Fuzzy Models within a Multi-Innovation Context [2023, IEEE Trans Fuzzy Syst.]

Hedge

Hedge is quite simple to explain in words: the algorithm combines the predictions of all the experts on a given round by taking their weighted average, where the weight of an expert exponentially decays according to the number of previous mistakes. The goal is to find a weight for the experts' predictions to make the final decision.

  • ActiveHedge: Hedge meets Active Learning [2022, ICML]

Human Learning

Information Retrieval

  • Efficient Test Collection Construction via Active Learning [2021, ICTIR]: The active selection here is for IR evaluation.
  • Are Binary Annotations Sufficient? Video Moment Retrieval via Hierarchical Uncertainty-based Active Learning [2023]
  • Active Learning with Crowd Sourcing Improves Information Retrieval [2023, ICML workshop]

Instance Search (INS)

  • Confidence-Aware Active Feedback for Efficient Instance Search [2021]

Generative Adversarial Network Training

AL could reduce the number of needed instances to train a GAN.

  • Learning Class-Conditional GANs with Active Sampling [2019, SIGKDD]

Knowledge Distillation

  • Robust Active Distillation [2022]
  • PVD-AL: Progressive Volume Distillation with Active Learning for Efficient Conversion Between Different NeRF Architectures [2023]
  • Stream-Based Active Distillation for Scalable Model Deployment [2023, CVPR workshop]

Label Enhancement

Label De-noising:

  • Active Deep Learning to Tune Down the Noise in Labels [2018, KDD]

Clean Label Uncertainties: Enhance the quality of generated from weakly supervised model.

Large Language Models (LLMs)

Prompt Engineering

  • Active Prompting with Chain-of-Thought for Large Language Models [2023]
  • Eliciting Human Preferences with Language Models [2023]

In-Context Learning

  • Active Learning Principles for In-Context Learning with Large Language Models [2023]

Finetuning

  • Interactive Multi-fidelity Learning for Cost-effective Adaptation of Language Model with Sparse Human Supervision [2023, NeurIPS]

Learning from Label Proportions (LLP)

  • Active learning from label proportions via pSVM [2021, Neurocomputing]

Model Selection

As active model selection, both pool based and stream based. Normally there is only model selection without training.

  • Online Active Model Selection for Pre-trained Classifiers [2020]

Multi-Fidelity Machine Learning

Multimodal Learning

  • Towards Balanced Active Learning for Multimodal Classification [2023]
  • Bad Students Make Great Teachers: Active Learning Accelerates Large-Scale Visual Understanding [2023]

Online Learning System

Optimization Problems

The background is that simulation (for evaluation) is quite expensive in many optimization problem. Utilize active sampling to reduce the optimization cost.

Works:

Multi-Objective Optimizations

Works:

Influence Maximization in Network

Works:

  • Near-optimal Batch Mode Active Learning and Adaptive Submodular Optimization [2013, ICML]

Predict-then-Optimize

The quality of the prediction is decided by the decision loss.

Works:

  • Active Learning in the Predict-then-Optimize Framework: A Margin-Based Approach [2023]

Ordinal Regression/Classification

Could be consider as a regression where the relative orders of the instances matter.

Positive and Unlabeled (PU) Learning

A special case of binary classification where a learner only has access to labeled positive examples and unlabeled examples.

  • Class Prior Estimation in Active Positive and Unlabeled Learning [2020, IJCAI]

Rank

  • A Bayesian Active Learning Approach to Comparative Judgement [2023]
  • Annotating Data for Fine-Tuning a Neural Ranker? Current Active Learning Strategies are not Better than Random Selection [2023, SIGIR-AP]

Reliability Analysis

  • Sequential active learning of low-dimensional model representations for reliability analysis [2021, Arxiv]

Sequence Labeling

Software Engineering

Software Defects Prediction:

Code Learning:

  • Active Code Learning: Benchmarking Sample-Efficient Training of Code Models [2023]

Spiking Neural Network

  • Bio-inspired Active Learning method in spiking neural network [2023, KBS]
  • Effective Active Learning Method for Spiking Neural Networks [2023, TNNLS]

Speech Recognition

Symbolic Regression

  • Online Symbolic Regression with Informative Query [2023]
  • Active Learning in Symbolic Regression Performance with Physical Constraints [2023]

Treatment Effect

  • Causal-BALD: Deep Bayesian Active Learning of Outcomes to Infer Treatment-Effects from Observational Data [2021, NeurIPS]