ICML-20 待读的 Paper
2020.9.12
花了一上午的时间,过了一遍 ICML-2020 Accepted Paper List, 挑出了自己想读的 Paper。
主要关注于自己的一些研究点。
Noisy Labels
- Beyond Synthetic Noise: Deep Learning on Controlled Noisy Labels
- SIGUA: Forgetting May Make Learning with Noisy Labels More Robust
- Error-Bounded Correction of Noisy Labels
- Does label smoothing mitigate label noise?
- Deep k-NN for Noisy Labels
- Improving generalization by controlling label-noise information in neural network weights
- Normalized Loss Functions for Deep Learning with Noisy Labels
- Variational Label Enhancement
- Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates
- Searching to Exploit Memorization Effect in Learning with Noisy Labels
- Label-Noise Robust Domain Adaptation
- Training Binary Neural Networks through Learning with Noisy Supervision
- Learning with Bounded Instance- and Label-dependent Label Noise
- Progressive Identification of True Labels for Partial-Label Learning
- Learning with Multiple Complementary Labels
- Unbiased Risk Estimators Can Mislead: A Case Study of Learning with Complementary Labels
Semi-Supervised Learning
- Semi-Supervised Learning with Normalizing Flows
- Negative Sampling in Semi-Supervised learning
- Poisson Learning: Graph Based Semi-Supervised Learning At Very Low Label Rates
- Time-Consistent Self-Supervision for Semi-Supervised Learning
- Safe Deep Semi-Supervised Learning for Unseen-Class Unlabeled Data
- Deep Streaming Label Learning
Domain Adaptation
- Continuously Indexed Domain Adaptation
- RIFLE: Backpropagation in Depth for Deep Transfer Learning through Re-Initializing the Fully-connected LayEr
- Implicit Class-Conditioned Domain Alignment for Unsupervised Domain Adaptation
- Understanding Self-Training for Gradual Domain Adaptation
- Do We Really Need to Access the Source Data? Source Hypothesis Transfer for Unsupervised Domain Adaptation
- LTF: A Label Transformation Framework for Correcting Label Shift
- Label-Noise Robust Domain Adaptation
Data Bias, Weighting
- Adversarial Filters of Dataset Biases
- Optimizing Data Usage via Differentiable Rewards
- Data preprocessing to mitigate bias: A maximum entropy based approach
- DeBayes: a Bayesian Method for Debiasing Network Embeddings
- A Distributional Framework For Data Valuation
Class-Imbalance
- Class-Weighted Classification: Trade-offs and Robust Approaches
- Online Continual Learning from Imbalanced Data
- Logistic Regression for Massive Data with Rare Events
MixUp, Interpolation, Extrapolation, etc.
- Puzzle Mix: Exploiting Saliency and Local Statistics for Optimal Mixup
- Learning Representations that Support Extrapolation
- Extrapolation for Large-batch Training in Deep Learning
- Training Neural Networks for and by Interpolation
- Sequence Generation with Mixed Representations
PU Learning
- Self-PU: Self Boosted and Calibrated Positive-Unlabeled Training
Active Learning
- Adaptive Region-Based Active Learning
GCN or Recommendation System
- Continuous Graph Neural Networks
- Simple and Deep Graph Convolutional Networks
- Graph Convolutional Network for Recommendation with Low-pass Collaborative Filters
- Generalization and Representational Limits of Graph Neural Networks
- Graph-based Nearest Neighbor Search: From Practice to Theory
- Ordinal Non-negative Matrix Factorization for Recommendation
- Improved Communication Cost in Distributed PageRank Computation – A Theoretical Study
- Optimization and Analysis of the pAp@k Metric for Recommender Systems
- Scalable and Efficient Comparison-based Search without Features
- Learning to Rank Learning Curves
- When Does Self-Supervision Help Graph Convolutional Networks?
Neural ODE
- Towards Adaptive Residual Network Training: A Neural-ODE Perspective
- Combining Differentiable PDE Solvers and Graph Neural Networks for Fluid Flow Prediction
- Approximation Capabilities of Neural ODEs and Invertible Residual Networks
Calibration, Confidence, Out-of-distribution
- Confidence-Aware Learning for Deep Neural Networks
- Being Bayesian, Even Just a Bit, Fixes Overconfidence in ReLU Networks
- SDE-Net: Equipping Deep Neural Networks with Uncertainty Estimates
- Detecting Out-of-Distribution Examples with Gram Matrices
- Interpretable, Multidimensional, Multimodal Anomaly Detection with Negative Sampling for Detection of Device Failure
- Uncertainty Estimation Using a Single Deep Deterministic Neural Network
- How Good is the Bayes Posterior in Deep Neural Networks Really?
Federated Learning, Fairness
- Fair k-Centers via Maximum Matching
- Federated Learning with Only Positive Labels
Interesting Problems, Settings
- Why Are Learned Indexes So Effective?
- Learning with Feature and Distribution Evolvable Streams
- Cost-effectively Identifying Causal Effects When Only Response Variable is Observable
- Rigging the Lottery: Making All Tickets Win
- Do We Need Zero Training Loss After Achieving Zero Training Error?
- Small Data, Big Decisions: Model Selection in the Small-Data Regime
- Why bigger is not always better: on finite and infinite neural networks
- On Learning Sets of Symmetric Elements
- Collaborative Machine Learning with Incentive-Aware Model Rewards
- Generalisation error in learning with random features and the hidden manifold model
- Sample Amplification: Increasing Dataset Size even when Learning is Impossible
- When are Non-Parametric Methods Robust?
- Performative Prediction
- Supervised learning: no loss no cry
- Teaching with Limited Information on the Learner's Behaviour
- Learning De-biased Representations with Biased Representations
- Do RNN and LSTM have Long Memory?
- It's Not What Machines Can Learn, It's What We Cannot Teach
- Enhancing Simple Models by Exploiting What They Already Know
Interesting Theory
- On the Generalization Benefit of Noise in Stochastic Gradient Descent
- Let's Agree to Agree: Neural Networks Share Classification Order on Real Datasets
- Optimizer Benchmarking Needs to Account for Hyperparameter Tuning
- On the Noisy Gradient Descent that Generalizes as SGD
- Rethinking Bias-Variance Trade-off for Generalization of Neural Networks
- Understanding and Mitigating the Tradeoff between Robustness and Accuracy
- The Implicit and Explicit Regularization Effects of Dropout
- Optimal Continual Learning has Perfect Memory and is NP-hard
- Curvature-corrected learning dynamics in deep neural networks
- Explainable k-Means and k-Medians Clustering
- Near-Tight Margin-Based Generalization Bounds for Support Vector Machines
- Uniform Convergence of Rank-weighted Learning
- Decision Trees for Decision-Making under the Predict-then-Optimize Framework
Interesting Algorithm
- SoftSort: A Continuous Relaxation for the argsort Operator
- Boosting Deep Neural Network Efficiency with Dual-Module Inference
- Circuit-Based Intrinsic Methods to Detect Overfitting
- Learning Similarity Metrics for Numerical Simulations
- Deep Divergence Learning
- Consistent Estimators for Learning to Defer to an Expert
- Smaller, more accurate regression forests using tree alternating optimization
- Learning To Stop While Learning To Predict
- DROCC: Deep Robust One-Class Classification
Point Cloud, 3 dimension
- Hypernetwork approach to generating point clouds