Online Learning for Resouce Sharing
2022-IEEE INFOCOM-Distributed Bandits with Heterogeneous Agents
2021-NeurIPS-Cooperative Stochastic Bandits with Asynchronous Agents and Constrained Feedback
2021-ICML- Incentivized bandit learning with self-reinforcing user preferences
2018-NeurIPS-Bandit learning with positive externalities
2022-AAAI-A online learning approach to sequential user-centric selection problem
2022-INFOCOM-Distributed Bandits with Heterogeneous Agents
2021-NeurIPS-Cooperative Stochastic Bandits with Asynchronous Agents and Constrained Feedback
2022-ICML-Multiple-Play Stochastic Bandits with Shareable Finite-Capacity Arms
2022-AAAI-Modeling Attrition in Recommender Systems with Departing Bandits
2018-NeurIPS-Bandit Learning with Positive Externalities
2018-IEEE Trans on Automatic Conctrol-The Multi-Armed Bandit With Stochastic Plays
2020-Arxiv-Bandit Learning in Decentralized Matching Markets
2013-FOCS-Bandits With Knapsacks
2014-ACM Conference on Economics and Computation-Incentivizing Exploration
2022-TECS-Edge Intelligence: Concepts, Architectures, Applications, and Future Directions
2022-ICML-Congested Bandits: Optimal Routing via Short-term Resets
2022-ICML-Online Learning and Pricing with Reusable Resources: Linear Bandits with Sub-Exponential Rewards
2022-ICML-Congested Bandits: Optimal Routing via Short-term Resets
2021-ICML-Online Learning for Load Balancing of Unknown Monotone Resource Allocation Games
2021-NIPS-Fast Routing under Uncertainty: Adaptive Learning in Congestion Games with Exponential Weights
2022-ICML-Online Learning and Pricing with Reusable Resources: Linear Bandits with Sub-Exponential Rewards
Resource allocation problem for cloud, ridesharing, car, etc
2022-ICML-Congested Bandits: Optimal Routing via Short-term Resets
2021-NIPS-Bandit Learning with Delayed Impact of Actions
Action of an arm impact the later reward of an arm
2021-NIPS-Stochastic Multi-Armed Bandits with Control Variates
auxiliary information about the arm rewards is available
Such as wireless networking
2021-NIPS-Bandit Quickest Changepoint Detection
Bandits optimization for sensor placement to changepoint detection
2021-NIPS-One More Step Towards Reality: Cooperative Bandits with Imperfect Communication
Bring communication and networking model into the cooperative bandits
2022-AISTAT-Solving Multi-Arm Bandit Using a Few Bits of Communication
Communication vs. regret tradeoff
2022-JMLR-Multi-Agent Multi-Armed Bandits with Limited Communication
Communication vs. regret tradeoff
2021-AISTAT-Combinatorial Gaussian Process Bandits with Probabilistically Triggered Arms
2021-COLT-Multiplayer Bandit Learning, from Competition to Cooperation
Cooperation vs. regret tradeoff, competition vs. regret tradeoff
2020-AISTATS-Competing Bandits in Matching Markets
2022-SIGMETRICS-Correlated Combinatorial Bandits for Online Resource Allocation
2021-SIGMETRICS-Social Learning in Multi Agent Multi Armed Bandits
2021-SIGMETRICS-Online Virtual Machine Allocation with Lifetime and Load Predictions
Resource fragmentation and resource utilization
Bandit and Causal Inference
2021-WWW-Unifying Offline Causal Inference and Online Bandit Learning for Data Driven Decision
2021-ICDM-A Robust Algorithm to Unifying Offline Causal Inference and Online Multi-armed Bandit Learning
2021-NIPS-Deep Proxy Causal Learning and its Application to Confounded Bandit Policy Evaluation
2019-AISTATS-Interpretable Almost-Exact Matching for Causal Inference
2017-AISTATS-A Framework for Optimal Matching for Causal Inference