Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings
#

bandit-algorithms

Here are 96 public repositories matching this topic...

SMPyBandits

🔬 Research Framework for Single and Multi-Players 🎰 Multi-Arms Bandits (MAB) Algorithms, implementing all the state-of-the-art algorithms for single-player (UCB, KL-UCB, Thompson...) and multi-player (MusicalChair, MEGA, rhoRand, MCTop/RandTopM etc).. Available on PyPI:https://pypi.org/project/SMPyBandits/ and documentation on

  • UpdatedApr 30, 2024
  • Jupyter Notebook

A hyperparameter optimization framework, inspired by Optuna.

  • UpdatedAug 12, 2025
  • Go

Yahoo! news article recommendation system by linUCB

  • UpdatedFeb 1, 2018
  • Python

Big Data's open seminars: An Interactive Introduction to Reinforcement Learning

  • UpdatedJun 7, 2021
  • Jupyter Notebook

My solutions to Yandex Practical Reinforcement Learning course in PyTorch and Tensorflow

  • UpdatedDec 22, 2021
  • Jupyter Notebook

A lightweight python library for bandit algorithms

  • UpdatedJul 21, 2022
  • Python

Code for our ACML and INTERSPEECH papers: "Speaker Diarization as a Fully Online Bandit Learning Problem in MiniVox".

  • UpdatedSep 20, 2021
  • Cuda

More about the exploration-exploitation tradeoff with harder bandits

  • UpdatedMay 12, 2019
  • Jupyter Notebook

A curated list on papers about combinatorial multi-armed bandit problems.

  • UpdatedMay 10, 2021

A comprehensive Python library implementing a variety of contextual and non-contextual multi-armed bandit algorithms—including LinUCB, Epsilon-Greedy, Upper Confidence Bound (UCB), Thompson Sampling, KernelUCB, NeuralLinearBandit, and DecisionTreeBandit—designed for reinforcement learning applications

  • UpdatedDec 31, 2024
  • Python

Building recommender Systems using contextual bandit methods to address cold-start issue and online real-time learning

  • UpdatedJul 1, 2021
  • Jupyter Notebook

🐯REPLICA of "Auction-based combinatorial multi-armed bandit mechanisms with strategic arms"

  • UpdatedDec 17, 2023
  • Python

This is a collection of interesting papers that I have read so far or want to read. Note that the list is not up-to-date. Topics: reinforcement learning, deep learning, mathematics, statistics, bandit algorithms, optimization.

  • UpdatedApr 3, 2025

Personalized and Interactive Music Recommendation with Bandit approach

  • UpdatedSep 15, 2019
  • Jupyter Notebook

Deep contextual bandits in PyTorch: Neural Bandits, Neural Linear, and Linear Full Posterior Sampling with comprehensive benchmarking on synthetic and real datasets

  • UpdatedJun 29, 2025
  • Python

This repository aims at learning most popular MAB and CMAB algorithms and watch how they run. It is interesting for those wishing to start learning these topics.

  • UpdatedDec 7, 2021
  • Python

Improve this page

Add a description, image, and links to thebandit-algorithms topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with thebandit-algorithms topic, visit your repo's landing page and select "manage topics."

Learn more


[8]ページ先頭

©2009-2025 Movatter.jp