Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

DBA: Dynamic Multi-Armed Bandit Algorithm

View through CrossRef
We introduce Dynamic Bandit Algorithm (DBA), a practical solution to improve the shortcoming of the pervasively employed reinforcement learning algorithm called Multi-Arm Bandit, aka Bandit. Bandit makes real-time decisions based on the prior observations. However, Bandit is heavily biased to the priors that it cannot quickly adapt itself to a trend that is interchanging. As a result, Bandit cannot, quickly enough, make profitable decisions when the trend is changing. Unlike Bandit, DBA focuses on quickly adapting itself to detect these trends early enough. Furthermore, DBA remains as almost as light as Bandit in terms of computations. Therefore, DBA can be easily deployed in production as a light process similar to The Bandit. We demonstrate how critical and beneficial is the main focus of DBA, i.e. the ability to quickly finding the most profitable option in real-time, over its stateof-the-art competitors. Our experiments are augmented with a visualization mechanism that explains the profitability of the decisions made by each algorithm in each step by animations. Finally we observe that DBA can substantially outperform the original Bandit by close to 3 times for a set Key Performance Indicator (KPI) in a case of having 3 arms.
Association for the Advancement of Artificial Intelligence (AAAI)
Title: DBA: Dynamic Multi-Armed Bandit Algorithm
Description:
We introduce Dynamic Bandit Algorithm (DBA), a practical solution to improve the shortcoming of the pervasively employed reinforcement learning algorithm called Multi-Arm Bandit, aka Bandit.
Bandit makes real-time decisions based on the prior observations.
However, Bandit is heavily biased to the priors that it cannot quickly adapt itself to a trend that is interchanging.
As a result, Bandit cannot, quickly enough, make profitable decisions when the trend is changing.
Unlike Bandit, DBA focuses on quickly adapting itself to detect these trends early enough.
Furthermore, DBA remains as almost as light as Bandit in terms of computations.
Therefore, DBA can be easily deployed in production as a light process similar to The Bandit.
We demonstrate how critical and beneficial is the main focus of DBA, i.
e.
the ability to quickly finding the most profitable option in real-time, over its stateof-the-art competitors.
Our experiments are augmented with a visualization mechanism that explains the profitability of the decisions made by each algorithm in each step by animations.
Finally we observe that DBA can substantially outperform the original Bandit by close to 3 times for a set Key Performance Indicator (KPI) in a case of having 3 arms.

Related Results

Case Report: Clinical management of a severe DBA patient with a novel RPS19 mutation
Case Report: Clinical management of a severe DBA patient with a novel RPS19 mutation
Diamond-Blackfan anemia (DBA) is a rare congenital bone marrow failure disorder characterized by defective erythropoiesis, typically caused by mutations in ribosomal protein (RP) g...
Contributions to high accuracy snapshot GNSS positioning
Contributions to high accuracy snapshot GNSS positioning
(English) Snapshot positioning is the technique to determine the position of a Global Navigation Satellite System (GNSS) receiver using only a very brief interval of the received ...
Childhood Dead-Before-Arrival at a Nigerian Tertiary Health Facility
Childhood Dead-Before-Arrival at a Nigerian Tertiary Health Facility
Background: Dead-before-arrival (DBA) is a term used to describe patients who had no sign of life at the time of presentation to the hospital. Little information exists...
Federated Bandit: A Gossiping Approach
Federated Bandit: A Gossiping Approach
We study Federated Bandit, a decentralized Multi-Armed Bandit (MAB) problem with a set of N agents, who can only communicate their local data with neighbors described by a connecte...
ARMED EXTORTION IN LIGHT OF THE PRINCIPLE OF CRIMINAL LEGALITY
ARMED EXTORTION IN LIGHT OF THE PRINCIPLE OF CRIMINAL LEGALITY
Furthermore, the DRC's military courts and tribunals fail to respect the principle of legality of offenses and penalties, in that they conflate the offense of armed robbery with th...
Multi-armed bandit games
Multi-armed bandit games
AbstractA sequential optimization model, known as the multi-armed bandit problem, is concerned with optimal allocation of resources between competing activities, in order to genera...
Ribosomal Protein S19 and Diamond Blackfan Anemia.
Ribosomal Protein S19 and Diamond Blackfan Anemia.
Abstract Diamond Blackfan Anemia (DBA) is one of several bone marrow failures that have been linked to defects in ribosome synthesis. 25% of DBA cases are linked to ...

Back to Top