Multi-Armed Bandit

Optimize
Price: $0.01/callLatency: <1msComplexity: O(n)

Quick scenarios

Choose the best option among 3 alternatives with uncertain payoffs. UCB1 balances exploitation of known-good options with exploration of uncertain ones.

Click "Run Algorithm" to see results
cURL
curl -X POST https://oraclaw-api.onrender.com/api/v1/optimize/bandit \
  -H "Content-Type: application/json" \
  -d '{"arms":[{"id":"A","name":"Option A","pulls":10,"totalReward":7},{"id":"B","name":"Option B","pulls":10,"totalReward":5},{"id":"C","name":"Option C","pulls":2,"totalReward":1.8}],"algorithm":"ucb1"}'

Input Schema

arms: Array of {id, name, pulls?, totalReward?}
algorithm: 'ucb1' | 'thompson' | 'epsilon-greedy'
config: {explorationConstant?, rewardDecay?}

Output Fields

selectedscorealgorithmexploitationexplorationregret