Multi-Armed Bandit
OptimizePrice: $0.01/callLatency: <1msComplexity: O(n)
Quick scenarios
Choose the best option among 3 alternatives with uncertain payoffs. UCB1 balances exploitation of known-good options with exploration of uncertain ones.
Click "Run Algorithm" to see results
cURL
curl -X POST https://oraclaw-api.onrender.com/api/v1/optimize/bandit \
-H "Content-Type: application/json" \
-d '{"arms":[{"id":"A","name":"Option A","pulls":10,"totalReward":7},{"id":"B","name":"Option B","pulls":10,"totalReward":5},{"id":"C","name":"Option C","pulls":2,"totalReward":1.8}],"algorithm":"ucb1"}'Input Schema
arms: Array of {id, name, pulls?, totalReward?}
algorithm: 'ucb1' | 'thompson' | 'epsilon-greedy'
config: {explorationConstant?, rewardDecay?}
Output Fields
selectedscorealgorithmexploitationexplorationregret