Joseph, Matthew, Michael J Kearns, Jamie Morgenstern, Seth Neel, and Aaron Leon Roth. "Fair Algorithms for Infinite and Contextual Bandits." Proceedings of the AAAI/ACM Conference on Artificial ...
How does a gambler maximize winnings from a row of slot machines? This is the inspiration for the "multi-armed bandit problem," a common task in reinforcement learning in which "agents" make choices ...
Chemists find the best working conditions for new reactions by experimenting with hundreds or thousands of combinations of parameters — such as catalysts, solvents and temperatures. This process, ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results