Transport gratuit la punctele de livrare Pick Up peste 299 lei
Packeta 15 lei Easybox 20 lei Cargus 25 lei FAN 25 lei

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems

Limba englezăengleză
Carte Carte broșată
Carte Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems Sebastian Bubeck
Codul Libristo: 04834934
Editura now publishers Inc, decembrie 2012
A multi-armed bandit problem - or, simply, a bandit problem - is a sequential allocation problem def... Descrierea completă
? points 242 b
489 lei
În depozitul extern Expediem în 15-20 zile

30 de zile pentru retur bunuri


Ar putea de asemenea, să te intereseze


top
Science of Cooking Stuart Farrimond / Copertă tare
common.buy 113 lei
top
Happiness 1 Shuzo Oshimi / Carte broșată
common.buy 65 lei
top
The Course of Love Alain de Botton / Carte broșată
common.buy 48 lei
top
Learn to Draw Manga Women / Carte broșată
common.buy 79 lei
Practical Veterinary Dental Radiography Brook A. Niemiec / Copertă tare
common.buy 667 lei
Colour Quest (R) Cityscapes John Woodcock / Carte broșată
common.buy 77 lei
Advanced Nutrition and Human Metabolism Sareen Gropper / Copertă tare
common.buy 629 lei
Dog Behaviour, Evolution, and Cognition Adam Miklosi / Carte broșată
common.buy 394 lei
Clean C++20 Stephan Roth / Carte broșată
common.buy 198 lei
Japanese Sake Bible Takashi Eguchi / Carte broșată
common.buy 83 lei
Albert Camus: A Life Olivier Todd / Carte broșată
common.buy 113 lei
Design Principles for Photography Jeremy Webb / Carte broșată
common.buy 224 lei
Battles of Tolkien David Day / Carte broșată
common.buy 73 lei

A multi-armed bandit problem - or, simply, a bandit problem - is a sequential allocation problem defined by a set of actions. At each time step, a unit resource is allocated to an action and some observable payoff is obtained. The goal is to maximize the total payoff obtained in a sequence of allocations. The name bandit refers to the colloquial term for a slot machine (a "one-armed bandit" in American slang). In a casino, a sequential allocation problem is obtained when the player is facing many slot machines at once (a "multi-armed bandit"), and must repeatedly choose where to insert the next coin. Multi-armed bandit problems are the most basic examples of sequential decision problems with an exploration-exploitation trade-off. This is the balance between staying with the option that gave highest payoffs in the past and exploring new options that might give higher payoffs in the future. Although the study of bandit problems dates back to the 1930s, exploration-exploitation trade-offs arise in several modern applications, such as ad placement, website optimization, and packet routing. Mathematically, a multi-armed bandit is defined by the payoff process associated with each option. In this book, the focus is on two extreme cases in which the analysis of regret is particularly simple and elegant: independent and identically distributed payoffs and adversarial payoffs. Besides the basic setting of finitely many actions, it also analyzes some of the most important variants and extensions, such as the contextual bandit model. This monograph is an ideal reference for students and researchers with an interest in bandit problems.

Informații despre carte

Titlu complet Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems
Limba engleză
Legare Carte - Carte broșată
Data publicării 2012
Număr pagini 138
EAN 9781601986269
ISBN 1601986262
Codul Libristo 04834934
Greutatea 208
Dimensiuni 234 x 159 x 8
Dăruiește această carte chiar astăzi
Este foarte ușor
1 Adaugă cartea în coș și selectează Livrează ca un cadou 2 Îți vom trimite un voucher în schimb 3 Cartea va ajunge direct la adresa destinatarului

Logare

Conectare la contul de utilizator Încă nu ai un cont Libristo? Crează acum!

 
obligatoriu
obligatoriu

Nu ai un cont? Beneficii cu contul Libristo!

Datorită contului Libristo, vei avea totul sub control.

Creare cont Libristo