InNetflix TechBlogbyNetflix Technology BlogRecommending for Long-Term Member Satisfaction at NetflixBy Jiangwei Pan, Gary Tang, Henry Wang, and Justin BasilicoAug 29, 20248Aug 29, 20248
InLyft EngineeringbyJonas TimmermannLyft’s Reinforcement Learning PlatformTackling decision making problems with a platform for developing & serving Reinforcement Learning models with a focus on Contextual BanditsMar 12, 2024Mar 12, 2024
Viet VoApplication of Multi-Armed Bandits to Promotion Ranking in MoMoMulti-Armed BanditsAug 26, 2023Aug 26, 2023
InWalmart Global Tech BlogbyAbhimanyu MitraLessons from Adopting Explore-Exploit Modeling in Industrial-Scale Recommender SystemsAuthored and contributed by: Abhimanyu Mitra & Afroza AliAug 21, 2023Aug 21, 2023
InTDS ArchivebyMassimiliano CostacurtaDynamic Pricing with Multi-Armed Bandit: Learning by DoingApplying Reinforcement Learning strategies to real-world use cases, especially in dynamic pricing, can reveal many surprisesAug 16, 20237Aug 16, 20237
Intech-at-instacartbyDavid VengerovUsing Contextual Bandit models in large action spaces at InstacartDavid Vengerov, Vinesh Gudla, Tejaswi Tenneti, Haixun Wang, Kourosh HakhamaneshiJun 15, 2023Jun 15, 2023
InLyft EngineeringbyYanqiao WangA Review of Multi-Armed Bandits Applications at LyftBy Sara Smoot, Alex Contryman and Yanqiao WangJan 23, 20231Jan 23, 20231
InNetflix TechBlogbyNetflix Technology BlogArtwork Personalization at NetflixArtwork is the first instance of personalizing not just what we recommend but also how we recommend.Dec 7, 2017125Dec 7, 2017125
InTDS ArchivebyPavel SurmenokContextual Bandits and Reinforcement LearningIf you develop personalization of user experience for your website or an app, contextual bandits can help you. Using contextual bandits…Aug 27, 201710Aug 27, 201710
InTDS ArchivebyAdrien BiarnesA Complete Tutorial on Off-Policy Evaluation for Recommender SystemsHow to reduce the offline-online evaluation gapMar 11, 2023Mar 11, 2023
InExpedia Group TechnologybyFedor ParfenovMulti-Variate Web Optimisation Using Linear Contextual BanditsOr how you can run full webpage optimisations with a context-aware outcomeSep 1, 20201Sep 1, 20201
InAdyen TechbyAdyenOptimizing payment conversion rates using contextual multi-armed banditsBy Rodel van Rooijen, Data Scientist, AdyenNov 3, 2020Nov 3, 2020
Arthur JulianiSimple Reinforcement Learning with Tensorflow Part 1.5: Contextual Bandits(Note: This post is designed as an additional tutorial to act as a bridge between Parts 1 & 2.)Sep 28, 201620Sep 28, 201620
InRed Ventures Data Science & EngineeringbyJoe Tenini, PhDHow to Make Good DecisionsUsing bandit policies to efficiently learn as you goAug 14, 2019Aug 14, 2019
InRed Ventures Data Science & EngineeringbyTim SchottHow To Build A Better BanditImproving Contextual Bandits with Rich Interaction FeaturesAug 15, 20221Aug 15, 20221
InInsightbyBrian O'GormanMulti-armed bandits for dynamic movie recommendationsMaking the best recommendations to anonymous audiencesApr 11, 20181Apr 11, 20181
InTDS ArchivebyAerin KimGamma Distribution — Intuition, Derivation, and Examplesand why does it matter?Oct 12, 201917Oct 12, 201917
Zaid GharaybehContextual Bandit Recommender System Using Vowpal WabbitContextual Bandits are a class of online learning algorithms that model an agent that learns to act optimally by efficiently acquiring new…Dec 16, 2021Dec 16, 2021
InN3TWORKbyJulian RungeHow We Boosted App Revenue by 10% with Real-time PersonalizationJulian Runge, Mark Williams, James Marr, Hernan Silberman, Yavuz Acikalin, Eric Seufert, Steve DetwilerFeb 20, 20191Feb 20, 20191