Sarıtaç, A. ÖmerTekin, Cem2019-02-212019-02-212017-11http://hdl.handle.net/11693/50176Date of Conference: 14-16 Nov. 2017Conference name: 2017 IEEE Global Conference on Signal and Information Processing (GlobalSIP)In this paper, we study the combinatorial multi-armed bandit problem (CMAB) with probabilistically triggered arms (PTAs). Under the assumption that the arm triggering probabilities (ATPs) are positive for all arms, we prove that a simple greedy policy, named greedy CMAB (G-CMAB), achieves bounded regret. This improves the result in previous work, which shows that the regret is O (log T) under no such assumption on the ATPs. Then, we numerically show that G-CMAB achieves bounded regret in a real-world movie recommendation problem, where the action corresponds to recommending a set of movies, arms correspond to the edges between movies and users, and the goal is to maximize the total number of users that are attracted by at least one movie. In addition to this problem, our results directly apply to the online influence maximization (OIM) problem studied in numerous prior works.EnglishBounded regretCombinatorial multi-armed banditOnline learningProbabilistically triggered armsCombinatorial multi-armed bandit problem with probabilistically triggered arms: a case with bounded regretConference Paper10.1109/GlobalSIP.2017.8308614