Online contextual influence maximization with costly observations

buir.contributor.authorSarıtaç, Anıl Ömer
buir.contributor.authorKarakurt, Altuğ
buir.contributor.authorTekin, Cem
dc.citation.epage289en_US
dc.citation.issueNumber2en_US
dc.citation.spage273en_US
dc.citation.volumeNumber5en_US
dc.contributor.authorSarıtaç, Anıl Ömeren_US
dc.contributor.authorKarakurt, Altuğen_US
dc.contributor.authorTekin, Cemen_US
dc.date.accessioned2020-01-29T10:11:34Z
dc.date.available2020-01-29T10:11:34Z
dc.date.issued2019-06
dc.departmentDepartment of Electrical and Electronics Engineeringen_US
dc.departmentDepartment of Industrial Engineeringen_US
dc.description.abstractIn the online contextual influence maximization problem with costly observations, the learner faces a series of epochs in each of which a different influence spread process takes place over a network. At the beginning of each epoch, the learner exogenously influences (activates) a set of seed nodes in the network. Then, the influence spread process takes place over the network, through which other nodes get influenced. The learner has the option to observe the spread of influence by paying an observation cost. The goal of the learner is to maximize its cumulative reward, which is defined as the expected total number of influenced nodes over all epochs minus the observation costs. We depart from the prior work in three aspects: 1) the learner does not know how the influence spreads over the network, i.e., it is unaware of the influence probabilities; 2) influence probabilities depend on the context; and 3) observing influence is costly. We consider two different influence observation settings: costly edge-level feedback, in which the learner freely observes the set of influenced nodes, but pays to observe the influence outcomes on the edges of the network; and costly node-level feedback, in which the learner pays to observe whether a node is influenced or not. Since the offline influence maximization problem itself is NP-hard, for these settings, we develop online learning algorithms that use an approximation algorithm as a subroutine to obtain the set of seed nodes in each epoch. When the influence probabilities are Hölder continuous functions of the context, we prove that these algorithms achieve sublinear regret (for any sequence of contexts) with respect to an approximation oracle that knows the influence probabilities for all contexts. Our numerical results on several networks illustrate that the proposed algorithms perform on par with the state-of-the-art methods even when the observations are cost free.en_US
dc.identifier.doi10.1109/TSIPN.2018.2866334en_US
dc.identifier.issn2373-776X
dc.identifier.urihttp://hdl.handle.net/11693/52895
dc.language.isoEnglishen_US
dc.publisherIEEEen_US
dc.relation.isversionofhttps://doi.org/10.1109/TSIPN.2018.2866334en_US
dc.source.titleIEEE Transactions on Signal and Information Processing over Networksen_US
dc.subjectInfluence maximizationen_US
dc.subjectCombinatorial banditsen_US
dc.subjectSocial networksen_US
dc.subjectApproximation algorithmsen_US
dc.subjectCostly observationsen_US
dc.subjectRegret boundsen_US
dc.titleOnline contextual influence maximization with costly observationsen_US
dc.typeArticleen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Online_Contextual_Influence_Maximization_with_Costly_Observations.pdf
Size:
1.01 MB
Format:
Adobe Portable Document Format
Description: