The influence maximization (IM) problem aims at finding a subset of seed nodes in a social network that maximize the spread of influence. In this study, we focus on a sub-class of IM problems, where whether the nodes are willing to be the seeds when being invited is uncertain, called contingency-aware IM. Such contingency aware IM is critical for applications for non-profit organizations in low resource communities (e.g., spreading awareness of disease prevention). Despite the initial success, a major practical obstacle in promoting the solutions to more communities is the tremendous runtime of the greedy algorithms and the lack of high performance computing (HPC) for the non-profits in the field -- whenever there is a new social network, the non-profits usually do not have the HPCs to recalculate the solutions. Motivated by this and inspired by the line of works that use reinforcement learning (RL) to address combinatorial optimization on graphs, we formalize the problem as a Markov Decision Process (MDP), and use RL to learn an IM policy over historically seen networks, and generalize to unseen networks with negligible runtime at test phase. To fully exploit the properties of our targeted problem, we propose two technical innovations that improve the existing methods, including state-abstraction and theoretically grounded reward shaping. Empirical results show that our method achieves influence as high as the state-of-the-art methods for contingency-aware IM, while having negligible runtime at test phase.
翻译:影响最大化(IM)问题的目的是在社会网络中找到一组种子节点,使影响力的传播最大化。在本研究中,我们侧重于一组IM问题,即节点在被邀请时是否愿意成为种子的不确定因素,称为应急意识IM。这种了解IM的应急意识对于低资源社区非营利组织的应用(例如,传播疾病预防意识)至关重要。尽管最初取得了成功,但促进解决更多社区问题的一个重大实际障碍是贪婪算法的巨大运行时间和外地非营利机构缺乏高性能计算(HPC)的问题 -- -- 只要有新的社会网络,非营利机构通常没有HPC来重新计算解决方案。受这种应急意识的影响,并受到利用强化学习(RL)解决图表上的组合优化问题的工作方针的启发,我们把问题正式化为Markawa决策进程(MDP),并使用RL来学习IM政策超越历史所见的网络,以及普遍化为无形网络,在测试阶段,非盈利组织通常没有HPC来重新计算解决方案。我们用微量的测试方法,在测试阶段,包括测试阶段,我们用目前的标准方法来彻底改进。