Parallel tempering (PT), also known as replica exchange, is the go-to workhorse for simulations of multi-modal distributions. The key to the success of PT is to adopt efficient swap schemes. The popular deterministic even-odd (DEO) scheme exploits the non-reversibility property and has successfully reduced the communication cost from $O(P^2)$ to $O(P)$ given sufficiently many $P$ chains. However, such an innovation largely disappears in big data due to the limited chains and few bias-corrected swaps. To handle this issue, we generalize the DEO scheme to promote non-reversibility and propose a few solutions to tackle the underlying bias caused by the geometric stopping time. Notably, in big data scenarios, we obtain an appealing communication cost $O(P\log P)$ based on the optimal window size. In addition, we also adopt stochastic gradient descent (SGD) with large and constant learning rates as exploration kernels. Such a user-friendly nature enables us to conduct approximation tasks for complex posteriors without much tuning costs.
翻译:平行温带(PT)也称为复制交换,是模拟多模式分布的“上到工作马”,是PT成功的关键,是采用高效互换计划。流行的确定性均价(DEO)计划利用了不可逆性财产,成功地将通信成本从2美元减为1美元(P),因为有相当多的P美元链。然而,这种创新在很大程度上在大数据中消失,因为链条有限,且很少有纠正偏差的互换。为了处理这一问题,我们推广DEO计划,以促进不可逆性,并提出一些解决方案,以解决几何停止时间造成的根本偏见。值得注意的是,在大数据假设中,我们获得了基于最佳窗口规模的具有吸引力的通信成本(PP)$(P),此外,我们还采用以大量和持续学习率的梯度梯度下降(SGD)作为勘探内核。这种方便用户的性质使我们能够在不需大量调整成本的情况下为复杂的远方远方远方远方的远方远方远方远方远方远方远方远方远方远方远方远方相。