<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Epidemic Modeling | Junfei Zhan's Website</title><link>https://junfei-z.github.io/tags/epidemic-modeling/</link><atom:link href="https://junfei-z.github.io/tags/epidemic-modeling/index.xml" rel="self" type="application/rss+xml"/><description>Epidemic Modeling</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Thu, 23 Apr 2026 00:00:00 +0000</lastBuildDate><image><url>https://junfei-z.github.io/media/icon_hu70bcee51a3cd7a7338014254a2e0c844_1401285_512x512_fill_lanczos_center_3.png</url><title>Epidemic Modeling</title><link>https://junfei-z.github.io/tags/epidemic-modeling/</link></image><item><title>Scalable Node-Level Vaccine Allocation on Contact Networks: Bridging Optimal Control and Reinforcement Learning</title><link>https://junfei-z.github.io/research/scalable-node-level-vaccine-allocation-on-contact-networks/</link><pubDate>Thu, 23 Apr 2026 00:00:00 +0000</pubDate><guid>https://junfei-z.github.io/research/scalable-node-level-vaccine-allocation-on-contact-networks/</guid><description>&lt;a href="https://junfei-z.github.io/vaccine_rl/" target="_blank">
&lt;img src="https://img.shields.io/badge/Interactive%20Demo-Open-2563eb?logo=googlechrome&amp;logoColor=white" alt="Demo">
&lt;/a>
&lt;p>📄 &lt;em>Master&amp;rsquo;s Thesis, University of Pennsylvania (2026). Advisor: Prof. Saswati Sarkar.&lt;/em>&lt;/p>
&lt;p>In the first weeks of a pandemic, vaccines must be allocated across a large, heterogeneous population under a tight daily dose budget and over a horizon of weeks to months. A deployable policy must name specific individuals — not group-level proportions — and cope with three structural difficulties: sequential decisions over a long horizon with a delayed reward signal, a combinatorial daily action space of size $\binom{N}{K}$, and individual network position that matters as much as demographic group.&lt;/p>
&lt;h2 id="interactive-demo">Interactive Demo&lt;/h2>
&lt;p>The companion demo walks through the thesis visually:&lt;/p>
&lt;ol>
&lt;li>&lt;strong>Three-group population model&lt;/strong> — baseline (X), high-risk elderly (Y), and high-contact hubs (Z), each with group-specific symptomatic, hospitalisation, and case-fatality rates.&lt;/li>
&lt;li>&lt;strong>10-compartment SEPAILHRVD disease model&lt;/strong> — latent, pre-symptomatic, asymptomatic, symptomatic, late-stage, hospitalised, recovered, vaccinated, and dead.&lt;/li>
&lt;li>&lt;strong>Barabási–Albert network construction&lt;/strong> — watch preferential attachment grow a scale-free contact graph and the characteristic power-law degree tail emerge.&lt;/li>
&lt;li>&lt;strong>Stochastic simulator&lt;/strong> — seed infections in any group mix and watch an unvaccinated outbreak unfold day by day, reporting cumulative deaths as the no-intervention baseline.&lt;/li>
&lt;li>&lt;strong>Method comparison&lt;/strong> &lt;em>(coming soon)&lt;/em> — OC-Random, OC-high, Naive RL, and Node RL on identical seeds.&lt;/li>
&lt;/ol>
&lt;p>👉 &lt;a href="https://junfei-z.github.io/vaccine_rl/">&lt;strong>Open the interactive demo&lt;/strong>&lt;/a>&lt;/p>
&lt;h2 id="contributions">Contributions&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>C1 — Stochastic node-level simulator&lt;/strong>: a high-fidelity environment integrating an explicit Barabási–Albert contact network with a 10-compartment SEPAILHRVD model, capturing intrinsic stochasticity of infection events and individual-level risk heterogeneity.&lt;/li>
&lt;li>&lt;strong>C2 — OC-high&lt;/strong>: augments principled group-level optimal control with a high-degree-first intra-group heuristic, bridging aggregate policy and individual action.&lt;/li>
&lt;li>&lt;strong>C3 — Node RL&lt;/strong>: an end-to-end actor–critic with a shared-parameter scoring MLP and Gumbel-Top-$K$ reparameterised sampling, yielding $O(K)$ gradient variance versus $\Theta(N)$ for independent Bernoulli baselines.&lt;/li>
&lt;li>&lt;strong>C4 — Regime map&lt;/strong>: systematic benchmarking across population size, horizon, and initial-infection ratio identifying when each method is preferable — and when the additional compute of node-level RL is justified.&lt;/li>
&lt;/ul>
&lt;h2 id="headline-findings">Headline Findings&lt;/h2>
&lt;ul>
&lt;li>OC-high matches or beats Node RL in most regimes at roughly &lt;strong>two orders of magnitude&lt;/strong> less preparation cost.&lt;/li>
&lt;li>Node RL&amp;rsquo;s advantage is real but &lt;strong>confined&lt;/strong> to short horizons and hub-heavy initial infections, where the mean-field assumption underlying OC-high structurally breaks down.&lt;/li>
&lt;li>The intra-group high-degree heuristic alone accounts for a &lt;strong>5–10% reduction in deaths&lt;/strong> on average, comparable to the contribution of the group-level OC rates themselves.&lt;/li>
&lt;/ul></description></item><item><title>RL for Stochastic Vaccine Allocation on Contact Networks</title><link>https://junfei-z.github.io/project/2_stock/</link><pubDate>Mon, 17 Mar 2025 00:00:00 +0000</pubDate><guid>https://junfei-z.github.io/project/2_stock/</guid><description>&lt;p>Bridged deterministic optimal control and reinforcement learning to develop a stochastic vaccine allocation strategy on individual-level contact networks, enabling robust pandemic response modeling.&lt;/p>
&lt;h2 id="highlights">Highlights&lt;/h2>
&lt;ul>
&lt;li>Modeled epidemic spread using a high-dimensional continuous-time Markov process (CTMP) on a contact graph.&lt;/li>
&lt;li>Designed a vaccination policy using policy gradient-based RL, warm-started from a mean-field ODE solution.&lt;/li>
&lt;li>Evaluated policies on metrics like mortality and hospitalizations across synthetic and real-world network topologies.&lt;/li>
&lt;/ul>
&lt;h2 id="tools">Tools&lt;/h2>
&lt;p>Python, PyTorch, NetworkX, OpenAI Gym&lt;/p></description></item><item><title>基于 Reinforcement Learning 的接触网络随机疫苗分配策略</title><link>https://junfei-z.github.io/zh/project/2_stock/</link><pubDate>Mon, 17 Mar 2025 00:00:00 +0000</pubDate><guid>https://junfei-z.github.io/zh/project/2_stock/</guid><description>&lt;p>将确定性最优控制与 Reinforcement Learning 相结合，开发了个体级接触网络上的随机疫苗分配策略，实现了鲁棒的疫情响应建模。&lt;/p>
&lt;h2 id="项目亮点">项目亮点&lt;/h2>
&lt;ul>
&lt;li>在接触图上使用高维连续时间马尔可夫过程 (CTMP) 对疫情传播进行建模。&lt;/li>
&lt;li>设计了基于 Policy Gradient 的 RL 疫苗接种策略，并以 Mean-Field ODE 解作为热启动。&lt;/li>
&lt;li>在合成和真实世界网络拓扑上评估了策略在死亡率和住院率等指标上的表现。&lt;/li>
&lt;/ul>
&lt;h2 id="工具">工具&lt;/h2>
&lt;p>Python, PyTorch, NetworkX, OpenAI Gym&lt;/p></description></item></channel></rss>