File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1017/S1748499523000027
- Scopus: eid_2-s2.0-85150355424
- Find via

Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Article: Pseudo-model-free hedging for variable annuities via deep reinforcement learning
| Title | Pseudo-model-free hedging for variable annuities via deep reinforcement learning |
|---|---|
| Authors | |
| Keywords | Hedging strategy self-revision Online learning phase Sequential anchor-hedging reward signals Single terminal reward signals Training phase Two-phase deep reinforcement learning Variable annuities hedging |
| Issue Date | 2023 |
| Citation | Annals of Actuarial Science, 2023, v. 18, n. 2 How to Cite? |
| Abstract | This paper proposes a two-phase deep reinforcement learning approach, for hedging variable annuity contracts with both GMMB and GMDB riders, which can address model miscalibration in Black-Scholes financial and constant force of mortality actuarial market environments. In the training phase, an infant reinforcement learning agent interacts with a pre-designed training environment, collects sequential anchor-hedging reward signals, and gradually learns how to hedge the contracts. As expected, after a sufficient number of training steps, the trained reinforcement learning agent hedges, in the training environment, equally well as the correct Delta while outperforms misspecified Deltas. In the online learning phase, the trained reinforcement learning agent interacts with the market environment in real time, collects single terminal reward signals, and self-revises its hedging strategy. The hedging performance of the further trained reinforcement learning agent is demonstrated via an illustrative example on a rolling basis to reveal the self-revision capability on the hedging strategy by online learning. |
| Persistent Identifier | http://hdl.handle.net/10722/363518 |
| ISSN | 2023 Impact Factor: 1.5 2023 SCImago Journal Rankings: 0.730 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Chong, Wing Fung | - |
| dc.contributor.author | Cui, Haoen | - |
| dc.contributor.author | Li, Yuxuan | - |
| dc.date.accessioned | 2025-10-10T07:47:31Z | - |
| dc.date.available | 2025-10-10T07:47:31Z | - |
| dc.date.issued | 2023 | - |
| dc.identifier.citation | Annals of Actuarial Science, 2023, v. 18, n. 2 | - |
| dc.identifier.issn | 1748-4995 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/363518 | - |
| dc.description.abstract | This paper proposes a two-phase deep reinforcement learning approach, for hedging variable annuity contracts with both GMMB and GMDB riders, which can address model miscalibration in Black-Scholes financial and constant force of mortality actuarial market environments. In the training phase, an infant reinforcement learning agent interacts with a pre-designed training environment, collects sequential anchor-hedging reward signals, and gradually learns how to hedge the contracts. As expected, after a sufficient number of training steps, the trained reinforcement learning agent hedges, in the training environment, equally well as the correct Delta while outperforms misspecified Deltas. In the online learning phase, the trained reinforcement learning agent interacts with the market environment in real time, collects single terminal reward signals, and self-revises its hedging strategy. The hedging performance of the further trained reinforcement learning agent is demonstrated via an illustrative example on a rolling basis to reveal the self-revision capability on the hedging strategy by online learning. | - |
| dc.language | eng | - |
| dc.relation.ispartof | Annals of Actuarial Science | - |
| dc.subject | Hedging strategy self-revision | - |
| dc.subject | Online learning phase | - |
| dc.subject | Sequential anchor-hedging reward signals | - |
| dc.subject | Single terminal reward signals | - |
| dc.subject | Training phase | - |
| dc.subject | Two-phase deep reinforcement learning | - |
| dc.subject | Variable annuities hedging | - |
| dc.title | Pseudo-model-free hedging for variable annuities via deep reinforcement learning | - |
| dc.type | Article | - |
| dc.description.nature | link_to_subscribed_fulltext | - |
| dc.identifier.doi | 10.1017/S1748499523000027 | - |
| dc.identifier.scopus | eid_2-s2.0-85150355424 | - |
| dc.identifier.volume | 18 | - |
| dc.identifier.issue | 2 | - |
| dc.identifier.eissn | 1748-5002 | - |
