Search

A Nonzero-sum game with reinforcement learning under mean-variance framework

<?xml version="1.0" encoding="UTF-8"?><modsCollection xmlns="http://www.loc.gov/mods/v3" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.loc.gov/mods/v3 http://www.loc.gov/standards/mods/v3/mods-3-8.xsd">
<mods version="3.8">
<titleInfo>
<nonSort xml:space="preserve">A  </nonSort>
<title>Nonzero-sum game with reinforcement learning under mean-variance framework</title>
</titleInfo>
<name type="personal" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080649876">
<namePart>Guo, Junyi</namePart>
<nameIdentifier>MAPA20080649876</nameIdentifier>
</name>
<name type="corporate" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20100017661">
<namePart>International Actuarial Association</namePart>
<nameIdentifier>MAPA20100017661</nameIdentifier>
</name>
<typeOfResource>text</typeOfResource>
<genre authority="marcgt">periodical</genre>
<originInfo>
<place>
<placeTerm type="code" authority="marccountry">bel</placeTerm>
</place>
<dateIssued encoding="marc">2026</dateIssued>
<issuance>serial</issuance>
</originInfo>
<language>
<languageTerm type="code" authority="iso639-2b">eng</languageTerm>
</language>
<physicalDescription>
<form authority="marcform">print</form>
</physicalDescription>
<abstract displayLabel="Summary">This paper examines a competitive setting in which two agents invest in a risk-free and a risky asset while considering both their own wealth and their wealth gap relative to each other. With market parameters partially or fully unknown, the problem is formulated as a nonzero-sum differential game within a reinforcement learning framework. Each agent seeks to optimize a Choquet-regularized, time-inconsistent mean-variance objective. Using dynamic programming, the authors derive a time-consistent Nash equilibrium in an incomplete market. Under a Gaussian mean-return assumption, they obtain an explicit analytical solution that enables the construction of a practical reinforcement learning algorithm. The algorithm shows uniform convergence, despite the absence of a traditional policy-improvement guarantee, and numerical experiments confirm its robustness and effectiveness</abstract>
<note type="statement of responsibility">Junyi Guo...[et al.]</note>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080597641">
<topic>Mercados financieros</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20250003316">
<topic>Gestión de riesgos</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080586447">
<topic>Modelo estocástico</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080576790">
<topic>Modelo Gaussiano</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080579258">
<topic>Cálculo actuarial</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080602437">
<topic>Matemática del seguro</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080592042">
<topic>Modelos matemáticos</topic>
</subject>
<classification authority="">6</classification>
<relatedItem type="host">
<titleInfo>
<title>Astin bulletin</title>
</titleInfo>
<originInfo>
<publisher>Belgium : ASTIN and AFIR Sections of the International Actuarial Association</publisher>
</originInfo>
<identifier type="issn">0515-0361</identifier>
<identifier type="local">MAP20077000420</identifier>
<part>
<text>19/01/2026 Volume 56 Issue 1 - January 2026 , p. 154 - 180</text>
</part>
</relatedItem>
<recordInfo>
<recordContentSource authority="marcorg">MAP</recordContentSource>
<recordCreationDate encoding="marc">260202</recordCreationDate>
<recordChangeDate encoding="iso8601">20260205101750.0</recordChangeDate>
<recordIdentifier source="MAP">MAP20260001968</recordIdentifier>
<languageOfCataloging>
<languageTerm type="code" authority="iso639-2b">spa</languageTerm>
</languageOfCataloging>
</recordInfo>
</mods>
</modsCollection>