Búsqueda

Using machine learning to better model long-term care insurance claims

<?xml version="1.0" encoding="UTF-8"?><modsCollection xmlns="http://www.loc.gov/mods/v3" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.loc.gov/mods/v3 http://www.loc.gov/standards/mods/v3/mods-3-8.xsd">
<mods version="3.8">
<titleInfo>
<title>Using machine learning to better model long-term care insurance claims</title>
</titleInfo>
<name type="personal" usage="primary" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20220008242">
<namePart>Cummings, Jared</namePart>
<nameIdentifier>MAPA20220008242</nameIdentifier>
</name>
<typeOfResource>text</typeOfResource>
<genre authority="marcgt">periodical</genre>
<originInfo>
<place>
<placeTerm type="code" authority="marccountry">esp</placeTerm>
</place>
<dateIssued encoding="marc">2022</dateIssued>
<issuance>serial</issuance>
</originInfo>
<language>
<languageTerm type="code" authority="iso639-2b">spa</languageTerm>
</language>
<physicalDescription>
<form authority="marcform">print</form>
</physicalDescription>
<abstract displayLabel="Summary">Long-term care insurance (LTCI) should be an essential part of a family financial plan. It could protect assets from the expensive and relatively common risk of needing disability assistance, and LTCI purchase rates are lower than expected. Though there are multiple reasons for this trend, it is partially due to the difficultly insurers have in operating profitably as LTCI providers. If LTCI providers were better able to forecast claim rates, they would have less difficulty maintaining profitability. In this article, we develop several models to improve upon those used by insurers to forecast claim rates. We find that standard logistic regression is outperformed by tree-based and neural network models. More modest improvements can be found by using a neighbor-based model. Of all of our tested models, the random forest models were the consistent top performers. Additionally, simple sampling techniques influence the performance of each of the models. This is especially true for the deep neural network, which improves drastically under oversampling. The effects of the sampling vary depending on the size of the available data. To better understand this relationship, we thoroughly examine three states with various amounts of available data as case studies.

</abstract>
<note type="statement of responsibility">Jared Cummings</note>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20170005476">
<topic>Machine learning</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080567118">
<topic>Reclamaciones</topic>
</subject>
<subject xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="MAPA20080573867">
<topic>Seguro de salud</topic>
</subject>
<classification authority="">6</classification>
<relatedItem type="host">
<titleInfo>
<title>North American actuarial journal</title>
</titleInfo>
<originInfo>
<publisher>Schaumburg : Society of Actuaries, 1997-</publisher>
</originInfo>
<identifier type="issn">1092-0277</identifier>
<identifier type="local">MAP20077000239</identifier>
<part>
<text>12/09/2022 Tomo 26 Número 3 - 2022 , p. 470-483</text>
</part>
</relatedItem>
<recordInfo>
<recordContentSource authority="marcorg">MAP</recordContentSource>
<recordCreationDate encoding="marc">220916</recordCreationDate>
<recordChangeDate encoding="iso8601">20220916085518.0</recordChangeDate>
<recordIdentifier source="MAP">MAP20220023795</recordIdentifier>
<languageOfCataloging>
<languageTerm type="code" authority="iso639-2b">spa</languageTerm>
</languageOfCataloging>
</recordInfo>
</mods>
</modsCollection>