guide:48fbf088c9: Difference between revisions

From Stochiki
No edit summary
mNo edit summary
Line 56: Line 56:
<div class="text-right">
<div class="text-right">


<proofs page = "guide_proofs:55aeb8bb60" section = "mgf/collective_risk_model"  label = "Moment Generating Function for Collective Risk Model" />  
<proofs page = "guide_proofs:48fbf088c9" section = "mgf/collective_risk_model"  label = "Moment Generating Function for Collective Risk Model" />  


</div>
</div>

Revision as of 20:05, 21 May 2023

An insurance company will sell insurance policies that enable the purchaser to make a claim when a certain well-defined event occurs within a specified time period. An actuary will be called upon to examine the distributional (stochastic) properties of portfolios of such policies. We consider two basic approaches: the individual risk model and the collective risk model. The individual risk model is less granular than the collective risk model since it doesn't try to model claim frequency and claim size (severity) separately. We will denote by [math]S[/math] the random variable representing the aggregate claims (total loss to the insurer) associated with a portfolio of policies.

The Individual Risk Model

In the individual risk model, we let [math]X_i[/math] represent the claim (loss) size associated with the [math]i\textrm{th}[/math] policy in a portfolio of [math]n[/math] policies:

[[math]]S = \sum_{i=1}^n X_i \, .[[/math]]

It is assumed that [math]X_1,\ldots,X_n[/math] are mutually independent but not necessarily identically distributed.

The Collective Risk Model

In the collective risk model, we let [math]N[/math] denote the random variable representing the number of individual claims and let [math]Y_i[/math] represent the claim (loss) size associated with the [math]i\textrm{th}[/math] individual claim:

[[math]]S = \sum_{i=1}^N Y_i \,.[[/math]]

The following additional constraints are also assumed:

  • [math]Y_1,\ldots,Y_n[/math] are mutually independent and identically distributed.
  • [math]N[/math] is independent of each [math]Y_i[/math].

Relation to Compound Distributions

A compound probability distribution is the probability distribution that results from assuming that a random variable is distributed according to some parametrized distribution, with the parameters of that distribution being assumed to be themselves random variables. The compound distribution is the result of marginalizing over the intermediate random variables that represent the parameters of the initial distribution. An important type of compound distribution occurs when the parameter being marginalized over represents the number of random variables in a summation of random variables as is the case with the collective risk model.

Mean and Variance

Mean and variance of the compound distribution derive in a simple way from the law of total expectation and the law of total variance. The mean equals

[[math]] \mu_S = \operatorname{E}[S] = \operatorname{E}[N]\operatorname{E}[Y_i] = \mu_N \mu_Y [[/math]]

and the variance equals

[[math]] \begin{align*} \sigma^2_S = \operatorname{Var}[S] &= \operatorname{E}[N]\operatorname{Var}[Y_i] + \operatorname{E}[Y_i]^2 \operatorname{Var}[N] \\ &=\mu_N \sigma^2_Y + \mu^2_Y \sigma_N^2. \end{align*} [[/math]]

Moment Generating Function

The moment generation function of [math]S[/math] can be expressed in terms of the moment generating functions of [math]Y[/math] and [math]N[/math]:

[[math]] M_{S}(t) = M_N\left(\ln\left(M_Y(t)\right)\right). [[/math]]

<proofs page = "guide_proofs:48fbf088c9" section = "mgf/collective_risk_model" label = "Moment Generating Function for Collective Risk Model" />

Compound Poisson Distribution

If the claim frequency [math]N[/math] has a poisson distribution with mean [math]\theta[/math], then [math]S[/math] is said to have a compound poisson distribution.

Properties

We have (see Mean and Variance)

[[math]] \mu_S = \mu_N \mu_Y,\,\sigma^2_S = \mu_N \sigma^2_Y + \mu^2_Y \sigma_N^2. [[/math]]

Since [math]\operatorname{E}[N] = \operatorname{Var}[N][/math], these formulae can be reduced to:

[[math]]\mu_S = \theta \mu_Y,\,\sigma^2_S = \theta \operatorname{E}\left[Y^2\right].[[/math]]

The probability generating function of [math]S[/math] has a simple representation in terms of the probability generation function of [math]Y[/math]:

[[math]]P_S(t) = \textrm{e}^{\lambda(P_Y(t) - 1)}.[[/math]]

Compound Negative Binomial Distribution

If the claim frequency [math]N[/math] has a negative binomial distribution with parameters [math]r[/math] and [math]\beta[/math], then [math]S[/math] is said to have a compound negative binomial distribution.

Properties

Since [math]\mu_N = r\beta[/math] and [math] \sigma^2_{N} = \mu_N (1 + \beta) [/math], then

[[math]]\mu_S = r\beta \mu_Y,\,\sigma^2_S = r\beta \left(\beta \operatorname{E}[Y]^2 + \operatorname{E}[Y^2] \right).[[/math]]

The probability generating function of [math]S[/math] equals

[[math]]P_S(t) = [1 - \beta(P_Y(t) - 1)]^{-r}.\,[[/math]]

Compound Binomial Distribution

If the claim frequency [math]N[/math] has a binomial distribution with size parameter [math]m[/math] and success parameter [math]q[/math], then [math]S[/math] is said to have a compound binomial distribution.

Properties

Since [math]\mu_N = pn[/math] and [math] \sigma^2_{N} = \mu_N (1-p) [/math], then

[[math]]\mu_S = np \mu_Y,\,\sigma^2_S = np \left(\operatorname{E}[Y^2] - p \operatorname{E}[Y]^2 \right).[[/math]]

The probability generating function of [math]S[/math] can be computed:

[[math]]P_S(t) = [1 + p(P_Y(t) - 1)]^{n}.\,[[/math]]

Normal Approximation

It would be useful to consider a simple method that approximates the distribution function for [math]S[/math], which is usually difficult to compute. One basic approach is to use a normal approximation for the distribution of [math]S[/math]. To understand why such an approximation is used, we first consider the classical central limit theorem.

Classical Central Limit Theorem

Suppose [math]X_1,X_2,\ldots[/math] is a sequence of independent and identically distributed random variables with [math]\operatorname{E}[X_i] = \mu[/math] and [math]\operatorname{Var}[X_i] = \sigma_2 \lt \infty[/math]. If [math]S_n = \sum_{i=1}^nX_i[/math] and [math]Z[/math] is a standard normal, then

[[math]] \begin{equation}\label{clt} \operatorname{P}\left( \frac{S_n - n\mu}{\sigma \sqrt{n}} \leq c \right) \rightarrow \operatorname{P}\left( Z \leq c \right)\quad \text{as} \,\,n\rightarrow \infty \end{equation} [[/math]]

In other words, if we take the random sum [math]S_n[/math], center it by subtracting by its expected value [math]n\mu[/math], and then divide by its standard deviation [math]\sigma \sqrt{n}[/math], we obtain a random variable which is approximately normally distributed.

The Method

The conditions required for the classical central limit theorem to hold aren't necessarily met in both the individual risk model and the collective risk model. The normal approximation method for [math]S[/math] tries to mimic the approach in the classical central limit theorem by centering [math]S[/math] and dividing by its standard deviation:

Normal Approximation
  1. Compute [math]\mu_S[/math] and [math] \sigma_S^2 [/math].
  2. Set [math] Z = \frac{S - \mu_S}{\sigma_S} [/math].
  3. Use the standard normal distribution table to get approximate percentiles for [math]Z[/math].

References

  • Wikipedia contributors. "Central limit theorem". Wikipedia. Wikipedia. Retrieved 8 June 2019.