Insurance

The insurance problem is a benchmark in stochastic optimal control with applications in actuarial science and financial planning. It models the dynamic interplay between insurance coverage, medical expenses, health, revenue, and utility over time. Both the state trajectory $x(\cdot)$ and the control $u(\cdot)$ are decision variables. The aim is to maximise the expected utility subject to constraints on insurance, expenses, health, revenue, and marginal utility.

The problem can be written as

\[\min_{x,\,u} J(x,u) = - \int_0^{t_f} U(t) \cdot f_x(t) \, dt\]

subject to the dynamics

\[\dot{x}_1(t) = \bigl(1 - \gamma \, t \, v'(m(t)) / (dU/dR)(t)\bigr) \, h(t), \quad \dot{x}_2(t) = h(t), \quad \dot{x}_3(t) = (1 + \sigma) \, I(t) \, f_x(t),\]

with boundary conditions

\[x_1(0) = 0, \quad x_2(0) = 0.001, \quad x_3(0) = 0, \qquad x_3(t_f) = P,\]

and the constraints

\[0 \le I(t) \le 1.5, \quad 0 \le m(t) \le 1.5, \quad 0 \le h(t) \le 25, \quad 0 \le R(t), H(t), U(t), \quad 0.001 \le dU/dR(t),\]

where $f_x(t)$ is the illness distribution and $v(m) = m^{\alpha/2} / (1 + m^{\alpha/2})$.

Qualitative behaviour

  • Insurance $I(t)$ is dynamically adjusted to balance risk coverage and costs.
  • Medical expenses $m(t)$ interact with health $h(t)$ and influence utility.
  • Utility $U(t)$ and revenue $R(t)$ are coupled through marginal utility $dU/dR(t)$.
  • The system illustrates the trade-offs between health, wealth, and insurance over time.

Characteristics

  • Nonlinear dynamics linking health, expenses, and utility.
  • Multiple control variables with inequality constraints.
  • Stochastic component in the illness distribution $f_x(t)$.
  • Useful as a benchmark for actuarial optimal control problems.

References

  • Schmidli, H. (2014). Stochastic Control for Insurance Companies. Wiley.
  • Li, W. (2023). Individual Insurance Choice: A Stochastic Control Approach. University of Waterloo.
  • Guerdouh, D. (2022). Optimal Control Strategies for the Premium Policy of an Insurance Firm. MDPI.

Packages

Import all necessary packages and define DataFrames to store information about the problem and resolution results.

using OptimalControlProblems    # to access the Beam model
using OptimalControl            # to import the OptimalControl model
using NLPModelsIpopt            # to solve the model with Ipopt
import DataFrames: DataFrame    # to store data
using NLPModels                 # to retrieve data from the NLP solution
using Plots                     # to plot the trajectories
using Plots.PlotMeasures        # for leftmargin, bottommargin
using JuMP                      # to import the JuMP model
using Ipopt                     # to solve the JuMP model with Ipopt

data_pb = DataFrame(            # to store data about the problem
    Problem=Symbol[],
    Grid_Size=Int[],
    Variables=Int[],
    Constraints=Int[],
)

data_re = DataFrame(            # to store data about the resolutions
    Model=Symbol[],
    Flag=Any[],
    Iterations=Int[],
    Objective=Float64[],
)

Initial guess

The initial guess (or first iterate) can be visualised by running the solver with max_iter=0. Here is the initial guess.

Click to unfold and see the code for plotting the initial guess.
function plot_initial_guess(problem)

    # dimensions
    x_vars = metadata[problem][:state_name]
    u_vars = metadata[problem][:control_name]
    n = length(x_vars) # number of states
    m = length(u_vars) # number of controls

    # import OptimalControl model
    docp = eval(problem)(OptimalControlBackend())
    nlp_oc = nlp_model(docp)

    # solve
    nlp_oc_sol = NLPModelsIpopt.ipopt(nlp_oc; max_iter=0)

    # build an optimal control solution
    ocp_sol = build_ocp_solution(docp, nlp_oc_sol)

    # plot the OptimalControl solution
    plt = plot(
        ocp_sol;
        state_style=(color=1,),
        costate_style=(color=1, legend=:none),
        control_style=(color=1, legend=:none),
        path_style=(color=1, legend=:none),
        dual_style=(color=1, legend=:none),
        size=(816, 220*(n+m)),
        label="OptimalControl",
        leftmargin=20mm,
    )
    for i in 2:n
        plot!(plt[i]; legend=:none)
    end

    # import JuMP model
    nlp_jp = eval(problem)(JuMPBackend())

    # solve
    set_optimizer(nlp_jp, Ipopt.Optimizer)
    set_optimizer_attribute(nlp_jp, "max_iter", 0)
    optimize!(nlp_jp)

    # plot
    t = time_grid(problem, nlp_jp)     # t0, ..., tN = tf
    x = state(problem, nlp_jp)         # function of time
    u = control(problem, nlp_jp)       # function of time
    p = costate(problem, nlp_jp)       # function of time

    for i in 1:n # state
        label = i == 1 ? "JuMP" : :none
        plot!(plt[i], t, t -> x(t)[i]; color=2, linestyle=:dash, label=label)
    end

    for i in 1:n # costate
        plot!(plt[n+i], t, t -> -p(t)[i]; color=2, linestyle=:dash, label=:none)
    end

    for i in 1:m # control
        plot!(plt[2n+i], t, t -> u(t)[i]; color=2, linestyle=:dash, label=:none)
    end

    return plt
end

plot_initial_guess(:insurance)
Example block output

Solve the problem

OptimalControl model

Import the OptimalControl model and solve it.

# import DOCP model
docp = insurance(OptimalControlBackend())

# get NLP model
nlp_oc = nlp_model(docp)

# solve
nlp_oc_sol = NLPModelsIpopt.ipopt(
    nlp_oc;
    print_level=4,
    tol=1e-8,
    mu_strategy="adaptive",
    sb="yes",
)
Total number of variables............................:     4009
                     variables with only lower bounds:     2005
                variables with lower and upper bounds:     1503
                     variables with only upper bounds:        0
Total number of equality constraints.................:     3508
Total number of inequality constraints...............:        0
        inequality constraints with only lower bounds:        0
   inequality constraints with lower and upper bounds:        0
        inequality constraints with only upper bounds:        0


Number of Iterations....: 96

                                   (scaled)                 (unscaled)
Objective...............:  -2.0592143332449284e+00   -2.0592143332449284e+00
Dual infeasibility......:   1.4381106078523287e-09    1.4381106078523287e-09
Constraint violation....:   7.0074213098791915e-10    7.0074213098791915e-10
Variable bound violation:   7.7670225963439427e-37    7.7670225963439427e-37
Complementarity.........:   1.0235615024240518e-11    1.0235615024240518e-11
Overall NLP error.......:   1.4381106078523287e-09    1.4381106078523287e-09


Number of objective function evaluations             = 98
Number of objective gradient evaluations             = 97
Number of equality constraint evaluations            = 98
Number of inequality constraint evaluations          = 0
Number of equality constraint Jacobian evaluations   = 97
Number of inequality constraint Jacobian evaluations = 0
Number of Lagrangian Hessian evaluations             = 96
Total seconds in IPOPT                               = 1.998

EXIT: Optimal Solution Found.

The problem has the following numbers of steps, variables and constraints.

push!(data_pb,
    (
        Problem=:insurance,
        Grid_Size=metadata[:insurance][:N],
        Variables=get_nvar(nlp_oc),
        Constraints=get_ncon(nlp_oc),
    )
)
1×4 DataFrame
RowProblemGrid_SizeVariablesConstraints
SymbolInt64Int64Int64
1insurance50040093508

JuMP model

Import the JuMP model and solve it.

# import model
nlp_jp = insurance(JuMPBackend())

# solve
set_optimizer(nlp_jp, Ipopt.Optimizer)
set_optimizer_attribute(nlp_jp, "print_level", 4)
set_optimizer_attribute(nlp_jp, "tol", 1e-8)
set_optimizer_attribute(nlp_jp, "mu_strategy", "adaptive")
set_optimizer_attribute(nlp_jp, "linear_solver", "mumps")
set_optimizer_attribute(nlp_jp, "sb", "yes")
optimize!(nlp_jp)
Total number of variables............................:     4009
                     variables with only lower bounds:     2005
                variables with lower and upper bounds:     1503
                     variables with only upper bounds:        0
Total number of equality constraints.................:     3508
Total number of inequality constraints...............:        0
        inequality constraints with only lower bounds:        0
   inequality constraints with lower and upper bounds:        0
        inequality constraints with only upper bounds:        0


Number of Iterations....: 87

                                   (scaled)                 (unscaled)
Objective...............:  -2.0592143331069113e+00   -2.0592143331069113e+00
Dual infeasibility......:   9.3461195306875605e-10    9.3461195306875605e-10
Constraint violation....:   2.3152775252555102e-09    2.3152775252555102e-09
Variable bound violation:   1.3176950434279923e-09    1.3176950434279923e-09
Complementarity.........:   1.2959183640751635e-11    1.2959183640751635e-11
Overall NLP error.......:   2.3152775252555102e-09    2.3152775252555102e-09


Number of objective function evaluations             = 89
Number of objective gradient evaluations             = 88
Number of equality constraint evaluations            = 89
Number of inequality constraint evaluations          = 0
Number of equality constraint Jacobian evaluations   = 88
Number of inequality constraint Jacobian evaluations = 0
Number of Lagrangian Hessian evaluations             = 87
Total seconds in IPOPT                               = 0.904

EXIT: Optimal Solution Found.

Numerical comparisons

Let's get the flag, the number of iterations and the objective value from the resolutions.

# from OptimalControl model
push!(data_re,
    (
        Model=:OptimalControl,
        Flag=nlp_oc_sol.status,
        Iterations=nlp_oc_sol.iter,
        Objective=nlp_oc_sol.objective,
    )
)

# from JuMP model
push!(data_re,
    (
        Model=:JuMP,
        Flag=termination_status(nlp_jp),
        Iterations=barrier_iterations(nlp_jp),
        Objective=objective_value(nlp_jp),
    )
)
2×4 DataFrame
RowModelFlagIterationsObjective
SymbolAnyInt64Float64
1OptimalControlfirst_order96-2.05921
2JuMPLOCALLY_SOLVED87-2.05921

We compare the OptimalControl and JuMP solutions in terms of the number of iterations, the $L^2$-norm of the differences in the state, control, and variable, as well as the objective values. Both absolute and relative errors are reported.

Click to unfold and get the code of the numerical comparison.
function L2_norm(T, X)
    # T and X are supposed to be one dimensional
    s = 0.0
    for i in 1:(length(T) - 1)
        s += 0.5 * (X[i]^2 + X[i + 1]^2) * (T[i + 1]-T[i])
    end
    return √(s)
end

function numerical_comparison(problem, docp, nlp_oc_sol, nlp_jp)

    # get relevant data from OptimalControl model
    ocp_sol = build_ocp_solution(docp, nlp_oc_sol) # build an ocp solution
    t_oc = time_grid(ocp_sol)
    x_oc = state(ocp_sol).(t_oc)
    u_oc = control(ocp_sol).(t_oc)
    v_oc = variable(ocp_sol)
    o_oc = objective(ocp_sol)
    i_oc = iterations(ocp_sol)

    # get relevant data from JuMP model
    t_jp = time_grid(problem, nlp_jp)
    x_jp = state(problem, nlp_jp).(t_jp)
    u_jp = control(problem, nlp_jp).(t_jp)
    o_jp = objective(problem, nlp_jp)
    v_jp = variable(problem, nlp_jp)
    i_jp = iterations(problem, nlp_jp)

    x_vars = metadata[problem][:state_name]
    u_vars = metadata[problem][:control_name]
    v_vars = metadata[problem][:variable_name]

    println("┌─ ", string(problem))
    println("│")

    # number of iterations
    println("├─  Number of iterations")
    println("│")
    println("│     OptimalControl : ", i_oc)
    println("│     JuMP           : ", i_jp)
    println("│")

    # state
    for i in eachindex(x_vars)
        xi_oc = [x_oc[k][i] for k in eachindex(t_oc)]
        xi_jp = [x_jp[k][i] for k in eachindex(t_jp)]
        L2_oc = L2_norm(t_oc, xi_oc)
        L2_jp = L2_norm(t_oc, xi_jp)
        L2_ae = L2_norm(t_oc, xi_oc-xi_jp)
        L2_re = L2_ae/(0.5*(L2_oc + L2_jp))

        println("├─  State $(x_vars[i]) (L2 norm)")
        println("│")
        #println("│     OptimalControl : ", L2_oc)
        #println("│     JuMP           : ", L2_jp)
        println("│     Absolute error : ", L2_ae)
        println("│     Relative error : ", L2_re)
        println("│")
    end

    # control
    for i in eachindex(u_vars)
        ui_oc = [u_oc[k][i] for k in eachindex(t_oc)]
        ui_jp = [u_jp[k][i] for k in eachindex(t_jp)]
        L2_oc = L2_norm(t_oc, ui_oc)
        L2_jp = L2_norm(t_oc, ui_jp)
        L2_ae = L2_norm(t_oc, ui_oc-ui_jp)
        L2_re = L2_ae/(0.5*(L2_oc + L2_jp))

        println("├─  Control $(u_vars[i]) (L2 norm)")
        println("│")
        #println("│     OptimalControl : ", L2_oc)
        #println("│     JuMP           : ", L2_jp)
        println("│     Absolute error : ", L2_ae)
        println("│     Relative error : ", L2_re)
        println("│")
    end

    # variable
    if !isnothing(v_vars)
        for i in eachindex(v_vars)
            vi_oc = v_oc[i]
            vi_jp = v_jp[i]
            vi_ae = abs(vi_oc-vi_jp)
            vi_re = vi_ae/(0.5*(abs(vi_oc) + abs(vi_jp)))

            println("├─  Variable $(v_vars[i])")
            println("│")
            #println("│     OptimalControl : ", vi_oc)
            #println("│     JuMP           : ", vi_jp)
            println("│     Absolute error : ", vi_ae)
            println("│     Relative error : ", vi_re)
            println("│")
        end
    end

    # objective
    o_ae = abs(o_oc-o_jp)
    o_re = o_ae/(0.5*(abs(o_oc) + abs(o_jp)))

    println("├─  objective")
    println("│")
    #println("│     OptimalControl : ", o_oc)
    #println("│     JuMP           : ", o_jp)
    println("│     Absolute error : ", o_ae)
    println("│     Relative error : ", o_re)
    println("│")
    println("└─")

    return nothing
end

numerical_comparison(:insurance, docp, nlp_oc_sol, nlp_jp)
┌─ insurance
│
├─  Number of iterations
│
│     OptimalControl : 96
│     JuMP           : 87
│
├─  State I (L2 norm)
│
│     Absolute error : 1.0178071985651289e-5
│     Relative error : 4.760426824017757e-6
│
├─  State m (L2 norm)
│
│     Absolute error : 1.49878052363808e-5
│     Relative error : 5.212944731938122e-6
│
├─  State x₃ (L2 norm)
│
│     Absolute error : 2.660211316609389e-8
│     Relative error : 2.3132337217319656e-8
│
├─  Control h (L2 norm)
│
│     Absolute error : 0.005048484374706771
│     Relative error : 0.0016007423542701133
│
├─  Control R (L2 norm)
│
│     Absolute error : 4.8103826241895204e-6
│     Relative error : 6.179972043877601e-6
│
├─  Control H (L2 norm)
│
│     Absolute error : 6.008317144880914e-6
│     Relative error : 1.861142343925355e-6
│
├─  Control U (L2 norm)
│
│     Absolute error : 5.3568892013885775e-8
│     Relative error : 8.933004907956516e-9
│
├─  Control dUdR (L2 norm)
│
│     Absolute error : 6.00920632370804e-5
│     Relative error : 1.6609392136291374e-5
│
├─  objective
│
│     Absolute error : 1.380171532616714e-10
│     Relative error : 6.702418055181658e-11
│
└─

Plot the solutions

Visualise states, costates, and controls from the OptimalControl and JuMP solutions:

# build an ocp solution to use the plot from OptimalControl package
ocp_sol = build_ocp_solution(docp, nlp_oc_sol)

# dimensions
n = state_dimension(ocp_sol)   # or length(metadata[:insurance][:state_name])
m = control_dimension(ocp_sol) # or length(metadata[:insurance][:control_name])

# from OptimalControl solution
plt = plot(
    ocp_sol;
    state_style=(color=1,),
    costate_style=(color=1, legend=:none),
    control_style=(color=1, legend=:none),
    path_style=(color=1, legend=:none),
    dual_style=(color=1, legend=:none),
    size=(816, 240*(n+m)),
    label="OptimalControl",
    leftmargin=20mm,
)
for i in 2:n
    plot!(plt[i]; legend=:none)
end

# from JuMP solution
t = time_grid(:insurance, nlp_jp)     # t0, ..., tN = tf
x = state(:insurance, nlp_jp)         # function of time
u = control(:insurance, nlp_jp)       # function of time
p = costate(:insurance, nlp_jp)       # function of time

for i in 1:n # state
    label = i == 1 ? "JuMP" : :none
    plot!(plt[i], t, t -> x(t)[i]; color=2, linestyle=:dash, label=label)
end

for i in 1:n # costate
    plot!(plt[n+i], t, t -> -p(t)[i]; color=2, linestyle=:dash, label=:none)
end

for i in 1:m # control
    plot!(plt[2n+i], t, t -> u(t)[i]; color=2, linestyle=:dash, label=:none)
end
Example block output