1. Carga de Paquetes

library(quantmod)
library(tseries)
library(PerformanceAnalytics)
library(fitdistrplus)
library(ggplot2)
library(readxl)
library(fGarch)

2. Carga y Preparación de Datos

# Seleccionar archivo interactivamente
datos <- read_excel(file.choose())

# Procesamiento inicial
datos$Fecha <- as.Date(datos$Fecha)
datos <- datos[order(datos$Fecha),]
precios <- xts(datos$`EMBONOR-B`, order.by = datos$Fecha)
rendimientos <- na.omit(diff(log(precios)))

3. Análisis Exploratorio

Evolución de Precios

ggplot(fortify(precios), aes(x = Index, y = precios)) +
  geom_line(color = "steelblue") +
  labs(title = "Serie Temporal de Precios EMBONOR-B", 
       x = "Fecha", y = "Precio (CLP)") +
  theme_minimal() +
  theme(plot.background = element_rect(fill = "white"))

Distribución de Rendimientos

ggplot(data.frame(Rendimientos = coredata(rendimientos)), aes(x = Rendimientos)) +
  geom_histogram(aes(y = ..density..), bins = 30, fill = "skyblue", alpha = 0.7) +
  stat_function(fun = dnorm, 
                args = list(mean = mean(rendimientos), sd = sd(rendimientos)), 
                color = "red", linetype = "dashed") +
  geom_density(color = "darkblue", linewidth = 1) +
  labs(title = "Distribución de Rendimientos Diarios",
       x = "Rendimiento Logarítmico", y = "Densidad") +
  theme_minimal()

Q-Q Plot

ggplot(data.frame(Rendimientos = coredata(rendimientos)), aes(sample = Rendimientos)) +
  stat_qq(color = "blue", alpha = 0.6) +
  stat_qq_line(color = "red", linetype = "dashed") +
  labs(title = "Q-Q Plot vs Distribución Normal",
       x = "Cuantiles Teóricos", y = "Cuantiles Muestrales") +
  theme_minimal()

4. Pruebas de Normalidad

cat("### Resultados de Pruebas de Normalidad:\n")
## ### Resultados de Pruebas de Normalidad:
if(length(rendimientos) < 5000) print(shapiro.test(coredata(rendimientos)))
## 
##  Shapiro-Wilk normality test
## 
## data:  coredata(rendimientos)
## W = 0.94717, p-value = 8.927e-08
print(jarque.bera.test(coredata(rendimientos)))
## 
##  Jarque Bera Test
## 
## data:  coredata(rendimientos)
## X-squared = 24.424, df = 2, p-value = 4.971e-06

5. Modelamiento con Distribución t-Student

# Definición de funciones de distribución
dstd_robust <- function(x, mu, sigma, nu) fGarch::dstd(x, mean = mu, sd = sigma, nu = nu)
pstd_robust <- function(q, mu, sigma, nu) fGarch::pstd(q, mean = mu, sd = sigma, nu = nu)
qstd_robust <- function(p, mu, sigma, nu) fGarch::qstd(p, mean = mu, sd = sigma, nu = nu)

# Ajuste de parámetros
fit_t <- fitdist(
  data = as.numeric(coredata(rendimientos)),
  distr = "std_robust",
  start = list(mu = 0, sigma = sd(rendimientos), nu = 5),
  lower = c(mu = -0.1, sigma = 0.001, nu = 2.1),
  upper = c(mu = 0.1, sigma = 0.5, nu = 30),
  control = list(trace = 1)
)
##   Nelder-Mead direct search function minimizer
## function value for initial parameters = -2.698600
##   Scaled convergence tolerance is 4.02123e-08
## Stepsize computed as 0.500000
## BUILD              4 99999999999999996864064668260046802.000000 -2.698600
## SHRINK             9 99999999999999996864064668260046802.000000 -2.698600
## SHRINK            14 99999999999999996864064668260046802.000000 -2.698600
## HI-REDUCTION      16 -0.668315 -2.698600
## LO-REDUCTION      18 -1.231120 -2.698600
## HI-REDUCTION      20 -1.678294 -2.698600
## HI-REDUCTION      22 -1.706665 -2.698600
## HI-REDUCTION      24 -1.989677 -2.698600
## HI-REDUCTION      26 -2.059652 -2.698600
## HI-REDUCTION      28 -2.238226 -2.698600
## HI-REDUCTION      30 -2.312195 -2.698600
## HI-REDUCTION      32 -2.422011 -2.698600
## LO-REDUCTION      34 -2.480441 -2.698600
## HI-REDUCTION      36 -2.619452 -2.698600
## HI-REDUCTION      38 -2.673745 -2.698600
## HI-REDUCTION      40 -2.691475 -2.698600
## HI-REDUCTION      42 -2.693892 -2.698600
## HI-REDUCTION      44 -2.696654 -2.698600
## REFLECTION        46 -2.697895 -2.698656
## LO-REDUCTION      48 -2.698103 -2.698834
## REFLECTION        50 -2.698600 -2.699093
## HI-REDUCTION      52 -2.698656 -2.699107
## REFLECTION        54 -2.698834 -2.699296
## HI-REDUCTION      56 -2.699093 -2.699296
## HI-REDUCTION      58 -2.699107 -2.699296
## EXTENSION         60 -2.699250 -2.699779
## LO-REDUCTION      62 -2.699265 -2.699779
## LO-REDUCTION      64 -2.699296 -2.699779
## EXTENSION         66 -2.699722 -2.700652
## LO-REDUCTION      68 -2.699764 -2.700652
## LO-REDUCTION      70 -2.699779 -2.700652
## EXTENSION         72 -2.700373 -2.701824
## LO-REDUCTION      74 -2.700383 -2.701824
## EXTENSION         76 -2.700652 -2.702957
## LO-REDUCTION      78 -2.701203 -2.702957
## EXTENSION         80 -2.701824 -2.704472
## LO-REDUCTION      82 -2.702835 -2.704472
## EXTENSION         84 -2.702957 -2.705736
## HI-REDUCTION      86 -2.703954 -2.705736
## EXTENSION         88 -2.704216 -2.706605
## LO-REDUCTION      90 -2.704472 -2.706605
## REFLECTION        92 -2.705736 -2.707162
## EXTENSION         94 -2.705819 -2.707794
## EXTENSION         96 -2.706605 -2.709523
## HI-REDUCTION      98 -2.707162 -2.709523
## LO-REDUCTION     100 -2.707695 -2.709523
## EXTENSION        102 -2.707794 -2.709552
## EXTENSION        104 -2.708545 -2.710788
## REFLECTION       106 -2.709523 -2.711164
## HI-REDUCTION     108 -2.709552 -2.711164
## LO-REDUCTION     110 -2.710530 -2.711317
## REFLECTION       112 -2.710788 -2.712583
## HI-REDUCTION     114 -2.711164 -2.712583
## HI-REDUCTION     116 -2.711317 -2.712583
## LO-REDUCTION     118 -2.711679 -2.712583
## HI-REDUCTION     120 -2.712090 -2.712583
## EXTENSION        122 -2.712171 -2.712785
## HI-REDUCTION     124 -2.712277 -2.712785
## LO-REDUCTION     126 -2.712470 -2.712785
## LO-REDUCTION     128 -2.712583 -2.712789
## REFLECTION       130 -2.712742 -2.712826
## HI-REDUCTION     132 -2.712785 -2.712826
## HI-REDUCTION     134 -2.712789 -2.712826
## LO-REDUCTION     136 -2.712813 -2.712839
## LO-REDUCTION     138 -2.712823 -2.712847
## HI-REDUCTION     140 -2.712826 -2.712847
## LO-REDUCTION     142 -2.712839 -2.712850
## HI-REDUCTION     144 -2.712845 -2.712850
## REFLECTION       146 -2.712847 -2.712852
## HI-REDUCTION     148 -2.712848 -2.712852
## EXTENSION        150 -2.712850 -2.712858
## HI-REDUCTION     152 -2.712851 -2.712858
## EXTENSION        154 -2.712852 -2.712865
## LO-REDUCTION     156 -2.712854 -2.712865
## EXTENSION        158 -2.712858 -2.712873
## EXTENSION        160 -2.712861 -2.712884
## EXTENSION        162 -2.712865 -2.712904
## LO-REDUCTION     164 -2.712873 -2.712904
## EXTENSION        166 -2.712884 -2.712936
## EXTENSION        168 -2.712903 -2.712966
## EXTENSION        170 -2.712904 -2.712990
## EXTENSION        172 -2.712936 -2.713035
## REFLECTION       174 -2.712966 -2.713056
## LO-REDUCTION     176 -2.712990 -2.713056
## LO-REDUCTION     178 -2.713035 -2.713056
## REFLECTION       180 -2.713046 -2.713073
## HI-REDUCTION     182 -2.713049 -2.713073
## LO-REDUCTION     184 -2.713056 -2.713073
## REFLECTION       186 -2.713060 -2.713076
## REFLECTION       188 -2.713069 -2.713083
## LO-REDUCTION     190 -2.713073 -2.713083
## EXTENSION        192 -2.713076 -2.713095
## LO-REDUCTION     194 -2.713082 -2.713095
## EXTENSION        196 -2.713083 -2.713111
## EXTENSION        198 -2.713094 -2.713129
## LO-REDUCTION     200 -2.713095 -2.713129
## EXTENSION        202 -2.713111 -2.713149
## LO-REDUCTION     204 -2.713117 -2.713149
## EXTENSION        206 -2.713129 -2.713177
## LO-REDUCTION     208 -2.713143 -2.713177
## HI-REDUCTION     210 -2.713149 -2.713177
## EXTENSION        212 -2.713163 -2.713197
## HI-REDUCTION     214 -2.713170 -2.713197
## LO-REDUCTION     216 -2.713177 -2.713197
## EXTENSION        218 -2.713178 -2.713212
## HI-REDUCTION     220 -2.713196 -2.713212
## HI-REDUCTION     222 -2.713197 -2.713212
## LO-REDUCTION     224 -2.713197 -2.713212
## LO-REDUCTION     226 -2.713202 -2.713212
## REFLECTION       228 -2.713211 -2.713214
## LO-REDUCTION     230 -2.713212 -2.713214
## HI-REDUCTION     232 -2.713212 -2.713214
## REFLECTION       234 -2.713214 -2.713215
## HI-REDUCTION     236 -2.713214 -2.713215
## LO-REDUCTION     238 -2.713214 -2.713215
## REFLECTION       240 -2.713215 -2.713216
## LO-REDUCTION     242 -2.713215 -2.713216
## LO-REDUCTION     244 -2.713215 -2.713216
## REFLECTION       246 -2.713216 -2.713216
## LO-REDUCTION     248 -2.713216 -2.713216
## EXTENSION        250 -2.713216 -2.713216
## LO-REDUCTION     252 -2.713216 -2.713216
## LO-REDUCTION     254 -2.713216 -2.713216
## REFLECTION       256 -2.713216 -2.713217
## HI-REDUCTION     258 -2.713216 -2.713217
## LO-REDUCTION     260 -2.713216 -2.713217
## LO-REDUCTION     262 -2.713216 -2.713217
## REFLECTION       264 -2.713216 -2.713217
## EXTENSION        266 -2.713216 -2.713217
## LO-REDUCTION     268 -2.713217 -2.713217
## EXTENSION        270 -2.713217 -2.713217
## LO-REDUCTION     272 -2.713217 -2.713217
## EXTENSION        274 -2.713217 -2.713217
## LO-REDUCTION     276 -2.713217 -2.713217
## LO-REDUCTION     278 -2.713217 -2.713217
## REFLECTION       280 -2.713217 -2.713217
## REFLECTION       282 -2.713217 -2.713217
## LO-REDUCTION     284 -2.713217 -2.713217
## Exiting from Nelder Mead minimizer
##     286 function evaluations used
##   Nelder-Mead direct search function minimizer
## function value for initial parameters = -2.714718
##   Scaled convergence tolerance is 4.04525e-08
## Stepsize computed as 0.222295
## BUILD              4 99999999999999996864064668260046802.000000 -2.714718
## SHRINK             9 99999999999999996864064668260046802.000000 -2.714718
## HI-REDUCTION      11 -0.291081 -2.714718
## LO-REDUCTION      13 -1.840761 -2.714718
## LO-REDUCTION      15 -1.986662 -2.714718
## HI-REDUCTION      17 -2.287098 -2.714718
## HI-REDUCTION      19 -2.332881 -2.714718
## HI-REDUCTION      21 -2.468087 -2.714718
## HI-REDUCTION      23 -2.518418 -2.714718
## LO-REDUCTION      25 -2.577670 -2.714718
## HI-REDUCTION      27 -2.661349 -2.714718
## HI-REDUCTION      29 -2.692866 -2.714718
## LO-REDUCTION      31 -2.695972 -2.714718
## LO-REDUCTION      33 -2.705094 -2.714718
## HI-REDUCTION      35 -2.709425 -2.714718
## HI-REDUCTION      37 -2.711291 -2.714718
## LO-REDUCTION      39 -2.713017 -2.714718
## HI-REDUCTION      41 -2.713320 -2.714718
## HI-REDUCTION      43 -2.714253 -2.714718
## HI-REDUCTION      45 -2.714321 -2.714718
## HI-REDUCTION      47 -2.714548 -2.714718
## LO-REDUCTION      49 -2.714580 -2.714718
## HI-REDUCTION      51 -2.714635 -2.714718
## LO-REDUCTION      53 -2.714667 -2.714718
## LO-REDUCTION      55 -2.714673 -2.714718
## EXTENSION         57 -2.714689 -2.714743
## REFLECTION        59 -2.714713 -2.714747
## HI-REDUCTION      61 -2.714718 -2.714747
## REFLECTION        63 -2.714741 -2.714779
## REFLECTION        65 -2.714743 -2.714783
## HI-REDUCTION      67 -2.714747 -2.714783
## EXTENSION         69 -2.714769 -2.714827
## LO-REDUCTION      71 -2.714779 -2.714827
## LO-REDUCTION      73 -2.714783 -2.714827
## LO-REDUCTION      75 -2.714802 -2.714827
## REFLECTION        77 -2.714820 -2.714841
## LO-REDUCTION      79 -2.714823 -2.714841
## LO-REDUCTION      81 -2.714827 -2.714841
## REFLECTION        83 -2.714835 -2.714850
## LO-REDUCTION      85 -2.714840 -2.714850
## HI-REDUCTION      87 -2.714841 -2.714850
## REFLECTION        89 -2.714846 -2.714853
## LO-REDUCTION      91 -2.714848 -2.714853
## EXTENSION         93 -2.714850 -2.714860
## LO-REDUCTION      95 -2.714852 -2.714860
## LO-REDUCTION      97 -2.714853 -2.714860
## EXTENSION         99 -2.714857 -2.714868
## LO-REDUCTION     101 -2.714857 -2.714868
## LO-REDUCTION     103 -2.714860 -2.714868
## REFLECTION       105 -2.714865 -2.714871
## EXTENSION        107 -2.714867 -2.714873
## EXTENSION        109 -2.714868 -2.714879
## HI-REDUCTION     111 -2.714871 -2.714879
## EXTENSION        113 -2.714873 -2.714883
## LO-REDUCTION     115 -2.714874 -2.714883
## REFLECTION       117 -2.714878 -2.714884
## LO-REDUCTION     119 -2.714879 -2.714884
## EXTENSION        121 -2.714883 -2.714890
## HI-REDUCTION     123 -2.714884 -2.714890
## LO-REDUCTION     125 -2.714884 -2.714890
## REFLECTION       127 -2.714886 -2.714890
## LO-REDUCTION     129 -2.714889 -2.714891
## LO-REDUCTION     131 -2.714890 -2.714891
## LO-REDUCTION     133 -2.714890 -2.714891
## HI-REDUCTION     135 -2.714891 -2.714891
## LO-REDUCTION     137 -2.714891 -2.714891
## HI-REDUCTION     139 -2.714891 -2.714891
## HI-REDUCTION     141 -2.714891 -2.714891
## REFLECTION       143 -2.714891 -2.714891
## LO-REDUCTION     145 -2.714891 -2.714891
## Exiting from Nelder Mead minimizer
##     147 function evaluations used
##   Nelder-Mead direct search function minimizer
## function value for initial parameters = -2.714977
##   Scaled convergence tolerance is 4.04563e-08
## Stepsize computed as 0.211081
## BUILD              4 99999999999999996864064668260046802.000000 -2.714977
## SHRINK             9 99999999999999996864064668260046802.000000 -2.714977
## HI-REDUCTION      11 -0.299531 -2.714977
## LO-REDUCTION      13 -1.858118 -2.714977
## HI-REDUCTION      15 -2.206114 -2.714977
## HI-REDUCTION      17 -2.300837 -2.714977
## LO-REDUCTION      19 -2.392366 -2.714977
## HI-REDUCTION      21 -2.551474 -2.714977
## HI-REDUCTION      23 -2.614225 -2.714977
## LO-REDUCTION      25 -2.629375 -2.714977
## LO-REDUCTION      27 -2.676589 -2.714977
## LO-REDUCTION      29 -2.684417 -2.714977
## HI-REDUCTION      31 -2.704297 -2.714977
## LO-REDUCTION      33 -2.706097 -2.714977
## LO-REDUCTION      35 -2.713061 -2.714977
## HI-REDUCTION      37 -2.713118 -2.714977
## HI-REDUCTION      39 -2.713796 -2.714977
## HI-REDUCTION      41 -2.714519 -2.714977
## HI-REDUCTION      43 -2.714551 -2.714977
## LO-REDUCTION      45 -2.714806 -2.714977
## HI-REDUCTION      47 -2.714826 -2.714977
## LO-REDUCTION      49 -2.714853 -2.714977
## HI-REDUCTION      51 -2.714921 -2.714977
## LO-REDUCTION      53 -2.714925 -2.714977
## LO-REDUCTION      55 -2.714933 -2.714977
## LO-REDUCTION      57 -2.714956 -2.714977
## LO-REDUCTION      59 -2.714957 -2.714977
## HI-REDUCTION      61 -2.714967 -2.714977
## LO-REDUCTION      63 -2.714970 -2.714977
## LO-REDUCTION      65 -2.714974 -2.714977
## EXTENSION         67 -2.714976 -2.714981
## LO-REDUCTION      69 -2.714977 -2.714981
## EXTENSION         71 -2.714977 -2.714983
## LO-REDUCTION      73 -2.714979 -2.714983
## REFLECTION        75 -2.714981 -2.714983
## REFLECTION        77 -2.714983 -2.714984
## HI-REDUCTION      79 -2.714983 -2.714984
## HI-REDUCTION      81 -2.714983 -2.714984
## HI-REDUCTION      83 -2.714984 -2.714985
## HI-REDUCTION      85 -2.714984 -2.714985
## REFLECTION        87 -2.714984 -2.714985
## HI-REDUCTION      89 -2.714985 -2.714985
## HI-REDUCTION      91 -2.714985 -2.714985
## HI-REDUCTION      93 -2.714985 -2.714985
## LO-REDUCTION      95 -2.714985 -2.714985
## LO-REDUCTION      97 -2.714985 -2.714985
## LO-REDUCTION      99 -2.714985 -2.714985
## EXTENSION        101 -2.714985 -2.714985
## HI-REDUCTION     103 -2.714985 -2.714985
## REFLECTION       105 -2.714985 -2.714985
## LO-REDUCTION     107 -2.714985 -2.714985
## Exiting from Nelder Mead minimizer
##     109 function evaluations used
##   Nelder-Mead direct search function minimizer
## function value for initial parameters = -2.714995
##   Scaled convergence tolerance is 4.04566e-08
## Stepsize computed as 0.210107
## BUILD              4 99999999999999996864064668260046802.000000 -2.714995
## SHRINK             9 99999999999999996864064668260046802.000000 -2.714995
## HI-REDUCTION      11 -0.300616 -2.714995
## LO-REDUCTION      13 -1.859281 -2.714995
## HI-REDUCTION      15 -2.231098 -2.714995
## HI-REDUCTION      17 -2.305065 -2.714995
## LO-REDUCTION      19 -2.404367 -2.714995
## HI-REDUCTION      21 -2.555663 -2.714995
## HI-REDUCTION      23 -2.613963 -2.714995
## LO-REDUCTION      25 -2.629830 -2.714995
## LO-REDUCTION      27 -2.674978 -2.714995
## LO-REDUCTION      29 -2.679875 -2.714995
## HI-REDUCTION      31 -2.701821 -2.714995
## LO-REDUCTION      33 -2.706042 -2.714995
## HI-REDUCTION      35 -2.712333 -2.714995
## LO-REDUCTION      37 -2.713017 -2.714995
## HI-REDUCTION      39 -2.713562 -2.714995
## HI-REDUCTION      41 -2.714415 -2.714995
## HI-REDUCTION      43 -2.714547 -2.714995
## HI-REDUCTION      45 -2.714768 -2.714995
## HI-REDUCTION      47 -2.714849 -2.714995
## HI-REDUCTION      49 -2.714865 -2.714995
## LO-REDUCTION      51 -2.714912 -2.714995
## HI-REDUCTION      53 -2.714958 -2.714995
## LO-REDUCTION      55 -2.714958 -2.714995
## LO-REDUCTION      57 -2.714969 -2.714995
## HI-REDUCTION      59 -2.714978 -2.714995
## LO-REDUCTION      61 -2.714980 -2.714995
## HI-REDUCTION      63 -2.714986 -2.714995
## LO-REDUCTION      65 -2.714987 -2.714995
## HI-REDUCTION      67 -2.714988 -2.714995
## HI-REDUCTION      69 -2.714990 -2.714995
## HI-REDUCTION      71 -2.714991 -2.714995
## REFLECTION        73 -2.714992 -2.714995
## HI-REDUCTION      75 -2.714993 -2.714995
## LO-REDUCTION      77 -2.714994 -2.714995
## HI-REDUCTION      79 -2.714994 -2.714995
## REFLECTION        81 -2.714995 -2.714995
## HI-REDUCTION      83 -2.714995 -2.714995
## HI-REDUCTION      85 -2.714995 -2.714995
## HI-REDUCTION      87 -2.714995 -2.714995
## LO-REDUCTION      89 -2.714995 -2.714995
## HI-REDUCTION      91 -2.714995 -2.714995
## HI-REDUCTION      93 -2.714995 -2.714995
## LO-REDUCTION      95 -2.714995 -2.714995
## HI-REDUCTION      97 -2.714995 -2.714995
## Exiting from Nelder Mead minimizer
##     99 function evaluations used

6. Cálculo de Medidas de Riesgo

nivel_confianza <- 0.95

# Método paramétrico
var_t <- qstd_robust(1 - nivel_confianza, 
                    mu = fit_t$estimate["mu"],
                    sigma = fit_t$estimate["sigma"], 
                    nu = fit_t$estimate["nu"])

cvar_t <- integrate(function(x) x * dstd_robust(x, 
                                               mu = fit_t$estimate["mu"],
                                               sigma = fit_t$estimate["sigma"], 
                                               nu = fit_t$estimate["nu"]), 
                   -Inf, var_t)$value / (1 - nivel_confianza)

# Método empírico
var_emp <- quantile(rendimientos, 1 - nivel_confianza)
cvar_emp <- mean(rendimientos[rendimientos <= var_emp])

7. Visualización Comparativa

df_comp <- data.frame(
  Metodo = rep(c("Paramétrico", "Empírico"), each = 2),
  Medida = rep(c("VaR", "CVaR"), 2),
  Valor = c(var_t, cvar_t, var_emp, cvar_emp)
)

ggplot(df_comp, aes(x = Medida, y = Valor, fill = Metodo)) +
  geom_col(position = position_dodge(), width = 0.7) +
  scale_fill_manual(values = c("Paramétrico" = "tomato", "Empírico" = "steelblue")) +
  geom_text(aes(label = round(Valor, 4)), 
            position = position_dodge(width = 0.7), 
            vjust = -0.5, size = 4) +
  labs(title = "Comparación de Medidas de Riesgo al 95% de Confianza",
       x = "Medida de Riesgo", y = "Valor (%)") +
  theme_bw()

8. Interpretación de Resultados

Resultados Finales

cat("**VaR t-Student:**", round(var_t, 5), "\n")
## **VaR t-Student:** -0.02761
cat("**CVaR t-Student:**", round(cvar_t, 5), "\n")
## **CVaR t-Student:** -0.0555
cat("**VaR Empírico:**", round(var_emp, 5), "\n")
## **VaR Empírico:** -0.02454
cat("**CVaR Empírico:**", round(cvar_emp, 5))
## **CVaR Empírico:** -0.03715

Análisis Clave

  1. No Normalidad de los Datos
    • Valores p < 0.05 en ambas pruebas confirman desviación de la normalidad
    • Los rendimientos muestran colas más pesadas que la distribución normal
    • Justifica el uso de distribuciones con colas pesadas como la t-Student
  2. Parámetros Estimados t-Student
    • σ ≈ 0.0441: Volatilidad diaria del 4.41%
    • ν ≈ 2.1: Grados de libertad indican colas 3x más pesadas que la normal
    • μ ≈ -3.7^{-4}: Tendencia central neutra en el periodo analizado
  3. Diferencias entre Métodos
    • VaR paramétrico más conservador (-2.76% vs -2.45%):
      • La t-Student captura mejor el riesgo de cola
      • Método empírico subestima riesgo en colas no observadas históricamente
    • Brecha en CVaR (-5.55% vs -3.72%):
      • Indica presencia de eventos extremos no modelados
      • Sugiere necesidad de análisis de estrés complementario