Tugas Akhir Anreg
Data
Inisialisasi Library
## Warning: package 'readxl' was built under R version 4.3.2
##
## Attaching package: 'dplyr'
## The following objects are masked from 'package:stats':
##
## filter, lag
## The following objects are masked from 'package:base':
##
## intersect, setdiff, setequal, union
## Warning: package 'plotly' was built under R version 4.3.2
## Loading required package: ggplot2
## Warning: package 'ggplot2' was built under R version 4.3.2
##
## Attaching package: 'plotly'
## The following object is masked from 'package:ggplot2':
##
## last_plot
## The following object is masked from 'package:stats':
##
## filter
## The following object is masked from 'package:graphics':
##
## layout
## Warning: package 'lmtest' was built under R version 4.3.2
## Loading required package: zoo
## Warning: package 'zoo' was built under R version 4.3.2
##
## Attaching package: 'zoo'
## The following objects are masked from 'package:base':
##
## as.Date, as.Date.numeric
## Warning: package 'car' was built under R version 4.3.3
## Loading required package: carData
##
## Attaching package: 'car'
## The following object is masked from 'package:dplyr':
##
## recode
##
## Attaching package: 'MASS'
## The following object is masked from 'package:plotly':
##
## select
## The following object is masked from 'package:dplyr':
##
## select
## Warning: package 'olsrr' was built under R version 4.3.2
##
## Attaching package: 'olsrr'
## The following object is masked from 'package:MASS':
##
## cement
## The following object is masked from 'package:datasets':
##
## rivers
## Warning: package 'ggcorrplot' was built under R version 4.3.3
## Warning: package 'GGally' was built under R version 4.3.2
## Registered S3 method overwritten by 'GGally':
## method from
## +.gg ggplot2
Import Data
## # A tibble: 22 × 13
## Y X1 X2 X3 X4 X5 X6 X7 X8 X9 X10 X11
## <dbl> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl> <dbl>
## 1 27.2 65.2 6.92 152414 11.9 0.39 58.8 3.52 10199 12696 67.6 3.93
## 2 28.1 67.0 7.57 255498 5.87 0.3 61.8 2.21 16617 36053 65.8 2.02
## 3 21.8 65.8 7.42 376837 6.17 1.18 76.4 3.22 13793 59622 65.6 4.2
## 4 25.2 63.6 6.97 474521 7.71 1.03 67.9 2.64 10953 57201 66.9 2.09
## 5 21.8 65.2 8.16 271277 3.98 0.45 80.6 1.96 11511 42634 67.6 1.2
## 6 14.3 63.8 7.39 231008 6.34 0.71 84.0 5.45 13997 45693 65.6 2.23
## 7 20.0 63.0 8.45 221536 1.75 0.63 81.6 2.52 9900 22483 62.4 0.78
## 8 24.8 66.1 8.26 141391 3.95 1.2 87.4 2.55 8767 21220 67.9 2.04
## 9 11.8 65.8 8.04 288310 3.69 0.13 93.0 3.79 12818 35172 66.0 1.66
## 10 12.6 66.9 6.98 335360 5.02 0.5 80.7 2.62 10797 58147 68.3 0.93
## # ℹ 12 more rows
## # ℹ 1 more variable: X12 <dbl>
Reduksi Variabel X
Eksplorasi Korelasi Data
Hipotesis: H0 : Tidak ada korelasi antarpeubah H1 : Ada korelasi antarpeubah
Dengan minimal ditandai oleh bintang satu (*) yang berarti variabel X tersebut tolak H0 atau ada korelasi dengan Y, maka hanya X1, X2, X4, X6, X7, X8, X9, dan X12 yang mempunyai korelasi. Sedangkan variabel X3, X5, X10, dan X11 memiliki korelasi yang kecil dan telah terwakili oleh variabel X yang lainnya sehingga dapat dihilangkan untuk kemudian diperiksa model dan multikolinearitasnya.
Model Reduksi Hasil Korelasi
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X6 + X7 + X8 + X9 + X12, data = dt)
##
## Residuals:
## Min 1Q Median 3Q Max
## -10.4006 -1.4476 0.5167 2.3660 6.7547
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 5.806e+01 5.405e+01 1.074 0.3023
## X1 -7.717e-01 6.726e-01 -1.147 0.2719
## X2 9.612e-01 3.126e+00 0.308 0.7633
## X4 4.282e-01 6.137e-01 0.698 0.4976
## X6 -1.058e-01 1.554e-01 -0.681 0.5078
## X7 -2.572e+00 1.415e+00 -1.817 0.0923 .
## X8 9.913e-04 5.710e-04 1.736 0.1062
## X9 -5.652e-05 7.422e-05 -0.762 0.4599
## X12 1.318e-01 2.054e-01 0.642 0.5321
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 4.906 on 13 degrees of freedom
## Multiple R-squared: 0.6744, Adjusted R-squared: 0.4741
## F-statistic: 3.366 on 8 and 13 DF, p-value: 0.02559
Diperoleh model sebagai berikut
\[
\hat Y = 0.5806 - 0.7717X_1 - 0.9612X_2 + 0.4282X_4 - 0.1058X_6 -
2.572X_7 + 0.0009X_8 - 0.00005X_9 + 0.1318X_{12} + e
\]
Pengecekan Multikolinearitas
## X1 X2 X4 X6 X7 X8 X9 X12
## 6.429485 9.363917 4.320674 3.777043 2.107861 11.289087 10.675684 11.701402
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X6 + X7 + X8 + X9, data = dt)
##
## Residuals:
## Min 1Q Median 3Q Max
## -10.6906 -2.1485 -0.1154 2.9997 6.5588
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 7.950e+01 4.158e+01 1.912 0.0766 .
## X1 -9.080e-01 6.246e-01 -1.454 0.1681
## X2 1.351e+00 3.001e+00 0.450 0.6594
## X4 4.124e-01 6.001e-01 0.687 0.5033
## X6 -1.562e-01 1.312e-01 -1.190 0.2538
## X7 -2.598e+00 1.385e+00 -1.876 0.0817 .
## X8 9.174e-04 5.474e-04 1.676 0.1159
## X9 -8.516e-05 5.805e-05 -1.467 0.1644
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 4.801 on 14 degrees of freedom
## Multiple R-squared: 0.6641, Adjusted R-squared: 0.4962
## F-statistic: 3.954 on 7 and 14 DF, p-value: 0.01375
## X1 X2 X4 X6 X7 X8 X9
## 5.788236 9.009798 4.313712 2.813403 2.106158 10.829677 6.816299
Model Hasil Reduksi Multikolinearitas
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X6 + X7 + X9, data = dt)
##
## Residuals:
## Min 1Q Median 3Q Max
## -10.2818 -2.6755 -0.1858 2.6222 8.0058
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 4.727e+01 3.903e+01 1.211 0.245
## X1 -6.125e-01 6.343e-01 -0.966 0.350
## X2 3.486e+00 2.876e+00 1.212 0.244
## X4 8.014e-01 5.859e-01 1.368 0.192
## X6 -1.330e-01 1.381e-01 -0.963 0.351
## X7 -2.103e+00 1.432e+00 -1.468 0.163
## X9 -2.833e-05 4.987e-05 -0.568 0.578
##
## Residual standard error: 5.083 on 15 degrees of freedom
## Multiple R-squared: 0.5967, Adjusted R-squared: 0.4354
## F-statistic: 3.699 on 6 and 15 DF, p-value: 0.01854
## X1 X2 X4 X6 X7 X9
## 5.327040 7.386697 3.668411 2.782125 2.010498 4.489353
Diperoleh model sebagai berikut
\[
\hat Y = 47,27 - 0.6125X_1 + 3.486X_2 + 0.8014X_4 - 0.1330X_6 - 2.103X_7
- 0.00002833
\] ——————————————————————–
Pencilan, Titik Leverage, dan Amatan Berpengaruh
Perhitungan ri, ci, Hi, DFFITSi
index <- c(1:22)
ri <- studres(model4)
ci <- cooks.distance(model4)
DFFITSi <- dffits(model4)
hi <- ols_hadi(model4)
hii <- hatvalues(model4)
hasil <- data.frame(index,ri,ci,DFFITSi,hi,hii); round(hasil,4)## index ri ci DFFITSi hadi potential residual hii
## 1 1 0.4282 0.0165 0.3305 0.6870 0.5958 0.0912 0.3734
## 2 2 1.0002 0.0263 0.4287 0.6784 0.1838 0.4947 0.1552
## 3 3 0.5437 0.0068 0.2128 0.3005 0.1532 0.1474 0.1328
## 4 4 0.5689 0.0152 0.3191 0.4755 0.3145 0.1610 0.2393
## 5 5 -0.2444 0.0032 -0.1452 0.3826 0.3528 0.0298 0.2608
## 6 6 -0.1956 0.0038 -0.1587 0.6775 0.6584 0.0191 0.3970
## 7 7 -0.6673 0.0349 -0.4853 0.7492 0.5289 0.2202 0.3460
## 8 8 0.8353 0.0404 0.5264 0.7411 0.3971 0.3440 0.2842
## 9 9 -1.0517 0.0323 -0.4770 0.7514 0.2057 0.5457 0.1706
## 10 10 -1.1523 0.1654 -1.0878 1.5267 0.8912 0.6355 0.4712
## 11 11 1.4434 0.0926 0.8339 1.3381 0.3337 1.0044 0.2502
## 12 12 -0.8916 0.0679 -0.6849 0.9794 0.5901 0.3892 0.3711
## 13 13 -0.4006 0.0061 -0.2008 0.3313 0.2512 0.0801 0.2008
## 14 14 1.7955 0.0528 0.6515 1.7015 0.1316 1.5698 0.1163
## 15 15 0.4072 0.0102 0.2593 0.4882 0.4056 0.0826 0.2886
## 16 16 0.2896 0.0055 0.1896 0.4704 0.4286 0.0419 0.3000
## 17 17 -0.5523 0.0246 -0.4056 0.6908 0.5395 0.1513 0.3504
## 18 18 -1.1497 0.0486 -0.5894 0.9110 0.2629 0.6482 0.2081
## 19 19 -0.1593 0.0047 -0.1746 1.2136 1.2009 0.0127 0.5456
## 20 20 1.2520 0.1391 1.0053 1.3956 0.6448 0.7508 0.3920
## 21 21 -2.6233 0.1469 -1.1963 3.3803 0.2080 3.1723 0.1722
## 22 22 0.2283 0.2990 1.4002 37.6507 37.6248 0.0260 0.9741
Plot ri, ci, Hi, DFFITSi, dan Potential Residual
Internally Studentized Residuals Plot (ri)
ggplot(hasil) +
geom_point(aes(y = `ri`, x = `index`),
color="blue",size=2) +
ylab("ri") +
xlab("index") +
ggtitle("Plot ri") +
theme_classic() +
theme(plot.title = element_text(hjust = 0.5))Cook Distance Plot (ci)
ggplot(hasil) +
geom_point(aes(y = `ci`, x = `index`),
color="blue",size=2) +
ylab("ci") +
xlab("index") +
ggtitle("Plot ci") +
theme_classic() +
theme(plot.title = element_text(hjust = 0.5))Difference in Fits Plot (DFFITSi)
ggplot(hasil) +
geom_point(aes(y = `DFFITSi`, x = `index`),
color="blue",size=2) +
ylab("DFFITSi") +
xlab("index") +
ggtitle("Plot DFFITSi") +
theme_classic() +
theme(plot.title = element_text(hjust = 0.5))Pendeteksian Nilai Tidak Biasa
Pencilan atau Outlier
## [1] 21
Pencilan adalah titik amatan yang menjauh pada tren Y walau X nya masih berada pada selang. Dalam data tersebut terdapat satu titik observasi yang terdeteksi sebagai sebuah pencilan, yaitu amatan ke-21 dengan ri bernilai -2.6233 yang absolutnya lebih besar dari 2..
Titik Leverage
titik_leverage <- vector("list", dim(hasil)[1])
for (i in 1:dim(hasil)[1]) {
cutoff <- 2 * 7 / 22
titik_leverage[[i]] <- which(hii > cutoff)
}
leverages <- unlist(titik_leverage)
titik_leverage <- sort(unique(leverages))
titik_leverage## [1] 22
Titik leverage adalah nilai amatan yang X nya jauh lebih besar dari X lainnya dan terletak hampir dalam garis regresi tetapi juga menjauh. Terdapat satu titik observasi yang termasuk sebagai titik leverage pada data tersebut, yaitu amatan ke-22 dengan nilai hii 0.9741. Kedua titik ini memiliki nilai hii yang lebih besar dari 2*p/n = 0,636, dengan p = 7 adalah banyaknya parameter dan n = 22 adalah banyaknya amatan.
Amatan Berpengaruh
amatan_berpengaruh <- vector("list", dim(hasil)[1])
for (i in 1:dim(hasil)[1]) {
cutoff <- 2 * sqrt((7 / 22))
amatan_berpengaruh[[i]] <- which(abs(DFFITSi) > cutoff)
}
berpengaruh <- unlist(amatan_berpengaruh)
amatan_berpengaruh <- sort(unique(berpengaruh))
amatan_berpengaruh## [1] 21 22
Amatan berpengaruh adalah nilai amatan yang memengaruhi hasil dugaan parameter regresi, R-square, dan uji hipotesis jika disisihkan. Hasil pada data tersebut menunjukkan bahwa amatan ke-10, 21, dan 22 termasuk sebagai amatan berpengaruh yang memiliki nilai DFFITSi -1.1963 dan 1,400. Absolut dari nilai tersebut lebih besar dari 2*sqrt(p/n) = 1,128.
data1 <- dt[-c(21, 22),]
data2 <- dt[-c(21),]
data3 <- dt[-c(22),]
model_berpengaruh1 <- lm(Y~X1+X2+X4+X6+X7+X9, data=data1)
model_berpengaruh2 <- lm(Y~X1+X2+X4+X6+X7+X9, data=data2)
model_berpengaruh3 <- lm(Y~X1+X2+X4+X6+X7+X9, data=data3)
summary(model4)##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X6 + X7 + X9, data = dt)
##
## Residuals:
## Min 1Q Median 3Q Max
## -10.2818 -2.6755 -0.1858 2.6222 8.0058
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 4.727e+01 3.903e+01 1.211 0.245
## X1 -6.125e-01 6.343e-01 -0.966 0.350
## X2 3.486e+00 2.876e+00 1.212 0.244
## X4 8.014e-01 5.859e-01 1.368 0.192
## X6 -1.330e-01 1.381e-01 -0.963 0.351
## X7 -2.103e+00 1.432e+00 -1.468 0.163
## X9 -2.833e-05 4.987e-05 -0.568 0.578
##
## Residual standard error: 5.083 on 15 degrees of freedom
## Multiple R-squared: 0.5967, Adjusted R-squared: 0.4354
## F-statistic: 3.699 on 6 and 15 DF, p-value: 0.01854
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X6 + X7 + X9, data = data1)
##
## Residuals:
## Min 1Q Median 3Q Max
## -5.658 -2.782 -0.081 2.215 6.994
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 6.321e+01 4.101e+01 1.541 0.1472
## X1 -9.305e-01 5.932e-01 -1.569 0.1408
## X2 3.616e+00 4.856e+00 0.745 0.4697
## X4 1.027e+00 7.174e-01 1.431 0.1759
## X6 -7.489e-02 1.478e-01 -0.507 0.6208
## X7 -2.482e+00 1.387e+00 -1.789 0.0969 .
## X9 -3.013e-05 8.413e-05 -0.358 0.7259
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 4.463 on 13 degrees of freedom
## Multiple R-squared: 0.6597, Adjusted R-squared: 0.5026
## F-statistic: 4.2 on 6 and 13 DF, p-value: 0.01441
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X6 + X7 + X9, data = data2)
##
## Residuals:
## Min 1Q Median 3Q Max
## -5.6614 -2.5983 0.1445 2.3220 7.2611
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 5.877e+01 3.337e+01 1.761 0.1000
## X1 -9.608e-01 5.537e-01 -1.735 0.1047
## X2 4.448e+00 2.465e+00 1.804 0.0927 .
## X4 1.124e+00 5.116e-01 2.198 0.0453 *
## X6 -9.156e-02 1.181e-01 -0.775 0.4512
## X7 -2.366e+00 1.218e+00 -1.942 0.0725 .
## X9 -1.571e-05 4.254e-05 -0.369 0.7174
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 4.308 on 14 degrees of freedom
## Multiple R-squared: 0.7177, Adjusted R-squared: 0.5968
## F-statistic: 5.933 on 6 and 14 DF, p-value: 0.002919
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X6 + X7 + X9, data = data3)
##
## Residuals:
## Min 1Q Median 3Q Max
## -10.2528 -2.6639 -0.4331 3.0228 7.6474
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 5.323e+01 4.802e+01 1.108 0.286
## X1 -5.730e-01 6.778e-01 -0.845 0.412
## X2 2.380e+00 5.684e+00 0.419 0.682
## X4 6.725e-01 8.277e-01 0.812 0.430
## X6 -1.106e-01 1.731e-01 -0.639 0.533
## X7 -2.258e+00 1.629e+00 -1.387 0.187
## X9 -4.753e-05 9.865e-05 -0.482 0.637
##
## Residual standard error: 5.251 on 14 degrees of freedom
## Multiple R-squared: 0.523, Adjusted R-squared: 0.3186
## F-statistic: 2.559 on 6 and 14 DF, p-value: 0.06921
Tetep model terbaiknya saat data ke 21 dikeluarkan
#pembentukan data baru tanpa amatan 9 dan 19
dt2 <- subset(dt, select = -c(X3, X5, X8, X10, X11, X12))
dt2 <- dt2[-c(21),]
#pemodelan rlb terbaru
model5 <- lm(Y~.,data= dt2)
summary(model5)##
## Call:
## lm(formula = Y ~ ., data = dt2)
##
## Residuals:
## Min 1Q Median 3Q Max
## -5.6614 -2.5983 0.1445 2.3220 7.2611
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 5.877e+01 3.337e+01 1.761 0.1000
## X1 -9.608e-01 5.537e-01 -1.735 0.1047
## X2 4.448e+00 2.465e+00 1.804 0.0927 .
## X4 1.124e+00 5.116e-01 2.198 0.0453 *
## X6 -9.156e-02 1.181e-01 -0.775 0.4512
## X7 -2.366e+00 1.218e+00 -1.942 0.0725 .
## X9 -1.571e-05 4.254e-05 -0.369 0.7174
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 4.308 on 14 degrees of freedom
## Multiple R-squared: 0.7177, Adjusted R-squared: 0.5968
## F-statistic: 5.933 on 6 and 14 DF, p-value: 0.002919
## Analysis of Variance Table
##
## Response: Y
## Df Sum Sq Mean Sq F value Pr(>F)
## X1 1 325.62 325.62 17.5467 0.00091 ***
## X2 1 26.04 26.04 1.4033 0.25590
## X4 1 131.44 131.44 7.0830 0.01861 *
## X6 1 97.50 97.50 5.2538 0.03791 *
## X7 1 77.48 77.48 4.1754 0.06031 .
## X9 1 2.53 2.53 0.1364 0.71745
## Residuals 14 259.80 18.56
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
Eksplorasi Kondisi
Plot Sisaan Vs Y duga
1. Sisaan menyebar di sekitar 0, sehingga nilai harapan galat sama
dengan nol
2. Lebar pita sama untuk setiap nilai dugaan, sehingga ragam
homogen
3. Plot sisaan vs y duga tidak berpola, model pas
Plot Sisaan Vs Urutan
Kondisi Gaus Markov
1. Nilai Harapan Sisaan sama dengan Nol
H0: Nilai harapan sisaan sama dengan nol
H1: Nilai harapan sisaan tidak sama dengan nol
##
## One Sample t-test
##
## data: model5$residuals
## t = -2.6763e-17, df = 20, p-value = 1
## alternative hypothesis: true mean is not equal to 0
## 95 percent confidence interval:
## -1.640606 1.640606
## sample estimates:
## mean of x
## -2.104901e-17
Uji t.tes tersebut menunjukkan hasil p-value = 1 > alpha = 0.05, maka tak tolak H0, nilai harapan sisaan sama dengan nol pada taraf nyata 5%. Asumsi terpenuhi.
2. Ragam Sisaan Homogen
\(H0:var[e]=sigma2I\) (ragam sisan
homogen)
\(H1:var[e] != sigma2I\) (ragam siaan
tidak homogen)
##
## studentized Breusch-Pagan test
##
## data: model5
## BP = 7.4095, df = 6, p-value = 0.2846
Uji ini sering disebut dengan uji homokesdasitas yang dilakukan dengan yang dilakukan dengan uji Breusch-Pagan. Karena p-value = 0.5789 > alpha = 0.05, maka tak tolak H0, ragam sisaan homogen pada taraf nyata 5%. Asumsi terpenuhi.
3. Sisaan Saling Bebas
\(H0:E[ei,ej]=0\) (sisaan saling
bebas/tidak ada autokorelasi)
\(H1:E[ei,ej] != 0\) (sisaan tidak
saling bebas/ada autokorelasi)
##
## Durbin-Watson test
##
## data: model5
## DW = 1.9585, p-value = 0.3501
## alternative hypothesis: true autocorrelation is greater than 0
Uji ini sering disebut dengan uji autokorelasi yang dilakukan dengan Durbin watson. Karena p-value = 0.634 (pada DW test) > alpha = 0.05, maka tak tolak H0, sisaan saling bebas pada taraf nyata 5%, sehingga asumsi terpenuhi.
Uji Formal Normalitas Sisaan
\(H_0 : N\) (sisaan menyebar Normal) \(H_1 : N\) (sisaan tidak menyebar Normal)
##
## Shapiro-Wilk normality test
##
## data: model5$residuals
## W = 0.96194, p-value = 0.5559
## Best Subsets Regression
## --------------------------------
## Model Index Predictors
## --------------------------------
## 1 X6
## 2 X4 X7
## 3 X1 X4 X7
## 4 X1 X2 X4 X7
## 5 X1 X2 X4 X6 X7
## 6 X1 X2 X4 X6 X7 X9
## --------------------------------
##
## Subsets Regression Summary
## ----------------------------------------------------------------------------------------------------------------------------------
## Adj. Pred
## Model R-Square R-Square R-Square C(p) AIC SBIC SBC MSEP FPE HSP APC
## ----------------------------------------------------------------------------------------------------------------------------------
## 1 0.4434 0.4141 0.3404 10.6049 132.6765 72.0095 135.8100 566.4822 29.5295 1.4979 0.6737
## 2 0.6017 0.5575 0.4955 4.7542 127.6493 68.4243 131.8274 429.2228 23.2752 1.1980 0.5310
## 3 0.6394 0.5757 0.3606 4.8873 127.5646 69.3007 132.7872 412.9505 23.2452 1.2204 0.5304
## 4 0.7034 0.6293 0.5274 3.7096 125.4571 69.7724 131.7243 362.2283 21.1228 1.1374 0.4819
## 5 0.7150 0.6200 0.4941 5.1364 126.6224 72.3819 133.9341 372.9779 22.4857 1.2492 0.5130
## 6 0.7177 0.5968 0.4386 7.0000 128.4189 75.3234 136.7750 397.7938 24.7432 1.4275 0.5645
## ----------------------------------------------------------------------------------------------------------------------------------
## AIC: Akaike Information Criteria
## SBIC: Sawa's Bayesian Information Criteria
## SBC: Schwarz Bayesian Criteria
## MSEP: Estimated error of prediction, assuming multivariate normality
## FPE: Final Prediction Error
## HSP: Hocking's Sp
## APC: Amemiya Prediction Criteria
null<-lm(Y~ 1, data=dt2) # 1 here means the intercept
full<-lm(Y~., data=dt2)
step(full, scope=list(lower=null, upper=full),data=datanew, direction='backward', trace=0)##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X7, data = dt2)
##
## Coefficients:
## (Intercept) X1 X2 X4 X7
## 59.761 -1.042 4.181 1.289 -2.955
##
## Call:
## lm(formula = Y ~ X6 + X1, data = dt2)
##
## Coefficients:
## (Intercept) X6 X1
## 84.3127 -0.2551 -0.6719
##
## Call:
## lm(formula = Y ~ X6 + X1, data = dt2)
##
## Coefficients:
## (Intercept) X6 X1
## 84.3127 -0.2551 -0.6719
modell1 <- lm(Y~X1+X2+X4+X7, data = dt2)
modell2 <- lm(Y~X1+X6, data = dt2)
summary(modell1)$adj.r.squared## [1] 0.6292841
## [1] 0.5351652
## [1] 125.4571
## [1] 128.6817
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X7, data = dt2)
##
## Residuals:
## Min 1Q Median 3Q Max
## -6.5916 -3.4506 0.4097 1.8438 7.0582
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 59.7614 18.7337 3.190 0.00570 **
## X1 -1.0423 0.4450 -2.342 0.03243 *
## X2 4.1812 2.2490 1.859 0.08150 .
## X4 1.2889 0.3934 3.276 0.00475 **
## X7 -2.9545 0.9519 -3.104 0.00683 **
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 4.13 on 16 degrees of freedom
## Multiple R-squared: 0.7034, Adjusted R-squared: 0.6293
## F-statistic: 9.487 on 4 and 16 DF, p-value: 0.0003966
##
## Call:
## lm(formula = Y ~ X1 + X2 + X4 + X7, data = dt2)
##
## Residuals:
## Min 1Q Median 3Q Max
## -6.5916 -3.4506 0.4097 1.8438 7.0582
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 59.7614 18.7337 3.190 0.00570 **
## X1 -1.0423 0.4450 -2.342 0.03243 *
## X2 4.1812 2.2490 1.859 0.08150 .
## X4 1.2889 0.3934 3.276 0.00475 **
## X7 -2.9545 0.9519 -3.104 0.00683 **
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 4.13 on 16 degrees of freedom
## Multiple R-squared: 0.7034, Adjusted R-squared: 0.6293
## F-statistic: 9.487 on 4 and 16 DF, p-value: 0.0003966
Plot Sisaan Vs Urutan
## Kondisi Gaus Markov
1. Nilai Harapan Sisaan sama dengan Nol
H0: Nilai harapan sisaan sama dengan nol
H1: Nilai harapan sisaan tidak sama dengan nol
##
## One Sample t-test
##
## data: model$residuals
## t = -6.6554e-18, df = 20, p-value = 1
## alternative hypothesis: true mean is not equal to 0
## 95 percent confidence interval:
## -1.681668 1.681668
## sample estimates:
## mean of x
## -5.36551e-18
2. Ragam Sisaan Homogen
\(H0:var[e]=sigma2I\) (ragam sisan
homogen)
\(H1:var[e] != sigma2I\) (ragam siaan
tidak homogen)
##
## studentized Breusch-Pagan test
##
## data: model
## BP = 2.3095, df = 4, p-value = 0.679
3. Sisaan Saling Bebas
\(H0:E[ei,ej]=0\) (sisaan saling
bebas/tidak ada autokorelasi)
\(H1:E[ei,ej] != 0\) (sisaan tidak
saling bebas/ada autokorelasi)
##
## Durbin-Watson test
##
## data: model
## DW = 1.7442, p-value = 0.2495
## alternative hypothesis: true autocorrelation is greater than 0