library(stats)
library(stats4)
library(dplyr)
library(readxl)
library(cowplot)
library(ggplot2)
library(viridis) # Para paletas de colores
library(hrbrthemes) # Temas, componentes de tema y utilidades adicionales para ggplot2
library(tidyverse)
library(knitr)
library(patchwork)
library(forecast)
library(TSstudio)
library(lubridate)
library(datasets)
library(plotly)
base<-read_excel("base_art.xlsx")
base$Año <- as.factor(base$Año)
"Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
[1] "Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
Tabla_1 <- base %>%
dplyr::group_by(Año) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(Año)
"Grafico"
[1] "Grafico"
G1<-ggplot(Tabla_1, aes(x =Año, y=Total) ) +
geom_bar(width = 0.7,stat="identity",
position = position_dodge(), fill="red3") +
ylim(c(0,6.5))+
#xlim(c(0,300)) +
#ggtitle("Un título") +
labs(x="Año", y= "Frecuencias \n (Porcentajes)") +
geom_text(aes(label=paste(Total," ", "", "(", Porcentaje, "%", ")")),
vjust=-0.9,
color="black",
hjust=0.5,
# define text position and size
position = position_dodge(0.9),
angle=0,
size=3.8) +
theme(axis.text.x = element_text(angle = 0, vjust = 1, hjust=1)) +
theme_bw(base_size = 16) +
#coord_flip() +
facet_wrap(~"Producción Cientifica Anual.")
#cargamos las librerias correspondientes
library(tm)
library(NLP)
library(SnowballC)
library(RColorBrewer)
library(wordcloud)
#Que palabras hay con mayor frecuencia en los comentarios para .....
#nuestro texto lo guardamos en bloc de notas en formato txt
texto <- readLines("P_CLAVES.txt")
texto = iconv(texto, to="ASCII//TRANSLIT")
texto = Corpus(VectorSource(texto))
########### LIMPIAMOS NUESTRO TEXTO CON EL COMANDO tm_map
#ponemos todos los datos a minuscula (A!=a)
discurso=tm_map(texto, tolower)
#quitamos los espacios en blanco
discurso =tm_map(discurso, stripWhitespace)
#quitamos la puntuacion
discurso = tm_map(discurso, removePunctuation)
#quitamos los numeros
discurso = tm_map(discurso, removeNumbers)
#mostramos palabras vacias y genericas
#stopwords("spanish")
#quitamos palabras genericas
discurso=tm_map(discurso, removeWords,stopwords("spanish"))
#tambien podemos tener nuestra propia lista de palabras a quitar
############### DATA FRAME DE PALABRAS CON SU FRECUENCIA
#Creamos matriz de letras
letras= TermDocumentMatrix(discurso)
findFreqTerms(letras, lowfreq=5)
[1] "image" "poisson" "denoising"
[4] "mixed" "method" "noise"
[7] "estimation" "poissongaussian" "and"
[10] "restoration" "imaging"
matrix=as.matrix(letras)
#lo ordenamos y sumamos las letras de nuestra matriz
vector <- sort(rowSums(matrix),decreasing=TRUE)
#creamos la data con las palabras y su frecuencia
dataletras <- data.frame(word= names(vector),frequencia=vector)
dataletras
word frequencia
image image 21
noise noise 20
poisson poisson 15
denoising denoising 13
mixed mixed 6
estimation estimation 6
method method 5
poissongaussian poissongaussian 5
and and 5
restoration restoration 5
imaging imaging 5
segmentation segmentation 4
photonlimited photonlimited 4
learning learning 4
likelihood likelihood 4
admm admm 3
deconvolution deconvolution 3
methods methods 3
representations representations 3
signal signal 3
estimate estimate 3
regularization regularization 3
filter filter 3
distribution distribution 3
selection selection 3
particle particle 3
maximum maximum 3
adaptive adaptive 2
denoiser denoiser 2
quantum quantum 2
decomposition decomposition 2
microscopy microscopy 2
gradient gradient 2
newtons newtons 2
pca pca 2
parameter parameter 2
stabilization stabilization 2
variance variance 2
risk risk 2
unbiased unbiased 2
iterative iterative 2
microscopic microscopic 2
based based 2
pde pde 2
total total 2
variation variation 2
enhancement enhancement 2
statistical statistical 2
fourier fourier 2
estimator estimator 2
loglikelihood loglikelihood 2
transformation transformation 2
wavelet wavelet 2
minimax minimax 2
lowlight lowlight 2
sparse sparse 2
plugandplay plugandplay 1
processing processing 1
copula copula 1
laser laser 1
multiphoton multiphoton 1
scanning scanning 1
singular singular 1
value value 1
variables variables 1
block block 1
matching matching 1
means means 1
nonlocal nonlocal 1
periodic periodic 1
precision precision 1
search search 1
shrinkage shrinkage 1
stem stem 1
transform transform 1
bioimaging bioimaging 1
carlo carlo 1
fluorescence fluorescence 1
monte monte 1
mse mse 1
pgure pgure 1
steins steins 1
sure sure 1
hard hard 1
thresholdingiht thresholdingiht 1
conjugate conjugate 1
gradientcg gradientcg 1
nonconvex nonconvex 1
removal removal 1
biopsy biopsy 1
images images 1
fourt fourt 1
nonlinear nonlinear 1
order order 1
cmeans cmeans 1
fuzzy fuzzy 1
medical medical 1
presence presence 1
cell cell 1
nuclei nuclei 1
nonblind nonblind 1
penalty penalty 1
splitting splitting 1
variable variable 1
calcifications calcifications 1
mammography mammography 1
masses masses 1
masking masking 1
unsharp unsharp 1
superresolution superresolution 1
reconstruction reconstruction 1
exploiting exploiting 1
prior prior 1
skellam skellam 1
conditional conditional 1
field field 1
random random 1
band band 1
fdm fdm 1
fibfs fibfs 1
function function 1
intrinsic intrinsic 1
mnure mnure 1
mpg mpg 1
cross cross 1
filters filters 1
framework framework 1
functions functions 1
generalized generalized 1
validations validations 1
variational variational 1
distributions distributions 1
gamma gamma 1
gaussian gaussian 1
minimization minimization 1
models models 1
normalization normalization 1
posterior posterior 1
priors priors 1
alternating alternating 1
direction direction 1
multipliers multipliers 1
primaldual primaldual 1
problem problem 1
algorithm algorithm 1
approximation approximation 1
genetic genetic 1
local local 1
optimization optimization 1
polynomial polynomial 1
swarm swarm 1
attention attention 1
dataset dataset 1
deep deep 1
guidance guidance 1
network network 1
neural neural 1
simulation simulation 1
synthetic synthetic 1
convolutional convolutional 1
networks networks 1
selfsupervised selfsupervised 1
unpaired unpaired 1
brownian brownian 1
data data 1
motion motion 1
tracking tracking 1
reinforcement reinforcement 1
star star 1
model model 1
optimal optimal 1
recovery recovery 1
regularized regularized 1
exponential exponential 1
gammadistributed gammadistributed 1
poissondistributed poissondistributed 1
sensitivity sensitivity 1
variability variability 1
atmospheric atmospheric 1
blind blind 1
optics optics 1
turbulence turbulence 1
frame frame 1
anscombe anscombe 1
filtering filtering 1
guassian guassian 1
haar haar 1
energies energies 1
log log 1
psnr psnr 1
wiener wiener 1
dictionary dictionary 1
modeling modeling 1
# lo nombra y le da formato de data.frame
############ GRAFICAMOS LA NUBE DE PALABRAS
#wordcloud(words = dataletras$word, freq = dataletras$freq, min.freq = 1,
#max.words=100000)
#en el centro la palabra mas importante,
wordcloud(words = dataletras$word, freq = dataletras$freq, min.freq = 1,
max.words=100000, random.order=FALSE, rot.per=0.2,
colors=brewer.pal(8, "Set2"))
[1] "zhang" "kumar" "srivastava"
word frequencia
zhang zhang 6
kumar kumar 5
srivastava srivastava 5
subodh subodh 3
feng feng 3
alessandro alessandro 3
deledalle deledalle 2
harmany harmany 2
salmon salmon 2
willett willett 2
wang wang 2
zhu zhu 2
ahmed ahmed 2
and and 2
dong dong 2
sun sun 2
amita amita 2
arvind arvind 2
dhaka dhaka 2
hamurabi hamurabi 2
nandal nandal 2
ashour ashour 2
cao cao 2
liu liu 2
ming ming 2
galvantejada galvantejada 2
foi foi 2
basarab basarab 1
dutta dutta 1
georgeot georgeot 1
kouame kouame 1
bajic bajic 1
jelenkovic jelenkovic 1
pantelic pantelic 1
skoric skoric 1
berkels berkels 1
binev binev 1
dahmen dahmen 1
mevenkamp mevenkamp 1
voyles voyles 1
yankovich yankovich 1
howard howard 1
nichols nichols 1
smith smith 1
andrea andrea 1
colao colao 1
dheera dheera 1
dongeek dongeek 1
franco franco 1
goyal goyal 1
jeffrey jeffrey 1
kirmani kirmani 1
shapiro shapiro 1
shin shin 1
venkatraman venkatraman 1
vivek vivek 1
wong wong 1
hirakawa hirakawa 1
jin jin 1
keigo keigo 1
xiaodan xiaodan 1
zhenyu zhenyu 1
angelini angelini 1
elsa elsa 1
jeanchristophe jeanchristophe 1
montagner montagner 1
olivomarin olivomarin 1
yoann yoann 1
fan fan 1
haimiao haimiao 1
qibin qibin 1
yichuan yichuan 1
rajee rajee 1
rajesh rajesh 1
shuyin shuyin 1
tang tang 1
tao tao 1
wende wende 1
zhenmin zhenmin 1
zhihai zhihai 1
abhinav abhinav 1
pradeep pradeep 1
kaicong kaicong 1
krawtschenko krawtschenko 1
roman roman 1
simon simon 1
sven sven 1
tran tran 1
trunghieu trunghieu 1
jun jun 1
linyan linyan 1
wei wei 1
zhao zhao 1
zhihui zhihui 1
badri badri 1
esakkirajan esakkirajan 1
narayan narayan 1
rasal rasal 1
sankaralingam sankaralingam 1
subudhi subudhi 1
thangaraj thangaraj 1
tushar tushar 1
veerakumar veerakumar 1
rajeev rajeev 1
eneldo eneldo 1
francisco francisco 1
gamboa gamboa 1
hasmat hasmat 1
lopez lopez 1
malik malik 1
martinezacuna martinezacuna 1
monica monica 1
monteagudo monteagudo 1
rosales rosales 1
satyendra satyendra 1
singh singh 1
chan chan 1
honfu honfu 1
raymond raymond 1
tieyong tieyong 1
wen wen 1
youwei youwei 1
zeng zeng 1
amira amira 1
beagum beagum 1
dacnhuong dacnhuong 1
dey dey 1
dimitra dimitra 1
fuqian fuqian 1
gia gia 1
nguyen nguyen 1
nhu nhu 1
nilanjan nilanjan 1
pistolla pistolla 1
samsad samsad 1
shi shi 1
sifaki sifaki 1
feifan feifan 1
dongwei dongwei 1
ren ren 1
wangmeng wangmeng 1
xiaohe xiaohe 1
yue yue 1
zuo zuo 1
benfenati benfenati 1
bonacci bonacci 1
bourouina bourouina 1
francesco francesco 1
hugues hugues 1
talbot talbot 1
tarik tarik 1
binjie binjie 1
chen chen 1
huajian huajian 1
huang huang 1
qin qin 1
shaosen shaosen 1
weiyue weiyue 1
yisong yisong 1
yong yong 1
kai kai 1
wenbo wenbo 1
xie xie 1
ying ying 1
zhenduo zhenduo 1
zheng zheng 1
delaram delaram 1
hossein hossein 1
mohammad mohammad 1
motamed motamed 1
rohban rohban 1
saligrama saligrama 1
vaziri vaziri 1
venkatesh venkatesh 1
arturo arturo 1
carlos carlos 1
celayapadilla celayapadilla 1
gamboarosales gamboarosales 1
huizilopoztli huizilopoztli 1
jorge jorge 1
jose jose 1
lunagarcia lunagarcia 1
marina marina 1
morenobaez morenobaez 1
ninoslav ninoslav 1
changming changming 1
dongming dongming 1
huan huan 1
jiaqi jiaqi 1
jinhua jinhua 1
lijuan lijuan 1
peng peng 1
yang yang 1
azzari azzari 1
lucio lucio 1
charlesalban charlesalban 1
joseph joseph 1
rebecca rebecca 1
zachary zachary 1
ajay ajay 1
boyat boyat 1
brijendra brijendra 1
joshi joshi 1
makitalo makitalo 1
markku markku 1
elad elad 1
giryes giryes 1
michael michael 1
raja raja 1
[1] "Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
[1] "Grafico"
Tabla_2 <- base %>%
dplyr::group_by(DIVULGACIÓN) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(DIVULGACIÓN)
"Grafico"
[1] "Grafico"
G2<-ggplot(Tabla_2, aes(x =DIVULGACIÓN, y=Total) ) +
geom_bar(width = 0.7,stat="identity",
position = position_dodge(), fill="cyan4") +
ylim(c(0,33))+
#xlim(c(0,300)) +
#ggtitle("Un título") +
labs(x="Divulgación", y= "Frecuencias \n (Porcentajes)") +
geom_text(aes(label=paste0(Total," ", "", "(", Porcentaje, "%", ")")),
vjust=-0.9,
color="black",
hjust=0.5,
# define text position and size
position = position_dodge(0.9),
angle=0,
size=4.5) +
theme(axis.text.x = element_text(angle = 0, vjust = 1, hjust=1)) +
theme_bw(base_size = 16) +
#coord_flip() +
facet_wrap(~"Distribución de Tipo de Divulgación.")
base$BASE <- as.factor(base$BASE)
"Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
[1] "Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
Tabla_4 <- base %>%
dplyr::group_by(BASE) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(BASE)
"Grafico"
[1] "Grafico"
G4<-ggplot(Tabla_4, aes(x =BASE, y=Total) ) +
geom_bar(width = 0.9,stat="identity",
position = position_dodge(), fill="pink3") +
ylim(c(0,11))+
#xlim(c(0,300)) +
#ggtitle("Un título") +
labs(x="Base de Datos", y= "Frecuencias \n (Porcentajes)") +
geom_text(aes(label=paste0(Total," ", "", "(", Porcentaje, "%", ")")),
vjust=-0.9,
color="black",
hjust=0.5,
# define text position and size
position = position_dodge(0.9),
angle=0,
size=4.5) +
theme(axis.text.x = element_text(angle = 0, vjust = 1, hjust=1)) +
theme_bw(base_size = 17) +
#coord_flip() +
facet_wrap(~"Distribución según Base de Datos.")
base$Ruido <- as.factor(base$Ruido)
"Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
[1] "Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
Tabla_5 <- base %>%
dplyr::group_by(Ruido) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(Ruido)
"Grafico"
[1] "Grafico"
G5<-ggplot(Tabla_5, aes(x =Ruido , y=Total) ) +
geom_bar(width = 0.9,stat="identity",
position = position_dodge(), fill="orange") +
ylim(c(0,25))+
#xlim(c(0,300)) +
#ggtitle("Un título") +
labs(x="Ruido", y= "Frecuencias \n (Porcentajes)") +
geom_text(aes(label=paste0(Total," ", "", "(", Porcentaje, "%", ")")),
vjust=-0.9,
color="black",
hjust=0.5,
# define text position and size
position = position_dodge(0.9),
angle=0,
size=4.5) +
theme(axis.text.x = element_text(angle = 0, vjust = 1, hjust=1)) +
theme_bw(base_size = 16) +
#coord_flip() +
facet_wrap(~"Distribución según el Tipo de Ruido.")
base$País <- as.factor(base$País)
"Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
[1] "Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
Tabla_6 <- base %>%
dplyr::group_by(País) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(País)
"Grafico"
[1] "Grafico"
G6<-ggplot(Tabla_6, aes(x =País, y=Total)) +
geom_bar(width = 0.9,stat="identity",
position = position_dodge(), fill="blue3") +
ylim(c(0,10))+
#xlim(c(0,300)) +
#ggtitle("Un título") +
labs(x="País", y= "Frecuencias \n (Porcentajes)") +
geom_text(aes(label=paste0(Total," ", "", "(", Porcentaje, "%", ")")),
vjust=-0.9,
color="black",
hjust=0.5,
# define text position and size
position = position_dodge(0.9),
angle=0,
size=4.0) +
theme(axis.text.x = element_text(angle = 0, vjust = 1, hjust=1)) +
theme_bw(base_size = 17) +
#coord_flip() +
facet_wrap(~"Distribución según el País.")
"Creación de Tabla Cruzada"
[1] "Creación de Tabla Cruzada"
Tabla7 <- base %>%
dplyr::group_by(BASE, Ruido) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(BASE)
"Grafico"
[1] "Grafico"
G7<-ggplot(Tabla7, aes(x = BASE, y=Total, fill=Ruido) ) +
geom_bar(width = 0.9,stat="identity",
position = position_dodge()) +
ylim(c(0,7.5))+
#xlim(c(0,300)) +
#ggtitle("Un título") +
labs(x="Base de Datos", y= "Frecuencias \n (Porcentajes)") +
labs(fill = "Ruido") +
scale_fill_manual(values = c("cyan", "red2")) +
geom_text(aes(label=paste0(Total," ", "", "(", Porcentaje, "%", ")")),
vjust=-0.9,
color="black",
hjust=0.5,
# define text position and size
position = position_dodge(0.9),
angle=0,
size=3.8)+
scale_fill_discrete(name = "Ruido", labels = c("POISSON", "POISSON-GAUSSIANO")) +
theme(axis.text.x = element_text(angle = 0, vjust = 1, hjust=1)) +
theme_bw(base_size = 17) +
#coord_flip() +
facet_wrap(~"Distribución de la Base de Datos según el Tipo de Ruido.")
"Creación de Tabla Cruzada"
[1] "Creación de Tabla Cruzada"
Tabla8 <- base %>%
dplyr::group_by(País, Ruido) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(País)
"Grafico"
[1] "Grafico"
G8<-ggplot(Tabla8, aes(x = País, y=Total, fill=Ruido) ) +
geom_bar(width = 0.9,stat="identity",
position = position_dodge()) +
ylim(c(0,7.5))+
#xlim(c(0,300)) +
#ggtitle("Un título") +
labs(x="", y= "Frecuencias") +
labs(fill = "Estimación") +
scale_fill_manual(values = c("cyan", "red2")) +
geom_text(aes(label=paste0(Total," ", "")),
vjust=-0.9,
color="black",
hjust=0.5,
# define text position and size
position = position_dodge(0.9),
angle=0,
size=5.0)+
scale_fill_discrete(name = "Ruido", labels = c("POISSON", "POISSON-GAUSSIANO")) +
theme(axis.text.x = element_text(angle = 0, vjust = 1, hjust=1)) +
theme_bw(base_size = 16) +
#coord_flip() +
facet_wrap(~"Distribución del País según el Tipo de Ruido.")
[1] "Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
[1] "Grafico"
[1] "Creación de tabla de frecuencia para hacer el grafico con fr y porcentaje"
base$EST <- as.factor(base$EST)
Tabla_10 <- base %>%
dplyr::group_by(EST, Ruido) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(EST)
Tabla_10<- na.omit(Tabla_10)
G9<-ggplot(Tabla_10, aes(x = EST, y = Total,
fill = Ruido)) +
geom_bar(stat = "identity")+
coord_flip() +
facet_wrap(~"Distribución de la Estimación según el Tipo de Ruido.") +
geom_text(aes(label=paste0(Total," ", "")),
vjust=-0.1,
color="black",
hjust=0.5,
# define text position and size
position = position_stack(0.9),
angle=0,
size=3.0)+
theme_bw(base_size = 14)
base$Año <- as.factor(base$Año)
Tabla_11 <- base %>%
dplyr::group_by(Año, Ruido) %>%
dplyr::summarise(Total = n()) %>%
dplyr::mutate(Porcentaje = round(Total/sum(Total)*100, 1)) %>%
dplyr::arrange(Año)
G10<-ggplot(Tabla_11, aes(x = Año, y = Total,
fill = Ruido)) +
geom_bar(stat = "identity")+
coord_flip() +
facet_wrap(~"Distribución del Año según el Tipo de Ruido.") +
geom_text(aes(label=paste0(Total," ", "", "(", Porcentaje, "%", ")")),
vjust=-0.1,
color="black",
hjust=0.5,
# define text position and size
position = position_stack(0.9),
angle=0,
size=3.0)+
theme_bw(base_size = 14)