blob_id
stringlengths 40
40
| directory_id
stringlengths 40
40
| path
stringlengths 3
327
| content_id
stringlengths 40
40
| detected_licenses
sequencelengths 0
91
| license_type
stringclasses 2
values | repo_name
stringlengths 5
134
| snapshot_id
stringlengths 40
40
| revision_id
stringlengths 40
40
| branch_name
stringclasses 41
values | visit_date
timestamp[us]date 2016-08-02 22:44:29
2023-09-06 08:39:28
| revision_date
timestamp[us]date 1977-08-08 00:00:00
2023-09-05 12:13:49
| committer_date
timestamp[us]date 1977-08-08 00:00:00
2023-09-05 12:13:49
| github_id
int64 19.4k
646M
⌀ | star_events_count
int64 0
40k
| fork_events_count
int64 0
32.4k
| gha_license_id
stringclasses 13
values | gha_event_created_at
timestamp[us]date 2012-09-05 15:55:17
2023-09-14 21:52:42
⌀ | gha_created_at
timestamp[us]date 2008-05-25 01:21:32
2023-05-27 23:04:20
⌀ | gha_language
stringclasses 57
values | src_encoding
stringclasses 24
values | language
stringclasses 1
value | is_vendor
bool 1
class | is_generated
bool 1
class | length_bytes
int64 7
9.18M
| extension
stringclasses 1
value | filename
stringlengths 2
141
| content
stringlengths 7
9.18M
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
bd81e37dfcf6b3cdb4c0bd715854b39667cedc7d | d6ff1e6257582f785915e3a0fad3d4896ebd9acb | /R_old/OVERALL_TRANSPIRATION.R | dd4c315e6edbe8f2886bcf7adad85997b5a0dd40 | [] | no_license | RemkoDuursma/Kelly2015NewPhyt | 355084d7d719c30b87200b75887f5521c270b1b5 | 447f263f726e68298ee47746b4de438fbc8fdebf | refs/heads/master | 2021-01-15T13:02:00.392770 | 2015-09-08T04:56:15 | 2015-09-08T04:56:15 | 42,089,956 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 3,247 | r | OVERALL_TRANSPIRATION.R | setwd("C:/Documents and Settings/Jeffrey Kelly/Desktop/EUC DATA/EUC OVERALL BIOMASS")
PILBIOMASS<-read.csv("PILTRANSAA.csv",sep=",", header=TRUE)
names(PILBIOMASS)
str(PILBIOMASS)
windows(width=8, height=4) #, pointsize=18)
par(xaxs="i",yaxs="i")
par(las=2)
par(mar=c(4.5,4.5,1,1))
par(xaxs="i",yaxs="i")
par(mfrow=c(1,2), cex.lab=1)
#PIL
par(las=1)
with(PILBIOMASS,plot(E[ST=="PILAD"]~D[ST=="PILAD"],col="blue",pch=1,
ylim=range(0,1.1*max(E)),xlim=range(0,1.05*max(D)),
ylab="",xlab=expression(bold(Day))))
title(main="Eucalyptus pilularis", font.main=4,cex.main=1)
with(PILBIOMASS,arrows(D[ST=="PILAD"],
ESE[ST=="PILAD"], D[ST=="PILAD"], LSE[ST=="PILAD"]
, length = .035, angle = 90, code = 3,col="blue"))
#or
with(PILBIOMASS,points(E[ST=="PILAND"]~D[ST=="PILAND"],col="blue",pch=16))
with(PILBIOMASS,arrows(D[ST=="PILAND"],
ESE[ST=="PILAND"], D[ST=="PILAND"], LSE[ST=="PILAND"]
, length = .035, angle = 90, code = 3,col="blue"))
with(PILBIOMASS,points(E[ST=="PILED"]~D[ST=="PILED"],col="red",pch=1))
with(PILBIOMASS,arrows(D[ST=="PILED"],
ESE[ST=="PILED"], D[ST=="PILED"], LSE[ST=="PILED"]
, length = .035, angle = 90, code = 3,col="red"))
with(PILBIOMASS,points(E[ST=="PILEND"]~D[ST=="PILEND"],col="red",pch=16))
with(PILBIOMASS,arrows(D[ST=="PILEND"],
ESE[ST=="PILEND"], D[ST=="PILEND"], LSE[ST=="PILEND"]
, length = .035, angle = 90, code = 3,col="red"))
par(las=3)
mtext(side = 2, text =expression(bold(Transpiration~~(l~week^-1))), line = 2.5,font=2, cex=1.0)
legend("topleft", expression(aC[a]~-~W, aC[a]~-~D,eC[a]~-~W ,eC[a]~-~D),
cex=0.75,bty="n",
pch = c(16,1,16,1), col=c("blue","blue","red","red"), #xjust = .5, yjust = .5,
)
par(las=1)
setwd("C:/Documents and Settings/Jeffrey Kelly/Desktop/EUC DATA/EUC OVERALL BIOMASS")
POPBIOMASS<-read.csv("POPTRANSAA.csv",sep=",", header=TRUE)
names(POPBIOMASS)
str(POPBIOMASS)
#POP
#bottom,left,top,right
par(xaxs="i",yaxs="i")
par(las=2)
par(mar=c(4.5,1,1,4.5))
par(las=1)
with(POPBIOMASS,plot(E[ST=="POPAD"]~D[ST=="POPAD"],col="blue",pch=1,yaxt="n",
ylim=c(0, 1.1*max(E)),xlim=c(0,1.05*max(D)),
ylab="",xlab=expression(bold(Day))))
title(main="Eucalyptus populnea", font.main=4,cex.main=1)
with(POPBIOMASS,arrows(D[ST=="POPAD"],
ESE[ST=="POPAD"], D[ST=="POPAD"], LSE[ST=="POPAD"]
, length = .035, angle = 90, code = 3,col="blue"))
axis(4,labels=TRUE,tcl=-0.5,cex.axis=1)
#or
with(POPBIOMASS,points(E[ST=="POPAND"]~D[ST=="POPAND"],col="blue",pch=16))
with(POPBIOMASS,arrows(D[ST=="POPAND"],
ESE[ST=="POPAND"], D[ST=="POPAND"], LSE[ST=="POPAND"]
, length = .035, angle = 90, code = 3,col="blue"))
with(POPBIOMASS,points(E[ST=="POPED"]~D[ST=="POPED"],col="red",pch=1))
with(POPBIOMASS,arrows(D[ST=="POPED"],
ESE[ST=="POPED"], D[ST=="POPED"], LSE[ST=="POPED"]
, length = .035, angle = 90, code = 3,col="red"))
with(POPBIOMASS,points(E[ST=="POPEND"]~D[ST=="POPEND"],col="red",pch=16))
with(POPBIOMASS,arrows(D[ST=="POPEND"],
ESE[ST=="POPEND"], D[ST=="POPEND"], LSE[ST=="POPEND"]
, length = .035, angle = 90, code = 3,col="red"))
# looks great on screeen / printer
dev.copy2pdf(file="somname.pdf")
# looks great printed, or after printing MS to PDF
# this is the one you paste in word
dev.copy2eps(file="fig19.eps")
|
7e8c94c982763d3b9a74d47bf81ecba200e74f3e | a47ce30f5112b01d5ab3e790a1b51c910f3cf1c3 | /A_github/sources/authors/2774/plotly/coord.R | c489eb9d4c358419e3fd6f91a129c297999fc8aa | [] | no_license | Irbis3/crantasticScrapper | 6b6d7596344115343cfd934d3902b85fbfdd7295 | 7ec91721565ae7c9e2d0e098598ed86e29375567 | refs/heads/master | 2020-03-09T04:03:51.955742 | 2018-04-16T09:41:39 | 2018-04-16T09:41:39 | 128,578,890 | 5 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,359 | r | coord.R | #' *** This won't be possible until plotly.js implements aspect ratios... ***
#'
#' #' Force the aspect ratio according to x and y scales
#' #'
#' #' When x and y are numeric variables measured on the same scale,
#' #' or are related in some meaningful way, forcing the aspect ratio of the
#' #' plot to be proportional to the ratio of a unit change in x versus y improves
#' #' our ability to correctly perceive the data.
#' #'
#' #' @param p a plotly object
#' #' @param ratio aspect ratio, expressed as y / x
#' #' @export
#' #' @examples
#' #'
#' #' canada <- map_data("world", "canada")
#' #'
#' #' canada %>%
#' #' group_by(group) %>%
#' #' plot_ly(x = ~long, y = ~lat, alpha = 0.2) %>%
#' #' add_polygons(hoverinfo = "none", color = I("black")) %>%
#' #' coord_fix()
#' #'
#' #' # works on (non-faceted) ggplot2 plots, too
#' #' gg <- ggplot(canada, aes(long, lat, group = group)) +
#' #' geom_polygon() + coord_fixed()
#' #'
#' #' gg %>%
#' #' ggplotly() %>%
#' #' coord_fix()
#' #'
#'
#' coord_fix <- function(p, ratio = 1) {
#' p <- plotly_build(p)
#' # this won't work for subplots, or categorical data
#' x <- grepl("^xaxis", names(p$x$layout))
#' y <- grepl("^yaxis", names(p$x$layout))
#' if (sum(x) > 1 || sum(y) > 1) {
#' stop("Can not impose aspect ratio a plot with more than one x/y axis", call. = FALSE)
#' }
#' xDat <- unlist(lapply(p$x$data, "[[", "x"))
#' yDat <- unlist(lapply(p$x$data, "[[", "y"))
#' if (!is.numeric(xDat) || !is.numeric(yDat)) {
#' stop("Must have numeric data on both x and y axes to enforce aspect ratios", call. = FALSE)
#' }
#'
#' # warn about any pre-populated domains, they will get squashed
#' xDom <- p$x$layout[["xaxis"]]$domain %||% c(0, 1)
#' yDom <- p$x$layout[["yaxis"]]$domain %||% c(0, 1)
#' if (!identical(yDom, c(0, 1)) || !identical(xDom, c(0, 1))) {
#' warning(
#' "coord_fix() won't respect prespecified axis domains (other than the default)",
#' call. = FALSE
#' )
#' }
#'
#' xRng <- range(xDat, na.rm = TRUE)
#' yRng <- range(yDat, na.rm = TRUE)
#' asp <- ratio * diff(yRng) / diff(xRng)
#' if (asp < 1) {
#' p$x$layout[["yaxis"]]$domain <- c(0 + asp / 2, 1 - asp / 2)
#' } else {
#' asp <- 1 / asp
#' p$x$layout[["xaxis"]]$domain <- c(0 + asp / 2, 1 - asp / 2)
#' }
#' p
#' }
|
87612036fd5fa980712ac1e05cfc398425c50685 | 86c0b4c6c1746ebf0441c62421748190d057067d | /plot/mass.R | 20769a023b830ddef7c35ab7c099b5ac260e9f87 | [
"MIT"
] | permissive | yufree/democode | 372f0684c49505965b0ba5abe0675c2b6f7fb3da | 0a332ac34a95677ce859b49033bdd2be3dfbe3c4 | refs/heads/master | 2022-09-13T11:08:55.152350 | 2022-08-28T23:09:00 | 2022-08-28T23:09:00 | 20,328,810 | 5 | 14 | null | 2017-01-06T16:07:25 | 2014-05-30T12:41:28 | HTML | UTF-8 | R | false | false | 1,185 | r | mass.R | source("http://bioconductor.org/biocLite.R")
biocLite("mzR")
library(mzR)
all <- openMSfile('./FULL200.CDF')
df <- header(all)
bb <- peaks(all)
aaaa <- sapply(bb,as.data.frame)
oddvals <- seq(1, ncol(aaaa), by=2)
aaaaa <- unlist(aaaa[oddvals])
ccc <- unique(c(aaaaa))
ccc <- ccc[order(ccc)]
# bbb <- sapply(bb, "[",250:700)
# ddd <- unique(c(bbb))
# dddd <- ddd[ddd<700]
time <- df$retentionTime
df2 <- matrix(0, nrow = length(ccc), ncol = length(time))
rownames(df2) <- ccc
colnames(df2) <- time
rm(aaaa)
rm(aaaaa)
rm(oddvals)
rm(df)
rm(all)
gc()
for(i in 1:length(time)){
temp <- bb[[i]]
index <- which(ccc%in%temp[,1])
df2[index,i] <- temp[,2]
}
ddd <- as.integer(ccc)
library(data.table)
dt = data.table(df2)
dt$fac <- ddd
df3 <- dt[,lapply(.SD, sum), by=ddd ]
df3 <- as.matrix(df3)
df7 <- df3[,2000:3000]
heatmap(df7)
library(rARPACK)
df4 <- svds(df3,2)
df5 <- df4$u %*% diag(df4$d) %*% t(df4$v)
rownames(df5) <- ddd
colnames(df5) <- time
df6 <- df5[,2000:3000]
heatmap(df6)
df8 <- as.data.frame(df5)
df9 <- as.data.frame(t(df8))
rownames(df8) <- ddd
colnames(df9) <- time
write.table(df3,'df3.txt')
|
99a524e8baa9751bbd5db7787f3567c66a6e8bee | 4450235f92ae60899df1749dc2fed83101582318 | /ThesisRpackage/R/3Article_old/GSE42861_function.R | 4e60f4de4028e99df28eb3e6e687f0b5409e866e | [
"MIT"
] | permissive | cayek/Thesis | c2f5048e793d33cc40c8576257d2c9016bc84c96 | 14d7c3fd03aac0ee940e883e37114420aa614b41 | refs/heads/master | 2021-03-27T20:35:08.500966 | 2017-11-18T10:50:58 | 2017-11-18T10:50:58 | 84,567,700 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 8,617 | r | GSE42861_function.R | #' main experiment
#'
#' @export
GSE42861_experiment <- function(s, save = TRUE) {
# glm
glm <- Method(name = "glm",
hypothesis.testing.method = phenotypeWayReg_glm_score(family = binomial,
factorized.X1 = TRUE),
impute.genotype.method = imputeByMean(),
nickname = "glm")
# glm + 2 PCs
glm_2PC <- PCAClassicLinearMethod(K = 2,
center = TRUE,
hypothesis.testing.method = phenotypeWayReg_glm_score(family = binomial,
factorized.X1 = TRUE),
nickname = "glm+2PCs",
assumingStructure = FALSE)
# glm + 6 pcs
glm_6PC <- PCAClassicLinearMethod(K = 6,
center = TRUE,
hypothesis.testing.method = phenotypeWayReg_glm_score(family = binomial,
factorized.X1 = TRUE),
nickname = "glm+6PCs",
assumingStructure = FALSE)
# glm + 6 refactor
glm_6refractor <- refractorMethod(K = 6,
verbose = FALSE,
t = 500,
nickname = "glm+6refractor")
# glm + 6 lfmm ridge
glm_6lfmm.ridge <- RidgeLFMMMethod(K = 6,
hypothesis.testing.method = phenotypeWayReg_glm_score(),
lambda = 1e6,
nickname = "glm+6lfmm")
# lfmm ridge
lfmm.ridge <- RidgeLFMMMethod(K = 6,
hypothesis.testing.method = lm_zscore(gif = FALSE),
lambda = 1e6,
nickname = "lfmm ridge")
# run exp
exp <- ComparisonExperiment(s,
glm,
glm_2PC,
glm_6PC,
glm_6refractor,
glm_6lfmm.ridge,
lfmm.ridge)
exp <- runExperiment(exp)
# save exp
if (save) {
dumpExperiment(exp)
}
exp
}
#' @export
GSE42861_plot <- function(exp) {
# Rmk: i am only interested in pvalue1 and score1 other pavalue was not computed
# with glm of lm
# qqplot
ggplot(exp$df.res %>% dplyr::filter(variable.name == "pvalue1")) +
stat_qq(aes(sample = -log10(estimate)),
distribution = stats::qexp, dparams = list(rate = log(10))) +
geom_abline(slope = 1, intercept = 0) +
facet_grid(method.name~.) +
ggtitle("-log10(pvalue) qqplot")
}
#' @export
GSE42861_get_RahmaniLoci <- function() {
table <- tabulizer::extract_tables("~/Projects/Biblio/org-ref-pdfs/SF_Rahmani_2016.pdf",
pages = 19, method = "data.frame")[[1]]
table
}
################################################################################
# Long running
#' Run of PCA
#'
#'
#' @export
long_GSE42861_PCA <- function() {
library(Article3Package)
G.file <- "~/Projects/Data2016_2017/GSE42861/betanormalized_metylationlvl.filtered.rds"
X.file <- "~/Projects/Data2016_2017/GSE42861/X.rds"
s <- TrueSampler(G.file = G.file,
X.file = X.file,
outlier.file = NULL,
n = NULL,
L = NULL)
exp <- HGDP_PCA(s, save = TRUE)
}
#' Run of LFMM
#'
#'
#' @export
long_GSE42861_LFMM <- function() {
cl <- parallel::makeCluster(2)
doParallel::registerDoParallel(cl)
library(Article3Package)
G.file <- "~/Projects/Data2016_2017/GSE42861/betanormalized_metylationlvl.filtered.rds"
X.file <- "~/Projects/Data2016_2017/GSE42861/X.rds"
s <- TrueSampler(G.file = G.file,
X.file = X.file,
outlier.file = NULL,
n = NULL,
L = NULL)
lambdas <- c(1e-10, 1e0, 1e2, 1e10)
Ks <- c(1,6,8,20)
HGDB_runs(s, Ks = Ks, lambdas = lambdas, save = TRUE)
}
#' Run of GSE42861_experiment
#'
#'
#' @export
long_GSE42861_exp <- function() {
library(Article3Package)
G.file <- "~/Projects/Data2016_2017/GSE42861/betanormalized_metylationlvl.rds"
X.file <- "~/Projects/Data2016_2017/GSE42861/X.rds"
s <- TrueSampler(G.file = G.file,
X.file = X.file,
outlier.file = NULL,
n = NULL,
L = NULL)
cl <- parallel::makeCluster(6)
doParallel::registerDoParallel(cl)
exp <- GSE42861_experiment(s, save = TRUE)
exp
}
#' cross validation
#'
#'
#' @export
long_GSE42861_CrossVal <- function(cluster.nb = NULL,
K = 6,
G.file = "~/Projects/Data2016_2017/GSE42861/betanormalized_metylationlvl.filtered.rds",
X.file = "~/Projects/Data2016_2017/GSE42861/X.rds",
lambdas = c(1e-10, 1e0, 1e2, 1e10),
rep = 5,
missing.prop = 0.5,
save = TRUE,
bypass = FALSE) {
KrakTest(bypass)
if (!is.null(cluster.nb)) {
cl <- parallel::makeCluster(cluster.nb)
doParallel::registerDoParallel(cl)
}
s <- TrueSampler(G.file = G.file,
X.file = X.file,
outlier.file = NULL,
n = NULL,
L = NULL)
dat <- sampl(s)
m <- finalLfmmRdigeMethod(K = K,
lambda = NULL)
description <- paste0("long_GSE42861_CrossVal with K=", K,
"and lambdas = ",paste(lambdas,collapse = '|'))
exp <- Experiment(name = "long_GSE42861_CrossVal", description = description)
exp$crossvalidation.res <- crossvalidation_kfold_missingvalue(m = m,
dat = dat,
rep = rep,
missing.prop = missing.prop,
lambdas = lambdas)
# save exp
if (save) {
dumpExperiment(exp)
}
exp
}
#' cross validation
#'
#'
#' @export
long_GSE42861_lfmm_glm <- function(K.lfmm = 6,
K.refactor = 6,
G.file = "~/Projects/Data2016_2017/GSE42861/betanormalized_metylationlvl.filtered.rds",
X.file = "~/Projects/Data2016_2017/GSE42861/X.rds",
lambda = 1e-10,
save = TRUE,
bypass = FALSE,
refactor = FALSE) {
KrakTest(bypass)
s <- TrueSampler(G.file = G.file,
X.file = X.file,
outlier.file = NULL,
n = NULL,
L = NULL)
dat <- sampl(s)
G <- dat$G
X <- dat$X
## other co.var correction
dat$G <- G
dat$X <- X[,-1]
m.lm <- finalLm()
m.lm <- fit(m.lm, dat)
m.lfmm <- finalLfmmRdigeMethod(K = K.lfmm,
lambda = lambda)
m.refactor <- finalRefactorMethod(K = K.refactor)
description <- paste0("long_GSE42861_lfmm_glm with K=", K.lfmm,
"and lambdas = ", lambda)
exp <- Experiment(name = "long_GSE42861_lfmm_glm", description = description)
# run of the method
dat$G <- m.lm$epsilon
dat$X <- X[,1, drop = FALSE]
exp$m.lfmm <- fit(m.lfmm, dat)
exp$m.refactor <- fit(m.refactor, dat)
# hypothesis testing
glm.aux <- function(m, name) {
glm.func <- phenotypeWayReg_glm_score(family = binomial,
factorized.X1 = TRUE)
glm.res <- glm.func$fun(m, dat)
df <- tibble(index = 1:length(glm.res$score), method.name = name,
estimate = glm.res$score[1,], variable.name = "score")
df <- tibble(index = 1:length(glm.res$pvalue), method.name = name,
estimate = glm.res$pvalue[1,], variable.name = "pvalue") %>%
rbind(df)
df
}
exp$df.res <- rbind(glm.aux(exp$m.refactor, "refactor"),
glm.aux(exp$m.lfmm, "lfmm"))
# save exp
if (save) {
dumpExperiment(exp)
}
exp
}
|
ac457a941d93eb56777aeb1bda10707ce8907e13 | c54d1c0a3d81bddb25f3f55078f305ad6c15997b | /R/get_internal_tree.R | ffd29651c660f728d71fcc716d3ca033637fb637 | [] | no_license | cran/genpathmox | dc065d3b5ea1c8632068fe3d9bfa7b063045bb2c | 517be94b39d8742cd3d39aedc152e026d865afd6 | refs/heads/master | 2023-01-12T03:39:55.183481 | 2022-12-22T10:00:12 | 2022-12-22T10:00:12 | 25,984,875 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 9,114 | r | get_internal_tree.R | #' ############################################################################################
#' @title Calculating size (numeber of individual of a node)
#' @details
#' Internal function
#' @param x matrix or dataframe with data.
#' @param size value indicating the minimun threshold of number of observations for a node
#' @return the number of observations in a node
#' @keywords internal
#' @export
#'
percent.node <- function(x,size)
{
indiv = nrow(x)
min.n.ind = trunc(indiv*size)
list(min.n.ind=min.n.ind)
}
#' ############################################################################################
#' @title Calculating Deepth stop criterion
#' @details
#' Internal function
#' @param node id that identifies a specicif node
#' @return deepth of the tree
#' @keywords internal
#' @export
#'
showDeepth=function(node)
{
return (trunc(log2(node@id)))
}
#' ############################################################################################
#' @title Observations belonging to the root node
#' @details
#' Internal function
#' @param moxtree class containing the moxtree elements
#' @return the observations belonging to the root node
#' @keywords internal
#' @export
#'
root.tree <- function(moxtree)
{
root = NULL
for (n in moxtree@nodes)
{
if (n@id == 1)
{
root=n@elements
}
}
root
}
#' ############################################################################################
#' @title Observations belonging to the terminal nodes
#' @details
#' Internal function
#' @param moxtree class containing the moxtree element.
#' @return the observations belonging to the terminal nodes
#' @keywords internal
#' @export
#'
terminal.tree <- function(moxtree)
{
terminal = list()
id = list()
if (length(moxtree@nodes) > 1)
{
for (n in moxtree@nodes)
{
if (n@id == 1)
{
terminal[[length(terminal)+1]] = n@elements
id[[length(id)+1]] = "Root"
}
if (length(n@childs) == 0)
{
terminal[[length(terminal)+1]] = n@elements
id[[length(id)+1]] = n@id
}
}
for (i in 1:length(terminal)){names(terminal) = paste("node",id)}
terminal
}
else
{
terminal = NULL
}
terminal
}
#' ############################################################################################
#' @title Observations belonging to the nodes
#' @details
#' Internal function
#' @param moxtree class containing the moxtree elements
#' @return the observations belonging to the nodes
#' @keywords internal
#' @export
#'
nodes.tree <- function(moxtree)
{
nodes = list()
id = list()
if (length(moxtree@nodes) > 1)
{
for (n in moxtree@nodes)
{
nodes[[length(nodes)+1]] = n@elements
id[[length(id)+1]] = n@id
}
for (i in 1:length(nodes)) {names(nodes) = paste("node",id)}
nodes
}
else
{
nodes = NULL
}
nodes
}
#' ############################################################################################
#' @title Posibble partions for each node of the tree
#' @details
#' Internal function
#' @param moxtree class containing the moxtree elements
#' @return the Posibble partions for each node of the tree
#' @keywords internal
#' @export
#'
candidates.tree <- function(moxtree)
{
candidates = list()
id = list()
if (length(moxtree@nodes) > 1)
{
for (n in moxtree@nodes)
{
if (length(n@childs)>0)
{
candidates[[length(candidates)+1]] = n@info@candidates
id[[length(id)+1]] = n@id
}
}
for (i in 1:length(candidates)) {names(candidates) = paste("node",id)}
candidates
}
else
{
candidates = NULL
}
candidates
}
#' ############################################################################################
#' @title F-global test results for each tree partition
#' @details
#' Internal function
#' @param moxtree class containing the moxtree elements
#' @return the F-global test results for each tree partition
#' @keywords internal
#' @export
#'
fglobal.tree <- function(moxtree)
{
fglobal = list()
fgtable = NULL
if (length(moxtree@nodes) > 1)
{
for (n in moxtree@nodes)
{
if (length(n@childs) > 0)
{
fglobal[[length(fglobal)+1]] = data.frame(n@id,n@info@fgstatistic,n@info@fpvalg,n@info@variable,t(n@info@level))
}
}
for (i in 1:length(fglobal)) {fgtable = rbind(fgtable,fglobal[[i]])}
colnames(fgtable) = c("node","F value","Pr(>F)","variable","g1.mod","g2.mod")
Fg.r = fgtable
}
else
{
Fg.r = NULL
}
Fg.r
}
#' ############################################################################################
#' @title F-coefficients test results for each tree partition
#' @details
#' Internal function
#' @param moxtree class containing the moxtree elements
#' @return the F-coefficients test results for each tree partition
#' @keywords internal
#' @export
fcoef.tree <- function(moxtree)
{
fc = list()
id = list()
fctable = NULL
if (length(moxtree@nodes) > 1)
{
for (n in moxtree@nodes)
{
if (length(n@childs) > 0)
{
id[[length(id)+1]] = n@id
fctable = data.frame(as.matrix(n@info@fcstatistic),as.matrix(n@info@fpvalc))
colnames(fctable) = c("F value","Pr(>F)")
fc[[length(fc)+1]] = fctable
}
}
names(fc) = paste("node",id,sep="")
Fc.r = fc
}
else
{
Fc.r=list(fc=NULL,Signif=NULL)
}
Fc.r
}
#' ############################################################################################
#' @title General information about the tree
#' @details
#' Internal function
#' @param moxtree class containing the tree elements
#' @return a dataframe containing information about the tree and its nodes
#' @keywords internal
#' @export
#'
mox.tree <- function(moxtree)
{
info.node = list()
type = NULL
terminal = NULL
perc = NULL
var = NULL
mox = NULL
if (length(moxtree@nodes)>1)
{
for (n in moxtree@nodes)
{
if (n@id == 1)
{
length.root = length(n@elements)
}
if (length(n@childs) > 0)
{
info.node[[length(info.node)+1]] = data.frame(n@info@variable,n@id,n@childs,n@info@level)
}
if (length(n@childs) == 0)
{
type = "least"
terminal = "yes"
}
if (n@father == 0)
{
type = "root"
terminal = "no"
}
if (n@father!=0 && length(n@childs) != 0)
{
type = "node"
terminal = "no"
}
perc = round((length(n@elements)/length.root)*100,2)
data = data.frame(n@id,n@father,showDeepth(n),type,terminal,length(n@elements),perc)
mox = rbind(mox,data)
}
data.info.node = NULL
for (i in 1:length(info.node)) {data.info.node = rbind(data.info.node,info.node[[i]])}
names(data.info.node)[2] = "n.father"
names(data.info.node)[3] = "n.id"
MOX =merge (mox, data.info.node,by="n.id",all.x=TRUE)[,-9]
names(MOX) = c("node","parent","depth","type","terminal","size","%","variable","category")
MOX
}
else
{
MOX = NULL
}
MOX
}
#' ############################################################################################
#' @title General information about the pathmox algorithm
#' @details
#' Internal function
#' @param signif stop condition 1: significance of the p-value
#' @param size stop condition 2: minimum number of individuals in a node
#' @param deep stop condition 3: maximum tree depth level
#' @param y: set of segmentation variables
#' @keywords internal
#' @export
info.mox <- function(signif,size,deep,y)
{
cat("\n")
cat("PLS-SEM PATHMOX ANALYSIS","\n")
cat("\n")
cat("---------------------------------------------")
cat("\n")
cat("Info parameters algorithm","\n")
info.value = rbind(signif,size,deep)
dimnames(info.value) = NULL
info.name = c("threshold signif.","node size limit(%)","tree depth level")
info.tree = data.frame(info.name,info.value)
names(info.tree) = c("parameters algorithm", "value")
print(info.tree)
cat("\n")
cat("---------------------------------------------")
cat("\n")
cat("Info segmentation variables","\n")
type.y = rep(0, ncol(y))
treat.y = rep("binary", ncol(y))
for (i in 1:length(type.y))
{
type.y[i] = ifelse(is.ordered(y[, i]), "ord","nom")
if (nlevels(y[, i]) > 2)
if (is.ordered(y[, i]))
treat.y[i] = "ordinal"
else treat.y[i] = "nominal"
}
df.y = data.frame(nlevels = unlist(lapply(y, nlevels)),ordered = unlist(lapply(y, is.ordered)),
treatment = treat.y)
if (y[1,1] == 1){
df.y = df.y[-1,]
}
else
{
df.y
}
print(df.y)
}
#' ############################################################################################
#' @title printing the tree structure
#' @details
#' Internal function.
#' @param moxtree moxtree object
#' @return the tree structure
#' @keywords internal
#' @export
#'
printTree <- function(moxtree)
{
for (n in moxtree@nodes){
print (n)
}
}
|
84afd0009d68337cd59225335f8ca45ec7753b3d | c2061964216f76ad0f440c76dbfe1119e0279a22 | /R/API-methods.R | 65f3d6cff7f46778421a4f00c57d3ebfa0b38824 | [] | no_license | cran/antaresRead | 046829e05e411adfb55fc652ad49ea84f2610264 | f6a182b21854e12c5c470afcd38c26f44fb2b8d5 | refs/heads/master | 2023-04-16T10:45:23.521378 | 2023-04-06T16:20:02 | 2023-04-06T16:20:02 | 87,090,660 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 4,663 | r | API-methods.R |
#' API methods
#'
#' @param endpoint API endpoint to interrogate, it will be added after `default_endpoint`.
#' Can be a full URL (by wrapping ìn [I()]), in that case `default_endpoint` is ignored.
#' @param ... Additional arguments passed to API method.
#' @param default_endpoint Default endpoint to use.
#' @param opts Antares simulation options or a `list` with an `host = ` slot.
#'
#' @return Response from the API.
#' @export
#'
#' @name API-methods
#'
#' @importFrom httr GET accept_json stop_for_status content add_headers timeout
#'
#' @examples
#' \dontrun{
#'
#' # List studies with local API
#' api_get(
#' opts = list(host = "http://0.0.0.0:8080"),
#' endpoint = NULL
#' )
#'
#' }
api_get <- function(opts, endpoint, ..., default_endpoint = "v1/studies") {
if (inherits(endpoint, "AsIs")) {
opts$host <- endpoint
endpoint <- NULL
default_endpoint <- NULL
}
if (is.null(opts$host))
stop("No host provided in `opts`: use a valid simulation options object or explicitly provide a host with opts = list(host = ...)")
config <- c(
opts$httr_config,
list(
accept_json()
)
)
if (!is.null(opts$token) && opts$token != "") {
config <- c(
config,
add_headers(Authorization = paste("Bearer ", opts$token))
)
}
if (is.null(opts$timeout))
opts$timeout <- 60
result <- GET(
url = URLencode(paste(c(opts$host, default_endpoint, endpoint), collapse = "/")),
config = config,
timeout(opts$timeout),
...
)
#fix for skipping 404 when some output is missing
url_elements <- strsplit(result$url, "%2F")[[1]]
condition_status_check <- !(!is.na(url_elements[4]) & url_elements[4] %in% c("economy","adequacy") & result$status_code == 404)
if (condition_status_check) stop_for_status(result) else warn_for_status(result)
content(result)
}
#' @export
#'
#' @rdname API-methods
#'
#' @importFrom httr POST accept_json content_type_json stop_for_status content add_headers
api_post <- function(opts, endpoint, ..., default_endpoint = "v1/studies") {
if (inherits(endpoint, "AsIs")) {
opts$host <- endpoint
endpoint <- NULL
default_endpoint <- NULL
}
if (is.null(opts$host))
stop("No host provided in `opts`: use a valid simulation options object or explicitly provide a host with opts = list(host = ...)")
config <- c(
opts$httr_config,
list(
accept_json(),
content_type_json()
)
)
if (!is.null(opts$token) && opts$token != "") {
config <- c(
config,
add_headers(Authorization = paste("Bearer ", opts$token))
)
}
result <- POST(
url = URLencode(paste(c(opts$host, default_endpoint, endpoint), collapse = "/")),
config = config,
...
)
stop_for_status(result)
content(result)
}
#' @export
#'
#' @rdname API-methods
#'
#' @importFrom httr PUT accept_json stop_for_status content add_headers
api_put <- function(opts, endpoint, ..., default_endpoint = "v1/studies") {
if (inherits(endpoint, "AsIs")) {
opts$host <- endpoint
endpoint <- NULL
default_endpoint <- NULL
}
if (is.null(opts$host))
stop("No host provided in `opts`: use a valid simulation options object or explicitly provide a host with opts = list(host = ...)")
if (!is.null(opts$token) && opts$token != "") {
config <- add_headers(Authorization = paste("Bearer ", opts$token), Accept = "application/json")
} else {
config <- add_headers(Accept = "application/json")
}
result <- PUT(
url = URLencode(paste(c(opts$host, default_endpoint, endpoint), collapse = "/")),
config,
...
)
stop_for_status(result)
content(result)
}
#' @export
#'
#' @rdname API-methods
#'
#' @importFrom httr DELETE accept_json stop_for_status content
api_delete <- function(opts, endpoint, ..., default_endpoint = "v1/studies") {
if (inherits(endpoint, "AsIs")) {
opts$host <- endpoint
endpoint <- NULL
default_endpoint <- NULL
}
if (is.null(opts$host))
stop("No host provided in `opts`: use a valid simulation options object or explicitly provide a host with opts = list(host = ...)")
config <- c(
opts$httr_config,
list(
accept_json()
)
)
if (!is.null(opts$token) && opts$token != "") {
config <- c(
config,
add_headers(Authorization = paste("Bearer ", opts$token))
)
}
result <- DELETE(
url = URLencode(paste(c(opts$host, default_endpoint, endpoint), collapse = "/")),
config = config,
...
)
stop_for_status(result)
content(result)
}
|
d29addc45ad1540ad95c8544e8002562baf29435 | d8affab3b21ca33c2b6397e28171c4ad69b03d98 | /regression.R | 471e4414ef889e20c3e50e5acbebf24faa2d7f99 | [] | no_license | nupurkok/analytics | 3e69e9eb88d9eb6cc4f33ae105b7993c46a69fce | b0b76dd306e443aae010cac55ffcda484c39ad42 | refs/heads/master | 2020-03-28T15:22:27.782207 | 2018-09-16T13:03:53 | 2018-09-16T13:03:53 | 148,586,169 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 546 | r | regression.R | women
str(women)
cor(women$height, women$weight)
cov(women$height, women$weight)
plot(women)
#create linear model
fit1 = lm (formula=weight ~ height,data = women)
summary(fit1)
fitted(fit1)
cbind(women, fitted(fit1), residuals(fit1))
ndata1 = data.frame(height = c(62.5, 63.5))
predict(fit1, newdata = ndata1)
#multiple linear regression
#predict mpg vs wt, hp
mtcars
fit2 = lm(mpg ~ wt + hp, data = mtcars)
summary(fit2)
range(mtcars$wt) ; range(mtcars$hp)
ndata2=data.frame(wt=c(2.5,3.4), hp=c(100,250))
predict(fit2, newdata = ndata2)
|
46c4e6309d7e779524b8b1a79263f38885577650 | ebb09f52b1ee12d8ae8d4c493e6f1079ee57868c | /ExploratoryDataAnalysis/Project2/plot1.R | 344f1ab64d1fa16fc56bc45754d6205e3ffc4c86 | [] | no_license | r6brian/datasciencecoursera | a1723f812a34eee7094dfaa0bfde6c618b349d6c | 548944d3ba68d302160f05158fb90859bc4c8bae | refs/heads/master | 2021-01-19T10:29:54.605308 | 2015-08-23T20:00:04 | 2015-08-23T20:00:04 | 26,268,379 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 634 | r | plot1.R | # 1. Have total emissions from PM2.5 decreased in the United States from 1999 to 2008?
# Read data files
NEI <- readRDS("data/exdata-data-NEI_data/summarySCC_PM25.rds")
SCC <- readRDS("data/exdata-data-NEI_data/Source_Classification_Code.rds")
# aggregrate based upon Emissions and Years
totalEmissions <- aggregate(Emissions ~ year, NEI, sum)
# plot a bar graph
png('plot1.png')
barplot(height=totalEmissions$Emissions/10^6,
names.arg=totalEmissions$year,
xlab="years",
ylab=expression('total PM'[2]*' emission(10^6 Tons)'),
main=expression('Total PM'[2]*' emissions at various years'))
dev.off() |
d016bf7c1cea2be45570d0826610230b375be3ce | 9bc17a169325375bc993b540d2ad0f0810ca0e76 | /R/twoway.plots.R | a98edb8797477c8f6316b7dfb57853a3015db298 | [] | no_license | alanarnholt/PASWR | 335b960db32232a19d08560938d26f168e43b0d6 | f11b56cff44d32c3683e29e15988b6a37ba8bfd4 | refs/heads/master | 2022-06-16T11:34:24.098378 | 2022-05-14T22:56:11 | 2022-05-14T22:56:11 | 52,523,116 | 2 | 1 | null | null | null | null | UTF-8 | R | false | false | 1,375 | r | twoway.plots.R | #' @title Exploratory Graphs for Two Factor Designs
#'
#' @description Function creates side-by-side boxplots for each factor, a design plot (means), and an interaction plot.
#'
#' @param Y response variable
#' @param fac1 factor one
#' @param fac2 factor two
#' @param COL a vector with two colors
#'
#' @author Alan T. Arnholt <arnholtat@@appstate.edu>
#'
#' @seealso \code{\link{oneway.plots}}, \code{\link{checking.plots}}
#'
#' @export
#'
#' @examples
#' with(data = TireWear, twoway.plots(Wear, Treat, Block))
#'
#' @keywords hplot
####################################################################
twoway.plots<-function(Y, fac1, fac2, COL=c("#A9E2FF", "#0080FF")){
opar <- par(no.readonly = TRUE)
par(mfrow=c(2, 2), mar = c(5.1, 4.1, 1.1, 1.1))
YL <- range(Y)
plot(Y ~ fac1, col = COL[1], xlab = deparse(substitute(fac1)),
ylab = deparse(substitute(Y)), ylim = YL)
plot(Y ~ fac2, col = COL[2], xlab = deparse(substitute(fac2)),
ylab = deparse(substitute(Y)), ylim = YL)
plot.design(Y ~ fac1 + fac2, fun = "mean",
ylab = deparse(substitute(Y)), ylim = YL)
interaction.plot(fac1, fac2, Y, xlab = deparse(substitute(fac1)),
trace.label = deparse(substitute(fac2)),
type = "b", legend = FALSE,
ylab = deparse(substitute(Y)), ylim = YL)
on.exit(par(opar))
}
|
b4e93e3bcccb0eb0d1014bd355bcfff5a5be6187 | 280019f481fe09da00296f45e5fa530051780756 | /ui.R | 10b50c14611abb664f6dbfc7ea4c164e2ac58b15 | [] | no_license | linareja/2017_Buenos_Aires_Elections | 1effb2b1d39bf660e9fa678a6a78ac3000f2122c | d500aaedb233fe541fe00dc63f0d488043467111 | refs/heads/master | 2021-09-12T16:33:33.715908 | 2018-04-18T18:19:56 | 2018-04-18T18:19:56 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,802 | r | ui.R |
library(shiny)
dashboardPage(
dashboardHeader(title = "2017 Elections in Buenos Aires Province"),
dashboardSidebar( sidebarMenu(
menuItem("Overview", tabName = "overview", icon = icon("globe")),
menuItem("Analysis", tabName = "analysis", icon = icon("bar-chart"))
)),
dashboardBody(
tabItems(
tabItem("overview",
fluidRow(
column(6,selectInput("charge", "Select charge to visualize", levels(raw_data$variable))
)),
fluidRow(
column(6, h3("Overview"))
),
fluidRow(
column(6,drawMapUI("map1")
),
column(6,drawTreeMapUI("treemap1")
)
),
fluidRow(
column(6, h3("Comparison"))
),
fluidRow(
column(6, selectInput("comparison_plot", "Select Comparison Plot", choices = c("Map", "Treemap")))
),
#Aca va a ir un selector para comparar por treemap o por mapa
conditionalPanel("input.comparison_plot == 'Map'",
fluidRow(
column(6,drawMapUI("map_compare1",is.multiple = F)),
column(6,drawMapUI("map_compare2",is.multiple = F))
)
),
conditionalPanel("input.comparison_plot == 'Treemap'",
fluidRow(
column(6,drawTreeMapUI("treemap_compare1",is.multiple = F)),
column(6,drawTreeMapUI("treemap_compare2",is.multiple = F))
)
)
),
tabItem("analysis",
drawHeatmapUI("heatmap"))
)
)
)
|
564a95d83be7184c25e4953fc74f13401f3970ba | b6ed5857732c3261abab33a6665e7193d6862aef | /tests/testthat/test-read-oneshot-eav.R | d2b16cc13808d4cf760c57f846c793651568b48e | [
"MIT"
] | permissive | cran/REDCapR | 5ac1ebdb03fbf7dfa1aab23a2c23f711adcd4847 | a1aa09eb27fb627207255018fa41e30fa5d4b0fc | refs/heads/master | 2022-08-27T14:49:33.798497 | 2022-08-10T15:10:18 | 2022-08-10T15:10:18 | 24,255,971 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 12,860 | r | test-read-oneshot-eav.R | library(testthat)
credential <- retrieve_credential_testing()
update_expectation <- FALSE
test_that("smoke test", {
testthat::skip_on_cran()
expect_message(
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token)
)
})
test_that("default", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/default.R"
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(
redcap_uri = credential$redcap_uri,
token = credential$token
)
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame) # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_true(returned_object$filter_logic=="", "A filter was not specified.")
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("specify-forms", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/specify-forms.R"
desired_forms <- c("demographics", "race_and_ethnicity")
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token, forms=desired_forms)
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame) # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_true(returned_object$filter_logic=="", "A filter was not specified.")
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("raw", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/raw.R"
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token, raw_or_label="raw")
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame, label="The returned data.frame should be correct") # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_true(returned_object$filter_logic=="", "A filter was not specified.")
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("raw-and-dag", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/raw-and-dag.R"
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token, raw_or_label="raw", export_data_access_groups=TRUE)
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame, label="The returned data.frame should be correct") # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_true(returned_object$filter_logic=="", "A filter was not specified.")
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("label-and-dag", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/label-and-dag.R"
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token, raw_or_label="label", export_data_access_groups=TRUE)
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame, label="The returned data.frame should be correct") # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_true(returned_object$filter_logic=="", "A filter was not specified.")
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("label-header", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/label-header.R"
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token, raw_or_label_headers="label")
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame, label="The returned data.frame should be correct", ignore_attr = TRUE) # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_true(returned_object$filter_logic=="", "A filter was not specified.")
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("filter-numeric", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/filter-numeric.R"
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
filter <- "[age] >= 61"
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token, filter_logic=filter)
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame, label="The returned data.frame should be correct") # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_equal(returned_object$filter_logic, filter)
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("filter-character", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/filter-character.R"
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
filter <- "[name_first] = 'John Lee'"
expect_message(
regexp = expected_outcome_message,
returned_object <- REDCapR:::redcap_read_oneshot_eav(redcap_uri=credential$redcap_uri, token=credential$token, filter_logic=filter)
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame, label="The returned data.frame should be correct") # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_equal(returned_object$filter_logic, filter)
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("date-range", {
testthat::skip_on_cran()
path_expected <- "test-data/specific-redcapr/read-oneshot-eav/default.R"
expected_outcome_message <- "\\d+ records and \\d+ columns were read from REDCap in \\d+(\\.\\d+\\W|\\W)seconds\\."
start <- as.POSIXct(strptime("2018-08-01 03:00", "%Y-%m-%d %H:%M"))
stop <- Sys.time()
expect_message(
regexp = expected_outcome_message,
returned_object <-
REDCapR:::redcap_read_oneshot_eav(
redcap_uri = credential$redcap_uri,
token = credential$token,
datetime_range_begin = start,
datetime_range_end = stop
)
)
if (update_expectation) save_expected(returned_object$data, path_expected)
expected_data_frame <- retrieve_expected(path_expected)
expect_equal(returned_object$data, expected=expected_data_frame, label="The returned data.frame should be correct", ignore_attr = TRUE) # dput(returned_object$data)
expect_equal(returned_object$status_code, expected=200L)
expect_equal(returned_object$raw_text, expected="", ignore_attr = TRUE) # dput(returned_object$raw_text)
expect_true(returned_object$records_collapsed=="", "A subset of records was not requested.")
expect_true(returned_object$fields_collapsed=="", "A subset of fields was not requested.")
expect_equal(returned_object$filter_logic, "")
expect_match(returned_object$outcome_message, regexp=expected_outcome_message, perl=TRUE)
expect_true(returned_object$success)
})
test_that("bad token -Error", {
testthat::skip_on_cran()
expected_outcome_message <- "The REDCapR record export operation was not successful\\."
expect_error(
regexp = expected_outcome_message,
REDCapR:::redcap_read_oneshot_eav(
redcap_uri = credential$redcap_uri,
token = "BAD00000000000000000000000000000"
)
)
})
rm(credential)
|
ae049e4f7dded0c1877205b17e89aab67356d759 | cf4263e82b2c118bc3ecea5dc62d561e7487cbd3 | /tests/testthat/test_flatten_data.R | 327e274c4b13ccbaaa4edf5a2d6be774fcc94394 | [
"MIT"
] | permissive | EDIorg/ecocomDP | 151a2d519ff740d466fafab74df5171a6ef196bf | 0554d64ce81f35ed59985d9d991203d88fe1621f | refs/heads/main | 2023-08-14T02:07:19.274860 | 2023-06-19T22:27:30 | 2023-06-19T22:27:30 | 94,339,321 | 26 | 10 | NOASSERTION | 2023-07-26T22:21:00 | 2017-06-14T14:22:43 | R | UTF-8 | R | false | false | 7,103 | r | test_flatten_data.R | context("flatten_data()")
# Compare L0 flat and L1 flat - The column names and values of the L0 flat and L1 flattened tables should match, with an exception:
# 1.) Primary keys, row identifiers, of the ancillary tables are now present.
# Column presence -------------------------------------------------------------
testthat::test_that("Column presence", {
for (i in c("df", "tbbl")) {
# Parameterize
if (i == "df") { # test w/data.frame
L0_flat <- as.data.frame(ants_L0_flat)
for (tbl in names(ants_L1$tables)) {
ants_L1$tables[[tbl]] <- as.data.frame(ants_L1$tables[[tbl]])
}
} else { # test w/tibble
L0_flat <- ants_L0_flat
}
crit <- read_criteria()
L1_flat <- flatten_data(ants_L1$tables)
# Adjust L0 flat to our expectations
L0_flat$location_name <- NA_character_ # Add exception
# TEST: All L0 flat columns (with above exceptions) should be in L1 flat
cols_missing_from_L1 <- base::setdiff(colnames(L0_flat), colnames(L1_flat))
expect_true(length(cols_missing_from_L1) == 0)
# TEST: All L1 flat columns should be in L0 flat
cols_missing_from_L0 <- base::setdiff(colnames(L1_flat), colnames(L0_flat))
expect_true(length(cols_missing_from_L0) == 0)
}
})
# Column classes --------------------------------------------------------------
testthat::test_that("Column classes", {
for (i in c("df", "tbbl")) {
# Parameterize
if (i == "df") { # test w/data.frame
L0_flat <- as.data.frame(ants_L0_flat)
for (tbl in names(ants_L1$tables)) {
ants_L1$tables[[tbl]] <- as.data.frame(ants_L1$tables[[tbl]])
}
} else { # test w/tibble
L0_flat <- ants_L0_flat
}
crit <- read_criteria()
L1_flat <- flatten_data(ants_L1$tables)
# TEST: flatten_data() applies a set of "smart" class coercions to return numeric values stored in the L1 as character back to their original numeric class. The following code tests that column classifications in L1 should be "similar" to those in L0.
L0_classes <- unlist(lapply(L0_flat, class))
L1_classes <- unlist(lapply(L1_flat, class))
# Harmonize classes (because there is some variation) before comparing
L0_classes[stringr::str_detect(names(L0_classes), "id")] <- "character" # identifiers should be character
L1_classes[stringr::str_detect(names(L1_classes), "id")] <- "character"
L0_classes[stringr::str_detect(L0_classes, "integer")] <- "numeric" # integer ~= numeric
L1_classes[stringr::str_detect(L1_classes, "integer")] <- "numeric"
# TEST: Compare col classes
for (c in seq(L1_classes)) {
col <- L1_classes[c]
if (names(col) %in% names(L0_classes)) {
use_i <- names(L0_classes) %in% names(col)
if (any(use_i)) {
expect_equal(L0_classes[use_i], col)
}
}
}
}
})
# Observations (rows) match ---------------------------------------------------
# TODO Implement this test?
# testthat::test_that("Observations (rows) match", {
# # Parameterize
# crit <- read_criteria()
# L0_flat <- ants_L0_flat
# L1_flat <- ecocomDP::flatten_data(ants_L1$tables)
# # Adjust L0 flat to our expectations
# L0_flat <- L0_flat %>%
# dplyr::select(-block) %>% # A higher level location lost when flattened
# dplyr::select(-author) %>% # Columns of NA are dropped when flattened
# dplyr::rename(location_name = plot) # Becomes "location_name" when flattened
# # TEST: Observation "A" in L0 flat has the same values in observation "A" of L1 flat
# # TODO observation_id are identical
# # TODO match cols and sort, then compare (some subset?)
# })
# Non-required columns --------------------------------------------------------
# Non-required columns of ecocomDP aren't required by flatten_data()
testthat::test_that("Non-required columns", {
for (i in c("df", "tbbl")) {
# Parameterize
if (i == "df") { # test w/data.frame
for (tbl in names(ants_L1$tables)) {
ants_L1$tables[[tbl]] <- as.data.frame(ants_L1$tables[[tbl]])
}
}
# Parameterize
crit <- read_criteria() %>%
dplyr::filter(required == TRUE, !is.na(column)) %>%
dplyr::select(table, column)
tbls <- ants_L1$tables
# Throw out all non-required columns
for (tname in names(tbls)) {
rqd <- crit$column[crit$table %in% tname]
tbls[[tname]] <- tbls[[tname]] %>% dplyr::select(dplyr::any_of(rqd))
}
# TEST: Missing non-required columns isn't an issue
L1_flat <- ecocomDP::flatten_data(tbls)
cols_in <- unname(unlist(lapply(tbls, colnames)))
cols_out <- colnames(L1_flat)
dif <- base::setdiff(cols_in, cols_out)
expect_equal(dif, # Difference is a set of cols that shouldn't be returned by anyway
c("location_ancillary_id", "taxon_ancillary_id", "observation_ancillary_id",
"variable_mapping_id", "table_name"))
}
})
# flatten_location() ----------------------------------------------------------
# location_name values are parsed into the original L0 column representation
testthat::test_that("flatten_location(): No nesting", {
loc <- tidyr::as_tibble( # A table demonstrating this use case
data.frame(
location_id = c("H1"),
location_name = c("Highest__1"),
latitude = 45,
longitude = 123,
elevation = 200,
parent_location_id = NA_character_,
stringsAsFactors = FALSE))
for (i in c("df", "tbbl")) {
# Parameterize
if (i == "df") { # test w/data.frame
loc <- as.data.frame(loc)
}
# Parameterize
res <- flatten_location(loc)
loc_flat <- res$location_flat
# TEST: Original columns of data are returned
expect_true(all(c("Highest") %in% colnames(loc_flat))) # column names
expect_equal(loc_flat$Highest, "1") # values
}
})
testthat::test_that("flatten_location(): 3 nested sites", {
loc <- tidyr::as_tibble( # A table demonstrating this use case
data.frame(
location_id = c("H1", "M2", "L3"),
location_name = c("Highest__1", "Middle__2", "Lowest__3"),
latitude = c(NA, NA, 45),
longitude = c(NA, NA, 123),
elevation = c(NA, NA, 200),
parent_location_id = c(NA_character_, "H1", "M2"),
stringsAsFactors = FALSE))
for (i in c("df", "tbbl")) {
# Parameterize
if (i == "df") { # test w/data.frame
loc <- as.data.frame(loc)
}
# Parameterize
res <- flatten_location(loc)
loc_flat <- res$location_flat
# TEST: Original columns of data are returned
expect_true(all(c("Highest", "Middle", "Lowest") %in% colnames(loc_flat))) # column names
expect_equal(loc_flat$Highest, "1") # values
expect_equal(loc_flat$Middle, "2")
expect_equal(loc_flat$Lowest, "3")
# TEST: Original columns are returned in the order of nesting
expect_equal(which(colnames(loc_flat) %in% "Highest"), 3)
expect_equal(which(colnames(loc_flat) %in% "Middle"), 4)
expect_equal(which(colnames(loc_flat) %in% "Lowest"), 5)
}
})
|
8ce7a9d3e16bf2b520b938c008850a5ca1577fb8 | 92456ce1d280dd99f0df1cc2a2567c5021286f03 | /R/prepare_data.R | 5c8fabf25b3ad3505598af1c3c14f7a6948f57d1 | [] | no_license | nzfarhad/AFG_MSNA_19_Analysis | 41643620a065ff3eaba40779624101b55562efe4 | 66b4cfe032b7665475606dcab5eae4fcacba0e9c | refs/heads/master | 2020-07-28T17:27:34.829098 | 2020-01-28T10:01:02 | 2020-01-28T10:01:02 | 209,478,936 | 0 | 2 | null | null | null | null | UTF-8 | R | false | false | 100,120 | r | prepare_data.R | # Title: Preparation of data for woa survey
# Authors: Sayed Nabizada, Jarod Lapp, Christopher Jarvis,
# Date created: 20/09/2019
# Date last changed: 25/09/2019
# Purpose: This script is for recoding variables in the whole of
# of Afghanistan survey data
# indicators and composite scores are created.
# setup analysis environment
source("./R/source.R")
library(msni19)
# character operation
ch<-as.character
chr<-as.character
coerc<-function(x){as.numeric(chr(x))}
# load data
# data <- read_excel(master_data, sheet = "MSNA_AFG_19_parent_sheet", na = c("","NA"), guess_max = 3000)
# overall_muac_data <- read_excel(master_data, sheet = "MSNA_AFG_19_muac" , na = c("","NA"))
# overall_hh_roster <- read_excel(master_data, sheet = "MSNA_AFG_19_hh_roster" , na = c("","NA"))
# overall_death_roster <- read_excel(master_data, sheet = "MSNA_AFG_19_hh_death_roster" , na = c("","NA"))
# overall_left_roster <- read_excel( master_data, sheet = "MSNA_AFG_19_hh_left_roster" , na = c("","NA"))
# data <- read.csv("input/data/clean/MSNA_AFG_19_parent_sheet.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
# overall_muac_data <- read.csv("input/data/clean/MSNA_AFG_19_muac.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
# overall_hh_roster <- read.csv("input/data/clean/MSNA_AFG_19_hh_roster.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
# overall_death_roster <- read.csv("input/data/clean/MSNA_AFG_19_hh_death_roster.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
# overall_left_roster <- read.csv("input/data/clean/MSNA_AFG_19_hh_left_roster.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
#
data <- read.csv("input/data/clean/complete_with_farah/MSNA_AFG_19_parent_sheet.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
overall_muac_data <- read.csv("input/data/clean/complete_with_farah/MSNA_AFG_19_muac.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
overall_hh_roster <- read.csv("input/data/clean/complete_with_farah/MSNA_AFG_19_hh_roster.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
overall_death_roster <- read.csv("input/data/clean/complete_with_farah/MSNA_AFG_19_hh_death_roster.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
overall_left_roster <- read.csv("input/data/clean/complete_with_farah/MSNA_AFG_19_hh_left_roster.csv",stringsAsFactors=F,na.strings = c("", "NA"), check.names = F)
# Temp for the data is exported out of kobo incorrectly.
rename1 <- function(d1) {
sub("/", ".", names(d1))
}
data$uuid <- data$`_uuid`
names(data) <- rename1(data)
names(overall_muac_data ) <- rename1(overall_muac_data )
names(overall_hh_roster ) <- rename1(overall_hh_roster )
names(overall_death_roster ) <- rename1(overall_death_roster)
names(overall_left_roster ) <- rename1(overall_left_roster)
# composite indicators #
# The composite indicators are a combination of different variables
# each value within a variable has a score and these need to be
# coded for the different categories.
# Then the variables can be summed in order to get the score
# This will be done for multiple sectors.
#### Composite indicators ############
### Food Security & Agriculture ####
# FCS
data <- data %>%
mutate(
# FCS
fcs_category_class = recode(
fcs_category,
"poor" = 4,
"borderline" = 2,
"acceptable" = 0
),
# HHS
hhs_category_class = recode(
hhs_category,
"severe_hunger" = 4,
"moderate_hunger" = 2,
"little_hunger" = 0
),
# Food Source
food_source_class = case_when(
food_source %in% c('gift', 'assistance') ~ 2,
food_source == 'borrowed' ~1,
TRUE ~ 0
),
# ag impact
ag_impact_class = case_when(
agricultural_impact_how == '76_100' ~ 3,
agricultural_impact_how == '51_75' ~ 1,
agricultural_impact %in% c('no', 'not_applicable') ~ 0,
agricultural_impact_how %in% c('0_25', '26_50' ) ~ 0
),
# livestock impact
ls_impact_class = case_when(
livestock_impact_how.livestock_died == 1 |
livestock_impact_how.left_unattended == 1 ~ 2,
livestock_impact_how.livestock_ill == 1 |
livestock_impact_how.less_milk == 1 ~ 1,
livestock_impact == 0 ~ 0,
TRUE ~ 0,
is.na(livestock_impact) ~ NA_real_
)
)
fsac_vars <- c("fcs_category_class", "hhs_category_class", "food_source_class", "ag_impact_class", "ls_impact_class")
data$fsac_score <- comp_score(data, fsac_vars)
data <- data %>%
mutate(
fsac_severity = case_when(
fsac_score <= 2 ~ 1,
fsac_score <= 5 ~ 2,
fsac_score <= 8 ~ 3,
fsac_score <= 16 ~ 4
),
fsac_sev_high = case_when(
fsac_severity <= 2 ~ 0,
fsac_severity <= 4 ~ 1
)
)
##################################################################
### Protection ####
# First setup the variables required to calculate the indicators and then calculate them
# This way around if the weights are changed then it's all in one place.
# protection incidents
severe_prot_incidents_vars <- c(
"adult_prot_incidents.assaulted_with_weapon",
"child_prot_incidents.assaulted_with_weapon",
"adult_prot_incidents.forced_work",
"child_prot_incidents.forced_work",
"adult_prot_incidents.forcibly_detained",
"child_prot_incidents.forcibly_detained",
"adult_prot_incidents.hindered_leave_settlement",
"child_prot_incidents.hindered_leave_settlement",
#### added from less_severe_prot_incidents
"adult_prot_incidents.verbally_threatened",
"child_prot_incidents.verbally_threatened",
"adult_prot_incidents.assaulted_without_weapon",
"child_prot_incidents.assaulted_without_weapon",
"adult_prot_incidents.hindered_leave_district",
"child_prot_incidents.hindered_leave_district"
)
# less_severe_prot_incidents_vars <-c(
# "adult_prot_incidents.verbally_threatened",
# "child_prot_incidents.verbally_threatened",
# "adult_prot_incidents.assaulted_without_weapon",
# "child_prot_incidents.assaulted_without_weapon",
# "adult_prot_incidents.hindered_leave_district",
# "child_prot_incidents.hindered_leave_district"
# )
data$severe_prot_incidents <- comp_score(data, severe_prot_incidents_vars)
# data$less_severe_prot_incidents <- comp_score(data, less_severe_prot_incidents_vars)
# protection concerns
severe_prot_concerns_vars <- c(
"prot_concerns.violence_maiming",
"prot_concerns.abduction",
"prot_concerns.explosive_hazards",
"prot_concerns.psych_wellbeing",
# added from less_severe_prot_concern
"prot_concerns.violence_injuries",
"prot_concerns.early_marriage",
"prot_concerns.destruction_property",
"prot_concerns.theft"
)
# less_severe_prot_concerns_vars <- c(
# "prot_concerns.violence_injuries",
# "prot_concerns.early_marriage",
# "prot_concerns.destruction_property",
# "prot_concerns.theft"
# )
data$severe_prot_concerns <- comp_score(data, severe_prot_concerns_vars)
# data$less_severe_prot_concerns <- comp_score(data, less_severe_prot_concerns_vars)
# explosive hazards
severe_explosive_hazards_vars <- c(
"explosive_impact.injury_death",
"explosive_impact.access_services",
"explosive_impact.relocation",
"explosive_impact.livelihoods_impact",
"explosive_impact.psych_impact"
)
less_severe_explosive_hazards_vars <- c(
"explosive_impact.restrict_recreation"
)
data$severe_explosive_hazards <- comp_score(data, severe_explosive_hazards_vars)
data$less_severe_explosive_hazards <- comp_score(data, less_severe_explosive_hazards_vars)
# tazkira
tazkira_total_vars <- c(
"adult_tazkira",
"child_tazkira")
data$tazkira_total <- comp_score(data, tazkira_total_vars)
children_working_yes_no_2 = case_when(
data$children_working == 0 ~ "0",
data$children_working >= 1 ~ "1 or more",
TRUE ~ NA_character_
)
# Protection Severity Score
## Weights
data <- data %>%
mutate(
prot_incident_class = case_when(
severe_prot_incidents >= 1 ~ 3,
# severe_prot_incidents == 0 & data$less_severe_prot_incidents >= 1 ~ 2,
TRUE ~ 0),
# violence targeting women, girls, boys
sgbv_incidents_class = case_when(
other_incidents.sgbv == 1 | other_concerns.sgbv == 1 ~ 2,
TRUE ~ 0
),
# children working unsafe conditions
children_work_safety_class = case_when(
children_working_yes_no_2 =='1 or more' ~ 1,
TRUE ~ 0
),
prot_concerns_class = case_when(
severe_prot_concerns >= 1 ~ 3,
# severe_prot_concerns == 0 & data$less_severe_prot_concerns >= 1 ~ 2,
TRUE ~ 0
),
# hh members injured conflict or nat disaster
injuries_class = case_when(
adult_injuries_cause %in% c('conflict', 'natural_disaster') |
child_injuries_cause %in% c('conflict', 'natural_disaster') ~ 3,
TRUE ~ 0
),
prot_explosive_hazards_class = case_when(
severe_explosive_hazards >= 1 ~ 3,
severe_explosive_hazards == 0 & less_severe_explosive_hazards >=1 ~ 2,
TRUE ~ 0
),
tazkira_class = case_when(
tazkira_total == 0 ~ 2,
tazkira_total > 0 & tazkira_total < hh_size ~ 1
)
)
# Score
prot_score_vars <- c(
"prot_incident_class",
"sgbv_incidents_class",
"children_work_safety_class",
"prot_concerns_class",
"injuries_class",
"prot_explosive_hazards_class",
"tazkira_class")
data$prot_score <- comp_score(data, prot_score_vars)
data <- data %>%
mutate(
prot_severity = case_when(
prot_score <= 2 ~ 1,
prot_score <= 5 ~ 2,
prot_score <= 8 ~ 3,
prot_score <= 18 ~ 4
),
prot_sev_high = case_when(
prot_severity >= 3 ~ 1,
TRUE ~ 0
)
)
################## protection new indicator 1 ######################
prot_all_indictors <- c(
"adult_prot_incidents.verbally_threatened",
"adult_prot_incidents.assaulted_without_weapon",
"adult_prot_incidents.assaulted_with_weapon",
"adult_prot_incidents.hindered_leave_settlement",
"adult_prot_incidents.hindered_leave_district",
"adult_prot_incidents.forced_work",
"adult_prot_incidents.forcibly_detained",
"child_prot_incidents.verbally_threatened",
"child_prot_incidents.assaulted_without_weapon",
"child_prot_incidents.assaulted_with_weapon",
"child_prot_incidents.hindered_leave_settlement",
"child_prot_incidents.hindered_leave_district",
"child_prot_incidents.forced_work",
"child_prot_incidents.forcibly_detained",
"other_incidents.sgbv",
"other_incidents.other",
"prot_concerns.violence_maiming",
"prot_concerns.violence_injuries",
"prot_concerns.psych_wellbeing",
"prot_concerns.abduction",
"prot_concerns.theft",
"prot_concerns.explosive_hazards",
"prot_concerns.destruction_property",
"prot_concerns.early_marriage",
"prot_concerns.other",
"other_concerns.sgbv",
"other_concerns.other"
)
data$prot_all_indictors_score <- comp_score(data, prot_all_indictors)
data <- data %>%
mutate(
prot_new_indicator_1 = case_when(
prot_all_indictors_score >= 1 ~ ">=1",
prot_all_indictors_score == 0 ~ "0",
TRUE ~ NA_character_
)
)
################## protection new indicator 2 ######################
data <- data %>%
mutate( displ_explosive_presence_na_to_0 = case_when(
displ_explosive_presence == "both" ~ 1,
displ_explosive_presence == "current" ~ 1,
displ_explosive_presence == "previous" ~ 1,
displ_explosive_presence == "no" ~ 0,
TRUE ~ 0
)
)
prot_all_indictors_2 <- c(
"adult_prot_incidents.verbally_threatened",
"adult_prot_incidents.assaulted_without_weapon",
"adult_prot_incidents.assaulted_with_weapon",
"adult_prot_incidents.hindered_leave_settlement",
"adult_prot_incidents.hindered_leave_district",
"adult_prot_incidents.forced_work",
"adult_prot_incidents.forcibly_detained",
"child_prot_incidents.verbally_threatened",
"child_prot_incidents.assaulted_without_weapon",
"child_prot_incidents.assaulted_with_weapon",
"child_prot_incidents.hindered_leave_settlement",
"child_prot_incidents.hindered_leave_district",
"child_prot_incidents.forced_work",
"child_prot_incidents.forcibly_detained",
"other_incidents.sgbv",
"other_incidents.other",
"prot_concerns.violence_maiming",
"prot_concerns.violence_injuries",
"prot_concerns.psych_wellbeing",
"prot_concerns.abduction",
"prot_concerns.theft",
"prot_concerns.explosive_hazards",
"prot_concerns.destruction_property",
"prot_concerns.early_marriage",
"prot_concerns.other",
"other_concerns.sgbv",
"other_concerns.other",
"displ_explosive_presence_na_to_0"
)
data$prot_all_indictors_score_2 <- comp_score(data, prot_all_indictors_2)
data <- data %>%
mutate(
prot_new_indicator_2 = case_when(
prot_all_indictors_score_2 >= 1 ~ ">=1",
prot_all_indictors_score_2 == 0 ~ "0",
TRUE ~ NA_character_
)
)
################## protection new indicator 3 ######################
data <- data %>%
mutate( nondispl_explosive_presence_na_to_0 = case_when(
nondispl_explosive_presence == "yes" ~ 1,
nondispl_explosive_presence == "no" ~ 0,
TRUE ~ 0
)
)
prot_all_indictors_3 <- c(
"adult_prot_incidents.verbally_threatened",
"adult_prot_incidents.assaulted_without_weapon",
"adult_prot_incidents.assaulted_with_weapon",
"adult_prot_incidents.hindered_leave_settlement",
"adult_prot_incidents.hindered_leave_district",
"adult_prot_incidents.forced_work",
"adult_prot_incidents.forcibly_detained",
"child_prot_incidents.verbally_threatened",
"child_prot_incidents.assaulted_without_weapon",
"child_prot_incidents.assaulted_with_weapon",
"child_prot_incidents.hindered_leave_settlement",
"child_prot_incidents.hindered_leave_district",
"child_prot_incidents.forced_work",
"child_prot_incidents.forcibly_detained",
"other_incidents.sgbv",
"other_incidents.other",
"prot_concerns.violence_maiming",
"prot_concerns.violence_injuries",
"prot_concerns.psych_wellbeing",
"prot_concerns.abduction",
"prot_concerns.theft",
"prot_concerns.explosive_hazards",
"prot_concerns.destruction_property",
"prot_concerns.early_marriage",
"prot_concerns.other",
"other_concerns.sgbv",
"other_concerns.other",
"displ_explosive_presence_na_to_0",
"nondispl_explosive_presence_na_to_0"
)
data$prot_all_indictors_score_3 <- comp_score(data, prot_all_indictors_3)
data <- data %>%
mutate(
prot_new_indicator_3 = case_when(
prot_all_indictors_score_3 >= 1 ~ ">=1",
prot_all_indictors_score_3 == 0 ~ "0",
TRUE ~ NA_character_
)
)
################## protection new indicator 4 ######################
data <- data %>%
mutate(
lcsi_category_class2 = case_when(
lcsi_category == "food_secure" | lcsi_category == "marginally_insecure" ~ 0,
lcsi_category == "moderately_insecure" | lcsi_category == "severely_insecure" ~ 1,
TRUE ~ 0
)
)
prot_all_indictors_4 <- c(
"adult_prot_incidents.verbally_threatened",
"adult_prot_incidents.assaulted_without_weapon",
"adult_prot_incidents.assaulted_with_weapon",
"adult_prot_incidents.hindered_leave_settlement",
"adult_prot_incidents.hindered_leave_district",
"adult_prot_incidents.forced_work",
"adult_prot_incidents.forcibly_detained",
"child_prot_incidents.verbally_threatened",
"child_prot_incidents.assaulted_without_weapon",
"child_prot_incidents.assaulted_with_weapon",
"child_prot_incidents.hindered_leave_settlement",
"child_prot_incidents.hindered_leave_district",
"child_prot_incidents.forced_work",
"child_prot_incidents.forcibly_detained",
"other_incidents.sgbv",
"other_incidents.other",
"prot_concerns.violence_maiming",
"prot_concerns.violence_injuries",
"prot_concerns.psych_wellbeing",
"prot_concerns.abduction",
"prot_concerns.theft",
"prot_concerns.explosive_hazards",
"prot_concerns.destruction_property",
"prot_concerns.early_marriage",
"prot_concerns.other",
"other_concerns.sgbv",
"other_concerns.other",
"displ_explosive_presence_na_to_0",
"nondispl_explosive_presence_na_to_0",
"lcsi_category_class2"
)
data$prot_all_indictors_score_4 <- comp_score(data, prot_all_indictors_4)
data <- data %>%
mutate(
prot_new_indicator_4 = case_when(
prot_all_indictors_score_4 >= 1 ~ ">=1",
prot_all_indictors_score_4 == 0 ~ "0",
TRUE ~ NA_character_
)
)
#################################################################
################## protection new indicator 5 ######################
prot_all_indictors_5 <- c(
"adult_prot_incidents.verbally_threatened",
"adult_prot_incidents.assaulted_without_weapon",
"adult_prot_incidents.assaulted_with_weapon",
"adult_prot_incidents.hindered_leave_settlement",
"adult_prot_incidents.hindered_leave_district",
"adult_prot_incidents.forced_work",
"adult_prot_incidents.forcibly_detained",
"child_prot_incidents.verbally_threatened",
"child_prot_incidents.assaulted_without_weapon",
"child_prot_incidents.assaulted_with_weapon",
"child_prot_incidents.hindered_leave_settlement",
"child_prot_incidents.hindered_leave_district",
"child_prot_incidents.forced_work",
"child_prot_incidents.forcibly_detained",
"other_incidents.sgbv",
"other_incidents.other",
"prot_concerns.violence_maiming",
"prot_concerns.violence_injuries",
"prot_concerns.psych_wellbeing",
"prot_concerns.abduction",
"prot_concerns.theft",
"prot_concerns.explosive_hazards",
"prot_concerns.destruction_property",
"prot_concerns.early_marriage",
"prot_concerns.other",
"other_concerns.sgbv",
"other_concerns.other",
"displ_explosive_presence_na_to_0",
"nondispl_explosive_presence_na_to_0",
"lcsi_category_class2",
"children_work_safety_class"
)
data$prot_all_indictors_score_5 <- comp_score(data, prot_all_indictors_5)
data <- data %>%
mutate(
prot_new_indicator_5 = case_when(
prot_all_indictors_score_5 >= 1 ~ ">=1",
prot_all_indictors_score_5 == 0 ~ "0",
TRUE ~ NA_character_
)
)
################## protection new indicator 6 ######################
data <- data %>%
mutate(
other_impact_class = case_when(
other_impact.injury_death == 1 | other_impact.new_mines == 1 ~ 1,
TRUE ~ 0
)
)
prot_all_indictors_6 <- c(
"adult_prot_incidents.verbally_threatened",
"adult_prot_incidents.assaulted_without_weapon",
"adult_prot_incidents.assaulted_with_weapon",
"adult_prot_incidents.hindered_leave_settlement",
"adult_prot_incidents.hindered_leave_district",
"adult_prot_incidents.forced_work",
"adult_prot_incidents.forcibly_detained",
"child_prot_incidents.verbally_threatened",
"child_prot_incidents.assaulted_without_weapon",
"child_prot_incidents.assaulted_with_weapon",
"child_prot_incidents.hindered_leave_settlement",
"child_prot_incidents.hindered_leave_district",
"child_prot_incidents.forced_work",
"child_prot_incidents.forcibly_detained",
"other_incidents.sgbv",
"other_incidents.other",
"prot_concerns.violence_maiming",
"prot_concerns.violence_injuries",
"prot_concerns.psych_wellbeing",
"prot_concerns.abduction",
"prot_concerns.theft",
"prot_concerns.explosive_hazards",
"prot_concerns.destruction_property",
"prot_concerns.early_marriage",
"prot_concerns.other",
"other_concerns.sgbv",
"other_concerns.other",
"displ_explosive_presence_na_to_0",
"nondispl_explosive_presence_na_to_0",
"lcsi_category_class2",
"other_impact_class"
)
data$prot_all_indictors_score_6 <- comp_score(data, prot_all_indictors_6)
data <- data %>%
mutate(
prot_new_indicator_6 = case_when(
prot_all_indictors_score_6 >= 1 ~ ">=1",
prot_all_indictors_score_6 == 0 ~ "0",
TRUE ~ NA_character_
)
)
###################################################end
### ESNFI ####
# shelter type
data$shelter_class<-ifelse(data$shelter == 'open_space',3,ifelse(data$shelter == 'tent' | data$shelter == 'makeshift_shelter' | data$shelter == 'collective_centre' | data$shelter == 'transitional',2,0))
# shelter damage
data$shelter_damage_class<-ifelse(data$shelter_damage_extent== 'fully_destroyed' & data$shelter_damage_repair == 'no',3,
ifelse(data$shelter_damage_extent== 'significant_damage' & data$shelter_damage_repair == 'no',2,
ifelse(data$shelter_damage_extent== 'partial_damage' & data$shelter_damage_repair == 'no',1,0)))
data$shelter_damage_class[is.na(data$shelter_damage_class)] <- 0
# TENENCY AGREEMENT
data$tenancy_class<-ifelse(data$tenancy == 'unofficial',3,ifelse(data$tenancy == 'own_home_without_doc' | data$tenancy == 'rental_verbal' | data$shelter_hosted == 'yes',2,0))
data$tenancy_class[is.na(data$tenancy_class)] <- 0
# blankets
data$blankets_class<-ifelse(data$blankets_number > data$hh_size,3,0)
data$blankets_class[is.na(data$blankets_class)] <- 0
# basic needs
data$sleeping_mats <- car::recode(data$sleeping_mats, " 'yes' = 1; 'no' = 0")
data$tarpaulin <- car::recode(data$tarpaulin, " 'yes' = 1; 'no' = 0")
data$cooking_pots <- car::recode(data$cooking_pots, " 'yes' = 1; 'no' = 0")
data$stainless_steel <- car::recode(data$stainless_steel, " 'yes' = 1; 'no' = 0")
data$water_storage <- car::recode(data$water_storage, " 'yes' = 1; 'no' = 0")
data$hygiene_sanitation <- car::recode(data$hygiene_sanitation, " 'yes' = 1; 'no' = 0")
data$basic_needs_total<-coerc(data[["sleeping_mats"]])+coerc(data[["tarpaulin"]])+coerc(data[["cooking_pots"]])+coerc(data[["stainless_steel"]])+coerc(data[["water_storage"]])+coerc(data[["hygiene_sanitation"]])
data$basic_needs_score<-car::recode(data$basic_needs_total,
"0:2=3;
3:5=2;
6=0")
# ESNFI Severity Score
data$esnfi_score<-coerc(data[["shelter_class"]])+coerc(data[["shelter_damage_class"]])+coerc(data[["tenancy_class"]])+coerc(data[["blankets_class"]])+coerc(data[["basic_needs_score"]])
data$esnfi_severity<-car::recode(data$esnfi_score,
"0:2='1';
3:6='2';
7:9='3';
10:16='4'")
data$esnfi_sev_high<-ifelse(data$esnfi_severity==3|data$esnfi_severity==4,1,0)
###################################################end
### <3 <3 <3 <3 <3 ESNFI 4 ARI <3 <3 <3 <3 <3 ####
# shelter type
data$shelter_class_4_ari<-case_when(data$shelter == 'open_space'| data$shelter == 'tent' | data$shelter == 'makeshift_shelter'| data$shelter == 'collective_centre' ~3, data$shelter == 'transitional'~2,
data$shelter=='permanent' & (data$shelter_hosted_why =='cash_rent'| data$shelter_hosted_why =='trans_shelter_host_family'| data$shelter_hosted_why =='materials_tools_extend') ~2, TRUE~ 0)
# shelter damage
data$shelter_damage_class_4_ari<-ifelse(data$shelter_damage_extent== 'fully_destroyed' & data$shelter_damage_repair == 'no',3,
ifelse(data$shelter_damage_extent== 'significant_damage' & data$shelter_damage_repair == 'no',2,
ifelse(data$shelter_damage_extent== 'partial_damage' & data$shelter_damage_repair == 'no',1,0)))
data$shelter_damage_class_4_ari[is.na(data$shelter_damage_class_4_ari)] <- 0
# TENENCY AGREEMENT
data$tenancy_class_4_ari<-ifelse(data$tenancy == 'unofficial',3,ifelse(data$tenancy == 'own_home_without_doc' | data$tenancy == 'rental_verbal' | data$shelter_hosted == 'yes',2,0))
data$tenancy_class_4_ari[is.na(data$tenancy_class_4_ari)] <- 0
# ESNFI Severity Score
data$esnfi_score_4_ari<-coerc(data[["shelter_class_4_ari"]])+coerc(data[["shelter_damage_class_4_ari"]])+coerc(data[["tenancy_class_4_ari"]])
data$esnfi_severity_4_ari<-car::recode(data$esnfi_score_4_ari,
"0:2='1';
3:4='2';
5:6='3';
7:10='4'")
data$esnfi_sev_high_4_ari<-ifelse(data$esnfi_severity_4_ari==3|data$esnfi_severity_4_ari==4,1,0)
#################################################################
### WASH ####
# water source #
data$water_source_class<-car::recode(data$water_source,
"'surface_water'=3;
'water_trucking'=2;
'spring_unprotected'=2;
'spring_protected'=0;
'handpump_private'=0;
'handpump_public'=0;
'piped_public'=0;
'other'=0")
# water barriers
data$water_barriers_class<-ifelse(data$water_sufficiency== 'insufficient' &
(data$water_barriers== 'too_far' |
data$water_barriers== 'high_risk' |
data$water_barriers== 'social_restrictions'),
3,ifelse(data$water_sufficiency== 'insufficient',2,
ifelse(data$water_sufficiency== 'barely_sufficient',1,0)))
data$water_barriers_class[is.na(data$water_barriers)] <- 0
# soap
data$soap_class<-ifelse(data$soap == 'yes_didnt_see' | data$soap == 'no', 1,0)
# latrines #
data$latrine_class<-ifelse(data$latrine == 'open' | data$latrine == 'public_latrine', 3,
ifelse(data$latrine == 'pit_latrine_uncovered',2,0))
# primary waste dispopsal #
data$waste_disposal_class<-ifelse(data$waste_disposal == 'open_space' | data$waste_disposal == 'burning', 2,0)
#distance to primary water source
data$water_distance_class<-ifelse(data$water_distance == 'over_1km'| data$water_distance == '500m_to_1km',3,0)
# WASH Severity Score
data$wash_score<-coerc(data[["water_source_class"]])+coerc(data[["water_barriers_class"]])+coerc(data[["soap_class"]])+coerc(data[["latrine_class"]])+coerc(data[["waste_disposal_class"]])+coerc(data[["water_distance_class"]])
data$wash_severity<-car::recode(data$wash_score,
"0:2='1';
3:5='2';
6:8='3';
9:16='4'")
data$wash_sev_high<-ifelse(data$wash_severity==3|data$wash_severity==4,1,0)
#################################################################
### Nutrition ####
muac_presence_analysis<-overall_muac_data %>%
group_by(`_submission__uuid`) %>%
filter(person_muac>=1) %>%
summarize(number_muac_person=sum(person_muac),
number_muac_mod_mal=sum(moderate_malnutrition),
number_muac_sev_mal=sum(severe_malnutrition),
number_muac_above_125 = sum(muac_measurement>=125, na.rm = T),
min_muac=min(muac_measurement),
ruft_reception_num = sum(rutf_reception== "yes"),
ruft_reception = sum(rutf_reception== "yes")>=1)
# Malnutrition present = 1, not present = 0
muac_presence_analysis$malnutrition_present<-ifelse(muac_presence_analysis$number_muac_mod_mal>=1 | muac_presence_analysis$number_muac_sev_mal>=1,1,0)
# join with parent table
data<-full_join(data, muac_presence_analysis, by = c("uuid"="_submission__uuid"))
# reported malnourishment (mod & sev muac)
data$muac_score<-ifelse(data$number_muac_sev_mal>1,7,ifelse(data$number_muac_sev_mal==1,6,ifelse(data$number_muac_sev_mal==0 & data$number_muac_mod_mal>1,4,ifelse(data$number_muac_sev_mal==0 & data$number_muac_mod_mal==1,3,0))))
# dietary diversity ---
# therefore nutrition compotite indicator will exclude hhs with children aged 2-5, since this they are not asked this question ###
data$dietary_div_count<-coerc(data[["minimum_dietary_diversity.staples"]])+coerc(data[["minimum_dietary_diversity.legumes"]])+coerc(data[["minimum_dietary_diversity.dairy"]])+coerc(data[["minimum_dietary_diversity.meat"]])+coerc(data[["minimum_dietary_diversity.eggs"]])+coerc(data[["minimum_dietary_diversity.vitamin_a_veg"]])+coerc(data[["minimum_dietary_diversity.other_veg"]])
data$dietary_div_score<-ifelse(data$dietary_div_count==0,4,ifelse(data$dietary_div_count==1,3,ifelse(data$dietary_div_count==2,2,ifelse(data$dietary_div_count==3,1,0))))
data$dietary_div_score[is.na(data$dietary_div_score)] <- 0
# Nutrition Severity Score
data$nut_score_hh_w_muac<-coerc(data[["muac_score"]])+coerc(data[["dietary_div_score"]])
data$nut_score<-data$nut_score_hh_w_muac
data$nut_score[is.na(data$nut_score)] <- 0
data$nut_severity<-car::recode(data$nut_score,
"0:2='1';
3:5='2';
6:8='3';
9:16='4'")
# data$nut_sev_high<-ifelse(data$nut_severity==3|data$nut_severity==4,1,0)
data$nut_sev_high<-ifelse(data$nut_severity==3|data$nut_severity==4 | data$nut_severity==2 ,1,0)
#################################################################
### Education EiE ####
education_analysis<-overall_hh_roster %>%
filter(!is.na(current_year_enrolled))
education_analysis$enrolled_and_attending<-ifelse(education_analysis$current_year_enrolled=='no',0,
ifelse(education_analysis$current_year_enrolled=='yes' & education_analysis$current_year_attending=='no',0,1))
education_analysis$total_schoolage_child<-1
#removal from school due to shock
education_analysis$shock_presence<-coerc(education_analysis[["edu_removal_shock.displacement"]])+coerc(education_analysis[["edu_removal_shock.conflict"]])+coerc(education_analysis[["edu_removal_shock.natural_disaster"]])
education_analysis$shock_presence[is.na(education_analysis$shock_presence)] <- 0
################## not part of composite #################################################
education_analysis$enrolled_1 <- if_else(education_analysis$current_year_enrolled=='no',0,1)
education_analysis$attending_1 <- if_else(education_analysis$current_year_attending=='no',0,1)
education_analysis <- education_analysis %>%
mutate(
attending_male = case_when(
current_year_attending == "yes" & hh_member_sex == "male" ~ 1,
TRUE ~ 0
),
attending_female = case_when(
current_year_attending == "yes" & hh_member_sex == "female" ~ 1,
TRUE ~ 0
),
enrolled_male = case_when(
current_year_enrolled == "yes" & hh_member_sex == "male" ~ 1,
TRUE ~ 0
),
enrolled_female = case_when(
current_year_enrolled == "yes" & hh_member_sex == "female" ~ 1,
TRUE ~ 0
),
shock_presence_male = case_when(
shock_presence > 0 & hh_member_sex == "male" ~ 1,
TRUE ~ 0
),
shock_presence_female = case_when(
shock_presence > 0 & hh_member_sex == "female" ~ 1,
TRUE ~ 0
)
)
####################################################################################
# group dataset into hh
education_analysis_hh<-education_analysis %>%
group_by(`_submission__uuid`) %>%
summarize(count_school_child=sum(total_schoolage_child),
count_enrolled_attending=sum(enrolled_and_attending),
count_current_enrolled = sum(enrolled_1, na.rm = T),
count_current_enrolled_male = sum(enrolled_male, na.rm = T),
count_current_enrolled_female = sum(enrolled_female, na.rm = T),
count_current_attending = sum(attending_1, na.rm = T),
count_current_attending_male = sum(attending_male, na.rm = T),
count_current_attending_female = sum(attending_female, na.rm = T),
count_shock=sum(shock_presence),
count_shock_male = sum(shock_presence_male, na.rm = T),
count_shock_female = sum(shock_presence_female, na.rm = T)
)
# shock weight
education_analysis_hh$shock_class<-ifelse(education_analysis_hh$count_shock >= 1, 5,0)
# percent children enrolled or attending
education_analysis_hh$percent_enrolled= coerc(education_analysis_hh[["count_enrolled_attending"]])/coerc(education_analysis_hh[["count_school_child"]])
education_analysis_hh$enroll_perc_class<-car::recode(education_analysis_hh$percent_enrolled,
"0:0.249=1;
0.25:0.499=2;
0.5:0.749=3;
0.75:1=4")
# greater than 3 children not attending
education_analysis_hh$count_not_enrolled<-coerc(education_analysis_hh[["count_school_child"]])-coerc(education_analysis_hh[["count_enrolled_attending"]])
education_analysis_hh$count_not_enrolled_class<-ifelse(education_analysis_hh$count_not_enrolled>=3,3,0)
# join with parent table
data<-full_join(data, education_analysis_hh,by = c("uuid"="_submission__uuid"))
# reasons not attending
data$severe_not_attending<-coerc(data[["boy_unattendance_reason.insecurity"]])+coerc(data[["boy_unattendance_reason.child_works_instead"]])+coerc(data[["girl_unattendance_reason.insecurity"]])+coerc(data[["girl_unattendance_reason.child_works_instead"]])
data$severe_not_attending[is.na(data$severe_not_attending)] <- 0
data$less_severe_not_attending<-coerc(data[["boy_unattendance_reason.lack_facilities"]])+coerc(data[["boy_unattendance_reason.lack_documentation"]])+coerc(data[["boy_unattendance_reason.too_expensive"]])+coerc(data[["girl_unattendance_reason.lack_facilities"]])+coerc(data[["girl_unattendance_reason.lack_documentation"]])+coerc(data[["girl_unattendance_reason.too_expensive"]])
data$less_severe_not_attending[is.na(data$less_severe_not_attending)] <- 0
data$not_attending_class<-ifelse(data$severe_not_attending >= 1,3, ifelse(data$severe_not_attending==0 & data$less_severe_not_attending >=1,2,0))
data$not_attending_class[is.na(data$not_attending_class)] <- 0
# Education Severity Score
data$edu_score_hh_w_schoolage<-coerc(data[["enroll_perc_class"]])+coerc(data[["shock_class"]])+coerc(data[["count_not_enrolled_class"]])+coerc(data[["not_attending_class"]])
data$edu_score<-data$edu_score_hh_w_schoolage
data$edu_score[is.na(data$edu_score)] <- 0
data$edu_severity<-car::recode(data$edu_score,
"0:2='1';
3:5='2';
6:8='3';
9:16='4'")
data$edu_sev_high<-ifelse(data$edu_severity==3|data$edu_severity==4,1,0)
#################################################################
### Health ####
#deaths under 5 years age
overall_death_roster$deaths_under5<-ifelse(overall_death_roster$hh_died_age<5,1,0)
# deaths >= 5
overall_death_roster$deaths_over5<-ifelse(overall_death_roster$hh_died_age>=5,1,0)
# group by hh
health_analysis<-overall_death_roster %>%
group_by(`_submission__uuid`) %>%
summarize(number_death_under5=sum(deaths_under5),
hh_member_died = sum(hh_member_died),
number_death_over5=sum(deaths_over5))
# join with parent dataset
data<-full_join(data, health_analysis,by = c("uuid"="_submission__uuid"))
data$number_death_under5[is.na(data$number_death_under5)] <- 0
# #deaths under 5 yrs age weight
# data$number_death_under5_class<-ifelse(data$number_death_under5 >= 1, 3,0)
# data$number_death_under5_class[is.na(data$number_death_under5_class)] <- 0
#
# # deaths >= 5 weight
# data$number_death_over5_class<-ifelse(data$number_death_over5 >= 1, 2,0)
# data$number_death_over5_class[is.na(data$number_death_over5_class)] <- 0
# health facility barriers
data$health_barriers_total<-coerc(data[["health_facility_barriers.unsafe"]])+coerc(data[["health_facility_barriers.cost_services"]])+coerc(data[["health_facility_barriers.cost_medicines"]])+coerc(data[["health_facility_barriers.too_far"]])+coerc(data[["health_facility_barriers.documentation_problems"]])+coerc(data[["health_facility_barriers.insufficient_female_staff"]])+coerc(data[["health_facility_barriers.treatment_refused"]])+coerc(data[["health_facility_barriers.other"]])
data$health_barriers_total[is.na(data$health_barriers_total)] <- 0
# data$health_facility_barriers_class<-ifelse(data$health_facility_access == 'no' & data$health_barriers_total>1,3,ifelse(data$health_facility_access == 'no' & data$health_barriers_total==1,2,0))
data$health_facility_barriers_class<-ifelse(data$health_facility_access == 'no' ,3,0)
# health facility distance
data$health_facility_dist_class<-ifelse(data$health_facility_distance == 'none' | data$health_facility_distance=='more_10km',3,ifelse(data$health_facility_distance=='6_10km',2,0))
# health facilities affected
data$health_facility_affected_class<-ifelse(data$health_facility_affected_how == 'forcibly_closed'|data$health_facility_affected_how == 'damaged_conflict'|data$health_facility_affected_how == 'damaged_natural_disasters',3, ifelse(data$health_facility_affected_how=='lack_staff'|data$health_facility_affected_how=='lack_medicine',2,0))
data$health_facility_affected_class[is.na(data$health_facility_affected_class)] <- 0
# health selected as priority need
data$health_priority_need_class<-ifelse(data$priority_needs.healthcare == 1, 3,0)
# behaviour changes as result of conflict
data$behavior_change_cause_class<-case_when(data$adult_behavior_change == 'yes'& data$behavior_change_cause=='yes'~ 3,
data$child_behavior_change == 'yes'& data$behavior_change_cause=='yes'~ 3, TRUE~ 0)
data$behavior_change_cause_class[is.na(data$behavior_change_cause_class)] <- 0
# birth location
# data$birth_location_class<-ifelse(data$birth_location == 'home'|data$birth_location == 'midwife_home'|data$birth_location == 'outside'|data$birth_location == 'other',1,0)
# data$birth_location_class[is.na(data$birth_location_class)] <- 0
# Health Severity Score
# data$health_score<- coerc(data[["health_facility_barriers_class"]])+coerc(data[["health_facility_dist_class"]])+coerc(data[["health_facility_affected_class"]])+coerc(data[["health_priority_need_class"]])+coerc(data[["behavior_change_cause_class"]])+coerc(data[["birth_location_class"]])
data$health_score<- coerc(data[["health_facility_barriers_class"]])+coerc(data[["health_facility_dist_class"]])+coerc(data[["health_facility_affected_class"]])+coerc(data[["health_priority_need_class"]])+coerc(data[["behavior_change_cause_class"]])
data$health_severity<-car::recode(data$health_score,
"0:2='1';
3:5='2';
6:8='3';
9:16='4'")
data$health_sev_high<-ifelse(data$health_severity==3|data$health_severity==4,1,0)
#################################################################
# number sectoral needs ####
data$total_sectoral_needs<-coerc(data[["fsac_sev_high"]])+coerc(data[["prot_sev_high"]])+coerc(data[["esnfi_sev_high"]])+coerc(data[["wash_sev_high"]])+coerc(data[["nut_sev_high"]])+coerc(data[["edu_sev_high"]])+coerc(data[["health_sev_high"]])
#################################################################
### LSCI - coping strategies ####
# coping severity
data$lcsi_severity<-car::recode(data$lcsi_category,
"'food_secure'='minimal';
'marginally_insecure'='stress';
'moderately_insecure'='severe';
'severely_insecure'='extreme'")
## Indicators ####
### Numerators
## Some numerators combine variables calcualte those here
data$edu_age_boys_girls_num <- comp_score(data, c("boys_ed","girls_ed"))
food_water_rent_vars <- c(
"food_exp",
"water_expt",
"rent_exp"
)
data$food_water_rent_num <- comp_score(data, food_water_rent_vars)
all_expenses_vars <- c(
"food_exp",
"water_expt",
"rent_exp",
"fuel_exp",
"debt_exp")
data$all_expenses <- comp_score(data, all_expenses_vars)
min_die_vars <- c(
"minimum_dietary_diversity.staples",
"minimum_dietary_diversity.legumes",
"minimum_dietary_diversity.dairy",
"minimum_dietary_diversity.meat",
"minimum_dietary_diversity.eggs",
"minimum_dietary_diversity.vitamin_a_veg",
"minimum_dietary_diversity.other_veg")
data$min_die_num <- comp_score(data, min_die_vars)
priority_nfi_vars <- c(
"sleeping_mats",
"tarpaulin",
"cooking_pots",
"stainless_steel",
"water_storage",
"hygiene_sanitation"
)
data$priority_nfi_num <- comp_score(data, priority_nfi_vars)
child_vars <- c(
"males_0_2_total",
"males_3_5_total",
"females_0_2_total",
"females_3_5_total")
data$children_under5 <- comp_score(data, child_vars)
comp_ind_vars <- c(
"prot_sev_high",
"fsac_sev_high",
"esnfi_sev_high",
"wash_sev_high",
"edu_sev_high",
"health_sev_high"
)
data$comp_ind_sev <- comp_score(data, comp_ind_vars)
comp_ind_vars_nut <- c(
"prot_sev_high",
"fsac_sev_high",
"esnfi_sev_high",
"wash_sev_high",
"edu_sev_high",
"health_sev_high",
"nut_sev_high"
)
data$comp_ind_sev_nut <- comp_score(data, comp_ind_vars_nut)
## Age categories hh
data <- data %>%
mutate(
age_0_4_hh = case_when(
hoh_age <=4 ~ 1,
TRUE ~ 0
),
age_0_17_hh = case_when(
hoh_age <=17 ~ 1,
TRUE ~ 0
) ,
age_0_14_hh = case_when(
hoh_age <=14 ~ 1,
TRUE ~ 0
) ,
age_10_17_hh = case_when(
hoh_age >= 10 & hoh_age <=17 ~ 1,
TRUE ~ 0
) ,
age_15_64_hh = case_when(
hoh_age >= 14 & hoh_age <=64 ~ 1,
TRUE ~ 0
) ,
age_18_59_hh = case_when(
hoh_age >= 18 & hoh_age <=59 ~ 1,
TRUE ~ 0
) ,
age_18_64_hh = case_when(
hoh_age >= 18 & hoh_age <=64 ~ 1,
TRUE ~ 0
) ,
age_60_and_more_hh = case_when(
hoh_age >= 60 ~ 1,
TRUE ~ 0
) ,
age_65_hh = case_when(
hoh_age >= 65 ~ 1,
TRUE ~ 0
) ,
testt = case_when(
hoh_age < 120 ~ 1,
TRUE ~ 0
)
)
## Age categories roster
hh_group <- overall_hh_roster %>%
mutate(
age_0_4 = hh_member_age <=4,
age_0_17 = hh_member_age <=17,
age_0_14 = hh_member_age <=14,
age_10_17 = hh_member_age >= 10 & hh_member_age <=17,
age_15_64 = hh_member_age >= 14 & hh_member_age <=64,
age_18_59 = hh_member_age >= 18 & hh_member_age <=59,
age_18_64 = hh_member_age >= 18 & hh_member_age <=64,
age_60_and_more = hh_member_age >= 60,
age_65 = hh_member_age >= 65
) %>%
group_by(`_submission__uuid`) %>%
summarise(
age_0_4 = sum(age_0_4, na.rm = TRUE),
age_0_17 = sum(age_0_17, na.rm = TRUE),
age_0_14 = sum(age_0_14, na.rm = TRUE),
age_10_17 = sum(age_10_17, na.rm = TRUE),
age_15_64 = sum(age_15_64, na.rm = TRUE),
age_18_59 = sum(age_18_59, na.rm = TRUE),
age_18_64 = sum(age_18_64, na.rm = TRUE),
age_60_and_more = sum(age_60_and_more, na.rm = T),
age_65 = sum(age_65, na.rm = TRUE)
)
# Age Cat Vars
age_0_4_vars <- c(
'age_0_4',
'age_0_4_hh'
)
age_0_17_vars <- c(
'age_0_17',
'age_0_17_hh'
)
age_0_14_vars <- c(
'age_0_14',
'age_0_14_hh'
)
age_10_17_vars <- c(
'age_10_17',
'age_10_17_hh'
)
age_15_64_vars <- c(
'age_15_64',
'age_15_64_hh'
)
age_18_59_vars <- c(
'age_18_59',
'age_18_59_hh'
)
age_18_64_vars <- c(
'age_18_64',
'age_18_64_hh'
)
age_60_and_more_var <- c(
'age_60_and_more',
'age_60_and_more_hh'
)
age_65_var <- c(
'age_65',
'age_65_hh'
)
data <- full_join(data, hh_group,by = c("uuid"="_submission__uuid"))
# Merge Age cat hh_roster and hh data
data$age_0_4_merged <- comp_score(data,age_0_4_vars)
data$age_0_17_merged <- comp_score(data,age_0_17_vars)
data$age_0_14_merged <- comp_score(data,age_0_14_vars)
data$age_10_17_merged <- comp_score(data,age_10_17_vars)
data$age_15_64_merged <- comp_score(data,age_15_64_vars)
data$age_18_59_merged <- comp_score(data,age_18_59_vars)
data$age_18_64_merged <- comp_score(data,age_18_64_vars)
data$age_60_and_more_merged <- comp_score(data,age_60_and_more_var)
data$age_65_merged <- comp_score(data,age_65_var)
# Adjust displacement status as more information in other data
non_displ_data <- read.csv("input/Non_Displaced_Host_List_v2.csv",stringsAsFactors=F,na.strings = c("", "NA"))
data<-full_join(data, non_displ_data,by = c("district"="district"))
data$final_displacement_status_non_displ<-ifelse(data$final_displacement_status=='non_displaced'|data$final_displacement_status=='host', data$non_displ_class,data$final_displacement_status)
# prev_displacement
data <- data %>%
mutate(
# prev_displacement_num
prev_displacement_num_class = case_when(
prev_displacement_num == 2 ~ "2",
prev_displacement_num == 3 ~ "3",
prev_displacement_num >3 ~ "4+"
),
# refugee_displace_year
refugee_displace_year_class = case_when(
refugee_displace_year == 0 ~ "0",
refugee_displace_year == 1 ~ "1",
refugee_displace_year == 2 ~ "2",
refugee_displace_year == 3 ~ "3",
refugee_displace_year > 3 ~ "4+"
),
# cb_return_displace_year
cb_return_displace_year_class = case_when(
cb_return_displace_year == 0 ~ "0",
cb_return_displace_year == 1 ~ "1",
cb_return_displace_year == 2 ~ "2",
cb_return_displace_year == 3 ~ "3",
cb_return_displace_year > 3 ~ "4+"
),
# cb_return_return_year
cb_return_return_year_call = case_when(
cb_return_return_year == 0 ~ "0",
cb_return_return_year == 1 ~ "1",
cb_return_return_year == 2 ~ "2",
cb_return_return_year == 3 ~ "3",
cb_return_return_year > 3 ~ "4+"
),
# idp_displ_year
idp_displ_year_class = case_when(
idp_displ_year == 0 ~ "0",
idp_displ_year == 1 ~ "1",
idp_displ_year == 2 ~ "2",
idp_displ_year == 3 ~ "3",
idp_displ_year > 3 ~ "4+"
),
# head of household age_group
hoh_age_group = case_when(
hoh_age >= 65 ~ "65+",
hoh_age < 65 ~ "<65"
),
# head of household disabled
hoh_disabled = case_when(
wg_walking == "yes" | wg_selfcare == "yes" ~ "disabled",
wg_walking == "no" | wg_selfcare == "no" ~ "not_disabled",
TRUE ~ NA_character_
),
pregnant_member = case_when(
pregnant > 0 ~ "at_least_one_mem_pregnant",
pregnant == 0 ~ "no_mem_pregnent",
TRUE ~ NA_character_
),
lactating_member = case_when(
lactating > 0 ~ "at_least_one_mem_lactating",
lactating == 0 ~ "no_mem_lactating",
TRUE ~ NA_character_
),
pregnant_lactating_member = case_when(
pregnant > 0 | lactating > 0 ~ "at_least_one_mem_pregnant_lactating",
pregnant == 0 & lactating == 0 ~ "no_mem_pregnent_lactating",
TRUE ~ NA_character_
),
female_literacy_yes_no = case_when(
female_literacy == 0 ~ "0",
female_literacy >= 1 ~ "1 or more",
TRUE ~ NA_character_
),
male_literacy_yes_no = case_when(
male_literacy == 0 ~ "0",
male_literacy >= 1 ~ "1 or more",
TRUE ~ NA_character_
),
# How many adults 18+ years worked outside of the household in the last 30 days?
adults_working_yes_no = case_when(
adults_working == 0 ~ "0",
adults_working >= 1 ~ "1 or more",
TRUE ~ NA_character_
),
children_working_yes_no = case_when(
children_working == 0 ~ "0",
children_working >= 1 ~ "1 or more",
TRUE ~ NA_character_
),
ag_income_cal = case_when(
ag_income == 0 ~ 0,
ag_income > 0 ~ ag_income / hh_size,
TRUE ~ NA_real_
),
livestock_income_cal = case_when(
livestock_income == 0 ~ 0,
livestock_income > 0 ~ livestock_income / hh_size,
TRUE ~ NA_real_
),
rent_income_cal = case_when(
rent_income == 0 ~ 0,
rent_income > 0 ~ rent_income / hh_size,
TRUE ~ NA_real_
),
small_business_income_cal = case_when(
small_business_income == 0 ~ 0,
small_business_income > 0 ~ small_business_income / hh_size,
TRUE ~ NA_real_
),
unskill_labor_income_cal = case_when(
unskill_labor_income == 0 ~ 0,
unskill_labor_income > 0 ~ unskill_labor_income / hh_size,
TRUE ~ NA_real_
),
skill_labor_income_cal = case_when(
skill_labor_income == 0 ~ 0,
skill_labor_income > 0 ~ skill_labor_income / hh_size,
TRUE ~ NA_real_
),
formal_employment_income_cal = case_when(
formal_employment_income == 0 ~ 0,
formal_employment_income > 0 ~ formal_employment_income / hh_size,
TRUE ~ NA_real_
),
gov_benefits_income_cal = case_when(
gov_benefits_income == 0 ~ 0,
gov_benefits_income > 0 ~ gov_benefits_income / hh_size,
TRUE ~ NA_real_
),
hum_assistance_income_cal = case_when(
hum_assistance_income == 0 ~ 0,
hum_assistance_income > 0 ~ hum_assistance_income / hh_size,
TRUE ~ NA_real_
),
remittance_income_cal = case_when(
remittance_income == 0 ~ 0,
remittance_income > 0 ~ remittance_income / hh_size,
TRUE ~ NA_real_
),
loans_income_cal = case_when(
loans_income == 0 ~ 0,
loans_income > 0 ~ loans_income / hh_size,
TRUE ~ NA_real_
),
asset_selling_income_cal = case_when(
asset_selling_income == 0 ~ 0,
asset_selling_income > 0 ~ asset_selling_income / hh_size,
TRUE ~ NA_real_
),
total_income_cal = case_when(
total_income == 0 ~ 0,
total_income > 0 ~ total_income / hh_size,
TRUE ~ NA_real_
),
# Debt level
debt_amount_cal = case_when(
debt_amount == 0 ~ 0,
debt_amount > 0 ~ debt_amount / hh_size,
TRUE ~ NA_real_),
food_exp_cal = case_when(
total_income == 0 ~ 0,
total_income > 0 ~ food_exp / total_income,
TRUE ~ NA_real_
),
water_expt_cal = case_when(
total_income == 0 ~ 0,
total_income > 0 ~ water_expt / total_income,
TRUE ~ NA_real_
),
rent_exp_cal = case_when(
total_income == 0 ~ 0,
total_income > 0 ~ rent_exp / total_income,
TRUE ~ NA_real_
),
fuel_exp_cal = case_when(
total_income == 0 ~ 0,
total_income > 0 ~ fuel_exp / total_income,
TRUE ~ NA_real_
),
debt_exp_cal = case_when(
total_income == 0 ~ 0,
total_income > 0 ~ debt_exp / total_income,
TRUE ~ NA_real_
),
basic_needs_cal = case_when(
total_income == 0 ~ 0,
food_water_rent_num > 0 ~ food_water_rent_num / total_income,
TRUE ~ NA_real_
),
minimum_dietary_diversity_cal = case_when(
min_die_num >= 4 ~ "4 food groups",
min_die_num < 4 ~ "<4 food groups",
TRUE ~ NA_character_
),
rooms_hh_cal = case_when(
rooms > 0 ~ hh_size / rooms,
TRUE ~ 0
),
blankets_people_cal = case_when(
blankets_number == 0 ~ 0,
blankets_number > 0 ~ blankets_number / hh_size,
TRUE ~ NA_real_
),
blankets_suff_cal = case_when(
blankets_people_cal < 1 ~ "<1",
blankets_people_cal >= 1 ~ "1+",
TRUE ~ NA_character_
),
priority_nfi_cal = case_when(
priority_nfi_num <= 1 ~ "0_1",
priority_nfi_num <= 3 ~ "2_3",
priority_nfi_num <= 5 ~ "4_5",
priority_nfi_num <= 6 ~ "6",
TRUE ~ NA_character_
),
imp_energy_source1_cal = case_when(
energy_source %in% c("wood" , "animal_waste" , "paper_waste") ~ 1,
TRUE ~ 0
),
imp_energy_source2_cal = case_when(
energy_source %in% c("coal" , "charcoal" , "lpg" , "electricity") ~ 1,
TRUE ~ 0
),
diarrhea_cases_cal = case_when(
diarrhea_cases == 0 ~ 0,
diarrhea_cases > 0 ~ diarrhea_cases / diarrhea_total,
TRUE ~ NA_real_
),
perc_diarrhea_cases_cal = case_when(
diarrhea_cases == 0 ~ "0",
diarrhea_cases > 0 ~ ">=1",
TRUE ~ NA_character_
),
imp_water_source1_cal = case_when(
water_source %in% c("handpump_private", "handpump_public",
"piped_public", "spring_protected") ~ 1,
TRUE ~ 0
),
imp_water_source2_cal = case_when(
water_source %in% c("spring_unprotected","surface_water"
, "water_trucking", "other") ~ 1,
TRUE ~ 0
),
imp_san_source1_cal = case_when(
water_source %in% c("open", "pit_latrine_uncovered",
"other") ~ 1,
TRUE ~ 0
),
imp_san_source2_cal = case_when(
water_source %in% c("public_latrine", "pit_latrine_covered",
"vip_latrine", "flush_toilet_open_drain",
"flush_toilet_septic") ~ 1,
TRUE ~ 0
),
comp_ind_sev_2_call = case_when(
comp_ind_sev >= 2 ~ ">=2",
comp_ind_sev <2 ~ "<2",
TRUE ~ NA_character_
),
comp_ind_sev_2_nut_call = case_when(
comp_ind_sev_nut >= 2 ~ ">=2",
comp_ind_sev_nut <2 ~ "<2",
TRUE ~ NA_character_
),
comp_ind_sev_3_call = case_when(
comp_ind_sev >= 3 ~ ">=3",
comp_ind_sev < 3 ~ "<3",
TRUE ~ NA_character_
),
dep_ratio_call = case_when(
age_0_4_merged == 0 & age_65_merged == 0 ~ 0,
(age_0_14_merged > 0 | age_65_merged > 0) ~
sum(age_0_14_merged,age_65_merged, na.rm = TRUE)/sum(age_15_64_merged,na.rm = TRUE),
TRUE ~ NA_real_
),
dep_ratio_call_2 = case_when(
age_0_17_merged == 0 & age_60_and_more_merged == 0 ~ 0,
(age_0_17_merged > 0 | age_60_and_more_merged > 0) ~
sum(age_0_17_merged,age_60_and_more_merged, na.rm = TRUE)/sum(age_18_59_merged ,na.rm = TRUE),
TRUE ~ NA_real_
),
female_lit_call = case_when(
female_literacy == 0 ~ 0,
female_literacy == 0 ~
female_literacy/sum(females_11_17_total,females_18_plus_total, na.rm=TRUE),
TRUE ~ NA_real_
),
male_lit_call = case_when(
male_literacy == 0 ~ 0,
male_literacy == 0 ~
male_literacy/sum(males_11_17_total,males_18_plus_total, na.rm=TRUE),
TRUE ~ NA_real_
),
adult_behavior_change_call = case_when(
adult_behavior_change == "yes" ~ 1,
adult_behavior_change == "no" ~ 0,
TRUE ~ NA_real_
),
child_behavior_change_call = case_when(
child_behavior_change == "yes" ~ 1,
child_behavior_change == "no" ~ 0,
TRUE ~ NA_real_
),
atleast_one_behav_change_call = case_when(
child_behavior_change_call == 0 & adult_behavior_change_call == 0 ~ 0,
child_behavior_change_call > 0 | adult_behavior_change_call > 0 ~ 1,
TRUE ~ NA_real_
),
adults_working_call = case_when(
adults_working == 0 ~ 0,
adults_working > 0 & age_18_64 > 0 ~ adults_working/age_18_64,
TRUE ~ NA_real_
),
child_working_call = case_when(
is.na(children_working) ~ 0,
children_working == 0 ~ 0,
children_working > 0 & age_10_17 > 0 ~ children_working/age_10_17,
TRUE ~ NA_real_
),
adult_tazkira_cal = case_when(
adult_tazkira == 0 ~ "0",
adult_tazkira >= 1 ~ ">=1",
TRUE ~ NA_character_
),
child_tazkira_cal = case_when(
child_tazkira == 0 ~ "0",
child_tazkira >= 1 ~ ">=1",
TRUE ~ NA_character_
),
child_tazkira_cal = case_when(
child_tazkira == 0 ~ "0",
child_tazkira >= 1 ~ ">=1",
TRUE ~ NA_character_
),
any_tazkira_cal = case_when(
adult_tazkira == 0 & child_tazkira == 0~ "0",
adult_tazkira >= 1 | child_tazkira >= 1~ ">=1",
TRUE ~ NA_character_
),
insuf_blank_energy = case_when(
blankets_suff_cal == "<1" & energy_source == "wood" | energy_source == "paper_waste" | energy_source == "animal_waste" ~ "yes",
TRUE ~ "no"
),
current_presence_mines = case_when(
displ_explosive_presence == "both" | displ_explosive_presence == "current" |
nondispl_explosive_presence == "yes" ~ "current_explosive_presence",
displ_explosive_presence == "previous" | displ_explosive_presence == "no" |
nondispl_explosive_presence == "no" ~ "no_explosive_presence",
TRUE ~ NA_character_
),
# child_working_call = case_when(
# children_working == 0 ~ 0,
# children_working > 0 ~ children_working/age_10_17,
# TRUE ~ NA_real_
# ),
count_current_enrolled_avg = count_current_enrolled / edu_age_boys_girls_num,
count_current_attending_avg = count_current_attending / edu_age_boys_girls_num
)
# Major events
major_events_vars <- c(
"major_events.avalanche",
"major_events.conflict",
"major_events.drought",
"major_events.earthquake",
"major_events.floods",
"major_events.other"
)
major_events_score <- (rowSums(data[major_events_vars]))
data <- data %>%
mutate(
major_events_cal = case_when(
major_events_score == 0 ~ "none",
major_events_score == 1 ~ "1",
major_events_score == 2 ~ "2",
major_events_score >= 3 ~ ">= 3",
TRUE ~ NA_character_
)
)
# hno_intersectoral analysis
data <- data %>%
mutate(
# GBV incident OR threat
gbv_incidents_threats = case_when(
other_incidents == "sgbv" | other_concerns == "sgbv" ~ ">=1",
(other_incidents == "no" | other_incidents == "other") &
(other_concerns == "no" | other_concerns == "other") ~ "0",
TRUE ~ NA_character_
),
# At least one protection incident for adult OR child
prot_incident_adult_child = case_when(
adult_prot_incidents != "none" | child_prot_incidents != "none" ~ ">=1",
adult_prot_incidents == "none" & child_prot_incidents == "none" ~ "0",
TRUE ~ NA_character_
),
# Total income per day per household member in USD
daily_income_hh_members = case_when(
(((total_income / 30) / hh_size) / 78.36) > 1.90 ~ 1,
(((total_income / 30) / hh_size) / 78.36) <= 1.90 ~ 0,
TRUE ~ NA_real_
),
# health
health_service_access_class = case_when(
health_facility_access == "no" ~ 1,
health_facility_access == "yes" ~ 0,
TRUE ~ NA_real_
),
#ESNFI
shelter_type_access_class = case_when(
shelter == "tent" | shelter == "makeshift_shelter" |
shelter == "collective_centre" | shelter == "open_space" ~ 1,
shelter == "transitional" | shelter == "permanent" ~ 0,
TRUE ~ NA_real_
),
#EiE
hh_level_school_attendance_class = case_when(
count_current_attending > 0 ~ 1,
TRUE ~ 0
),
#FSA No data for Farah paper interviews
market_service_access_class = case_when(
market_access == "no" ~ 1,
market_access == "yes" ~ 0,
TRUE ~ NA_real_
),
#protection
identity_ownership_class = case_when(
child_tazkira == 0 & adult_tazkira == 0 ~ 1,
child_tazkira >=1 | adult_tazkira >=1 ~ 0,
TRUE ~ NA_real_
),
#WASH
access_to_water_class = case_when(
water_source == "handpump_private" | water_source == "handpump_public" |
water_source == "piped_public" | water_source == "spring_protected" ~ 1,
water_source == "spring_unprotected" | water_source == "surface_water" |
water_source == "water_trucking" | water_source == "other" ~ 0,
TRUE ~ NA_real_
)
)
access_services_vars <- c("health_service_access_class", "shelter_type_access_class",
"hh_level_school_attendance_class", "market_service_access_class",
"identity_ownership_class","access_to_water_class")
data$services_score <- comp_score(data, access_services_vars)
data <- data %>%
mutate(
comp_ind_access_services = case_when(
services_score <= 2 ~ 0,
services_score >= 3 ~ 1
)
)
#Recoding new variables
data$hh_no_tazkira <- ifelse(data$tazkira_total < 1, "Tazkira_No", "Tazkira_Yes")
data$muac_yes_no <- ifelse(data$muac_total > 0 & !is.na(data$min_muac) ,"Yes","No")
data$recent_non_recent <- ifelse(data$final_displacement_status_non_displ == "recent_idps", "recent_idps",
ifelse(data$final_displacement_status_non_displ == "non_recent_idps", "non_recent_idps", NA ))
data$edu_removal_shock_cal <- ifelse(data$shock_class == 5, "Yes", "No")
data$enrolled_attending <- ifelse(data$count_enrolled_attending > 0, "Enrolled_and_Attending", "Not" )
data$schoo_age_boys_girls <- coerc(data$boys_ed) + coerc(data$girls_ed)
## source disaggs
source("r/prepare_disagg.R")
################ MEB analysis ###########################################################
# sustainable income vars
sustainable_income_vars <- c(
'ag_income',
'livestock_income',
'rent_income',
'small_business_income',
'unskill_labor_income',
'skill_labor_income',
'formal_employment_income',
'gov_benefits_income'
)
# sustainable income per HH
data$sustainable_income <- comp_score(data, sustainable_income_vars)
# sustainable income per HH member
data$sustainable_income_per_mem <- data$sustainable_income / data$hh_size
# net inocome per hh
data$net_income <- data$sustainable_income - data$all_expenses
# total Exp per hh memeber
data$total_exp_per_mem <- data$all_expenses / data$hh_size
data <- data %>%
mutate(
food_exp_per_mem = food_exp / hh_size,
water_expt_per_mem = water_expt / hh_size,
rent_exp_per_mem = rent_exp / hh_size,
fuel_exp_per_mem = fuel_exp / hh_size,
debt_exp_per_mem = debt_exp / hh_size,
food_exp_spent = case_when(
food_exp > 0 ~ 1,
food_exp == 0 ~ 0,
TRUE ~ NA_real_
),
water_expt_spent = case_when(
water_expt > 0 ~ 1,
water_expt == 0 ~ 0,
TRUE ~ NA_real_
),
rent_exp_spent = case_when(
rent_exp > 0 ~ 1,
rent_exp == 0 ~ 0,
TRUE ~ NA_real_
),
fuel_exp_spent = case_when(
fuel_exp > 0 ~ 1,
fuel_exp == 0 ~ 0,
TRUE ~ NA_real_
),
debt_exp_spent = case_when(
debt_exp > 0 ~ 1,
debt_exp == 0 ~ 0,
TRUE ~ NA_real_
)
)
# sustainable income 2 vars
sustainable_income_2_vars <- c(
'ag_income',
'livestock_income',
'rent_income',
'small_business_income',
'unskill_labor_income',
'skill_labor_income',
'formal_employment_income'
)
# sustainable income per HH
data$sustainable_income_2 <- comp_score(data, sustainable_income_2_vars)
# sustainable income per HH member
data$sustainable_income_2_per_mem <- data$sustainable_income_2 / data$hh_size
# unskilled_labor_income per hh member
data$unskill_labor_income_per_mem <- data$unskill_labor_income / data$hh_size
# unskilled + agriculture + livestock income vars
unskill_ag_live_income_vars <- c(
'ag_income',
'livestock_income',
'unskill_labor_income'
)
# unskilled + agriculture + livestock income
data$unskill_ag_live_income_income <- comp_score(data, unskill_ag_live_income_vars)
# unskilled + agriculture + livestock income per HH member
data$unskill_ag_live_income_income_per_mem <- data$unskill_ag_live_income_income / data$hh_size
########################################################################################
############################# Vulnerablity composites ###################################
data <- data %>%
mutate(
hoh_disabled_vul_class = case_when(
data$hoh_disabled == "disabled" ~ 1,
data$hoh_disabled == "not_disabled" ~ 0,
TRUE ~ 0
),
hoh_debt_disagg_vul_class = case_when(
hoh_debt_disagg == "high_debt" ~ 1,
hoh_debt_disagg == "low_debt" ~ 0,
hoh_debt_disagg == "medium_debt" ~ 0,
hoh_debt_disagg == "no_debt" ~ 0,
TRUE ~ 0
),
tazkira_disagg_vul_class = case_when(
tazkira_disagg == "non_have_tazkira" ~ 1,
tazkira_disagg == "all_have_tazkira" ~ 0,
TRUE ~ 0
),
hoh_age_group_vul_class = case_when(
hoh_age_group == "65+" ~ 1,
hoh_age_group == "<65" ~ 0,
TRUE ~ 0
),
hoh_sex_disagg_vul_class = case_when(
hoh_sex_disagg == "female" ~ 1,
hoh_sex_disagg == "male" ~ 0,
TRUE ~ 0
),
pregnant_lactating_member_vul_class = case_when(
pregnant_lactating_member == "at_least_one_mem_pregnant_lactating" ~ 1,
pregnant_lactating_member == "no_mem_pregnent_lactating" ~ 0,
TRUE ~ 0
),
chronic_illness_vul_class = case_when(
chronic_illness == "yes" ~ 1,
chronic_illness == "no " ~ 0,
chronic_illness == "no_answer" ~ 0,
TRUE ~ 0
),
literacy_vul_class = case_when(
female_literacy_yes_no == "0" & male_literacy_yes_no == "0" ~ 1,
female_literacy_yes_no == "1 or more" ~ 0,
male_literacy_yes_no == "1 or more" ~ 0,
TRUE ~ 0
),
behav_change_disagg_vul_class = case_when(
behav_change_disagg == "yes" ~ 1,
behav_change_disagg == "no" ~ 0,
TRUE ~ 0
),
female_literacy_yes_no_class = case_when(
female_literacy_yes_no == "0" ~ 1,
female_literacy_yes_no == "1 or more" ~ 0,
TRUE ~ 0
),
behavior_change_cause_class2 = case_when(
behavior_change_cause == "yes" ~ 1,
TRUE ~ 0
),
child_behavior_change_class2 = case_when(
child_behavior_change == "yes" ~ 1,
TRUE ~ 0
),
adult_behavior_change_class2 = case_when(
adult_behavior_change == "yes" ~ 1,
TRUE ~ 0
),
adult_behavior_only_by_conflict = case_when(
adult_behavior_change == "yes" & behavior_change_cause == "yes" ~ 1,
TRUE ~ 0
)
)
## Vulnerable_group_1
Vulnerable_group_1_vars <- c(
"hoh_disabled_vul_class",
"hoh_debt_disagg_vul_class",
"tazkira_disagg_vul_class",
"hoh_age_group_vul_class",
"hoh_sex_disagg_vul_class",
"chronic_illness_vul_class",
"literacy_vul_class"
)
data$Vulnerable_group_1_vars_score <- comp_score(data, Vulnerable_group_1_vars)
data <- data %>%
mutate(
vulnerable_group_1 = case_when(
Vulnerable_group_1_vars_score >= 1 ~ "vulnerable",
Vulnerable_group_1_vars_score == 0 ~ "not_vulnerable",
TRUE ~ NA_character_
)
)
## vulnerable_group_4
Vulnerable_group_4_vars <- c(
"hoh_disabled_vul_class",
"hoh_debt_disagg_vul_class",
"tazkira_disagg_vul_class",
"hoh_age_group_vul_class",
"hoh_sex_disagg_vul_class",
"behav_change_disagg_vul_class"
)
data$Vulnerable_group_4_vars_score <- comp_score(data, Vulnerable_group_4_vars)
data <- data %>%
mutate(
vulnerable_group_4 = case_when(
Vulnerable_group_4_vars_score >= 1 ~ "vulnerable",
Vulnerable_group_4_vars_score == 0 ~ "not_vulnerable",
TRUE ~ NA_character_
)
)
## Vulnerable_group_5
Vulnerable_group_5_vars <- c(
"hoh_disabled_vul_class",
"hoh_debt_disagg_vul_class",
"tazkira_disagg_vul_class",
"hoh_age_group_vul_class",
"hoh_sex_disagg_vul_class",
"pregnant_lactating_member_vul_class",
"chronic_illness_vul_class",
"behav_change_disagg_vul_class",
"literacy_vul_class"
)
data$Vulnerable_group_5_vars_score <- comp_score(data, Vulnerable_group_5_vars)
data <- data %>%
mutate(
vulnerable_group_5 = case_when(
Vulnerable_group_5_vars_score >= 1 ~ "vulnerable",
Vulnerable_group_5_vars_score == 0 ~ "not_vulnerable",
TRUE ~ NA_character_
)
)
## Vulnerable_group_6
Vulnerable_group_6_vars <- c(
"hoh_disabled_vul_class",
"hoh_debt_disagg_vul_class",
"tazkira_disagg_vul_class",
"hoh_age_group_vul_class",
"hoh_sex_disagg_vul_class",
"behav_change_disagg_vul_class",
"female_literacy_yes_no_class"
)
data$Vulnerable_group_6_vars_score <- comp_score(data, Vulnerable_group_6_vars)
data <- data %>%
mutate(
vulnerable_group_6 = case_when(
Vulnerable_group_6_vars_score >= 1 ~ "vulnerable",
Vulnerable_group_6_vars_score == 0 ~ "not_vulnerable",
TRUE ~ NA_character_
)
)
## Vulnerable_group_7
## vulnerable_group_7
Vulnerable_group_7_vars <- c(
"hoh_disabled_vul_class",
"hoh_debt_disagg_vul_class",
"tazkira_disagg_vul_class",
"hoh_age_group_vul_class",
"hoh_sex_disagg_vul_class",
"behavior_change_cause_class2"
)
data$Vulnerable_group_7_vars_score <- comp_score(data, Vulnerable_group_7_vars)
data <- data %>%
mutate(
vulnerable_group_7 = case_when(
Vulnerable_group_7_vars_score >= 1 ~ "vulnerable",
Vulnerable_group_7_vars_score == 0 ~ "not_vulnerable",
TRUE ~ NA_character_
)
)
## vulnerable_group_8
Vulnerable_group_8_vars <- c(
"hoh_disabled_vul_class",
"hoh_debt_disagg_vul_class",
"tazkira_disagg_vul_class",
"hoh_age_group_vul_class",
"hoh_sex_disagg_vul_class",
"adult_behavior_only_by_conflict"
)
data$Vulnerable_group_8_vars_score <- comp_score(data, Vulnerable_group_8_vars)
data <- data %>%
mutate(
vulnerable_group_8 = case_when(
Vulnerable_group_8_vars_score >= 1 ~ "vulnerable",
Vulnerable_group_8_vars_score == 0 ~ "not_vulnerable",
TRUE ~ NA_character_
)
)
###############################################end
#########################
## esnfi_new_indicator_1
data <- data %>%
mutate(
shelter_class2 = case_when(
shelter == "tent" | shelter == "collective_centre" | shelter == "makeshift_shelter" |
shelter == "open_space" ~ 1,
shelter == "transitional" | shelter == "permanent" ~ 0,
TRUE ~ 0
),
shelter_damage_and_repair_class = case_when(
(shelter_damage.due_to_conflict == 1 | shelter_damage.due_to_natural_disaster == 1) &
shelter_damage_repair == "no" ~ 1,
shelter_damage.no == 1 | shelter_damage_repair == "yes" ~ 0,
TRUE ~ 0
),
priority_nfi_cal_class = case_when(
priority_nfi_cal == "0_1" | priority_nfi_cal == "2_3" ~ 1,
priority_nfi_cal == "4_5" | priority_nfi_cal == "6" ~ 0,
TRUE ~ 0
)
)
esnfi_new_indicator_1_vars <- c(
"shelter_class2",
"shelter_damage_and_repair_class",
"priority_nfi_cal_class"
)
esnfi_new_indicator_1_vars_score <- comp_score(data, esnfi_new_indicator_1_vars)
data <- data %>%
mutate(
esnfi_new_indicator_1 = case_when(
esnfi_new_indicator_1_vars_score >= 1 ~ 1,
esnfi_new_indicator_1_vars_score == 0 ~ 0,
TRUE ~ 0
)
)
###################################end
######## wash_new_indicator 1#######
data <- data %>%
mutate(
water_source_class2 = case_when(
water_source == "spring_unprotected" | water_source == "surface_water" | water_source == "water_trucking" |
water_source == "other" ~ 1,
water_source == "handpump_private" | water_source == "handpump_public" | water_source == "handpump_public" |
water_source == "spring_protected" ~ 0,
TRUE ~ 0
),
latrine_class2 = case_when(
latrine == "open" | latrine == "pit_latrine_uncovered" | latrine == "other" ~ 1,
latrine == "public_latrine" | latrine == "pit_latrine_covered" | latrine == "vip_latrine" |
latrine == "flush_toilet_open_drain" | latrine == "flush_toilet_septic" ~ 0,
TRUE ~ 0
),
soap_class2 = case_when(
soap == "no" ~ 1,
soap == "yes_didnt_see" | soap == "yes_saw" ~ 0,
TRUE ~ 0
),
diarrhea_cases_class = case_when(
diarrhea_cases >= 1 ~ 1,
diarrhea_cases == 0 ~ 0,
TRUE ~ 0
),
diarrhea_cases_class_children = case_when(
diarrhea_cases >= 1 ~ 1,
diarrhea_cases == 0 ~ 0,
TRUE ~ NA_real_
)
)
wash_new_indicator_1_vars <- c(
"water_source_class2",
"latrine_class2",
"soap_class2"
)
wash_new_indicator_1_vars_score <- comp_score(data, wash_new_indicator_1_vars)
data <- data %>%
mutate(
wash_new_indicator_1 = case_when(
wash_new_indicator_1_vars_score >= 2 ~ 1,
wash_new_indicator_1_vars_score == 0 ~ 0,
TRUE ~ 0
)
)
##############################################end
######## wash_new_indicator 2 #######
wash_new_indicator_2_vars <- c(
"water_source_class2",
"latrine_class2",
"soap_class2",
"diarrhea_cases_class"
)
wash_new_indicator_2_vars_score <- comp_score(data, wash_new_indicator_2_vars)
data <- data %>%
mutate(
wash_new_indicator_2 = case_when(
wash_new_indicator_2_vars_score >= 2 ~ 1,
wash_new_indicator_2_vars_score == 0 ~ 0,
TRUE ~ 0
)
)
############# winterization_indicator ##################
data <- data %>%
mutate(
shelter_class3 = case_when(
shelter == "tent" | shelter == "collective_centre" | shelter == "makeshift_shelter" |
shelter == "open_space" ~ 1,
shelter == "transitional" | shelter == "permanent" ~ 0,
TRUE ~ 0
),
blankets_suff_cal_class = case_when(
blankets_suff_cal == "<1" ~ 1,
blankets_suff_cal == "1+" ~ 0,
TRUE ~ 0
),
energy_source_class = case_when(
energy_source == "animal_waste" | energy_source == "charcoal" | energy_source == "paper_waste" |
energy_source == "wood" ~ 1,
energy_source == "coal" | energy_source == "lpg" | energy_source == "electricity" |
energy_source == "other" ~ 0,
TRUE ~ 0
)
)
winterization_indicator_vars <- c(
"shelter_class3",
"blankets_suff_cal_class",
"energy_source_class"
)
winterization_indicator_vars_score <- comp_score(data, winterization_indicator_vars)
data <- data %>%
mutate(
winterization_indicator = case_when(
winterization_indicator_vars_score >= 2 ~ 1,
winterization_indicator_vars_score == 0 ~ 0,
TRUE ~ 0
)
)
################################################end
#################### dip push factors ############
ipd_push_factors_vars <- c(
'idp_push_factors.active_conflict',
'idp_push_factors.anticipated_conflict',
'idp_push_factors.earthquake',
'idp_push_factors.floods',
'idp_push_factors.avalanche',
'idp_push_factors.drought',
'idp_push_factors.poverty',
'idp_push_factors.service_access',
'idp_push_factors.other'
)
ipd_push_factors_vars_short <- c(
'idp_push_factors.active_conflict',
'idp_push_factors.anticipated_conflict',
'idp_push_factors.earthquake',
'idp_push_factors.floods',
'idp_push_factors.avalanche',
'idp_push_factors.drought'
)
data$ipd_push_factors_vars_score <- comp_score(data, ipd_push_factors_vars)
data$ipd_push_factors_vars_score_short <- comp_score(data, ipd_push_factors_vars_short)
data <- data %>%
mutate(
idp_push_factors_cat = case_when(
ipd_push_factors_vars_score == 1 ~ "1_event",
ipd_push_factors_vars_score < 3 ~ "2_events",
ipd_push_factors_vars_score >=3 ~ "3_or_more_events",
TRUE ~ NA_character_
),
idp_push_factors_cat_short = case_when(
ipd_push_factors_vars_score_short == 1 ~ "1_event",
ipd_push_factors_vars_score_short < 3 ~ "2_events",
ipd_push_factors_vars_score_short >=3 ~ "3_or_more_events",
TRUE ~ NA_character_
)
)
##################### MSNI #######################
#### IMPACT
data <- data %>%
mutate(
major_events_impc = case_when(
major_events_cal == ">= 3" | major_events_cal == "2" ~ 3,
major_events_cal == "1" ~ 1,
TRUE ~ 0
),
agricultural_impact_how_impc = case_when(
agricultural_impact_how == "51_75" ~ 1,
agricultural_impact_how == "76_100" ~ 2,
TRUE ~ 0
),
livestock_impact_how_impc = case_when(
livestock_impact_how.livestock_died == 1 ~ 1,
livestock_impact_how.left_unattended == 1 ~ 1,
TRUE ~ 0
),
explosive_impact_death_impc = case_when(
explosive_impact.injury_death == 1 ~ 3,
TRUE ~ 0
),
explosive_impact_others_impc = case_when(
explosive_impact.psych_impact == 1 | explosive_impact.relocation == 1 |
explosive_impact.access_services == 1 | explosive_impact.restrict_recreation == 1 |
explosive_impact.livelihoods_impact == 1 | explosive_impact.other == 1 &
explosive_impact.injury_death != 1 ~ 2,
TRUE ~ 0
),
adult_injuries_cause_impc = case_when(
adult_injuries_cause == "conflict" | adult_injuries_cause == "natural_disaster" |
child_injuries_cause == "conflict" | child_injuries_cause == "natural_disaster" ~ 3,
TRUE ~ 0
),
shelter_damage_impc = case_when(
shelter_damage == "due_to_conflict" | shelter_damage == "due_to_natural_disaster" ~ 2,
TRUE ~ 0
),
edu_removal_shock_impc = case_when(
count_shock >= 1 ~ 1,
TRUE ~ 0
),
health_facility_reopened_impc = case_when(
health_facility_reopened == "remain_closed" ~ 1,
TRUE ~ 0
),
water_damaged_cause_impc = case_when(
water_damaged_cause == "conflict" | water_damaged_cause == "natural_disaster" |
water_damaged_cause == "drought" ~ 2,
TRUE ~ 0
),
aid_access_issue_2_impc = case_when(
aid_access_issue == "yes" & aid_access_issue_type == "insecurity" |
aid_access_issue_type == "explosive_hazards" ~ 2,
TRUE ~ 0
),
aid_access_issue_1_impc = case_when(
aid_access_issue == "yes" & aid_access_issue_type == "distance" |
aid_access_issue_type == "social_restrictions" ~ 1,
TRUE ~ 0
)
)
# impact class vars
msni_impact_score_vars <- c(
"major_events_impc",
"agricultural_impact_how_impc",
"livestock_impact_how_impc",
"explosive_impact_death_impc",
"explosive_impact_others_impc",
"adult_injuries_cause_impc",
"shelter_damage_impc",
"edu_removal_shock_impc",
"health_facility_reopened_impc",
"water_damaged_cause_impc",
"aid_access_issue_2_impc",
"aid_access_issue_1_impc"
)
# impact score
data$msni_impact_score <- comp_score(data, msni_impact_score_vars)
# impact severity
data <- data %>%
mutate(
impact = case_when(
msni_impact_score < 3 ~ 1,
msni_impact_score > 2 & msni_impact_score < 6 ~ 2,
msni_impact_score > 5 & msni_impact_score < 9 ~ 3,
msni_impact_score >= 9 ~ 4,
TRUE ~ 0
)
)
#### End IMPACT
#### Capacity gaps
data <- data %>%
mutate(
capacity_gaps = case_when(
lcsi_severity == "minimal" ~ 1,
lcsi_severity == "stress" ~ 2,
lcsi_severity == "severe" ~ 3,
lcsi_severity == "extreme" ~ 4,
TRUE ~ NA_real_
)
)
#### End Capacity gaps
# HC-LSG/ESNFI - shelter_lsg
data <- data %>%
mutate(
shelter_type_lsg = case_when(
shelter == "open_space" ~ 3,
shelter == "tent" | shelter == "makeshift_shelter" | shelter == "collective_centre" ~ 2,
# shelter == "transitional" ~ 1,
TRUE ~ 0
),
shelter_damage_lsg = case_when(
shelter_damage_extent == "fully_destroyed" & shelter_damage_repair == "no" ~ 3,
shelter_damage_extent == "significant_damage" & shelter_damage_repair == "no" ~ 2,
TRUE ~ 0
),
winterisation_lsg = case_when(
blankets_suff_cal == "<1" & (energy_source == "animal_waste" | energy_source == "paper_waste" |
energy_source == "wood") ~ 3,
TRUE ~ 0
),
access_nfi_lsg = case_when(
priority_nfi_num < 3 ~ 3,
priority_nfi_num > 2 & priority_nfi_num < 6 ~ 2,
TRUE ~ 0
)
)
# shelter_lsg class vars
msni_shelter_lsg_vars <- c(
"shelter_type_lsg",
"shelter_damage_lsg",
"winterisation_lsg",
"access_nfi_lsg"
)
# shelter_lsg score
data$msni_shelter_lsg_score <- comp_score(data, msni_shelter_lsg_vars)
# shelter_lsg severity
data <- data %>%
mutate(
shelter_lsg = case_when(
msni_shelter_lsg_score < 3 ~ 1,
msni_shelter_lsg_score > 2 & msni_shelter_lsg_score < 6 ~ 2,
msni_shelter_lsg_score > 5 & msni_shelter_lsg_score < 9 ~ 3,
msni_shelter_lsg_score >= 9 ~ 4
)
)
#### end shelter_lsg
# HC-LSG/FSA - fsl_lsg
data <- data %>%
mutate(
fcs_lsg = case_when(
fcs_category == "poor" ~ 3,
fcs_category == "borderline" ~ 2,
TRUE ~ 0
),
hhs_lsg = case_when(
hhs_category == "severe_hunger" ~ 3,
hhs_category == "moderate_hunger" ~ 2,
TRUE ~ 0
),
food_source_lsg = case_when(
food_source == "assistance" | food_source == "gift" | food_source == "borrowed" ~ 3,
# food_source == "borrowed" ~ 2,
TRUE ~ 0
),
market_access_lsg = case_when(
market_access == "no" ~ 3,
TRUE ~ 0
),
market_distance_lsg = case_when(
market_distance == "6_10km" ~ 2,
TRUE ~ 0
)
)
# fsl_lsg class vars
msni_fsl_lsg_vars <- c(
"fcs_lsg",
"hhs_lsg",
"food_source_lsg",
"market_access_lsg",
"market_distance_lsg"
)
# fsl_lsg score
data$msni_fsl_lsg_score <- comp_score(data, msni_fsl_lsg_vars)
# fsl_lsg severity
data <- data %>%
mutate(
fsl_lsg = case_when(
msni_fsl_lsg_score < 3 ~ 1,
msni_fsl_lsg_score > 2 & msni_fsl_lsg_score < 6 ~ 2,
msni_fsl_lsg_score > 5 & msni_fsl_lsg_score < 9 ~ 3,
msni_fsl_lsg_score >= 9 ~ 4
)
)
# fsl_lsg severity 2
data <- data %>%
mutate(
fsl_lsg_2 = case_when(
msni_fsl_lsg_score < 3 ~ 1,
msni_fsl_lsg_score > 2 & msni_fsl_lsg_score < 7 ~ 2,
msni_fsl_lsg_score > 6 & msni_fsl_lsg_score < 9 ~ 3,
msni_fsl_lsg_score >= 9 ~ 4
)
)
# fsl_lsg severity 3
data <- data %>%
mutate(
fsl_lsg_3 = case_when(
msni_fsl_lsg_score < 3 ~ 1,
msni_fsl_lsg_score > 2 & msni_fsl_lsg_score < 8 ~ 2,
msni_fsl_lsg_score > 7 & msni_fsl_lsg_score < 10 ~ 3,
msni_fsl_lsg_score >= 10 ~ 4
)
)
#### end fsl_lsg
# HC-LSG/Health - health_lsg
data <- data %>%
mutate(
access_health_center_lsg = case_when(
health_facility_access == "no" ~ 3,
TRUE ~ 0
),
health_facility_distance_lsg = case_when(
health_facility_distance == "none" | health_facility_distance == "more_10km" ~ 3,
health_facility_distance == "6_10km" ~ 2,
TRUE ~ 0
),
behav_change_lsg = case_when(
behav_change_disagg == "yes" ~ 3,
TRUE ~ 0
),
birth_location_lsg = case_when(
birth_location == "outside" | diarrhea_cases_class == 1 ~ 3,
birth_location == "home" | birth_location == "midwife_home" | birth_location == "other" ~ 2,
TRUE ~ 0
)
)
# health_lsg class vars
msni_health_lsg_vars <- c(
"access_health_center_lsg",
"health_facility_distance_lsg",
"behav_change_lsg",
"birth_location_lsg"
)
# health_lsg score
data$msni_health_lsg_score <- comp_score(data, msni_health_lsg_vars)
# health_lsg severity
data <- data %>%
mutate(
health_lsg = case_when(
msni_health_lsg_score < 3 ~ 1,
msni_health_lsg_score > 2 & msni_health_lsg_score < 6 ~ 2,
msni_health_lsg_score > 5 & msni_health_lsg_score < 9 ~ 3,
msni_health_lsg_score >= 9 ~ 4
)
)
#### end health_lsg
# HC-LSG/Protection - protection_lsg
data <- data %>%
mutate(
prot_incidents_4_lsg = case_when(
adult_prot_incidents.assaulted_with_weapon == 1 | adult_prot_incidents.hindered_leave_settlement == 1 |
adult_prot_incidents.forced_work == 1 | adult_prot_incidents.forcibly_detained == 1 |
child_prot_incidents.assaulted_with_weapon == 1 | child_prot_incidents.hindered_leave_settlement == 1 |
child_prot_incidents.forced_work == 1 | child_prot_incidents.forcibly_detained == 1 ~ 4,
TRUE ~ 0
),
prot_incidents_3_lsg = case_when(
adult_prot_incidents.verbally_threatened == 1 | adult_prot_incidents.assaulted_without_weapon == 1 |
adult_prot_incidents.hindered_leave_district == 1 | child_prot_incidents.verbally_threatened == 1 |
child_prot_incidents.assaulted_without_weapon == 1 | child_prot_incidents.hindered_leave_district == 1 &
(adult_prot_incidents.assaulted_with_weapon == 0 | adult_prot_incidents.hindered_leave_settlement == 0 |
adult_prot_incidents.forced_work == 0 | adult_prot_incidents.forcibly_detained == 0 |
child_prot_incidents.assaulted_with_weapon == 0 | child_prot_incidents.hindered_leave_settlement == 0 |
child_prot_incidents.forced_work == 0 | child_prot_incidents.forcibly_detained == 0) ~ 3,
TRUE ~ 0
),
other_incidents_lsg = case_when(
other_incidents == "sgbv" | other_concerns == "sgbv" | boy_marriage == "yes" | girl_marriage == "yes" ~ 3,
TRUE ~ 0
),
prot_concerns_lsg = case_when(
prot_concerns.violence_maiming == 1 | prot_concerns.violence_injuries == 1 | prot_concerns.psych_wellbeing == 1 |
prot_concerns.abduction == 1 | prot_concerns.theft == 1 | prot_concerns.explosive_hazards == 1 |
prot_concerns.destruction_property == 1 | prot_concerns.early_marriage == 1 | prot_concerns.other == 1 ~ 3,
TRUE ~ 0
),
safety_lsg = case_when(
safety == "poor" | safety == "very_poor" ~ 2,
TRUE ~ 0
)
)
# protection_lsg class vars
msni_protection_lsg_vars <- c(
"prot_incidents_4_lsg",
"prot_incidents_3_lsg",
"other_incidents_lsg",
"prot_concerns_lsg",
"safety_lsg"
)
# protection_lsg score
data$msni_protection_lsg_score <- comp_score(data, msni_protection_lsg_vars)
# protection_lsg severity
data <- data %>%
mutate(
protection_lsg = case_when(
msni_protection_lsg_score < 3 ~ 1,
msni_protection_lsg_score > 2 & msni_protection_lsg_score < 6 ~ 2,
msni_protection_lsg_score > 5 & msni_protection_lsg_score < 9 ~ 3,
msni_protection_lsg_score >= 9 ~ 4
)
)
#### end protection_lsg
# HC-LSG/WASH wash_lsg
data <- data %>%
mutate(
water_source_lsg = case_when(
water_source == "surface_water" ~ 3,
water_source == "water_trucking" | water_source == "spring_unprotected" ~ 2,
TRUE ~ 0
),
soap_lsg = case_when(
soap == "no" ~ 3,
TRUE ~ 0
),
latrine_lsg = case_when(
latrine == "open" | latrine == "public_latrine" | waste_disposal == "open_space" ~ 3,
latrine == "pit_latrine_uncovered" | waste_disposal == "burning" ~ 2,
TRUE ~ 0
),
water_distance_lsg = case_when(
water_distance == "over_1km" ~ 3,
water_distance == "500m_to_1km" ~ 2,
TRUE ~ 0
)
)
# wash_lsg class vars
msni_wash_lsg_vars <- c(
"water_source_lsg",
"soap_lsg",
"latrine_lsg",
"water_distance_lsg"
)
# wash_lsg score
data$msni_wash_lsg_score <- comp_score(data, msni_wash_lsg_vars)
# wash_lsg severity
data <- data %>%
mutate(
wash_lsg = case_when(
msni_wash_lsg_score < 3 ~ 1,
msni_wash_lsg_score > 2 & msni_wash_lsg_score < 6 ~ 2,
msni_wash_lsg_score > 5 & msni_wash_lsg_score < 9 ~ 3,
msni_wash_lsg_score >= 9 ~ 4
)
)
# wash_lsg severity 2
data <- data %>%
mutate(
wash_lsg_2 = case_when(
msni_wash_lsg_score < 3 ~ 1,
msni_wash_lsg_score > 2 & msni_wash_lsg_score < 7 ~ 2,
msni_wash_lsg_score > 6 & msni_wash_lsg_score < 9 ~ 3,
msni_wash_lsg_score >= 9 ~ 4
)
)
# wash_lsg severity 3
data <- data %>%
mutate(
wash_lsg_3 = case_when(
msni_wash_lsg_score < 3 ~ 1,
msni_wash_lsg_score > 2 & msni_wash_lsg_score < 8 ~ 2,
msni_wash_lsg_score > 7 & msni_wash_lsg_score < 10 ~ 3,
msni_wash_lsg_score >= 10 ~ 4
)
)
#### end wash_lsg
# HC-LSG/EiE - education_lsg
data <- data %>%
mutate(
not_attending_lsg = case_when(
percent_enrolled >= 0.75 & percent_enrolled <= 1 ~ 4,
percent_enrolled >= 0.5 & percent_enrolled <= 0.749 ~ 3,
percent_enrolled >= 0.25 & percent_enrolled <= 0.449 ~ 2,
percent_enrolled >= 0 & percent_enrolled <= 0.249 ~ 1,
TRUE ~ 0
),
education_level_lsg = case_when(
highest_edu == "none" ~ 2,
highest_edu == "primary" ~ 1,
TRUE ~ 0
),
unattending_security_lsg = case_when(
boy_unattendance_reason.insecurity == 1 | boy_unattendance_reason.child_works_instead == 1 |
girl_unattendance_reason.insecurity == 1 | girl_unattendance_reason.child_works_instead == 1 ~ 3,
TRUE ~ 0
),
unattending_cultural_lsg = case_when(
boy_unattendance_reason.cultural_reasons == 1 | girl_unattendance_reason == 1 |
boy_unattendance_reason.lack_facilities == 1 | girl_unattendance_reason.lack_facilities == 1 ~ 2,
TRUE ~ 0
),
unattending_finance_doc_lsg = case_when(
boy_unattendance_reason.lack_documentation == 1 | boy_unattendance_reason.too_expensive == 1 |
boy_unattendance_reason.lack_teachers == 1 | girl_unattendance_reason.lack_documentation ==1 |
girl_unattendance_reason.too_expensive == 1 | girl_unattendance_reason.lack_teachers == 1 ~ 1,
TRUE ~ 0
)
)
# education_lsg class vars
msni_education_lsg_vars <- c(
"not_attending_lsg",
"unattending_security_lsg",
"unattending_cultural_lsg",
"unattending_finance_doc_lsg",
"education_level_lsg"
)
# education_lsg score
data$msni_education_lsg_score <- comp_score(data, msni_education_lsg_vars)
# education_lsg severity
data <- data %>%
mutate(
education_lsg = case_when(
msni_education_lsg_score < 3 ~ 1,
msni_education_lsg_score > 2 & msni_education_lsg_score < 6 ~ 2,
msni_education_lsg_score > 5 & msni_education_lsg_score < 9 ~ 3,
msni_education_lsg_score >= 9 ~ 4
)
)
#### end education_lsg
#################################################
data <- data %>% filter(!is.na(province))
# fliter prolematic feilds
uuid_filter <- c("ac3e8430-ba88-497b-9895-c1bd8da7f79e",
"8ac61e9b-8ff8-4e4a-9619-1dc0ab31f396",
"7171e0a8-3a40-4c57-b84d-a65f08115994",
"596c244b-ea20-48ef-8218-023ac3f2831f")
`%notin%` <- Negate(`%in%`)
data <- data %>% filter(uuid %notin% uuid_filter )
# MSNI Indicator
data$msni <- msni(education_lsg = data$education_lsg,
fsl_lsg = data$fsl_lsg,
health_lsg = data$health_lsg,
protection_lsg = data$protection_lsg,
shelter_lsg = data$shelter_lsg,
wash_lsg = data$wash_lsg,
capacity_gaps = data$capacity_gaps,
impact = data$impact)
data$msni2 <- msni(education_lsg = data$education_lsg,
fsl_lsg = data$fsl_lsg_2,
health_lsg = data$health_lsg,
protection_lsg = data$protection_lsg,
shelter_lsg = data$shelter_lsg,
wash_lsg = data$wash_lsg_2,
capacity_gaps = data$capacity_gaps,
impact = data$impact)
data$msni3 <- msni(education_lsg = data$education_lsg,
fsl_lsg = data$fsl_lsg_3,
health_lsg = data$health_lsg,
protection_lsg = data$protection_lsg,
shelter_lsg = data$shelter_lsg,
wash_lsg = data$wash_lsg_3,
capacity_gaps = data$capacity_gaps,
impact = data$impact)
data$msni_sev_high <- ifelse(data$msni==3|data$msni==4,1,0)
# HHs found to have severe or extreme sectoral needs in one or more sectors
# lsg_needs_2_cal
data <- data %>%
mutate(
shelter_lsg_class = case_when(
shelter_lsg == 3 | shelter_lsg == 4 ~ 1,
shelter_lsg == 1 | shelter_lsg == 2 ~ 0,
TRUE ~ NA_real_
),
fsl_lsg_class = case_when(
fsl_lsg == 3 | fsl_lsg == 4 ~ 1,
fsl_lsg == 1 | fsl_lsg == 2 ~ 0,
TRUE ~ NA_real_
),
health_lsg_class = case_when(
health_lsg == 3 | health_lsg == 4 ~ 1,
health_lsg == 1 | health_lsg == 2 ~ 0,
TRUE ~ NA_real_
),
protection_lsg_class = case_when(
protection_lsg == 3 | protection_lsg == 4 ~ 1,
protection_lsg == 1 | protection_lsg == 2 ~ 0,
TRUE ~ NA_real_
),
wash_lsg_class = case_when(
wash_lsg == 3 | wash_lsg == 4 ~ 1,
wash_lsg == 1 | wash_lsg == 2 ~ 0,
TRUE ~ NA_real_
),
education_lsg = case_when(
education_lsg == 3 | education_lsg == 4 ~ 1,
education_lsg == 1 | education_lsg == 2 ~ 0,
TRUE ~ NA_real_
)
)
lsg_needs_2_cal_vars <- c(
"shelter_lsg_class",
"fsl_lsg_class",
"health_lsg_class",
"protection_lsg_class",
"wash_lsg_class",
"education_lsg"
)
# lsg_needs_2_cal score
data$lsg_needs_2_cal_score <- comp_score(data, lsg_needs_2_cal_vars)
# lsg_needs_2_cal
data <- data %>%
mutate(
lsg_needs_2_cal = case_when(
lsg_needs_2_cal_score == 0 ~ "no_need",
lsg_needs_2_cal_score == 1 ~ "one_need",
lsg_needs_2_cal_score > 1 ~ "two_or_more_need"
)
)
# msni drivers
data <- data %>%
mutate(
fsl_wash_driver = case_when(
fsl_lsg == 3 | fsl_lsg == 4 | wash_lsg == 3 | wash_lsg == 4 ~ "sectoral_need",
fsl_lsg == 1 | fsl_lsg == 2 | wash_lsg == 1 | wash_lsg == 2 ~ "no_need",
TRUE ~ NA_character_
),
impact_driver = case_when(
((impact == 3 | impact == 4) & (health_lsg == 3 | health_lsg == 4)) |
((impact == 3 | impact == 4) & (shelter_lsg == 3 | shelter_lsg == 4)) |
((impact == 3 | impact == 4) & (protection_lsg == 3 | protection_lsg == 4)) ~ "sectoral_need",
TRUE ~ "no_need"
),
shelter_driver_class = case_when(
shelter_lsg == 3 | shelter_lsg == 4 ~ 1,
TRUE ~ 0,
),
protection_driver_class = case_when(
protection_lsg == 3 | protection_lsg == 4 ~ 1,
TRUE ~ 0
),
health_driver_class = case_when(
health_lsg == 3 | health_lsg == 4 ~ 1,
TRUE ~ 0
),
capacity_gaps_sev = case_when(
capacity_gaps >=3 ~ "high",
capacity_gaps <=2 ~ "low",
TRUE ~ NA_character_
)
)
# esnfi_prot_health_driver
esnfi_prot_health_driver_vars <- c(
"shelter_driver_class",
"protection_driver_class",
"health_driver_class"
)
# esnfi_prot_health_driver score
data$esnfi_prot_health_driver_score <- comp_score(data, esnfi_prot_health_driver_vars)
# lsg_needs_2_cal
data <- data %>%
mutate(
esnfi_prot_health_driver = case_when(
esnfi_prot_health_driver_score <= 1 ~ "no_need",
esnfi_prot_health_driver_score >= 2 ~ "sectoral_need",
)
)
############### MSNI TEST ###############
data <- data %>%
mutate(
hh_msni_one = case_when(
education_lsg == 1 & fsl_lsg == 1 & health_lsg == 1 & protection_lsg == 1 & shelter_lsg == 1 & wash_lsg == 1 &
capacity_gaps == 1 & impact == 1 ~ "1",
TRUE ~ "1+"
),
hh_msni_one_only_sectors = case_when(
education_lsg == 1 & fsl_lsg == 1 & health_lsg == 1 & protection_lsg == 1 & shelter_lsg == 1 & wash_lsg == 1 ~ "1",
TRUE ~ "1+"
)
)
#########################################
#join main dataset var to hh roster
data_sub <- data %>% select(final_displacement_status_non_displ, region_disagg, urban_disagg,
hoh_sex_disagg, hoh_disabled_disagg, hoh_chronic_illness_disagg, hoh_elderly_disagg,
displacements_disagg, literate_adult_disagg, lcsi_disagg, host_disagg, disp_length_disagg, hoh_sex2_disagg,
behav_change_disagg, child_behav_change_disagg,
tazkira_disagg3, hoh_debt_disagg , vulnerable_group_4, vulnerable_group_7, registered_dissagg, informal_settlement,
child_tazkira_disagg, uuid)
overall_hh_roster <- overall_hh_roster %>%
mutate(
school_age = case_when(
hh_member_age >=6 & hh_member_age <= 18 ~ "school_age",
TRUE ~ "not_school_age"
),
current_year_attending_na_no = case_when(
current_year_attending == "no" ~ "no",
current_year_attending == "yes" ~ "yes",
TRUE & school_age == "school_age" ~ "no"
),
edu_removal_shock.no_sch_age = case_when(
edu_removal_shock.no == 1 ~ 1,
TRUE & school_age == "school_age" ~ 0
),
edu_removal_shock.conflict_sch_age = case_when(
edu_removal_shock.conflict == 1 ~ 1,
TRUE & school_age == "school_age" ~ 0
),
edu_removal_shock.displacement_sch_age = case_when(
edu_removal_shock.displacement == 1 ~ 1,
TRUE & school_age == "school_age" ~ 0
),
edu_removal_shock.natural_disaster_sch_age = case_when(
edu_removal_shock.natural_disaster == 1 ~ 1,
TRUE & school_age == "school_age" ~ 0
),
edu_removal_shock_sch_age = case_when(
edu_removal_shock.no == 1 ~ "yes",
TRUE & school_age == "school_age" ~ "no"
),
hh_member_age_cat = case_when(
hh_member_age >= 0 & hh_member_age < 6 ~ "0_5",
hh_member_age > 5 & hh_member_age < 19 ~ "6_18",
hh_member_age > 18 & hh_member_age < 60 ~ "19_59",
hh_member_age > 59 ~ "60+"
),
# demographic hh roster data
hh_member_age_cat_gender = case_when(
hh_member_age >= 0 & hh_member_age < 6 & hh_member_sex == "female" ~ "female_0_5",
hh_member_age >= 0 & hh_member_age < 6 & hh_member_sex == "male" ~ "male_0_5",
hh_member_age > 5 & hh_member_age < 19 & hh_member_sex == "female" ~ "female_6_18",
hh_member_age > 5 & hh_member_age < 19 & hh_member_sex == "male" ~ "male_6_18",
hh_member_age > 18 & hh_member_age < 60 & hh_member_sex == "female" ~ "female_19_59",
hh_member_age > 18 & hh_member_age < 60 & hh_member_sex == "male" ~ "male_19_59",
hh_member_age > 59 & hh_member_sex == "female" ~ "female_60+",
hh_member_age > 59 & hh_member_sex == "male" ~ "male_60+"
),
male_female_perc = case_when(
hh_member_sex == "female" ~ "female",
hh_member_sex == "male" ~ "male"
),
## request # 30
school_age_cat_gender = case_when(
hh_member_age > 5 & hh_member_age < 13 & hh_member_sex == "female" ~ "female_6_12",
hh_member_age > 12 & hh_member_age < 19 & hh_member_sex == "female" ~ "female_13_18",
hh_member_age > 5 & hh_member_age < 13 & hh_member_sex == "male" ~ "male_6_12",
hh_member_age > 12 & hh_member_age < 19 & hh_member_sex == "male" ~ "male_13_18",
TRUE ~ NA_character_
)
)
############## demographic hoh data #####################
hoh_data <- data %>%
select(
hh_member_sex = hoh_sex,
hh_member_age = hoh_age,
`_submission__uuid` = uuid,
province,
survey_village) %>%
mutate(
hh_member_age_cat_gender = case_when(
hh_member_age >= 0 & hh_member_age < 6 & hh_member_sex == "female" ~ "female_0_5",
hh_member_age >= 0 & hh_member_age < 6 & hh_member_sex == "male" ~ "male_0_5",
hh_member_age > 5 & hh_member_age < 19 & hh_member_sex == "female" ~ "female_6_18",
hh_member_age > 5 & hh_member_age < 19 & hh_member_sex == "male" ~ "male_6_18",
hh_member_age > 18 & hh_member_age < 60 & hh_member_sex == "female" ~ "female_19_59",
hh_member_age > 18 & hh_member_age < 60 & hh_member_sex == "male" ~ "male_19_59",
hh_member_age > 59 & hh_member_sex == "female" ~ "female_60+",
hh_member_age > 59 & hh_member_sex == "male" ~ "male_60+"
),
male_female_perc = case_when(
hh_member_sex == "female" ~ "female",
hh_member_sex == "male" ~ "male"
),
school_age_cat_gender = case_when(
hh_member_age > 5 & hh_member_age < 13 & hh_member_sex == "female" ~ "female_6_12",
hh_member_age > 12 & hh_member_age < 19 & hh_member_sex == "female" ~ "female_13_18",
hh_member_age > 5 & hh_member_age < 13 & hh_member_sex == "male" ~ "male_6_12",
hh_member_age > 12 & hh_member_age < 19 & hh_member_sex == "male" ~ "male_13_18",
TRUE ~ NA_character_
)
)
hoh_data <- hoh_data %>%
select(
hh_member_sex,
hh_member_age,
hh_member_age_cat_gender,
male_female_perc,
school_age_cat_gender,
province,
survey_village,
`_submission__uuid`
)
roster_data <- overall_hh_roster %>%
select(
hh_member_sex,
hh_member_age,
hh_member_age_cat_gender,
male_female_perc,
school_age_cat_gender,
province,
survey_village,
`_submission__uuid`
)
combined_hoh_and_roster <- rbind(roster_data, hoh_data)
# for demographic
combined_hoh_and_roster_joined <- koboloops::add_parent_to_loop(combined_hoh_and_roster, data_sub, uuid.name.loop = "_submission__uuid", uuid.name.parent = "uuid")
combined_hoh_and_roster_joined <- combined_hoh_and_roster_joined %>%
mutate(
hh_member_under_over_15 = case_when(
hh_member_age <= 15 ~ "15_and_under",
hh_member_age > 15 ~ "over_15",
TRUE ~ NA_character_
)
)
write.csv(combined_hoh_and_roster_joined, "./input/data/recoded/hh_roster_hoh_demographic.csv", row.names = F)
# for education questions
hh_roster_joined <- koboloops::add_parent_to_loop(overall_hh_roster, data_sub, uuid.name.loop = "_submission__uuid", uuid.name.parent = "uuid")
write.csv(hh_roster_joined, "./input/data/recoded/hh_roster.csv", row.names = F)
write.csv(data, "./input/data/recoded/data_with_strata2.csv", row.names = F)
## Test
|
17d0b4508a89eda9690757bbd1a506dc8eba11fb | de83a2d0fef79a480bde5d607937f0d002aa879e | /P2C2M.SNAPP/R/draw.samples2.R | 4afd6ee40fe1adb1f7db29b2654b926047494a2b | [] | no_license | P2C2M/P2C2M_SNAPP | 0565abc0ea93195c9622dc5d4e693ccde17bebc7 | 94cd62285419a79f5d03666ec2ea3e818803d0db | refs/heads/master | 2020-05-07T18:54:40.440682 | 2020-01-10T15:59:45 | 2020-01-10T15:59:45 | 180,788,408 | 2 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,099 | r | draw.samples2.R | ##### Randomly sample from posterior #####
draw.samples <- function(num_sims, gens_run, sample_unif){ # num.sims = user input # of simulations to perform; gens_run = # of markov steps saved; sample_unif = if true, sample posterior uniformly. Otherwise sample randomly
burnin <- ceiling(gens_run * 0.10)
non_burnin <- seq(burnin + 1, gens_run, 1) # get sequence of step numbers in non burnin posterior
if (num_sims > length(non_burnin)){ # if # of simulations input is greater than the number of steps in the posterior
post_samples <- non_burnin # use all non_burnin steps
} else{
if (sample_unif == TRUE){
interval <- length(non_burnin) / num_sims # get interval to sample
post_samples <- non_burnin[1]
while (post_samples[length(post_samples)] + interval <= gens_run){
post_samples = c(post_samples, post_samples[length(post_samples)] + interval)
}
post_samples <- sapply(post_samples, floor) # round down
} else{
post_samples <- sort(sample(non_burnin, num_sims)) # randomly sample steps
}
}
return(post_samples)
}
|
5e6123c9c6678ffff155f6d6bb0973954d846370 | 925c515b771a8ea7ca31cc530308d594c30fba07 | /code/TableS3.R | 3591bcb04a019b009fd1c4d141478c8c465a6176 | [] | no_license | melofton/freshwater-forecasting-review | 41ba42f0aee6180d7a731fcf838dccc8f7590588 | c06097cbab6d88c1dc30d0f2c3cf8a3baddaeacc | refs/heads/main | 2023-07-06T21:54:48.183725 | 2023-06-27T20:18:46 | 2023-06-27T20:18:46 | 478,673,588 | 0 | 1 | null | 2022-07-08T19:45:20 | 2022-04-06T18:05:25 | R | UTF-8 | R | false | false | 541 | r | TableS3.R | #Matrix analysis
#Author: Mary Lofton
#Date: 06JUL22
#clear environment
rm(list = ls())
#set-up
pacman::p_load(tidyverse, lubridate, cowplot,ggbeeswarm, viridis)
#read in data
dat5 <- read_csv("./data/cleaned_matrix.csv")
##Table 3 ####
dat10 <- dat5 %>%
mutate(ecosystem_type = ifelse(ecosystem == "river" | grepl("basin",other_ecosystem),"Lotic","Lentic"))
colnames(dat10)
tab3 <- dat10[,c(2,4,3,27,11,12,13,14,18,19,20,21,16,17,23)] %>%
arrange(Year)
tab3[16,"Year"] <- 2022
write.csv(tab3,"Table3.csv",row.names = FALSE)
|
1bc891cc48422875088ad36e2f4ff1053e811f2d | 218aae83a9d0994561991ba8affe528f1e381457 | /R/edgepoints.R | e7cee35ceeef11b83b9b9d0b6abfaa1dc7d09ef5 | [] | no_license | cran/edci | 4efcf830e8cec5d1522397140afd5650655b66b3 | d24ed3f7d6bd543f5b1fa07b8db821d42c8fe795 | refs/heads/master | 2020-12-25T16:56:26.204461 | 2018-05-16T20:49:37 | 2018-05-16T20:49:37 | 17,718,677 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 3,698 | r | edgepoints.R | edgepoints = function(data, h1n, h2n, asteps = 4, estimator = "kernel", kernel = "mean",
score = "gauss", sigma = 1, kernelfunc = NULL, margin = FALSE) {
epDelta = function(x) {
if (x < 0)
-1
else
1
}
epAt = function(x, y) {
if (x == 0) {
if (y >= 0)
pi/2
else
pi/2
} else {
atan(y/x)
}
}
epR1 = function(theta, x, y) {
sqrt(x^2 + y^2) * epDelta(x) * cos(epAt(x, y) - theta)
}
epR2 = function(theta, x, y) {
sqrt(x^2 + y^2) * epDelta(x) * sin(epAt(x, y) - theta)
}
angle = matrix(double(length(data)), nrow=nrow(data))
value = matrix(double(length(data)), nrow=nrow(data))
es = NULL
sc = NULL
ms = sigma * max(data)
if (estimator == "kernel")
es = 0
else if (estimator == "median")
es = 1
else if (estimator == "M_mean")
es = 2
else if (estimator == "M_median")
es = 3
else if (estimator == "test_mean")
es = 5
else if (estimator == "test_median")
es = 6
else
stop("estimator \"", estimator, "\" unknown.")
if (es==2 || es ==3) {
if (score == "gauss") {
sc = 0
}
if (score == "huber") {
sc = 1
#ms = sigma/2
}
if (score == "mean") {
sc = 9
}
}
env = ceiling(sqrt((h1n * nrow(data))^2 + (h2n * ncol(data))^2))
kernmat = NULL
if (kernel == "mean" || es >= 5) {
kern = 0
} else if (kernel == "linear") {
kern = 1
} else if (kernel == "linear2") {
kern = 2
} else if (kernel == "gauss") {
kern = 3
} else if (kernel == "func") {
kern = 4
kernmat = double(asteps * (2 * env + 1)^2)
for (i in ((-env):env)) {
for(j in ((-env):env)) {
for (k in (0:(asteps - 1))) {
theta = -pi/2 + (k * pi/asteps)
x = epR1(theta, i/nrow(data), j/ncol(data))/h1n
y = epR2(theta, i/nrow(data), j/ncol(data))/h2n
kernmat[k * (2 * env + 1)^2 + (i + env) * (2 * env + 1) + (j + env) + 1] =
kernelfunc(2 * x, y)
}
}
}
} else {
stop("kernel \"",kernel,"\" unknown.")
}
if (es == 1)
kern = 0
if (!is.null(es)) {
result = .C("c_edgepoints",
as.double(data),
nrow(data),
ncol(data),
as.integer(kern), # kernel
as.double(h1n),
as.double(h2n),
as.integer(es),
as.integer(sc), # Typ der Scorefunktion
as.double(sigma), # Sigma
as.double(kernmat), # Gewichtsmatrix
as.double(ms), # Max_Schritt
as.integer(asteps),
angle = angle,
value = value,
PACKAGE = "edci")
}
value = result$value
angle = result$angle
if (es == 5 || es == 6)
value = -value
if (margin == FALSE) {
if (es == 5 || es == 6)
v = 1
else
v = 0
value[c(1:env,(nrow(value) - env + 1):nrow(value)), ] = v
value[, c(1:env, (ncol(value) - env+1):ncol(value))] = v
} else if (margin == "cut") {
value = value[(env + 1):(nrow(value) - env), (env + 1):(ncol(value) - env)]
angle = angle[(env + 1):(nrow(angle) - env), (env + 1):(ncol(angle) - env)]
}
list(value = value, angle = angle)
}
eplist = function(data, maxval, test = FALSE, xc = NULL, yc = NULL) {
if (test == TRUE) {
data[[1]] = -data[[1]]
maxval = -maxval
}
n = sum(data[[1]] > maxval)
if (is.null(xc))
xc = seq(1/nrow(data[[1]]), 1, 1/nrow(data[[1]]))
if (is.null(yc))
yc = seq(1/ncol(data[[1]]), 1, 1/ncol(data[[1]]))
o = order(data[[1]], decreasing = TRUE)[1:n]
result = cbind(xc[(o - 1) %% nrow(data[[1]]) + 1], yc[(o - 1) %/% nrow(data[[1]]) +1 ],
data[[2]][o])
colnames(result) = c("x", "y", "angle")
result
}
|
4e4ff604aaf7b5ff470c8227b043cf073c00c388 | d3fdbf9442b8e0ffbc208ad50087f0ece05f405e | /Modulo 3- Resampling-Bayesianos-Markov/Ejercicio 3.3/Ejercicio3_3_MarianaSilvera.R | c9857fb242ead0825c194c8628a35108f8e2f36e | [] | no_license | msilvera/R-DataAnalysis2021-OTGA | b176f5f48076ce57ed1c7935fbe37ada31f21bda | 1bc03219b4d36c73d2196534c111878476d4373d | refs/heads/main | 2023-06-14T21:55:42.036064 | 2021-07-04T18:38:51 | 2021-07-04T18:38:51 | 380,082,123 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 947 | r | Ejercicio3_3_MarianaSilvera.R | library(FSAdata)
library(MASS)
library(dplyr)
#library(help="FSAdata")
#cargo los datos
#data <- WalleyeErie2
summary(WalleyeErie2)
data <-subset(x=WalleyeErie2, subset = !is.na(w)) #elimino los datos incompletos
summary(data)
set.seed(1) # semilla para el random
data <- data %>% mutate_at(vars("age"), factor) # transformo en factor la comlumna edad
#extraigo el 80% de los datos para entrenamiento
intrain <- sample(1:nrow(data), size = round(0.8*nrow(data)))
#genero modelo para la edad en base a las demas variables
lda.fit <-lda(age~. , data= data, subset= intrain)
lda.fit
#verifico que tan bien se comporta el discriminante lineal generado
lda.pred <- predict(lda.fit, data)
names(lda.pred)
#obtenfo la clase
lda.class <- lda.pred$class
#construyo la matriz
table(lda.class, data$age)
#veo que tan bien se ajusta, utilizando la media
mean(lda.class==data$age)
#resultado, desempeño de : 0.6571231 |
eb3d9c97b02f6f8d4ca16e857d987432473f6d4c | 89d2d6b83bb0fcad3db66b139a617b0cc40bf34a | /R3-Aliona.R | dfc622532aca8311dc0e2430e94dcfdf29a65c9b | [] | no_license | alionahst/R3 | 5e6760cab681ab10149267ed31884ccb16cc6eb5 | d980eddc32efd762b3178bc3933b8ba486929944 | refs/heads/master | 2023-01-03T05:25:17.275377 | 2020-10-20T22:13:42 | 2020-10-20T22:13:42 | 305,684,425 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,799 | r | R3-Aliona.R | #Chapter 3: Basic graphics and data - Aliona Hoste
demo(graphics)
plot(iris)
#1. Plot a cheat-sheet with values of color and point type (col = , and pch = ) from 1 to 25, and export it as a jpeg of 15 cm wide, 6 cm high and resolution 100 points per cm.
plot(0, 0, xlim = c(0, 26), ylim = c(0.5, 1.5)
, ylab = "color", xlab = "color number", yaxt = "n")
for (i in 1:25) {
points(i, 1, pch = i, col = i, cex = 1.5)
}
jpeg(filename = "firstplot.jpeg", width = 15, height = 6, units = "cm", res = 100)
plot(0, 0, xlim = c(0, 26), ylim = c(0.5, 1.5)
, ylab = "colors & sign", xlab = "color number", yaxt = "n")
for (i in 1:25) {
points(i, 1, pch = i, col = i, cex = 1.5)
}
dev.off()
#2. Plot into a graph ten Poisson distributions with lambda ranging from 1 to 10. Put legend and title. Export it as a .tiff file with size of 15x15 cm.
x <- seq(-1, 20, 1) # Sequence
y <- dpois(x, lambda = 1) # densities for x
plot(x, y, type = "n") # Empty plot (type = "n")
for(i in 1:10){
y <- dpois(x, lambda = i)
lines(x, y, col = i)
}
title(main="Poisson distribution, lambda = 1:10")
legend("topright", legend = paste("lambda =", 1:10),lty = 1, col = 1:10)
#export into tiff plot
tiff("Plot1_poisson_1to10.tiff", width = 15, height = 15, units = "cm",
bg = "transparent", res = 150) # Open the device "Plot1.tiff"
x <- seq(-1, 20, 1) # Sequence
y <- dpois(x, lambda = 1) # densities for x
plot(x, y, type = "n") # Empty plot (type = "n")
for(i in 1:10){
y <- dpois(x, lambda = i)
lines(x, y, col = i)
}
title(main="Poisson distribution, lambda = 1:10")
legend("topright", legend = paste("lambda =", 1:10),lty = 1, col = 1:10)
dev.off()
#3. Import data from this article: https://peerj.com/articles/328/
Webcsv <- "https://dfzljdn9uc3pi.cloudfront.net/2014/328/1/Appendix1.csv"
Data <- read.table(Webcsv, header = T, sep = ",", skip = 2)
str(Data)
#Be careful importing the data. Notice that you have to skip two first lines using “skip = 2”13.
#With these data, using for(), plot graphs to represent the effect of all the numerical variables, from “richness” to “mean_quality” on “yield”. Choose the type of graph that you think better represents this effect for the different species. Create only one pdf with all the graphs inside.
#To find the best graph for each type of data, a very helpful web is from Data to Viz https://www.data-to-viz.com/.
plot(Data[-1])
plot(Data$mean_yield ~ Data$richness)
for(i in names(Data[6:12]))
{
plot(Data$mean_yield ~ Data[[i]], ylab = "Mean yields", xlab = as.character(names(Data[i])))
title(main= paste("Mean yield in function of", as.character(names(Data[i]))))
}
|
7015d5870ad5056141a600ab0b532cfd67a48a59 | e56da52eb0eaccad038b8027c0a753d9eb2ff19e | /man-roxygen/tipsForTreeGeneration.R | b3874469bd8aa861c1cbae942f72fce3a7ff9898 | [] | no_license | ms609/TreeTools | fb1b656968aba57ab975ba1b88a3ddf465155235 | 3a2dfdef2e01d98bf1b58c8ee057350238a02b06 | refs/heads/master | 2023-08-31T10:02:01.031912 | 2023-08-18T12:21:10 | 2023-08-18T12:21:10 | 215,972,277 | 16 | 5 | null | 2023-08-16T16:04:19 | 2019-10-18T08:02:40 | R | UTF-8 | R | false | false | 174 | r | tipsForTreeGeneration.R | #' @param tips An integer specifying the number of tips, or a character vector
#' naming the tips, or any other object from which [`TipLabels()`] can
#' extract leaf labels.
|
15f17c33f851b0ab97d37c7507f338f9cc08551e | d30fa10aa7b3837145a1d1f0bcff6a55372ea4eb | /plot_kmer_dist.R | a39c14daba7632aabe23b7da8c1d0a54f095915a | [] | no_license | mborche2/Matts_Satellite_Size_Code | 541bfdada9a61238ecb6c59594dbfd5e60766e97 | 824fcf6e8f4ab555df774baa9cd8caf6dd8200ae | refs/heads/master | 2023-03-28T07:29:31.677930 | 2021-03-23T18:57:52 | 2021-03-23T18:57:52 | 348,837,905 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 689 | r | plot_kmer_dist.R | library(ggplot2)
setwd("/n/core/bigDataAI/Genomics/Gerton/jennifer_gerton/jeg10/")
for (i in 1:21){
filenam <- paste("plots/kmer_frequency_asp/kmer_frequency_",toString(i),"_array.tsv",sep = "")
array_specifics <- read.table(filenam,header = FALSE)
freq_table <- table(array_specifics[,2])
freq_df <- as.data.frame(freq_table)
freq_df[,2] <- log(freq_df[,2])
filenam2 <- paste("kmer_frequency_",toString(i),"_array_standard_axes.png",sep="")
png(filenam2)
plot(freq_df,xlab="Number of Occurences of Kmer in Array",ylab="Log (ln) Frequency of Kmer Occurence Number",axes=FALSE)
axis(side=1, at=seq(0,4000,by=25))
axis(side=2, at=seq(0, 10, by=1))
box()
dev.off()
}
|
393e68c42ae3b36432c1265386c913a44b8e6d7e | c97fa9aadc45c44fad6433ae10c772060bde355c | /MyNotes/03 - Geting and Cleaning Data/01 Class_Data.Table_Package.R | 41cd46ae55c1ae3679c91b186b783fad89090d5a | [] | no_license | vitorefigenio/datasciencecoursera | 9866816242d39fa9fc9520bc4d543efc815afeb5 | 03722d0c7c6d219ec84f48e02065493f6657cc0a | refs/heads/master | 2021-01-17T11:17:58.099767 | 2016-02-28T03:06:37 | 2016-02-28T03:06:37 | 29,034,385 | 0 | 0 | null | null | null | null | ISO-8859-1 | R | false | false | 943 | r | 01 Class_Data.Table_Package.R | #data.table package
# Create Data.Table
install.packages("data.table")
library(data.table)
DF = data.frame(x=rnorm(9), y=rep(c("a","b","c"), each=3), z=rnorm(9))
head(DF,3)
DT = data.table(x=rnorm(9), y=rep(c("a","b","c"), each=3), z=rnorm(9))
head(DT,3)
# comando ara ver tdas as abelas criadas na memória
tables()
# Subsetting rows
DT[2,]
DT[DT$y=="a"]
DT[c(2,3)]
# Subseting columns
DT[,c(2,3)]
# É comum o uso de expressões
{
x=1
y=2
}
k = {print(10);5}
print(k)
# Calculating values for variables with expressions
DT[,list(mean(x),sum(z))]
DT[,table(y)]
# Adding new column
DT[,w:=z^2]
DT
# Multiple Operations
DT[,m:={tmp <- (x+z); log2(tmp+5)}]
plot(DT[,m])
# plyr like operations
DT[, a:=x>0]
DT[,b:= mean(x+w), by=a]
DT
# Special Variables
set.seed(123)
DT <- data.table(x=sample(letters[1:3], 1E5, TRUE))
DT[, .N, by=x]
# keys
DT = data.table(x=rep(c("a","b","c"), each=100), z=rnorm(300))
setkey(DT,x)
DT['a'] |
12d5a52eb7e5fb10a0b5d87bdc8740c29b7c2a5a | 39315660a0226ae527ec8e0c7e6ae866df675b5f | /exercise1/computeCost.R | 5057e5c02d42dd31073fb2393dff4c7ded690bc3 | [] | no_license | Lemmawool/R-Practice | 28a7ce208f7d012eb4bc886fdb27b72754a171e9 | 0c3bed53e27953e9f19f92fd6e7b595a7e379262 | refs/heads/master | 2021-05-14T13:44:03.051151 | 2018-01-22T02:02:51 | 2018-01-22T02:02:51 | 115,955,944 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 105 | r | computeCost.R | computeCost <- function(X, y, theta){
m = length(y)
return ((1/(2*m)) * sum((X %*% theta - y) ^ 2))
} |
e662f9c90536aa7a7802ef2046cda55ac460d02e | 63227ea5a4085bb789824448502c95a98d8f375f | /cachematrix.R | 4e87ebeb3d25a82fe63b07127301dae99ce920d6 | [] | no_license | lfdelama/ProgrammingAssignment2 | f81f6ae4cf9246cc21a2fce019bc59a04949303d | 417909969f9fdd8c4d23700e1fcf535237a2c2ec | refs/heads/master | 2020-12-24T14:18:50.589091 | 2014-05-22T21:32:07 | 2014-05-22T21:32:07 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,314 | r | cachematrix.R | ## These two functions below are used to cache the inverse of a square matrix,
## so every time the same inverse is required, it doesn't need to be recomputed.
## This function creates a special matrix which
## contains a list of the following functions:
## - set, to set the value of the matrix
## - get, to get the value of the matrix
## - setinverse, to set the value of the inverse of the matrix
## - getinverse, to get the value of the inverse of the matrix
makeCacheMatrix<- function(x = matrix()) {
inv <- NULL
set <- function(y) {
x <<- y
inv <<- NULL
}
get <- function() x
setinverse <- function(inverse) inv <<- inverse
getinverse <- function() inv
list(set = set, get = get,
setinverse = setinverse,
getinverse = getinverse)
}
## This function calculates and returns the inverse of the special matrix that
## it was created with the above function.
## If the inverse of the matrix has previously calculated, this function will return
## directly the value stored in the cache of the makeCacheMatrix function.
cacheSolve <- function(x, ...) {
inv <- x$getinverse()
if(!is.null(inv)) {
message("getting cached data")
return(inv)
}
data <- x$get()
inv <- solve(data, ...)
x$setinverse(inv)
inv
}
|
48f4c3afd8bf9957f151bbbad760e9b7f9c317fe | 64e7ac1d0437b1d874b4ed070e6bda152decddee | /plot2.R | e2892d66195304ed5b560890e85b152215d7920e | [] | no_license | mooctus/ExData_Plotting1 | 072db8facebd27a8a8aab985be057b9b2c2b8122 | 005cba7dd9d88e94113a57eb6f8d77b9a3618811 | refs/heads/master | 2021-01-12T20:07:17.994147 | 2014-05-09T15:54:17 | 2014-05-09T15:54:17 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 569 | r | plot2.R | Sys.setlocale(category = "LC_ALL", locale = "C")
df <- read.table(file="household_power_consumption.txt", sep=";", na.strings="?", header=TRUE)
df$Time <- strptime(
paste0(df$Date, " ", df$Time),
format=paste0("%d/%m/%Y %H:%M:%S")
)
df$Date <- as.Date(df$Date,format="%d/%m/%Y")
df1 <- df[df$Date %in% as.Date(c('2007-02-01', '2007-02-02')),]
png(filename="plot2.png",width=480, height=480)
with (df1,
plot(Time, Global_active_power, type="n", xlab="", ylab="Global Active Power (kilowatts)")
)
with (df1,
lines(Time, Global_active_power)
)
dev.off()
|
ec9f4ad17398e6d6778438d88eaed81be1b890ff | e5a584e854ce025a135511f692dfc8e7ec178d49 | /grid.R | 07406eefe438ffe694bf2a7367e2ff8229779ff3 | [] | no_license | statspheny/sta242hw2 | 0fdb99c18f21c91dc990fd345c63daaf51067daa | 20aab766d40e91c16f3742f80bc3a953dc0c8846 | refs/heads/master | 2021-01-01T18:11:27.640548 | 2013-02-12T06:40:39 | 2013-02-12T06:40:39 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,593 | r | grid.R | newTrafficGrid = function(nrow,ncol) {
## This is a function that will
x = matrix(0,nrow=nrow,ncol=ncol)
class(x) = "trafficGrid"
return(x)
}
generateBMLgrid = function(nrow,ncol,nred,nblue) {
## the dimension of the grid
bmldim = c(nrow,ncol)
## randomly sample from nrow*ncol to get the correct number of red and blue cars
n = nrow*ncol
ncars = nred+nblue
## 1. randomly get ncars 2. of ncars, select the red ones 3. set
## the rest to be blue
cars = sample(1:n,ncars)
redcars = sample(cars,nred)
bluecars = cars[!(cars %in% redcars)]
obj = list(dim=bmldim,redcars=redcars,bluecar=bluecars)
class(obj) = "BMLgrid"
return(obj)
}
## generateBMLgrid = function(nrow,ncol,prob) {
## n = nrow*ncol
## data = sample(c(0,1,-1),size=n,prob=c(1-prob,prob/2,prob/2),replace=TRUE)
## x = matrix(data,nrow=nrow,ncol=ncol)
## redval = which(x==-1)
## blueval = which(x==1)
## obj = list()
## obj$matrix = x
## obj$red = redval
## obj$blue = blueval
## class(obj) = "BMLgrid"
## return(obj)
## }
changeCarProb = function(bgrid,prob) {
## This is a function that changes the probability of cars on the
## grid
return(bgrid)
}
changeCarNumber = function(bgrid,nred=NULL,nblue=NULL) {
## This is a function that changes the number of red or blue cars
## on the grid. If NULL, then the number of cars stays the same
return(bgrid)
}
## Methods
summary.BMLgrid = function(BMLgrid) {
print(BMLgrid)
}
plot.BMLgrid = function(obj) {
red = "#FF0000FF"
white = "#FFFFFFFF"
blue = "#0000FFFF"
shiftedmat = t(apply(obj$matrix,2,rev))
image(shiftedmat,col=c(red,white,blue))
}
checkIfCarStuck = function(toMove,inPlace) {
## This function checks if the blue car is stopped behind a red car
return(toMove %in% inPlace)
}
updateBlueCar = function(obj) {
## Each blue car moves up
mat = obj$matrix
nrow = nrow(mat) # number of rows
blueOld = obj$blue
## blueOld-1 moves the each car back one space
## !(blueOld%%ncol-1) is a logical for the cars at the end that reset
## add ncol to the cars that are reset
blueMove = (blueOld-1) + as.numeric(blueOld%%nrow==1)*nrow
## keep the old indices for the cars that are stuck behind old cars
stuckCars = checkIfCarStuck(blueMove,obj$red)
blueMove[stuckCars] = blueOld[stuckCars]
mat[blueOld]=0 #set old blue to zero
mat[blueMove]=1 #set new blue to 1
obj$matrix = mat
obj$blue = blueMove
return(obj)
}
|
7c95eaaba2e639e23869e5b4d852212db33c02c7 | 1ea27108545233075e57b2cc5c3b0ceeeb0c76d9 | /R_model_garch.R | e47c8a221f9921a9953a323155a4a2c5882370dc | [] | no_license | zhen-yang8/Stats451_group | a2af5cd72a2bede65e98d14df8eb41d5b07ac2fb | d8f3a2b44352498433e99fc72e2af53b3b322fed | refs/heads/main | 2023-01-22T13:16:31.244087 | 2020-12-05T01:05:49 | 2020-12-05T01:05:49 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 918 | r | R_model_garch.R | library(loo)
library(rstan)
garch11_setup <- rstan::stan_model(file = './Stan/model_garch.stan')
dat = read.csv("./data/bitcoin_train.csv")
test = read.csv("./data/bitcoin_test.csv")
y = dat$log_return
N = length(y)
y_test = test$log_return
J = length(y_test)
stan_data <- list(y = y,N = N,sigma1 = 0.01, J = J, y_test = y_test)
garch11 <- rstan::sampling(garch11_setup, data = stan_data)
p1 = stan_plot(garch11, pars = c("mu", "alpha0", "alpha1", "beta1")) +
ggtitle("GARCH(1,1)")
p1
p2 = stan_trace(garch11,
pars = c("mu", "alpha0", "alpha1", "beta1")) +
ggtitle("GARCH(1,1)")
p2
garch11_fit <- rstan::extract(garch11, permuted = TRUE)
mean(garch11_fit$mu)
# Compute the loglikelihood
log_lik <- extract_log_lik(garch11, merge_chains = FALSE)
r_eff <- relative_eff(exp(log_lik), cores = 2)
loo <- loo(log_lik, r_eff = r_eff, cores = 2)
print(loo)
plot(loo, main = "GARCH PSIS diagnostic plot")
|
cce30da1726ce79e12a8071b14d1336ed9eafb45 | 7c32bd1a1ea4b9a9bab53dcd206e9154206e7bab | /samples/R_Models/LogisticReg_Rmodel/training.R | e09007a64eaa9d9b28bb13e47e141645067d10df | [
"Apache-2.0"
] | permissive | paataugrekhelidze/model-management-resources | 55d92159fb5ffd97460c44f0d495cdb8308d96da | e3cc8719f349f9755690a4cf87f7e75574966e9c | refs/heads/main | 2023-08-21T11:56:30.560413 | 2021-09-23T18:02:59 | 2021-09-23T18:02:59 | 424,690,327 | 0 | 0 | Apache-2.0 | 2021-11-04T17:57:08 | 2021-11-04T17:57:08 | null | UTF-8 | R | false | false | 622 | r | training.R | # Copyright (c) 2020, SAS Institute Inc., Cary, NC, USA. All Rights Reserved.
# SPDX-License-Identifier: Apache-2.0
inputdata <- read.csv(file="hmeq_train.csv", header=TRUE, sep=",")
attach(inputdata)
# -----------------------------------------------
# FIT THE LOGISTIC MODEL
# -----------------------------------------------
reg<- glm(BAD ~ VALUE + factor(REASON) + factor(JOB) + DEROG + CLAGE + NINQ + CLNO , family=binomial)
# -----------------------------------------------
# SAVE THE OUTPUT PARAMETER ESTIMATE TO LOCAL FILE OUTMODEL.RDA
# -----------------------------------------------
save(reg, file="reg.rda")
|
473c2e4ddd8a5eda52aa13b7c5dd97e6401b60c7 | 4487f71ef15b6712e60cc28a6e6e4918abf612fa | /Popgen_HW1.R | 2994020562d6e41de964c2af4352d54ccc0e6aaa | [] | no_license | maccwinter/Genanalyse | 8cf3d5e4c370da0489b33f39d8920f7016a5e014 | 757073491f7e062ecda1c3b009dc2924ef4e163a | refs/heads/master | 2020-07-31T03:54:31.650358 | 2019-09-25T18:03:13 | 2019-09-25T18:03:13 | 210,476,455 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 5,498 | r | Popgen_HW1.R | #Popgen HW 1
#1B
#The genotype frequences can be represented as by the following, where the allele frequencies of S, I and G are represented by fs, fi, and fg, repectively.
#GSS = fs^2
#GFF =ff^2
#GII = fi^2
#GSF = 2fs*ff
#GSI = 2fs*fi
#GFI = 2ff*fi
#1B
#tot represents the total population number
tot <- 141 + 111 + 15 + 28 + 32 + 5
tot
#The population total is 332 individuals
#allelefreq is a function to calculate the allele frequencies where x represents homozygotes where y and z represent the two different heterozygotes.
allelefreq <- function(x,y,z) {(x + (y + z)/2)/tot}
#fs represents the allele frequency of S
fs <- allelefreq(141,111,15)
fs
#fs is 0.6144578
#ff is the allele frequency for the F allele
ff <- allelefreq(28,32,111)
ff
#ff is 0.2996988
#fi is the allele frequency of I
fi <- allelefreq(5,32,15)
fi
#fi is 0.08584337
#fs, fi, and ff need to add to 1
fs + fi + ff
#and they do :)
#And now to calculate genotype frequencies.
GSS = fs^2
GSS
#GSS is 0.3775584
GFF =ff^2
GFF
#GFF is 0.08981937
GII = fi^2
GII
#GII is 0.007369085
GSF = 2*fs*ff
GSF
#GSF is 0.3683045
GSI = 2*fs*fi
GSI
#GSI is 0.1054943
GFI = 2*ff*fi
GFI
#GFI is 0.05145431
#These genotype frequences should add to 1
GSS + GFF + GII + GFI +GSF +GSI
#AND THEY DO!!
#ge is the expected genotype totals of all individuals in the population
ge <- tot*c(GSS,GFF,GII,GSF,GSI,GFI)
ge
#egenotypenames is an object I'm going to use just to symbolically represent each expected genotype.
egenotypenames <- c("SS","FF","II","SF","SI","FI")
egenotypenames
egnames <- as.list(egenotypenames)
#Now I am going to represent the total amount of individuals per genotype with the following array: expectedgenotypes
expectedgenotypes <- data.frame(ge, row.names = egnames)
expectedgenotypes
# Expected Genotypes
#SS 125.349398
#FF 29.820030
#II 2.446536
#SF 122.277108
#SI 35.024096
#FI 17.082831
#go is the observed genotypes
go <- c(141,28,5,111,15,32)
go
#dif is the difference between and expected genotypes
dif <- go-ge
dif
#dif 2 is the square values of the differences between observed and expected genotypes
dif2 <- dif^2
#prechi is the division for genotypes of (observed - expected) by the corresponding expected values
prechi <- dif2/ge
prechi
#chi2 is the sum of all the values in prechi (aka the chi squared value)
chi2 <- sum(prechi)
chi2
#The chi squared value is 30.24456
#There are 2 non-independent variables and 6 possible genotypes. So k, the degrees of freedome is:
k <- 6-1-2
k
#There are 3 degrees of freedom.
#This chi squared value at a k of 3 statistically deviates from Hardy Weinberg.
#Question 2
#homo represents the frequency of homozygotes (based on the Hardy-Weinberg model) and hete represents the heterozygote frequency
homo<- function(p,q){(p^2)+(q^2)}
hete <- function(p,q){2*p*q}
homop0 <- homo(0,1)
homop0
homop0.25 <- homo(0.25,0.75)
homop0.25
homop0.5 <- homo(0.5,0.5)
homop0.5
homop0.75 <- homo(0.75,0.25)
homop0.75
homop1 <- homo(1,0)
homozygotes <-c(homop0,homop0.25,homop0.5,homop0.75,homop1)
homozygotes
heteq1 <- hete(0,1)
heteq0.75 <- hete(0.25,0.75)
heteq0.5 <- hete(0.5,0.5)
heteq0.25 <- hete(0.75,0.25)
heteq0 <- hete(1,0)
heterozygotes <- c(heteq0,heteq0.25,heteq0.5,heteq0.75,heteq1)
heterozygotes
pfreq <- list("0","0.25","0.5","0.75","1")
pfreq
genotype_frequencies <- data.frame(homozygotes, heterozygotes, row.names = pfreq)
genotype_frequencies
#p homozygotes heterozygotes
#0 1.000 0.000
#0.25 0.625 0.375
#0.5 0.500 0.500
#0.75 0.625 0.375
#1 1.000 0.000
#Heterozygote frequencies are maximized when p (and also q) is 0.5.
#Question 3
#D = gOD-pOpD where D is the non-d allele.
#gOD = 0.1 +(0.67)(0.4)
gOD <- 0.1 +0.67*0.4
gOD
#gOD = 0.368. O- individuals are homozygotes for O and the non-d allele (D). And gOD is 0.368 so:
Omindividuals <- gOD^2
Omindividuals
#There is a frequency of 0.135424 O- individuals in the population.
#Question 4A
# RF = (sum of recombinant progeny/total progeny)*100
RF <- ((7+12)/(63+7+12+58))*100
RF
#RF = 13.57143
#r = 0.13
r<-0.13
#4B
#Dt = Do((1-r)^t)
#Algebra --- t = ln(Dt/Do)/ln(1-r)
#t = ln(0.05/0.23)/ln(1-r)
t = log(0.05/0.23)/log(1-r)
t
#After 10.95816 generations of random mating, so 11 generations, D degrades to 0.05.
#Problem 5A
# pi = (#differences/#combinations)
pi <- (3+2+2+1+3+3+2+3+3)/10
pi
#pi = 2.2
#5B is written out on the paper. I expect pi to increase in subsequent generations after inbreeding.
#pi should increase faster for an inbreeding population, while it decreases for an outbreeding population.
#Problem 6
# f = 1 - (observed het/expected het)
# expected heterozygosity = 2 pq
q <- function(p){1-p}
expectedhet<- function(p){2*p*q(p)}
expectedhet(0.1)
#observed heterozygosity (GAa): observedhet = expectedhet(1-f).
GAa <- function(p,f){expectedhet(p)*(1-f)}
GAa(0.1,0.5)
#There are two types of homozygotes GAA and Gaa.
#p = GAA + 0.5*GAa
#q = GAa + 0.5*GAa
#GAA = p - 0.5*GAa
#Gaa = q - 0.5*GAa
GAA <- function(p,f){p -0.5*GAa(p,f)}
Gaa <- function(p,f){q(p)-0.5*GAa(p,f)}
gametestot <- function(p,f){GAA(p,f)+GAa(p,f)+Gaa(p,f)}
#6a
GAa(0.1,0.5)
#GAa = 0.09
GAA(0.1,0.5)
#GAA = 0.055
Gaa(0.1,0.5)
#Gaa = 0.855
gametestot(0.1,0.5)
#They add to 1!
#6b
GAa(0.3,0.02)
#GAa = 0.4116
GAA(0.3,0.02)
#GAA = 0.0942
Gaa(0.3,0.02)
#Gaa = 0.4942
gametestot(0.3,0.02)
#Adds to 1
#6c
GAa(0.5,-0.3)
#GAa = 0.65
GAA(0.5,-0.3)
#GAA = 0.175
Gaa(0.5,-0.3)
#Gaa = 0.175
gametestot(0.5,-0.3)
#Adds to 1
|
029876e4c43604c12ff1493065a5d9dac214c441 | 38e6bf92a54267ad564bcfc2550f49d807b11686 | /src/QC/2_proteinGroups_QC.R | 4e5743b081b061f95b613b6fab46638f538e920a | [] | no_license | JoWatson2011/APEX2_Analysis_Watson_Ferguson_2022 | 997972a98a1ff4b4d35f7bc0a6cd4bad1566bc89 | 1ffb39f59a7f57a14ed08164742fc2bccb5a3399 | refs/heads/master | 2023-04-10T19:58:05.692462 | 2022-10-14T09:52:03 | 2022-10-14T09:52:03 | 490,328,742 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 9,314 | r | 2_proteinGroups_QC.R | library(dplyr)
library(tidyr)
library(gplots)
library(data.table)
library(patchwork)
# library(readr)
library(RColorBrewer)
library(ggplot2)
# Read proteinGroups
proteinGroups <- readRDS("data/proteinGroups.RDS")
experiment_cols <- grep("LFQ intensity .*_P_",
colnames(proteinGroups),
value = T)
# Convert Intensity columns to 'numeric'
proteinGroups[,experiment_cols] <- apply(proteinGroups[,experiment_cols],
2,
as.numeric)
# Summary of variables proteinGroups will be filtered on.
# After running the script this information will be stored in the
# variable named report.
report <- vector()
report["Total Proteins Identified"] <- nrow(proteinGroups)
report["Potential Contaminants"] <- sum(proteinGroups$`Potential contaminant` == "+")
report["Reverse"] <- sum(proteinGroups$Reverse == "+")
report["Only Identified by Site"] <- sum(proteinGroups$`Only identified by site` == "+")
report["No Quantitative Data (incomplete cases)"] <- proteinGroups %>%
select(grep("LFQ intensity" , experiment_cols, value = T)) %>%
filter(rowSums(. == 0) == ncol(.)) %>% nrow()
# Filter proteinGroups.txt to remove contaminants, reverse.
proteinGroups_flt <- proteinGroups %>%
filter(`Potential contaminant` != "+",
`Reverse` != "+",
`Only identified by site` != "+",
`Razor + unique peptides` > 1,
`Unique + razor sequence coverage [%]` >= 5
)
# Add info to report
report["Proteins remaining following filtering"] <- nrow(proteinGroups_flt)
report["Incomplete cases in filtered dataset"] <- proteinGroups_flt %>%
select(all_of(experiment_cols)) %>%
filter(rowSums(. == 0) == ncol(.)) %>%
nrow()
# Filter rows with no quantitative information
proteinGroups_flt_cc <- proteinGroups_flt[rowSums(proteinGroups_flt[experiment_cols] > 0) >= 1,]
#Collapse gene/protein names
proteinGroups_flt_cc <- proteinGroups_flt_cc %>%
mutate(`Gene names` = gsub(";.*", "", `Gene names`),
`Majority protein IDs` = gsub(";.*", "", `Majority protein IDs`),
`Protein names` = gsub(";.*", "", `Protein names`))
# Transform
proteinGroups_log <- apply(proteinGroups_flt_cc[,experiment_cols], 2, function(i){
tmp <- ifelse(i == 0, NA, i)
tmp2 <- log10(tmp)
# tmp3 <- limma::normalizeBetweenArrays(tmp2, "cyclicloess")
return(tmp2)
})
# Normalise
proteinGroups_norm <- cbind(
proteinGroups_flt_cc[,!(names(proteinGroups_flt_cc) %in% c(experiment_cols))],
limma::normalizeBetweenArrays(proteinGroups_log, "quantile")
)
# Visualise normalilsation
proIntUnnorm <- proteinGroups_flt_cc %>%
select(`id`, all_of(experiment_cols)) %>%
pivot_longer(cols = -c(`id`),
names_to = "experiment",
values_to = "intensity") %>%
mutate(
intensity = ifelse(intensity == 0, NA, intensity),
intensity = log10(intensity),
rep = substr(experiment,
nchar(experiment)-1,
nchar(experiment))) %>%
ggplot(aes(x = intensity, color = rep, group = experiment)) +
geom_density() +
theme(legend.position = 'none')
proIntNorm <- proteinGroups_norm %>%
select(`id`, all_of(experiment_cols)) %>%
pivot_longer(cols = -c(`id`),
names_to = "experiment",
values_to = "intensity") %>%
mutate(rep = substr(experiment,
nchar(experiment)-1,
nchar(experiment))) %>%
filter(intensity < 10) %>%
ggplot(aes(x = intensity, color = rep, group = experiment)) +
geom_density() +
theme(legend.position = 'none') +
ggtitle("", subtitle = "Normalised Intensities")
proIntUnnorm / proIntNorm
ggsave("results/figs/QC/proNormCurve.tiff",
proIntUnnorm / proIntNorm,
width = 210, height = 297, units = "mm")
saveRDS(proteinGroups_norm, "data/proteinGroups_Flt.rds")
write.csv(proteinGroups_norm, "data/proteinGroups_Flt.csv")
#### FIGURES
# Pie chart to visualise propoprtion of missing values across all experiments.
pie(table(proteinGroups_flt_cc[,experiment_cols] == 0 |
is.na(proteinGroups_flt_cc[,experiment_cols])),
labels = c("Not NA", "NA"),
main = "Missing Ratios in Filtered Data")
#Heatmap correlation
cors <- proteinGroups %>%
select(c(grep("_A_", colnames(.)),
grep("_T_", colnames(.)))
) %>%
cor(use = "pairwise.complete.obs")
rownames(cors) <- colnames(cors) <- gsub("LFQ intensity ",
"",
colnames(cors))
cor_hm <- cors %>%
as.data.frame() %>%
tibble::rownames_to_column("Exp1") %>%
pivot_longer(-Exp1, names_to = "Exp2") %>%
# mutate(Exp1 = factor(Exp1, levels = rownames(cors)),
# Exp2, factor(Exp2, levels = colnames(cors))
# ) %>%
ggplot(aes(x = Exp1, y = Exp2, fill = value)) +
geom_tile() +
geom_text(aes(label =round(value,2)),size = 2) +
theme(axis.text.x = element_text(size = 5, angle = 45, hjust = 1),
axis.text.y = element_text(size = 5, angle = 45),
axis.title = element_blank(),
legend.key.size = unit(.25, "cm"),
legend.margin = margin(0,0,0,0, "cm"),
plot.margin = margin(0,0,0,0, "cm"),
legend.title = element_text(size = 5),
legend.text = element_text(size = 5)) +
scale_x_discrete(limits=rownames(cors)) +
scale_y_discrete(limits=colnames(cors)) +
scale_fill_gradientn(colours = c("#2C7BB6",
"#ABD9E9",
"#FFFFBF",
"#FDAE61",
"#D7191C"),
limits = c(0,1),
breaks = seq(0,1, 0.2)
)
ggsave("results/figs/forPaper/gg_PRO_CorHM.pdf", cor_hm)
### PCA
pca <- prcomp(t(na.omit(proteinGroups_norm[ , experiment_cols])
)
)
pca_df <- as.data.frame(pca$x)
pca_df$run <- gsub("LFQ intensity ", "",
gsub("_P", "", rownames(pca_df)
)
) %>%
substr(0, nchar(.) - 3)
# pca_df$rep <- gsub("LFQ intensity ", "", rownames(pca_df)) %>%
# substr(nchar(.) - 1, nchar(.))
pca_df$bait <- ifelse(grepl("R2A", pca_df$run), "FGFR2",
ifelse(grepl("R11A", pca_df$run), "RAB11", "GFP")
)
eigs <- pca$sdev^2
pc1<-signif(100*(eigs[1] / sum(eigs)), 4)
pc2<-signif(100*(eigs[2] / sum(eigs)), 4)
pca_g <- ggplot(pca_df, aes(PC1, PC2, color= run, shape = bait)) +
#ggforce::geom_mark_ellipse(aes(group = run, fill = bait), linetype = 0, alpha = 0.2) +
geom_point(size = 3, alpha = 0.7)+
scale_x_continuous(paste("PC1 (", pc1, "%)", sep=""))+
scale_y_continuous(paste("PC2 (", pc2, "%)", sep=""))+
scale_color_brewer("Run", palette = "Paired") +
guides(color = "none", shape = "none") +
#scale_fill_discrete(type = c("#D73027", "#4575B4", "#838B8B")) +
theme(
plot.title = element_text(size=20, face="bold",hjust = 0.5),
panel.background = element_blank(),
panel.grid = element_line("grey", linetype="dashed"),
legend.key = element_blank(),
axis.line = element_line("black"),
axis.text = element_text(size=6),
axis.title = element_text(size=6, face="bold")
)
ggsave("results/figs/forPaper/proPCA.pdf", pca_g,
width = 60, height = 60, units = "mm", dpi = 300)
###
# Number identified in
# each experiment
###
proNo <- proteinGroups_flt_cc %>%
select(`id`, all_of(experiment_cols)) %>%
pivot_longer(cols = -`id`,
names_to = "experiment",
values_to = "intensity") %>%
filter(intensity != 0) %>%
mutate(experiment = gsub("LFQ intensity ", "", experiment)) %>%
mutate(rep = substr(experiment, nchar(experiment) - 2, nchar(experiment)),
experiment = substr(experiment, 0, nchar(experiment) - 3)
) %>%
group_by(experiment,rep) %>%
summarise(n = n(), .groups = "keep") %>%
group_by(experiment) %>%
summarise(sd = sd(n, na.rm = T),
n = mean(n), .groups = "keep") %>%
ungroup() %>%
unique() %>%
mutate(apex = ifelse(grepl("_A_", experiment),
"APEX", "Total"),
experiment = gsub("_[A|T]", "", experiment)) %>%
ggplot(aes(x = experiment, y = n, fill = experiment,
ymax = n+sd, ymin = n-sd
)) +
geom_col() +
facet_wrap(~ apex, ncol = 1) +
scale_fill_discrete(type = c("#838B8B",
"#D73027",
"#838B8B",
"#ABD9E9",
"#4575B4")
) +
geom_errorbar(width = 0.5) +
theme(
panel.background = element_blank(),
panel.grid = element_line("grey", linetype="dashed"),
legend.key = element_blank(),
axis.line = element_line("black"),
# Text sizes may need modifying based on fig. sizes
axis.text = element_text(size=12),
axis.title = element_text(size=14, face="bold"),
legend.position = "none",
axis.text.x = element_text(angle = 45, hjust = 1)
) +
ggtitle("Quantified Protein Groups")
ggsave("results/figs/forPaper/proNumQuant.pdf", proNo,
width = 7, height = 7)
|
b7e1117a806ad701ebde8e552a73573769a5ea2b | 8c2253bd47fd3d76f28950d1ef24450b24c4a0d7 | /R/extract_timeseries_annual_landings.R | 3cf07b590b7a7c0115215fcba4003ec7175a9a33 | [] | no_license | cran/StrathE2E2 | bc63d4f0dffdde94da1c7ea41133c09033c0cd4e | 629dc5e7f2e323752349352bb2d651a56c6f4447 | refs/heads/master | 2023-02-25T13:18:59.217896 | 2021-01-22T21:40:05 | 2021-01-22T21:40:05 | 278,343,976 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 7,487 | r | extract_timeseries_annual_landings.R | #
# extract_timeseries_annual_landings.R
#
#' read designated model
#'
#' returns a model object with run and data slots
#'
#' @param model current model
#' @param build model build object
#' @param out model output
#'
#' @return inshore/offshore annual landings
#'
#' @noRd
#
# ------------------------------------------------------------------------------
extract_timeseries_annual_landings <- function(model, build, out) {
setup <- elt(model, "setup")
identifier <- elt(setup, "model.ident")
resultsdir <- elt(setup, "resultsdir")
run <- elt(build, "run")
nyears <- elt(run, "nyears")
#Print some of the full time series data to a csv file
#-----------------------------------------------------------------
offshore_annual_group_land_disc<-data.frame(year=seq(1,nyears))
offshore_annual_group_land_disc$PFland<-rep(0,nyears)
offshore_annual_group_land_disc$DFQland<-rep(0,nyears)
offshore_annual_group_land_disc$DFNQland<-rep(0,nyears)
offshore_annual_group_land_disc$MFland<-rep(0,nyears)
offshore_annual_group_land_disc$SBland<-rep(0,nyears)
offshore_annual_group_land_disc$CBland<-rep(0,nyears)
offshore_annual_group_land_disc$CZland<-rep(0,nyears)
offshore_annual_group_land_disc$BDland<-rep(0,nyears)
offshore_annual_group_land_disc$SLland<-rep(0,nyears)
offshore_annual_group_land_disc$CTland<-rep(0,nyears)
offshore_annual_group_land_disc$KPland<-rep(0,nyears)
offshore_annual_group_land_disc$PFdisc<-rep(0,nyears)
offshore_annual_group_land_disc$DFQdisc<-rep(0,nyears)
offshore_annual_group_land_disc$DFNQdisc<-rep(0,nyears)
offshore_annual_group_land_disc$MFdisc<-rep(0,nyears)
offshore_annual_group_land_disc$SBdisc<-rep(0,nyears)
offshore_annual_group_land_disc$CBdisc<-rep(0,nyears)
offshore_annual_group_land_disc$CZdisc<-rep(0,nyears)
offshore_annual_group_land_disc$BDdisc<-rep(0,nyears)
offshore_annual_group_land_disc$SLdisc<-rep(0,nyears)
offshore_annual_group_land_disc$CTdisc<-rep(0,nyears)
offshore_annual_group_land_disc$KPdisc<-rep(0,nyears)
inshore_annual_group_land_disc <- offshore_annual_group_land_disc
for(ik in 1:nyears){
offshore_annual_group_land_disc$PFland[ik] <- out$landp_o[ (1+(ik*360)) ] - out$landp_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$DFQland[ik] <- out$landd_quota_o[ (1+(ik*360)) ] - out$landd_quota_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$DFNQland[ik] <- out$landd_nonquota_o[ (1+(ik*360)) ] - out$landd_nonquota_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$MFland[ik] <- out$landm_o[ (1+(ik*360)) ] - out$landm_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$SBland[ik] <- out$landsb_o[ (1+(ik*360)) ] - out$landsb_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$CBland[ik] <- out$landcb_o[ (1+(ik*360)) ] - out$landcb_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$CZland[ik] <- out$landcz_o[ (1+(ik*360)) ] - out$landcz_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$BDland[ik] <- out$landbd_o[ (1+(ik*360)) ] - out$landbd_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$SLland[ik] <- out$landsl_o[ (1+(ik*360)) ] - out$landsl_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$CTland[ik] <- out$landct_o[ (1+(ik*360)) ] - out$landct_o[ (1+(ik-1)*360) ]
#No offshore landings of kelp
offshore_annual_group_land_disc$PFdisc[ik] <- out$discpel_o[ (1+(ik*360)) ] - out$discpel_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$DFQdisc[ik] <- out$discdem_quota_o[ (1+(ik*360)) ] - out$discdem_quota_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$DFNQdisc[ik] <- out$discdem_nonquota_o[ (1+(ik*360)) ] - out$discdem_nonquota_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$MFdisc[ik] <- out$discmig_o[ (1+(ik*360)) ] - out$discmig_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$SBdisc[ik] <- out$discsb_o[ (1+(ik*360)) ] - out$discsb_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$CBdisc[ik] <- out$disccb_o[ (1+(ik*360)) ] - out$disccb_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$CZdisc[ik] <- out$disccz_o[ (1+(ik*360)) ] - out$disccz_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$BDdisc[ik] <- out$discbd_o[ (1+(ik*360)) ] - out$discbd_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$SLdisc[ik] <- out$discsl_o[ (1+(ik*360)) ] - out$discsl_o[ (1+(ik-1)*360) ]
offshore_annual_group_land_disc$CTdisc[ik] <- out$discct_o[ (1+(ik*360)) ] - out$discct_o[ (1+(ik-1)*360) ]
#No offshore discards of kelp
inshore_annual_group_land_disc$PFland[ik] <- out$landp_i[ (1+(ik*360)) ] - out$landp_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$DFQland[ik] <- out$landd_quota_i[ (1+(ik*360)) ] - out$landd_quota_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$DFNQland[ik] <- out$landd_nonquota_i[ (1+(ik*360)) ] - out$landd_nonquota_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$MFland[ik] <- out$landm_i[ (1+(ik*360)) ] - out$landm_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$SBland[ik] <- out$landsb_i[ (1+(ik*360)) ] - out$landsb_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$CBland[ik] <- out$landcb_i[ (1+(ik*360)) ] - out$landcb_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$CZland[ik] <- out$landcz_i[ (1+(ik*360)) ] - out$landcz_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$BDland[ik] <- out$landbd_i[ (1+(ik*360)) ] - out$landbd_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$SLland[ik] <- out$landsl_i[ (1+(ik*360)) ] - out$landsl_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$CTland[ik] <- out$landct_i[ (1+(ik*360)) ] - out$landct_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$KPland[ik] <- out$landkp_i[ (1+(ik*360)) ] - out$landkp_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$PFdisc[ik] <- out$discpel_i[ (1+(ik*360)) ] - out$discpel_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$DFQdisc[ik] <- out$discdem_quota_i[ (1+(ik*360)) ] - out$discdem_quota_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$DFNQdisc[ik] <- out$discdem_nonquota_i[ (1+(ik*360)) ] - out$discdem_nonquota_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$MFdisc[ik] <- out$discmig_i[ (1+(ik*360)) ] - out$discmig_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$SBdisc[ik] <- out$discsb_i[ (1+(ik*360)) ] - out$discsb_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$CBdisc[ik] <- out$disccb_i[ (1+(ik*360)) ] - out$disccb_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$CZdisc[ik] <- out$disccz_i[ (1+(ik*360)) ] - out$disccz_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$BDdisc[ik] <- out$discbd_i[ (1+(ik*360)) ] - out$discbd_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$SLdisc[ik] <- out$discsl_i[ (1+(ik*360)) ] - out$discsl_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$CTdisc[ik] <- out$discct_i[ (1+(ik*360)) ] - out$discct_i[ (1+(ik-1)*360) ]
inshore_annual_group_land_disc$KPdisc[ik] <- out$disckp_i[ (1+(ik*360)) ] - out$disckp_i[ (1+(ik-1)*360) ]
}
filename = csvname(resultsdir, "model_inshore_annual_landings_discards", identifier)
writecsv(inshore_annual_group_land_disc, filename, row.names=FALSE)
filename = csvname(resultsdir, "model_offshore_annual_landings_discards", identifier)
writecsv(offshore_annual_group_land_disc, filename, row.names=FALSE)
list(
offshore_annual_group_land_disc = offshore_annual_group_land_disc,
inshore_annual_group_land_disc = inshore_annual_group_land_disc
)
}
|
c484eb8d4c0cfa39aa4c9cd14eed0f904cb5df74 | 4ce2d115fc47d9ae734d2bbb54382cdcc820a658 | /BuildComponentLambMortRateCovs/BuildComponentLambMortRateCovs.R | 0f9cc3c85dcad391ab1de5aaba048c232519c962 | [] | no_license | kmanlove/ClusterAssocDataPrep | 051765049c1975b51ea3c7ed2021249f0bcb3a60 | 236d0c7512a8b924e8986a20e5e690af72c0fd0c | refs/heads/master | 2016-09-06T16:17:42.915787 | 2013-10-08T14:05:33 | 2013-10-08T14:05:33 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 4,050 | r | BuildComponentLambMortRateCovs.R | #-- this script reads in the same data as the variance decomposition --#
#-- coxme models --#
filepath <-
"~/work/Kezia/Research/EcologyPapers/ClustersAssociations/Data/RevisedData_11Sept2013/"
relocdata <- read.csv(paste(filepath,
"RelocsWithNetworkMeasures/FullEweDataAllSummerRelocs_MinEdge.1_18Sept2013.csv", sep = ""), header = T)
relocdata$component.ind <- paste(relocdata$Pop, "_", relocdata$Year, "_",
relocdata$res.component, sep = "")
relocdata$popear.ind <- paste(relocdata$Pop, "_", relocdata$Year, sep = "")
lambdata <- read.csv(paste(filepath,
"CleanLambSurvData/FullEweRelocsLambSurvDat_MinEdge.1_allewerelocs_18Sept2013.csv",
sep = ""), header = T)
lambdata$LastYearKnownCompMort <- lambdata$LastYearLambStatus <-
lambdata$LastYearCompLambDiedOrNoLamb <- lambdata$ThisYearCompKnownMort <-
lambdata$ThisYearCompLambDiedOrNoLamb <-
lambdata$ThisPopyrLambDiedOrNoLamb <- lambdata$ThisPopyrKnownMort <- rep(NA, dim(lambdata)[1])
for(i in 1:dim(lambdata)[1]){
ewedat <- subset(relocdata, as.character(EWEID) ==
as.character(lambdata$EWEID)[i] &
as.numeric(as.character(Year)) ==
as.numeric(as.character(lambdata$Year[i])) -
1)[1, ]
EwesLastYearComponent <- as.character(ewedat$component.ind)
if(length(EwesLastYearComponent) == 0){
lambdata$LastYearKnownCompMort[i] <- lambdata$LastYearLambStatus[i] <-
lambdata$LastYearCompLambDiedOrNoLamb[i] <- NA
} else {
LastYearCompEwes <- levels(factor(subset(relocdata,
as.character(component.ind) ==
as.character(EwesLastYearComponent))$EWEID))
LastYearCompLambs <- subset(lambdata, as.numeric(as.character(Year)) == as.numeric(as.character(lambdata$Year[i])) - 1 & as.character(EWEID) %in% LastYearCompEwes)
lambdata$LastYearKnownCompMort[i] <- sum(LastYearCompLambs$CENSOR2) /
dim(LastYearCompLambs)[1]
lambdata$LastYearCompLambDiedOrNoLamb[i] <- sum(LastYearCompLambs$CENSOR2) /
length(LastYearCompEwes)
lambdata$LastYearLambStatus[i] <- ifelse(as.character(ewedat$HasLamb) ==
"NoLamb", "NoLamb",
ifelse(ewedat$CENSOR2 == 0,
"LambDied",
ifelse(ewedat$CENSOR2 == 1,
"LambSurvived", NA)))
}
ewedatnow <- subset(relocdata, as.character(EWEID) ==
as.character(lambdata$EWEID)[i] &
as.numeric(as.character(Year)) ==
as.numeric(as.character(lambdata$Year[i])))[1, ]
EwesThisYearComponent <- as.character(ewedatnow$component.ind)
ThisYearCompEwes <- levels(factor(subset(relocdata,
as.character(component.ind) ==
as.character(EwesThisYearComponent))$EWEID))
ThisYearCompLambs <- subset(lambdata, as.numeric(as.character(Year)) ==
as.numeric(as.character(lambdata$Year[i])) &
as.character(EWEID) %in% ThisYearCompEwes)
lambdata$ThisYearKnownCompMort[i] <- sum(ThisYearCompLambs$CENSOR2) /
dim(ThisYearCompLambs)[1]
lambdata$ThisYearCompLambDiedOrNoLamb[i] <- sum(ThisYearCompLambs$CENSOR2) /
length(ThisYearCompEwes)
#-- extract mort levels for this popyear --#
EwesThisPopyr <- as.character(ewedatnow$popear.ind)
ThisPopyrEwes <- levels(factor(subset(relocdata,
as.character(popear.ind) ==
as.character(EwesThisPopyr))$EWEID))
ThisPopyrLambs <- subset(lambdata, as.numeric(as.character(Year)) ==
as.numeric(as.character(lambdata$Year[i])) &
as.character(EWEID) %in% ThisPopyrEwes)
lambdata$ThisPopyrKnownMort[i] <- sum(ThisPopyrLambs$CENSOR2) /
dim(ThisPopyrLambs)[1]
lambdata$ThisPopyrLambDiedOrNoLamb[i] <- sum(ThisPopyrLambs$CENSOR2) /
length(ThisPopyrEwes)
}
write.path <- "~/work/Kezia/Research/EcologyPapers/ClustersAssociations/Data/RevisedData_11Sept2013/LambSurvDatWithLastYearCompCovs/"
write.csv(lambdata, paste(write.path,
"LambDataWithLastYearCompCovs_19Sept2013.csv", sep =
""))
|
92c314d901aa115f62b732ba0a5207c4460f1d33 | acf25199f5311f05b2d3a5119fe2a2e06fb82901 | /analysis.R | 6dca4128a89724d6a24d772d9f48c69aec95f017 | [] | no_license | AndrMenezes/mm2017 | 5dc6e9a6ac1c983b1e570a11fdb4144a446b8beb | 2756e0a2171ca943cc78ab771c63832727e2892c | refs/heads/master | 2021-04-15T09:28:56.306028 | 2019-06-25T22:11:27 | 2019-06-25T22:11:27 | 126,765,309 | 0 | 0 | null | null | null | null | ISO-8859-1 | R | false | false | 6,409 | r | analysis.R | # Definições gerais -------------------------------------------------------
setwd('C:/Users/User/Dropbox/4° Série/Modelos Mistos/Trabalho')
# setwd('C:/Users/André Felipe/Dropbox/4° Série/Modelos Mistos/Trabalho')
rm(list = ls(all.names = TRUE))
bib <- c('lme4', 'lmerTest', 'lsmeans', 'hnp', 'dplyr', 'ggplot2', 'RLRsim', 'nlme', 'xtable', 'influence.ME')
sapply(bib, require, character.only = T)
dados <- read.table(file = 'planta-final.txt', sep = ',', header = T)
dados$tempo_f <- factor(dados$tempo_f)
dados$arvore <- factor(dados$arvore)
head(dados)
str(dados)
# setwd('C:/Users/André Felipe/Dropbox/4° Série/Modelos Mistos/Trabalho/Relatório')
setwd('C:/Users/User/Dropbox/4° Série/Modelos Mistos/Trabalho/Relatório')
# Descritiva --------------------------------------------------------------
pdf(file = "boxplot-trat.pdf", width = 10.5, height = 6.5)
par(mar = c(3.2, 3.2, 1.5, 1.5), cex = 1.6)
boxplot(diametro ~ trat, data = dados, xlab = '', ylab = '', cex = 0.6, col = 'gray')
mtext("Dose", side = 1, line = 2.0, cex = 1.8)
mtext("Diâmetro (mm)", side = 2, line = 2, cex = 1.8)
graphics.off()
pdf(file = "boxplot-tempo.pdf", width = 10.5, height = 6.5)
par(mar = c(3.2, 3.2, 1.5, 1.5), cex = 1.6)
boxplot(diametro ~ tempo_f, data = dados, xlab = '', ylab = '', cex = 0.6, col = 'gray')
points(x = unique(dados$tempo_f), y = tapply(dados$diametro, dados$tempo_f, mean), pch = 16, col = 'red', cex = 0.8)
mtext("Dias após a avaliação", side = 1, line = 2.0, cex = 1.8)
mtext("Diâmetro (mm)", side = 2, line = 2, cex = 1.8)
graphics.off()
x11()
dados %>% ggplot(aes(x = tempo_f, y = diametro, group = interaction(tempo_f, trat))) +
geom_boxplot(aes(fill = factor(trat)), color = 'black') +
stat_summary(aes(group = 1), fun.y = mean, geom="line", color = 'black') +
stat_summary(aes(group = 1), fun.y = mean, geom="point", color = 'gold') +
labs(y = 'Diâmetro (mm)', x = 'Dias após a avaliação', fill = 'Dose: ') +
theme_bw() +
theme(text = element_text(size=20), panel.grid.minor = element_blank(), legend.position="top",
panel.grid.major = element_line(size = 0.4, linetype = 'dotted', colour = 'gray'))
ggsave(filename = 'boxplot-tempo-trat.pdf', width = 9, height = 6)
# Ajuste do modelo --------------------------------------------------------
mod1 <- lme(fixed = diametro ~ trat + tempo_n + trat*tempo_n, data = dados, random = ~ 1 | arvore)
mod2 <- lme(fixed = diametro ~ trat + tempo_n, data = dados, random = ~ 1 | arvore)
mod3 <- lme(fixed = diametro ~ tempo_n, data = dados, random = ~ 1 | arvore)
anova(mod1, mod2, mod3)
2 * (logLik(mod2) - logLik(mod1))
print(xtable(anova(mod2), digits = 4))
# Comparações das estrutura de correlação ---------------------------------
m.CS <- lme(fixed = diametro ~ trat + tempo_n + trat * tempo_n, data = dados, random = ~ 1 | arvore,
correlation = corCompSymm(form = ~tempo_n|arvore))
m.Exp <- lme(fixed = diametro ~ trat + tempo_n + trat * tempo_n, data = dados, random = ~ 1 | arvore,
correlation = corExp(form = ~tempo_n|arvore, nugget = T))
m.Gaus <- lme(fixed = diametro ~ trat + tempo_n + trat * tempo_n, data = dados, random = ~ 1 | arvore,
correlation = corGaus(form = ~tempo_n|arvore, nugget = T))
anova(m.CS, m.Exp)
anova(m.CS, m.Gaus)
anova(m.Exp, m.Gaus)
# Modelo escolhido --------------------------------------------------------
mod3 <- lme(fixed = diametro ~ tempo_n, data = dados, random = ~ 1 | arvore)
summary(mod3)
intervals(mod3)
mod3 <- lmer(diametro ~ tempo_n + (1 | arvore), data = dados)
res <- summary(mod3)
round(as.data.frame(res$coefficients)[, -3], 4)
round(confint(mod3), 4)
rand(mod3)
# Resíduos ----------------------------------------------------------------
## Resíduos marginais (erro aleatório)
my.hnp <- hnp(mod3, halfnormal = T, how.many.out = T, paint.out = T, plot = T)
pdf(file = "hnp.pdf", width = 11, height = 7)
par(mar = c(3.5, 3.5, 1.2, 0.6), cex = 1.8)
plot(my.hnp, xaxt = 'n', yaxt = 'n', xlab = '', ylab = '', cex = 0.6, ylim = c(0, 10))
mtext("Percentil da N(0, 1)", side = 1, line = 2.0, cex = 1.8)
mtext("Resíduos marginais", side = 2, line =2, cex = 1.8)
abline(h = seq(0, 10, l = 5), v=seq(0, 3, l = 5), col = "gray", lty = "dotted")
axis(1, seq(0, 3, l = 5))
axis(2, seq(0, 10, l = 5), FF(seq(0, 10, l = 5), 1))
graphics.off()
## resíduos de efeitos aleatórios
r2 <- random.effects(mod3)$arvore
pdf(file = "qq-ranef.pdf", width = 11, height = 7)
par(mar = c(3.5, 3.5, 1.2, 0.6), cex = 1.8)
qqnorm(r2[, 1], xaxt = 'n', yaxt = 'n', xlab = '', ylab = '', cex = 0.6, main = ""); qqline(r2[, 1])
mtext("Percentil da N(0, 1)", side = 1, line = 2.0, cex = 1.8)
mtext("Resíduos de efeitos aleatórios", side = 2, line =2, cex = 1.8)
abline(h = seq(-1.5, 1.5, l = 5), v=seq(-2, 2, l = 5), col = "gray", lty = "dotted")
axis(2, seq(-1.5, 1.5, l = 5))
axis(1, seq(-2, 2, l = 5), FF(seq(-2, 2, l = 5), 1))
graphics.off()
## ajustado versus residuo
x = fitted(mod3); y = residuals(mod3); Rx = range(x); Ry = range(y)
pdf(file = "pred.pdf", width = 11, height = 7)
par(mar = c(3.5, 3.5, 1.2, 0.6), cex = 1.8)
plot(y ~ x, xlab = '', ylab = '', cex = 0.8, xaxt = 'n', yaxt = 'n')
mtext("Valores ajustados", side = 1, line = 2.0, cex = 1.8)
mtext("Resíduos marginais", side = 2, line =2, cex = 1.8)
abline(h = seq(Ry[1], Ry[2], l = 5), v=seq(Rx[1], Rx[2], l = 5), col = "gray", lty = "dotted")
axis(2, seq(Ry[1], Ry[2], l = 5), FF(seq(Ry[1], Ry[2], l = 5), 1))
axis(1, seq(Rx[1], Rx[2], l = 5), FF(seq(Rx[1], Rx[2], l = 5), 1))
graphics.off()
## influencia
lmer3.infl <- influence(mod3, obs=TRUE)
cook <- cooks.distance(lmer3.infl)
x = 1:nrow(dados); y = cooks.distance(lmer3.infl); Rx = range(x); Ry = range(y)
pdf(file = "cook.pdf", width = 11, height = 7)
par(mar = c(3.5, 3.5, 1.2, 0.6), cex = 1.8)
plot(y, xlab = '', ylab = '', cex = 0.8, xaxt = 'n', yaxt = 'n')
mtext("Índice das observações", side = 1, line = 2.0, cex = 1.8)
mtext("Distância de Cook", side = 2, line = 2, cex = 1.8)
abline(h = seq(Ry[1], Ry[2], l = 5), v=seq(Rx[1], Rx[2], l = 5), col = "gray", lty = "dotted")
axis(2, seq(Ry[1], Ry[2], l = 5), FF(seq(Ry[1], Ry[2], l = 5), 3))
axis(1, seq(Rx[1], Rx[2], l = 5), FF(seq(Rx[1], Rx[2], l = 5), 0))
graphics.off()
|
ad8c47cf5866a4e8ae0b9980486cf83f6692621f | 04a98a7e184fd449985628ac7b8a92f19c1785a4 | /R/clsd.R | 26dbaf3caa1bc277896344527c47f9a940da1738 | [] | no_license | JeffreyRacine/R-Package-crs | 3548a0002f136e9e7c1d5c808f6a3867b20b417e | 6112a3914e65f60a45c8bcfc4076e9b7ea1f8e7a | refs/heads/master | 2023-01-09T18:23:59.615927 | 2023-01-03T16:20:22 | 2023-01-03T16:20:22 | 1,941,853 | 12 | 6 | null | 2023-01-03T16:20:23 | 2011-06-23T14:11:06 | C++ | UTF-8 | R | false | false | 38,963 | r | clsd.R | ## These functions are for (currently univariate) logspline density
## estimation written by [email protected] (Jeffrey S. Racine). They
## make use of spline routines in the crs package (available on
## CRAN). The approach involves joint selection of the degree and
## knots in contrast to the typical approach (e.g. Kooperberg and
## Stone) that sets the degree to 3 and optimizes knots only. Though
## more computationally demanding, the estimators are more efficient
## on average.
par.init <- function(degree,segments,monotone,monotone.lb) {
## This function initializes parameters for search along with upper
## and lower bounds if appropriate.
dim.p <- degree+segments
## The weights for the linear tails must be non-positive. The lower
## bound places a maximum bound on how quickly the tails are allowed
## to die off. Trial and error suggests the values below seem to be
## appropriate for a wide range of (univariate)
## distributions. Kooperberg suggests that in order to get the
## constraint theta < 0 use theta <= -epsilon for some small epsilon
## > 0. We therefore use sqrt machine epsilon.
par.ub <- - sqrt(.Machine$double.eps)
par.lb <- monotone.lb
par.init <- c(runif(1,-10,par.ub),rnorm(dim.p-2),runif(1,-10,par.ub))
par.upper <- c(par.ub,rep(Inf,dim.p-2),par.ub)
par.lower <- if(monotone){rep(-Inf,dim.p)}else{c(par.lb,rep(-Inf,dim.p-2),par.lb)}
return(list(par.init=par.init,
par.upper=par.upper,
par.lower=par.lower))
}
gen.xnorm <- function(x=NULL,
xeval=NULL,
lbound=NULL,
ubound=NULL,
er=NULL,
n.integrate=NULL) {
er <- extendrange(x,f=er)
if(!is.null(lbound)) er[1] <- lbound
if(!is.null(ubound)) er[2] <- ubound
if(min(x) < er[1] | max(x) > er[2]) warning(" data extends beyond the range of `er'")
xint <- sort(as.numeric(c(seq(er[1],er[2],length=round(n.integrate/2)),
quantile(x,seq(sqrt(.Machine$double.eps),1-sqrt(.Machine$double.eps),length=round(n.integrate/2))))))
if(is.null(xeval)) {
xnorm <- c(x,xint)
} else {
xnorm <- c(xeval,x,xint)
if(min(xeval) < er[1] | max(xeval) > er[2]) warning(" evaluation data extends beyond the range of `er'")
}
## Either x will be the first 1:length(x) elements in
## object[rank.xnorm] or xeval will be the first 1:length(xeval)
## elements in object[rank.xnorm]
return(list(xnorm=xnorm[order(xnorm)],rank.xnorm=rank(xnorm)))
}
density.basis <- function(x=NULL,
xeval=NULL,
xnorm=xnorm,
degree=NULL,
segments=NULL,
basis="tensor",
knots="quantiles",
monotone=TRUE) {
## To obtain the constant of integration for B-spline bases, we need
## to compute log(integral exp(P%*%beta)) so we take an equally
## spaced extended range grid of length n plus the sample
## realizations (min and max of sample therefore present for what
## follows), and evaluation points xeval if they exist.
## Charles Kooperberg has a manuscript "Statistical Modeling with
## Spline Functions', Jan 5 2006 on his web page. Chapter 6, page
## 286, figure 6.7 reveals a hybrid spline basis that in essence
## appears to drop two columns from my B-spline with 2 segments
## added artificially. This has the effect of removing the two bases
## that were delivering weight in tails leading to `kinks'. Hat-tip
## to Charles for his clear descriptions. Note we require the same
## for the derivatives below. Note that this logspline basis does
## not have the B-spline property that the pointwise sum of the
## bases is 1 everywhere.
suppressWarnings(Pnorm <- prod.spline(x=x,
xeval=xnorm,
K=cbind(degree,segments+if(monotone){2}else{0}),
knots=knots,
basis=basis))
if(monotone) Pnorm <- Pnorm[,-c(2,degree+segments+1)]
## Compute the normalizing constant so that the estimate integrates
## to one. We append linear splines to the B-spline basis to
## generate exponentially declining tails (K=cbind(1,1) creates the
## linear basis).
suppressWarnings(P.lin <- prod.spline(x=x,
xeval=xnorm,
K=cbind(1,1),
knots=knots,
basis=basis))
## We append the linear basis to the left and rightmost polynomial
## bases. We match the slope of the linear basis to that of the
## polynomial basis at xmin/xmax (note that
## Pnorm[xnorm==max(x),-ncol(Pnorm)] <- 0 is there because the
## gsl.bspline values at the right endpoint are very small but not
## exactly zero but want to rule out any potential issues hence set
## them correctly to zero)
Pnorm[xnorm<min(x),] <- 0
Pnorm[xnorm>max(x),] <- 0
Pnorm[xnorm==max(x),-ncol(Pnorm)] <- 0
P.left <- as.matrix(P.lin[,1])
P.right <- as.matrix(P.lin[,2])
## We want the linear segment to have the same slope as the
## polynomial segment it connects with and to match at the joint
## hence conduct some carpentry at the left boundary.
index <- which(xnorm==min(x))
index.l <- index+1
index.u <- index+5
x.l <- xnorm[index.l]
x.u <- xnorm[index.u]
slope.poly.left <- as.numeric((Pnorm[index.u,1]-Pnorm[index.l,1])/(x.u-x.l))
index.l <- index+1
index.u <- index+5
x.l <- xnorm[index.l]
x.u <- xnorm[index.u]
slope.linear.left <- as.numeric((P.left[index.u]-P.left[index.l])/(x.u-x.l))
## Complete carpentry at the right boundary.
index <- which(xnorm==max(x))
index.l <- index-1
index.u <- index-5
x.l <- xnorm[index.l]
x.u <- xnorm[index.u]
slope.poly.right <- as.numeric((Pnorm[index.u,ncol(Pnorm)]-Pnorm[index.l,ncol(Pnorm)])/(x.u-x.l))
index.l <- index-1
index.u <- index-5
x.l <- xnorm[index.l]
x.u <- xnorm[index.u]
slope.linear.right <- as.numeric((P.right[index.u]-P.right[index.l])/(x.u-x.l))
P.left <- as.matrix(P.left-1)*slope.poly.left/slope.linear.left+1
P.right <- as.matrix(P.right-1)*slope.poly.right/slope.linear.right+1
P.left[xnorm>=min(x),1] <- 0
P.right[xnorm<=max(x),1] <- 0
Pnorm[,1] <- Pnorm[,1]+P.left
Pnorm[,ncol(Pnorm)] <- Pnorm[,ncol(Pnorm)]+P.right
return(Pnorm)
}
density.deriv.basis <- function(x=NULL,
xeval=NULL,
xnorm=xnorm,
degree=NULL,
segments=NULL,
basis="tensor",
knots="quantiles",
monotone=TRUE,
deriv.index=1,
deriv=1) {
suppressWarnings(Pnorm.deriv <- prod.spline(x=x,
xeval=xnorm,
K=cbind(degree,segments+if(monotone){2}else{0}),
knots=knots,
basis=basis,
deriv.index=deriv.index,
deriv=deriv))
if(monotone) Pnorm.deriv <- Pnorm.deriv[,-c(2,degree+segments+1)]
suppressWarnings(P.lin <- prod.spline(x=x,
xeval=xnorm,
K=cbind(1,1),
knots=knots,
basis=basis,
deriv.index=deriv.index,
deriv=deriv))
## For the derivative bases on the extended range `xnorm', above
## and below max(x)/min(x) we assign the bases to constants
## (zero). We append the linear basis to the left and right of the
## bases. The left basis takes on linear values to the left of
## min(x), zero elsewhere, the right zero to the left of max(x),
## linear elsewhere.
Pnorm.deriv[xnorm<min(x),] <- 0
Pnorm.deriv[xnorm>max(x),] <- 0
P.left <- as.matrix(P.lin[,1])
P.left[xnorm>=min(x),1] <- 0
P.right <- as.matrix(P.lin[,2])
P.right[xnorm<=max(x),1] <- 0
Pnorm.deriv[,1] <- Pnorm.deriv[,1]+P.left
Pnorm.deriv[,ncol(Pnorm.deriv)] <- Pnorm.deriv[,ncol(Pnorm.deriv)]+P.right
return(Pnorm.deriv)
}
clsd <- function(x=NULL,
beta=NULL,
xeval=NULL,
degree=NULL,
segments=NULL,
degree.min=2,
degree.max=25,
segments.min=1,
segments.max=100,
lbound=NULL,
ubound=NULL,
basis="tensor",
knots="quantiles",
penalty=NULL,
deriv.index=1,
deriv=1,
elastic.max=TRUE,
elastic.diff=3,
do.gradient=TRUE,
er=NULL,
monotone=TRUE,
monotone.lb=-250,
n.integrate=500,
nmulti=1,
method = c("L-BFGS-B", "Nelder-Mead", "BFGS", "CG", "SANN"),
verbose=FALSE,
quantile.seq=seq(.01,.99,by=.01),
random.seed=42,
maxit=10^5,
max.attempts=25,
NOMAD=FALSE) {
if(elastic.max && !NOMAD) {
degree.max <- 3
segments.max <- 3
}
ptm <- system.time("")
if(is.null(x)) stop(" You must provide data")
## If no er is provided use the following ad-hoc rule which attempts
## to ensure we cover the support of the variable for distributions
## with moments. This gets the chi-square, t, and Gaussian for n >=
## 100 with all degrees of freedom and df=1 is perhaps the worst
## case scenario. This rule delivers er = 0.43429448, 0.21714724,
## 0.14476483, 0.10857362, 0.08685890, and 0.0723824110, for n = 10,
## 10^2, 10^3, 10^4, 10^5, and 10^6. It is probably too aggressive
## for the larger samples but one can override - the code traps for
## non-finite integration and issues a message when this occurs
## along with a suggestion.
if(!is.null(er) && er < 0) stop(" er must be non-negative")
if(is.null(er)) er <- 1/log(length(x))
if(is.null(penalty)) penalty <- log(length(x))/2
method <- match.arg(method)
fv <- NULL
gen.xnorm.out <- gen.xnorm(x=x,
lbound=lbound,
ubound=ubound,
er=er,
n.integrate=n.integrate)
xnorm <- gen.xnorm.out$xnorm
rank.xnorm <- gen.xnorm.out$rank.xnorm
if(is.null(beta)) {
## If no parameters are provided presume intention is to run
## maximum likelihood estimation to obtain the parameter
## estimates.
ptm <- ptm + system.time(ls.ml.out <- ls.ml(x=x,
xnorm=xnorm,
rank.xnorm=rank.xnorm,
degree.min=degree.min,
segments.min=segments.min,
degree.max=degree.max,
segments.max=segments.max,
lbound=lbound,
ubound=ubound,
elastic.max=elastic.max,
elastic.diff=elastic.diff,
do.gradient=do.gradient,
maxit=maxit,
nmulti=nmulti,
er=er,
method=method,
n.integrate=n.integrate,
basis=basis,
knots=knots,
penalty=penalty,
monotone=monotone,
monotone.lb=monotone.lb,
verbose=verbose,
max.attempts=max.attempts,
random.seed=random.seed,
NOMAD=NOMAD))
beta <- ls.ml.out$beta
degree <- ls.ml.out$degree
segments <- ls.ml.out$segments
fv <- ls.ml.out$fv
}
if(!is.null(xeval)) {
gen.xnorm.out <- gen.xnorm(x=x,
xeval=xeval,
lbound=lbound,
ubound=ubound,
er=er,
n.integrate=n.integrate)
xnorm <- gen.xnorm.out$xnorm
rank.xnorm <- gen.xnorm.out$rank.xnorm
}
if(is.null(degree)) stop(" You must provide spline degree")
if(is.null(segments)) stop(" You must provide number of segments")
ptm <- ptm + system.time(Pnorm <- density.basis(x=x,
xeval=xeval,
xnorm=xnorm,
degree=degree,
segments=segments,
basis=basis,
knots=knots,
monotone=monotone))
if(ncol(Pnorm)!=length(beta)) stop(paste(" Incompatible arguments: beta must be of dimension ",ncol(Pnorm),sep=""))
Pnorm.beta <- as.numeric(Pnorm%*%as.matrix(beta))
## Compute the constant of integration to normalize the density
## estimate so that it integrates to one.
norm.constant <- integrate.trapezoidal.sum(xnorm,exp(Pnorm.beta))
log.norm.constant <- log(norm.constant)
if(!is.finite(log.norm.constant))
stop(paste(" integration not finite - perhaps try reducing `er' (current value = ",round(er,3),")",sep=""))
## For the distribution, compute the density over the extended
## range, then return values corresponding to either the sample x or
## evaluation x (xeval) based on integration over the extended range
## for the xnorm points (xnorm contains x and xeval - this ought to
## ensure integration to one).
## f.norm is the density evaluated on the extended range (including
## sample observations and evaluation points if the latter exist),
## F.norm the distribution evaluated on the extended range.
f.norm <- exp(Pnorm.beta-log.norm.constant)
F.norm <- integrate.trapezoidal(xnorm,f.norm)
if(deriv > 0) {
ptm <- ptm + system.time(Pnorm.deriv <- density.deriv.basis(x=x,
xeval=xeval,
xnorm=xnorm,
degree=degree,
segments=segments,
basis=basis,
knots=knots,
monotone=monotone,
deriv.index=deriv.index,
deriv=deriv))
f.norm.deriv <- as.numeric(f.norm*Pnorm.deriv%*%beta)
} else {
f.deriv <- NULL
f.norm.deriv <- NULL
}
## Compute quantiles using the the quasi-inverse (Definition 2.3.6,
## Nelson (2006))
quantile.vec <- numeric(length(quantile.seq))
for(i in 1:length(quantile.seq)) {
if(quantile.seq[i]>=0.5) {
quantile.vec[i] <- max(xnorm[F.norm<=quantile.seq[i]])
} else {
quantile.vec[i] <- min(xnorm[F.norm>=quantile.seq[i]])
}
}
## Next, strip off the values of the distribution corresponding to
## either sample x or evaluation xeval
if(is.null(xeval)) {
f <- f.norm[rank.xnorm][1:length(x)]
F <- F.norm[rank.xnorm][1:length(x)]
f.norm <- f.norm[rank.xnorm][(length(x)+1):length(f.norm)]
F.norm <- F.norm[rank.xnorm][(length(x)+1):length(F.norm)]
xnorm <- xnorm[rank.xnorm][(length(x)+1):length(xnorm)]
if(deriv>0) {
f.deriv <- f.norm.deriv[rank.xnorm][1:length(x)]
f.norm.deriv <- f.norm.deriv[rank.xnorm][(length(x)+1):length(f.norm.deriv)]
}
P <- Pnorm[rank.xnorm,][1:length(x),]
P.beta <- Pnorm.beta[rank.xnorm][1:length(x)]
} else {
f <- f.norm[rank.xnorm][1:length(xeval)]
F <- F.norm[rank.xnorm][1:length(xeval)]
f.norm <- f.norm[rank.xnorm][(length(x)+length(xeval)+1):length(f.norm)]
F.norm <- F.norm[rank.xnorm][(length(x)+length(xeval)+1):length(F.norm)]
xnorm <- xnorm[rank.xnorm][(length(x)+length(xeval)+1):length(xnorm)]
if(deriv>0) {
f.deriv <- f.norm.deriv[rank.xnorm][1:length(xeval)]
f.norm.deriv <- f.norm.deriv[rank.xnorm][(length(x)+length(xeval)+1):length(f.norm.deriv)]
}
P <- Pnorm[rank.xnorm,][1:length(xeval),]
P.beta <- Pnorm.beta[rank.xnorm][1:length(xeval)]
}
clsd.return <- list(density=f,
density.deriv=f.deriv,
distribution=F,
density.er=f.norm,
density.deriv.er=f.norm.deriv,
distribution.er=F.norm,
xer=xnorm,
Basis.beta=P.beta,
Basis.beta.er=Pnorm.beta,
P=P,
Per=Pnorm,
logl=sum(P.beta-log.norm.constant),
constant=norm.constant,
degree=degree,
segments=segments,
knots=knots,
basis=basis,
nobs=length(x),
beta=beta,
fv=fv,
er=er,
penalty=penalty,
nmulti=nmulti,
x=x,
xq=quantile.vec,
tau=quantile.seq,
ptm=ptm)
class(clsd.return) <- "clsd"
return(clsd.return)
}
sum.log.density <- function(beta=NULL,
P=NULL,
Pint=NULL,
xint=NULL,
length.x=NULL,
penalty=NULL,
complexity=NULL,
...) {
return(2*sum(P%*%beta)-2*length.x*log(integrate.trapezoidal.sum(xint,exp(Pint%*%beta)))-penalty*complexity)
}
sum.log.density.gradient <- function(beta=NULL,
colSumsP=NULL,
Pint=NULL,
xint=NULL,
length.x=NULL,
penalty=NULL,
complexity=NULL,
...) {
exp.Pint.beta <- as.numeric(exp(Pint%*%beta))
exp.Pint.beta.Pint <- exp.Pint.beta*Pint
int.exp.Pint.beta.Pint <- numeric()
for(i in 1:complexity) int.exp.Pint.beta.Pint[i] <- integrate.trapezoidal.sum(xint,exp.Pint.beta.Pint[,i])
return(2*(colSumsP-length.x*int.exp.Pint.beta.Pint/integrate.trapezoidal.sum(xint,exp.Pint.beta)))
}
ls.ml <- function(x=NULL,
xnorm=NULL,
rank.xnorm=NULL,
degree.min=NULL,
segments.min=NULL,
degree.max=NULL,
segments.max=NULL,
lbound=NULL,
ubound=NULL,
elastic.max=FALSE,
elastic.diff=NULL,
do.gradient=TRUE,
maxit=NULL,
nmulti=NULL,
er=NULL,
method=NULL,
n.integrate=NULL,
basis=NULL,
knots=NULL,
penalty=NULL,
monotone=TRUE,
monotone.lb=NULL,
verbose=NULL,
max.attempts=NULL,
random.seed=NULL,
NOMAD=FALSE) {
## This function conducts log spline maximum
## likelihood. Multistarting is supported as is breaking out to
## potentially avoid wasted computation (be careful when using this,
## however, as it is prone to stopping early).
## Save seed prior to setting
if(exists(".Random.seed", .GlobalEnv)) {
save.seed <- get(".Random.seed", .GlobalEnv)
exists.seed = TRUE
} else {
exists.seed = FALSE
}
set.seed(random.seed)
if(missing(x)) stop(" You must provide data")
if(!NOMAD) {
## We set some initial parameters that are placeholders to get
## things rolling.
d.opt <- Inf
s.opt <- Inf
par.opt <- Inf
value.opt <- -Inf
length.x <- length(x)
length.xnorm <- length(xnorm)
## Loop through all degrees for every segment starting at
## segments.min.
d <- degree.min
while(d <= degree.max) {
## For smooth densities one can simply restrict degree to at least
## 2 (or 3 to be consistent with cubic splines)
s <- segments.min
while(s <= segments.max) {
if(options('crs.messages')$crs.messages) {
if(verbose) cat("\n")
cat("\r ")
cat("\rOptimizing, degree = ",d,", segments = ",s,", degree.opt = ",d.opt, ", segments.opt = ",s.opt," ",sep="")
}
## Generate objects that need not be recomputed for a given d
## and s
Pnorm <- density.basis(x=x,
xnorm=xnorm,
degree=d,
segments=s,
basis=basis,
knots=knots,
monotone=monotone)
P <- Pnorm[rank.xnorm,][1:length.x,]
colSumsP <- colSums(P)
Pint <- Pnorm[rank.xnorm,][(length.x+1):nrow(Pnorm),]
xint <- xnorm[rank.xnorm][(length.x+1):nrow(Pnorm)]
complexity <- d+s
## Multistart if desired.
for(n in 1:nmulti) {
## Can restart to see if we can improve on min... note initial
## values totally ad-hoc...
par.init.out <- par.init(d,s,monotone,monotone.lb)
par.init <- par.init.out$par.init
par.upper <- par.init.out$par.upper
par.lower <- par.init.out$par.lower
## Trap non-convergence, restart from different initial
## points, display message if needed (trace>0 up to 6 provides
## ever more detailed information for L-BFGS-B)
optim.out <- list()
optim.out[[4]] <- 9999
optim.out$value <- -Inf
m.attempts <- 0
while(tryCatch(suppressWarnings(optim.out <- optim(par=par.init,
fn=sum.log.density,
gr=if(do.gradient){sum.log.density.gradient}else{NULL},
upper=par.upper,
lower=par.lower,
method=method,
penalty=penalty,
P=P,
colSumsP=colSumsP,
Pint=Pint,
length.x=length.x,
xint=xint,
complexity=complexity,
control=list(fnscale=-1,maxit=maxit,if(verbose){trace=1}else{trace=0}))),
error = function(e){return(optim.out)})[[4]]!=0 && m.attempts < max.attempts){
## If optim fails to converge, reset initial parameters and
## try again.
if(options('crs.messages')$crs.messages) {
if(verbose && optim.out[[4]]!=0) {
if(!is.null(optim.out$message)) cat("\n optim message = ",optim.out$message,sep="")
cat("\n optim failed (degree = ",d,", segments = ",s,", convergence = ", optim.out[[4]],") re-running with new initial values",sep="")
}
}
par.init.out <- par.init(d,s,monotone,monotone.lb)
par.init <- par.init.out$par.init
par.upper <- par.init.out$par.upper
par.lower <- par.init.out$par.lower
m.attempts <- m.attempts+1
}
## Check for a new optimum, overwrite existing values with
## new values.
if(optim.out$value > value.opt) {
if(options('crs.messages')$crs.messages) {
if(verbose && n==1) cat("\n optim improved: d = ",d,", s = ",s,", old = ",formatC(value.opt,format="g",digits=6),", new = ",formatC(optim.out$value,format="g",digits=6),", diff = ",formatC(optim.out$value-value.opt,format="g",digits=6),sep="")
if(verbose && n>1) cat("\n optim improved (ms ",n,"/",nmulti,"): d = ",d,", s = ",s,", old = ",formatC(value.opt,format="g",digits=6),", new = ",formatC(optim.out$value,format="g",digits=6),", diff = ",formatC(optim.out$value-value.opt,format="g",digits=6),sep="")
}
par.opt <- optim.out$par
d.opt <- d
s.opt <- s
value.opt <- optim.out$value
}
}
if(!(segments.min==segments.max) && elastic.max && s.opt == segments.max) segments.max <- segments.max+elastic.diff
if(!(segments.min==segments.max) && elastic.max && s.opt < segments.max+elastic.diff) segments.max <- s.opt+elastic.diff
s <- s+1
}
d <- d+1
if(!(degree.min==degree.max) && elastic.max && d.opt == degree.max) degree.max <- degree.max+elastic.diff
if(!(degree.min==degree.max) && elastic.max && d.opt < degree.max+elastic.diff) degree.max <- d.opt+elastic.diff
}
} else {
eval.f <- function(input, params) {
sum.log.density <- params$sum.log.density
sum.log.density.gradient <- params$sum.log.density.gradient
method <- params$method
penalty <- params$penalty
x <- params$x
xnorm <- params$xnorm
knots <- params$knots
basis <- params$basis
monotone <- params$monotone
monotone.lb <- params$monotone.lb
rank.xnorm <- params$rank.xnorm
do.gradient <- params$do.gradient
maxit <- params$maxit
max.attempts <- params$max.attempts
verbose <- params$verbose
length.x <- length(x)
length.xnorm <- length(xnorm)
d <- input[1]
s <- input[2]
complexity <- d+s
Pnorm <- density.basis(x=x,
xnorm=xnorm,
degree=d,
segments=s,
basis=basis,
knots=knots,
monotone=monotone)
P <- Pnorm[rank.xnorm,][1:length.x,]
colSumsP <- colSums(P)
Pint <- Pnorm[rank.xnorm,][(length.x+1):nrow(Pnorm),]
xint <- xnorm[rank.xnorm][(length.x+1):nrow(Pnorm)]
## NOMAD minimizes only
optim.out <- list()
optim.out[[4]] <- 9999
optim.out$value <- -Inf
m.attempts <- 0
while(tryCatch(suppressWarnings(optim.out <- optim(par=par.init,
fn=sum.log.density,
gr=if(do.gradient){sum.log.density.gradient}else{NULL},
upper=par.upper,
lower=par.lower,
method=method,
penalty=penalty,
P=P,
colSumsP=colSumsP,
Pint=Pint,
length.x=length.x,
xint=xint,
complexity=complexity,
control=list(fnscale=-1,maxit=maxit,if(verbose){trace=1}else{trace=0}))),
error = function(e){return(optim.out)})[[4]]!=0 && m.attempts < max.attempts){
## If optim fails to converge, reset initial parameters and
## try again.
if(verbose && optim.out[[4]]!=0) {
if(options('crs.messages')$crs.messages) {
if(!is.null(optim.out$message)) cat("\n optim message = ",optim.out$message,sep="")
cat("\n optim failed (degree = ",d,", segments = ",s,", convergence = ", optim.out[[4]],") re-running with new initial values",sep="")
}
}
par.init.out <- par.init(d,s,monotone,monotone.lb)
par.init <- par.init.out$par.init
par.upper <- par.init.out$par.upper
par.lower <- par.init.out$par.lower
m.attempts <- m.attempts+1
}
if(options('crs.messages')$crs.messages) {
if(verbose) cat("\n")
cat("\r ")
cat("\rOptimizing, degree = ",d,", segments = ",s,", log likelihood = ",optim.out$value,sep="")
}
fv <- -optim.out$value
}
## Initial values
x0 <- c(degree.min,segments.min)
## Types of variables
bbin <-c(1, 1)
## Bounds
lb <- c(degree.min,segments.min)
ub <- c(degree.max,segments.max)
## Type of output
bbout <- c(0, 2, 1)
## Options
opts <-list("MAX_BB_EVAL"=10000)
## Generate params
params <- list()
params$sum.log.density <- sum.log.density
params$sum.log.density.gradient <- sum.log.density.gradient
params$method <- method
params$penalty <- penalty
params$x <- x
params$xnorm <- xnorm
params$knots <- knots
params$basis <- basis
params$monotone <- monotone
params$monotone.lb <- monotone.lb
params$rank.xnorm <- rank.xnorm
params$do.gradient <- do.gradient
params$maxit <- maxit
params$max.attempts <- max.attempts
params$verbose <- verbose
solution <- snomadr(eval.f=eval.f,
n=2,## number of variables
x0=x0,
bbin=bbin,
bbout=bbout,
lb=lb,
ub=ub,
nmulti=nmulti,
print.output=FALSE,
opts=opts,
params=params)
value.opt <- solution$objective
d <- solution$solution[1]
s <- solution$solution[2]
## Final call to optim to retrieve beta
length.x <- length(x)
length.xnorm <- length(xnorm)
complexity <- d+s
par.init.out <- par.init(d,s,monotone,monotone.lb)
par.init <- par.init.out$par.init
par.upper <- par.init.out$par.upper
par.lower <- par.init.out$par.lower
Pnorm <- density.basis(x=x,
xnorm=xnorm,
degree=d,
segments=s,
basis=basis,
knots=knots,
monotone=monotone)
P <- Pnorm[rank.xnorm,][1:length.x,]
colSumsP <- colSums(P)
Pint <- Pnorm[rank.xnorm,][(length.x+1):nrow(Pnorm),]
xint <- xnorm[rank.xnorm][(length.x+1):nrow(Pnorm)]
optim.out <- list()
optim.out[[4]] <- 9999
optim.out$value <- -Inf
m.attempts <- 0
while(tryCatch(suppressWarnings(optim.out <- optim(par=par.init,
fn=sum.log.density,
gr=if(do.gradient){sum.log.density.gradient}else{NULL},
upper=par.upper,
lower=par.lower,
method=method,
penalty=penalty,
P=P,
colSumsP=colSumsP,
Pint=Pint,
length.x=length.x,
xint=xint,
complexity=complexity,
control=list(fnscale=-1,maxit=maxit,if(verbose){trace=1}else{trace=0}))),
error = function(e){return(optim.out)})[[4]]!=0 && m.attempts < max.attempts){
## If optim fails to converge, reset initial parameters and
## try again.
if(verbose && optim.out[[4]]!=0) {
if(options('crs.messages')$crs.messages) {
if(!is.null(optim.out$message)) cat("\n optim message = ",optim.out$message,sep="")
cat("\n optim failed (degree = ",d,", segments = ",s,", convergence = ", optim.out[[4]],") re-running with new initial values",sep="")
}
}
par.init.out <- par.init(d,s,monotone,monotone.lb)
par.init <- par.init.out$par.init
par.upper <- par.init.out$par.upper
par.lower <- par.init.out$par.lower
m.attempts <- m.attempts+1
}
d.opt <- d
s.opt <- s
par.opt <- optim.out$par
value.opt <- optim.out$value
}
if(options('crs.messages')$crs.messages) {
cat("\r ")
if(!(degree.min==degree.max) && (d.opt==degree.max)) warning(paste(" optimal degree equals search maximum (", d.opt,"): rerun with larger degree.max",sep=""))
if(!(segments.min==segments.max) && (s.opt==segments.max)) warning(paste(" optimal segment equals search maximum (", s.opt,"): rerun with larger segments.max",sep=""))
if(par.opt[1]>0|par.opt[length(par.opt)]>0) warning(" optim() delivered a positive weight for linear segment (supposed to be negative)")
if(!monotone&&par.opt[1]<=monotone.lb) warning(paste(" optimal weight for left nonmonotone basis equals search minimum (",par.opt[1],"): rerun with smaller monotone.lb",sep=""))
if(!monotone&&par.opt[length(par.opt)]<=monotone.lb) warning(paste(" optimal weight for right nonmonotone basis equals search minimum (",par.opt[length(par.opt)],"): rerun with smaller monotone.lb",sep=""))
}
## Restore seed
if(exists.seed) assign(".Random.seed", save.seed, .GlobalEnv)
return(list(degree=d.opt,segments=s.opt,beta=par.opt,fv=value.opt))
}
summary.clsd <- function(object,
...) {
cat("\nCategorical Logspline Density\n",sep="")
cat(paste("\nModel penalty: ", format(object$penalty), sep=""))
cat(paste("\nModel degree/segments: ", format(object$degree),"/",format(object$segments), sep=""))
cat(paste("\nKnot type: ", format(object$knots), sep=""))
cat(paste("\nBasis type: ",format(object$basis),sep=""))
cat(paste("\nTraining observations: ", format(object$nobs), sep=""))
cat(paste("\nLog-likelihood: ", format(object$logl), sep=""))
cat(paste("\nNumber of multistarts: ", format(object$nmulti), sep=""))
cat(paste("\nEstimation time: ", formatC(object$ptm[1],digits=1,format="f"), " seconds",sep=""))
cat("\n\n")
}
print.clsd <- function(x,...)
{
summary.clsd(x)
}
plot.clsd <- function(x,
er=TRUE,
distribution=FALSE,
derivative=FALSE,
ylim,
ylab,
xlab,
type,
...) {
if(missing(xlab)) xlab <- "Data"
if(missing(type)) type <- "l"
if(!er) {
order.x <- order(x$x)
if(distribution){
y <- x$distribution[order.x]
if(missing(ylab)) ylab <- "Distribution"
if(missing(ylim)) ylim <- c(0,1)
}
if(!distribution&&!derivative) {
y <- x$density[order.x]
if(missing(ylab)) ylab <- "Density"
if(missing(ylim)) ylim <- c(0,max(y))
}
if(derivative) {
y <- x$density.deriv[order.x]
if(missing(ylab)) ylab <- "Density Derivative"
if(missing(ylim)) ylim <- c(min(y),max(y))
}
x <- x$x[order.x]
} else {
order.xer <- order(x$xer)
if(distribution){
y <- x$distribution.er[order.xer]
if(missing(ylab)) ylab <- "Distribution"
if(missing(ylim)) ylim <- c(0,1)
}
if(!distribution&&!derivative) {
y <- x$density.er[order.xer]
if(missing(ylab)) ylab <- "Density"
if(missing(ylim)) ylim <- c(0,max(y))
}
if(derivative){
y <- x$density.deriv.er[order.xer]
if(missing(ylab)) ylab <- "Density Derivative"
if(missing(ylim)) ylim <- c(min(y),max(y))
}
x <- x$xer[order.xer]
}
x <- plot(x,
y,
ylim=ylim,
ylab=ylab,
xlab=xlab,
type=type,
...)
}
coef.clsd <- function(object, ...) {
tc <- object$beta
return(tc)
}
fitted.clsd <- function(object, ...){
object$density
}
|
34045744083b6a451af5c1e46238c58314740b16 | 928176f46b5551d2e0af8ca160f06caae49b2303 | /get_mle_pure_r_code.R | 6f985c3fbd127ab96d38b2c97f06e9bee72527f9 | [] | no_license | tianqinglong/bootstrap_prediction | 8632c54c76f9635614084157a80cc4d487bdb0ee | 4a2ba7db7b57a6694cdda8d0912d8109aae40fdd | refs/heads/master | 2020-07-03T17:59:00.034956 | 2019-08-23T01:20:28 | 2019-08-23T01:20:28 | 201,996,897 | 1 | 0 | null | null | null | null | UTF-8 | R | false | false | 563 | r | get_mle_pure_r_code.R | get_Weibull_mle_R <- function(censor_data) {
n_minus_r <- censor_data[[4]] - censor_data[[1]]
sobj <- Surv(time = c( censor_data[[3]],
rep(censor_data[[2]], n_minus_r)
),
event = c( rep(1, censor_data[[1]] ),
rep(0, n_minus_r)
),
type = 'right'
)
sfit <- survreg(sobj~1, dist = 'weibull')
return( list( Shape_mle = 1/sfit$scale,
Scale_mle = as.numeric( exp(sfit$coefficients) )))
} |
f982dd38b90c24f19f65c15eb8b122712fb30623 | db4d0b8b4fe2601054f07bffe7f86c252e4a0e99 | /explorar_SpatialPolygonsDataFrame.r | bb1d4d9aabe26e3e5c3a6848e054b9fa708e57ea | [] | no_license | manuelcampagnolo/vector_datasets_R | f5aaf1fb58d87e8d85143135147f141e10538110 | ca22950ba05ac00fab78140c5408ea58ae9107ae | refs/heads/master | 2016-09-03T07:40:26.430196 | 2015-06-22T01:36:26 | 2015-06-22T01:36:26 | 37,262,455 | 0 | 1 | null | null | null | null | UTF-8 | R | false | false | 1,849 | r | explorar_SpatialPolygonsDataFrame.r | # ler shapefile de areas protegidas (ICNF)
icnf<-readOGR(dsn=getwd(),layer="AP_JUL_2014",encoding="ISO8859-1")
plot(icnf)
# qual é o CRS do cdg?
icnf@proj4string
# quantos multi-poligonos há?
length(icnf@polygons)
# como são as primeiras linhas da tabela de atributos?
head(icnf@data)
# procurar a linha da tabela de atributos que tem Montejunto no atributo Nome
indice<-which(grepl(pattern="Montejunto",icnf@data$NOME,ignore.case = TRUE))
# verificar que é a 12a linha
icnf@data[12,]
# representar esse 12o multi-polígono de icnf
# selecciona-se um subconjunto dos polígono seleccionando a(s) linha(s) como para a tabela de atributos
class(icnf[12,]) # ainda é SpatialPolygonsDataFrame
# obter informação sobre o 12o multi-polígono:
# ID do multi-polígono
icnf@polygons[[12]]@ID
# area do multi-polígono
icnf@polygons[[12]]@area
# quantas partes tem o 12o multi-poligono de icnf?
length(icnf@polygons[[12]]@Polygons) #tem 6 partes
#qual é o tipo de cada parte: "hole" ou não?
for (i in 1:6) print(icnf@polygons[[12]]@Polygons[[i]]@hole)
# qual é a área de cada parte do 12o multi-poligono?
for (i in 1:6) print(icnf@polygons[[12]]@Polygons[[i]]@area)
# quais são as coordenadas dos pontos que delimitam a 3a parte do 12o multi-polígono?
pol3 <- icnf@polygons[[12]]@Polygons[[3]]@coords # matriz com 2 colunas
# construir imagem
if (export) png(paste(aulas,"ap_montejunto.png",sep="\\"), width=800, height=600, res=120)
plot(icnf[12,])
text(x=icnf[12,]@bbox["x","min"], y=icnf[12,]@bbox["y","max"], as.character(icnf@data[12,"NOME"]),pos=4,cex=.9)
for (i in 1:length(icnf@polygons[[12]]@Polygons))
{
aux<-icnf@polygons[[12]]@Polygons[[i]];
if (aux@hole) polygon(aux@coords,col="yellow")
text(x=aux@labpt[1], y=aux@labpt[2],paste(round(aux@area/10000,1),"ha",sep=""),cex=.7,pos=4)
}
if (export) graphics.off()
|
ddc2d036394d28351b0b08666555189595167a36 | 8b6da8afa2945d53aea5380957b5714b74e732b6 | /plot3.R | 561dcab8559a392e4533ea6755a363abb7890889 | [] | no_license | neoeahit/ExData_Plotting1 | a9cf4c7b95e56bd5695f5fdd0a9524c89a167d5e | 42063482f7e35778c6cc6ddcabf7940d4c449e78 | refs/heads/master | 2021-01-22T05:43:25.381771 | 2017-02-13T06:36:59 | 2017-02-13T06:36:59 | 81,690,572 | 0 | 0 | null | 2017-02-11T23:45:30 | 2017-02-11T23:45:29 | null | UTF-8 | R | false | false | 392 | r | plot3.R | data=read_data()
png("plot3.png", height=480, width=480)
with(data, plot(Time, Sub_metering_1, type="l",ylab="Energy sub metering", xlab=" ", col="black"))
with(data, lines(Time, Sub_metering_2, col="red"))
with(data, lines(Time, Sub_metering_3, col="blue"))
legend(x="topright", lwd=1, legend=c("Sub_metering_1", "Sub_metering_2", "Sub_metering_3"), col=c("black", "red", "blue"))
dev.off()
|
be3be50699979a29d74330b627a0d96fa4c94f08 | 6e5d78bb8fe6d0026e110a6c29c60a012f16e1ff | /Data Mining Course/9. support vector machines.R | ff88d3f146493b9636db76720843dfad3c7e08f5 | [] | no_license | richarddeng88/Advanced_Data_Mining | b2d2b91e9a6f100fc4db8be0c6a53140ee64e6fe | ef386c9fa5293ad0ea69b779b36251b15f8b59f0 | refs/heads/master | 2021-01-15T15:45:23.576894 | 2016-10-22T22:02:42 | 2016-10-22T22:02:42 | 47,933,660 | 0 | 1 | null | null | null | null | UTF-8 | R | false | false | 569 | r | 9. support vector machines.R | #================================= support vector classifier========================================
set.seed(1)
x <- matrix(rnorm(20*2), ncol=2)
y <- c(rep(-1,10), rep(1,10))
x[y==1,] <- x[y==1,]+1
# we check weather the classes are linearly seperable.
plot(x, col=c(4,2))
## we encode the response as a factor and conbine the variable and response together.
dat <- data.frame(x=x, y=as.factor(y))
library(e1071)
sumfit <- svm(y~., data=dat, kernel="linear", cost=1000, scale=F)
# let us plot the support vector classifier obtained above.
plot(sumfit, dat)
|
668ff8306e879494bacdc420772840578efbe1d2 | 7c084e50e556bc0468b4dde2852d65f44df13e41 | /in_progress/models/costBenefitAnalysis/scriptsForPaper/interventionGroupsPresenter.R | 781c717a62ab17ec7de8479fd129a2b96ba8abc5 | [] | no_license | mmcdermott/disease-modeling | 0d2379bb2d2a41ecf120fd5476b8768c76a10fd0 | 2d0eb0caba95216718d60ab9a2b6706021121f3d | refs/heads/master | 2016-09-06T09:18:23.086497 | 2014-08-03T23:23:22 | 2014-08-03T23:23:22 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 15,860 | r | interventionGroupsPresenter.R | library(ggplot2)
source('interventionGroups.R')
source('deSolConstants.R')
#Title Generation:
plotTitle <- function(base,interventionName,final="") {
if (final != "") {
return(ggtitle(paste(c(base,interventionName,final),collapse=" ")))
} else {
return(ggtitle(paste(c(base,interventionName),collapse=" ")))
}
}
baseData <- read.csv(baseFile)
#Base Incidence
baseInc <- generateIncidence(baseData)
#US Health Care System (HCS) TB costs due to base system
baseHCSCost <- (baseData$cN0 + baseData$cN1)/1e9
baseCasesD <- 1e6*(baseData$progTotalD0 + baseData$progTotalD1)
#Data Labels
USB <- rep("USB", totT)
FB <- rep("FB", totT)
all <- rep("All", totT)
noInt <- rep("No Intervention", totT)
int <- rep("Intervention", totT)
int10 <- rep("10% Cured", totT)
int25 <- rep("25% Cured", totT)
int50 <- rep("50% Cured", totT)
int100 <- rep("100% Cured", totT)
savings <- rep("Savings", totT)
costs <- rep("Implementation Cost", totT)
totalCosts <- rep("US HCS Cost", totT)
averted <- rep("Cases Averted", totT)
TBdeathsAverted <- rep("TB Deaths Averted", totT)
redEnLTBI100L <- rep("100% reduction", totT)
redEnLTBI75L <- rep("75% reduction", totT)
redEnLTBI50L <- rep("50% reduction", totT)
redEnLTBI25L <- rep("25% reduction", totT)
redEnLTBI10L <- rep("10% reduction", totT)
#Aesthetics
USBC <- 'blue'
FBC <- 'green'
allC <- 'red'
noIntC <- 'black'
intC <- 'blue'
savingsC <- '#24913C'
costsC <- '#9F0013'
avertedC <- '#24913C'
TBdeathsAvertedC <- '#24913C'
# Creating our data containers.
rawData <- as.list(paperRedEnLTBIInts)
incidence <- as.list(paperRedEnLTBIInts)
HCSCost <- as.list(paperRedEnLTBIInts)
costOfInter <- as.list(paperRedEnLTBIInts)
saveOfInter <- as.list(paperRedEnLTBIInts)
interTot <- as.list(paperRedEnLTBIInts)
totSpent <- as.list(paperRedEnLTBIInts)
cpca <- as.list(paperRedEnLTBIInts)
casesAverted <- as.list(paperRedEnLTBIInts)
# Naming them for convenience.
names(rawData) <- paperRedEnLTBIInts
names(incidence) <- paperRedEnLTBIInts
names(HCSCost) <- paperRedEnLTBIInts
names(costOfInter) <- paperRedEnLTBIInts
names(saveOfInter) <- paperRedEnLTBIInts
names(interTot) <- paperRedEnLTBIInts
names(totSpent) <- paperRedEnLTBIInts
names(cpca) <- paperRedEnLTBIInts
names(casesAverted) <- paperRedEnLTBIInts
for (intervention in paperRedEnLTBIInts) {
# First, grab the base data and incidence
rawData[[intervention]] <-
read.csv(paste(c(intFilePrefix,intervention, intFileSuffix),collapse=""))
incidence[[intervention]] <-
generateIncidence(rawData[[intervention]])
#HCS cost borne by intervention
HCSCost[[intervention]] <- (rawData[[intervention]]$cN0 +
rawData[[intervention]]$cN1)/1e9
#Implementation cost of intervention
costOfInter[[intervention]] <-
(rawData[[intervention]]$interventionCost)/1e9
#Savings from intervention
saveOfInter[[intervention]] <- baseHCSCost - HCSCost[[intervention]]
#Total US HCS cost due to intervention
interTot[[intervention]] <- HCSCost[[intervention]] +
costOfInter[[intervention]]
#Total additional spent by US HCS due to intervention
totSpent[[intervention]] <- interTot[[intervention]] - baseHCSCost
#Cost per cases averted
intCasesD <- 1e6*(rawData[[intervention]]$progTotalD0 + rawData[[intervention]]$progTotalD1)
casesAverted[[intervention]] <- baseCasesD - intCasesD
cpca[[intervention]] <-
1e9*totSpent[[intervention]]/casesAverted[[intervention]]
}
#Incidence Reports: Comparing Baseline Incidence against Intervention Incidence
incData <- data.frame(year = years,
baseUSB=baseInc$IN0,
baseFB =baseInc$IN1,
baseAll=baseInc$INall,
redEnLTBI10USB=incidence[["redEnLTBI10"]]$IN0,
redEnLTBI10FB =incidence[["redEnLTBI10"]]$IN1,
redEnLTBI10All=incidence[["redEnLTBI10"]]$INall,
redEnLTBI25USB=incidence[["redEnLTBI25"]]$IN0,
redEnLTBI25FB =incidence[["redEnLTBI25"]]$IN1,
redEnLTBI25All=incidence[["redEnLTBI25"]]$INall,
redEnLTBI50USB=incidence[["redEnLTBI50"]]$IN0,
redEnLTBI50FB =incidence[["redEnLTBI50"]]$IN1,
redEnLTBI50All=incidence[["redEnLTBI50"]]$INall,
redEnLTBI100USB=incidence[["redEnLTBI100"]]$IN0,
redEnLTBI100FB =incidence[["redEnLTBI100"]]$IN1,
redEnLTBI100All=incidence[["redEnLTBI100"]]$INall)
incPlot <- ggplot(incData, aes(x=year)) +
scale_y_log10(breaks=c(1,2,5,10,25,50,100,200),
labels=c("Elimination (1)",2,5,10,25,50,100,200),
limits=c(0.5,250)) +
labs(x="Years", y="Incidence/Million", color="Population",
linetype="Intervention Status") +
ggtitle("Incidence/Million with Various Immigrating LTBI Cure Rates") +
geom_line(aes(y=baseUSB, color=USB, linetype=noInt)) +
geom_line(aes(y=redEnLTBI10USB, color=USB, linetype=int10)) +
geom_line(aes(y=redEnLTBI25USB, color=USB, linetype=int25)) +
geom_line(aes(y=redEnLTBI50USB, color=USB, linetype=int50)) +
geom_line(aes(y=redEnLTBI100USB, color=USB, linetype=int100)) +
geom_line(aes(y=baseFB, color=FB, linetype=noInt)) +
geom_line(aes(y=redEnLTBI10FB, color=FB, linetype=int10)) +
geom_line(aes(y=redEnLTBI25FB, color=FB, linetype=int25)) +
geom_line(aes(y=redEnLTBI50FB, color=FB, linetype=int50)) +
geom_line(aes(y=redEnLTBI100FB, color=FB, linetype=int100)) +
geom_line(aes(y=baseAll, color=all, linetype=noInt)) +
geom_line(aes(y=redEnLTBI10All, color=all, linetype=int10)) +
geom_line(aes(y=redEnLTBI25All, color=all, linetype=int25)) +
geom_line(aes(y=redEnLTBI50All, color=all, linetype=int50)) +
geom_line(aes(y=redEnLTBI100All, color=all, linetype=int100)) +
theme(axis.title=element_text(size=16),axis.text=element_text(size=15),
plot.title=element_text(size=18))#,legend.key.height =
#unit(1.8,'line'))
#Total Costs Excluding Sticker Price: Comparing costs of various interventions
#US Health Care System (HCS) TB costs due to base system
# baseCost <- (baseData$cN0 + baseData$cN1)/1e9
# #US HCS TB costs due to intervenvtion
# interCost <- (rawData$cN0 + rawData$cN1)/1e9
# #US HCS TB savings due to intervention
# totSaved <- baseCost - interCost
#
# savingsData <- data.frame(year=years, baseCost=baseCost, interCost=interCost,
# totSaved=totSaved)
# yrange <- round(seq(min(savingsData$baseCost),max(savingsData$baseCost),by=0.5),1)
# savingsPlot <- ggplot(savingsData,aes(x=year)) +
# labs(x="Years", y="Billions of USD", color="Intervention Status") +
# scale_y_continuous(breaks=yrange) +
# plotTitle("Total Saved by US Health Care System given
# Intervention",interventionName,
# "ignoring intervention cost") +
# geom_ribbon(aes(ymin=interCost,ymax=baseCost,fill=savings, alpha=0.2)) +
# geom_line(aes(y=baseCost, color=noInt)) +
# geom_line(aes(y=interCost, color=int)) +
# geom_line(aes(y=totSaved, color=savings)) +
# scale_fill_manual(values=c(savingsC)) +
# scale_color_manual(values=c(intC,noIntC,savingsC)) +
# guides(fill=F, alpha=F)
#
# #Total Costs Including Sticker Price: Comparing costs of various interventions
# #Implementation cost of intervention
# costInter <- (rawData$interventionCost)/1e9
# #Total US HCS cost due to intervention
# interTot <- interCost + costInter
# #Total additional spent by US HCS due to intervention
# totSpent <- interTot - baseCost
#
# costData <- data.frame(year=years, baseCost=baseCost, interCost=interTot,
# totSpent=totSpent)
# yrange <- round(seq(min(costData$interCost),max(costData$interCost)+0.5,by=0.5),1)
# costsPlot <- ggplot(costData,aes(x=year)) +
# labs(x="Years", y="Billions of USD", color="Intervention Status") +
# scale_y_continuous(breaks=yrange) +
# plotTitle("Total Spent by US Health Care System given
# Intervention",interventionName,
# "given presumed intervention cost") +
# geom_ribbon(aes(ymin=baseCost,ymax=interCost,fill=costs, alpha=1)) +
# geom_line(aes(y=baseCost, color=noInt)) +
# geom_line(aes(y=interCost, color=int)) +
# geom_line(aes(y=totSpent, color=costs)) +
# scale_fill_manual(values=c(costsC)) +
# scale_color_manual(values=c(costsC,intC,noIntC)) +
# guides(fill=F, alpha=F)
#
# #Total Cases Averted
# baseCases <- 1e6*(baseData$progAcute0 + baseData$progAcute1 +
# baseData$progChron0 + baseData$progChron1)
# baseCasesD <- 1e6*(baseData$progTotalD0 + baseData$progTotalD1)
# intCases <- 1e6*(rawData$progAcute0 + rawData$progAcute1 +
# rawData$progChron0 + rawData$progChron1)
# intCasesD <- 1e6*(rawData$progTotalD0 + rawData$progTotalD1)
# casesAverted <- baseCases - intCases
# casesAvertedD <- baseCasesD - intCasesD
#
# casesAvertedData <- data.frame(year=years,baseCases=baseCases,
# intCases=intCases,
# casesAverted=casesAverted)
# casesAvertedDataD <- data.frame(year=years,baseCases=baseCasesD,
# intCases=intCasesD,
# casesAverted=casesAvertedD)
#
# yrange <- round(seq(min(casesAvertedData$baseCases),
# max(casesAvertedData$baseCases),by=1e5),1)
# casesAvertedPlot <-
# ggplot(casesAvertedData,aes(x=year)) +
# labs(x="Years", y="Cases of TB", color="Intervention Status") +
# scale_y_continuous(breaks=yrange) +
# plotTitle("Total Cases of TB Averted given Intervention",interventionName) +
# geom_ribbon(aes(ymin=intCases,ymax=baseCases,fill=averted, alpha=0.2)) +
# geom_line(aes(y=baseCases, color=noInt)) +
# geom_line(aes(y=intCases, color=int)) +
# geom_line(aes(y=casesAverted, color=averted)) +
# scale_fill_manual(values=c(avertedC)) +
# scale_color_manual(values=c(avertedC,intC,noIntC)) +
# guides(fill=F, alpha=F)
#
# yrange <- round(seq(min(casesAvertedDataD$baseCases),
# max(casesAvertedDataD$baseCases),by=1e5),1)
# casesAvertedPlotD <-
# ggplot(casesAvertedDataD,aes(x=year)) +
# labs(x="Years", y="Discounted Cases of TB",
# color="Intervention Status") +
# scale_y_continuous(breaks=yrange) +
# plotTitle("Discounted Cases of TB Averted given Intervention",
# interventionName) +
# geom_ribbon(aes(ymin=intCases,ymax=baseCases,fill=averted, alpha=0.2)) +
# geom_line(aes(y=baseCases, color=noInt)) +
# geom_line(aes(y=intCases, color=int)) +
# geom_line(aes(y=casesAverted, color=averted)) +
# scale_fill_manual(values=c(avertedC)) +
# scale_color_manual(values=c(avertedC,intC,noIntC)) +
# guides(fill=F, alpha=F)
#
# #Cost per cases averted graph:
# cpcaData <- data.frame(year=yearsPC,cpca=1e9*totSpent[cutoffT:totT]/casesAverted[cutoffT:totT])
# cpcaDataD <- data.frame(year=yearsPC,cpca=1e9*totSpent[cutoffT:totT]/casesAvertedD[cutoffT:totT])
#
# cpcaPlot <-
# ggplot(cpcaData,aes(x=year)) +
# labs(x="Years", y="USD") +
# scale_x_continuous(breaks=c(initialYr,cutoffYr,seq(initialYr,finalYr,25))) +
# #scale_y_log10() +
# plotTitle("Cost per Raw TB Case Averted due to Intervention",
# interventionName) +
# geom_line(aes(y=cpca))
#
# cpcaPlotD <-
# ggplot(cpcaDataD,aes(x=year)) +
# labs(x="Years", y="USD") +
# scale_x_continuous(breaks=c(initialYr,cutoffYr,seq(initialYr,finalYr,25))) +
# #scale_y_log10() +
# plotTitle("Cost per Discounted TB Case Averted due to Intervention",
# interventionName) +
# geom_line(aes(y=cpca))
#
# #TB Deaths:
# baseDeaths <- 1e6*(baseData$tbdeath0 + baseData$tbdeath1)
# baseDeathsD <- 1e6*(baseData$tbdeathD0 + baseData$tbdeathD1)
# intDeaths <- 1e6*(rawData$tbdeath0 + rawData$tbdeath1)
# intDeathsD <- 1e6*(rawData$tbdeathD0 + rawData$tbdeathD1)
# deathsAverted <- baseDeaths - intDeaths
# deathsAvertedD <- baseDeathsD - intDeathsD
#
# deathsAvertedData <- data.frame(year=years,baseDeaths=baseDeaths,
# intDeaths=intDeaths,
# deathsAverted =deathsAverted)
# deathsAvertedDataD <- data.frame(year=years,baseDeaths=baseDeathsD,
# intDeaths=intDeathsD,
# deathsAverted=deathsAvertedD)
#
# yrange <- round(seq(min(deathsAvertedData$baseDeaths),
# max(deathsAvertedData$baseDeaths),by=5e3),1)
# deathsAvertedPlot <-
# ggplot(deathsAvertedData,aes(x=year)) +
# labs(x="Years", y="TB Deaths", color="Intervention Status") +
# scale_y_continuous(breaks=yrange) +
# plotTitle("Total TB Lives Saved Given Intervention",interventionName) +
# geom_ribbon(aes(ymin=intDeaths,ymax=baseDeaths,fill=averted, alpha=0.2)) +
# geom_line(aes(y=baseDeaths, color=noInt)) +
# geom_line(aes(y=intDeaths, color=int)) +
# geom_line(aes(y=deathsAverted, color=TBdeathsAverted)) +
# scale_fill_manual(values=c(TBdeathsAvertedC)) +
# scale_color_manual(values=c(intC,noIntC,TBdeathsAvertedC)) +
# guides(fill=F, alpha=F)
#
# yrange <- round(seq(min(deathsAvertedDataD$baseDeaths),
# max(deathsAvertedDataD$baseDeaths),by=5e3),1)
# deathsAvertedPlotD <-
# ggplot(deathsAvertedDataD,aes(x=year)) +
# labs(x="Years", y="Discounted TB Deaths", color="Intervention Status") +
# scale_y_continuous(breaks=yrange) +
# plotTitle("Discounted TB Lives Saved Given Intervention",interventionName) +
# geom_ribbon(aes(ymin=intDeaths,ymax=baseDeaths,fill=averted, alpha=0.2)) +
# geom_line(aes(y=baseDeaths, color=noInt)) +
# geom_line(aes(y=intDeaths, color=int)) +
# geom_line(aes(y=deathsAverted, color=TBdeathsAverted)) +
# scale_fill_manual(values=c(TBdeathsAvertedC)) +
# scale_color_manual(values=c(intC,noIntC,TBdeathsAvertedC)) +
# guides(fill=F, alpha=F)
#
#
#
# ggsave('paperRedEnLTBI.pdf',x,width=15,height=12)
fileName <- "redEnLTBI.pdf"
pdf(fileName,onefile=T)
print(incPlot)
#print(savingsPlot)
#print(costsPlot)
#print(casesAvertedPlot)
#print(casesAvertedPlotD)
#print(cpcaPlot)
#print(cpcaPlotD)
#print(deathsAvertedPlot)
#print(deathsAvertedPlotD)
dev.off()
|
9e6bb20d8f5f8e3c6c993d70755bde46c6950f81 | fed4b7e86cea3d0bd3f25449135f397f6e1bd9c9 | /PairwiseTTests.R | 292268ec0234f0038ae2e31f924445acb392bf18 | [] | no_license | pkiekel/CentraliaCollegeMath246 | a55589f3ed3ebdc855f00e2ff12a23cfebb5bc1e | 7183404a11621ae24dc9aa4d579cf167be025ba9 | refs/heads/master | 2021-01-21T13:33:44.419092 | 2018-08-07T17:17:46 | 2018-08-07T17:17:46 | 55,106,389 | 3 | 1 | null | null | null | null | UTF-8 | R | false | false | 365 | r | PairwiseTTests.R | hsb2<-read.table("http://www.ats.ucla.edu/stat/data/hsb2.csv", sep=",", header=T)
attach(hsb2)
tapply(write, ses, mean)
tapply(write, ses, sd)
a1 <- aov(write ~ ses)
summary(a1)
pairwise.t.test(write, ses, p.adj = "none")
pairwise.t.test(write, ses, p.adj = "bonf")
TukeyHSD(a1)
a2 <- aov(write ~ ses + female)
summary(a2)
TukeyHSD(a2, "ses")
|
27c82966efdf77ed3395561674ffab9ad03992a0 | 7ebe092c7171d9c370b7c89995bc00f6ffa305cd | /lib/model.fitting.functions.R | e8ee93b743d7ce90c167869d0de8b1269d2ead2c | [
"MIT"
] | permissive | stoufferlab/annual-plant-dynamics | df25eb9b4d3c4cfe70858ffb7cbd2b3f8d813aa8 | 2849ef4cffae362dcf156c97a6600a90847f0d2c | refs/heads/master | 2023-04-16T18:11:45.500162 | 2022-08-01T01:30:40 | 2022-08-01T01:30:40 | 516,149,479 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 3,617 | r | model.fitting.functions.R |
##################################
# model fitting functions
##################################
# produce two-species model prediction given:
# 1. model parameters
# 2. a set of initial conditions for all state variables
fecundity.model.predict = function(params, plants.i, plants.j, seeds.i, seeds.j, time, focal, verbose=FALSE){
# set the parameters in the ode solver
fecundity_dynamics_set_params(
gamma_i = params["gamma_i"],
mu_i = params["mu_i"],
nu_i = params["nu_i"],
r_i = exp(params["log_r_i"]),
K_i = params["K_i"],
beta_i = params["beta_i"],
gamma_j = params["gamma_j"],
mu_j = params["mu_j"],
nu_j = params["nu_j"],
r_j = exp(params["log_r_j"]),
K_j = params["K_j"],
beta_j = params["beta_j"],
alpha_ij = params["alpha_ij"],
alpha_ji = params["alpha_ji"]
)
# for tracing purposes
if(verbose>1){
print(params)
message("predicting")
flush(stdout())
}
# container for the predicted outputs
predicted.fecundity <- numeric(length(plants.i))
# iterate over all observations
for(i in seq.int(length(plants.i))){
# for tracing purposes
if(verbose>2){
message(i)
flush(stdout())
}
# starting conditions are viable seeds in seed bank, plants, and plant biomass
x0 <- c(
seeds.i[i], plants.i[i], plants.i[i] * params["beta_i"],
seeds.j[i], plants.j[i], plants.j[i] * params["beta_j"]
)
# output order matches the conditions above but with time elapsed in first column
growing.season <- fecundity_dynamics(x0, time[i], time[i]/1000.)
colnames(growing.season) <- c(
"time.elapsed",
"seeds.i",
"plants.i",
"biomass.i",
"seeds.j",
"plants.j",
"biomass.j"
)
# we only want the final values to make our prediction
growing.season <- growing.season[nrow(growing.season),]
# convert biomass to per capita fecundity
growing.season$fecundity.i <- exp(params["log_phi_i"]) * growing.season$biomass.i / growing.season$plants.i
growing.season$fecundity.j <- exp(params["log_phi_j"]) * growing.season$biomass.j / growing.season$plants.j
# select the fecundity corresponding to the focal species
predicted.fecundity[i] <- growing.season[,paste0("fecundity.",focal[i])]
}
# for tracing purposes
if(verbose>1){
message("predicted")
flush(stdout())
}
return(predicted.fecundity)
}
# calculate the negative loglikelihood of a set of observations given:
# 1. model parameters
# 2. a set of initial conditions for all state variables
# 3. observed fecundities
fecundity.model.NLL = function(params, plants.i, plants.j, seeds.i, seeds.j, time, focal, fecundity, verbose=0){
# calculate the vector of predicted values using function above
predicted.fecundity <- fecundity.model.predict(
params,
plants.i = plants.i,
plants.j = plants.j,
seeds.i = seeds.i,
seeds.j = seeds.j,
time = time,
focal = focal,
verbose = verbose
)
# anywhere in parameter space that is non-biolgical or uninformative should be avoided
# otherwise we treat observed fecundities as Poisson observations to calculate the log-likelihood
if(!all(predicted.fecundity > 0) || any(!is.finite(predicted.fecundity))){
return(Inf)
}else{
nll <- -sum(dpois(fecundity, predicted.fecundity, log=TRUE))
}
# for tracing purposes
if(verbose>0){
print(nll)
flush(stdout())
}
return(nll)
}
# define the order of parameters ; this is a requirement of mle2 to use a parameter vector (like optim)
parnames(fecundity.model.NLL) <- c(
"gamma_i",
"gamma_j",
"mu_i",
"mu_j",
"nu_i",
"nu_j",
"log_r_i",
"log_r_j",
"K_i",
"K_j",
"beta_i",
"beta_j",
"log_phi_i",
"log_phi_j",
"alpha_ij",
"alpha_ji"
)
|
fd2af82179b5d33372f94e17b08b4e623a57821d | f885f99d0090261317b8528128a1a72958760610 | /R/case_id.r | 274fa4184b0086795ca57376400478a236da5c85 | [] | no_license | BijsT/bupaR | 7a78d0d15655866264bab2bb7882602804303272 | 19f5e63c7393be690addf3c3977f1d00d0cdbfaf | refs/heads/master | 2021-08-26T06:40:32.388974 | 2017-11-21T23:12:47 | 2017-11-21T23:12:47 | 111,611,796 | 0 | 0 | null | 2017-11-21T23:11:10 | 2017-11-21T23:11:09 | null | UTF-8 | R | false | false | 524 | r | case_id.r | #' @title Case classifier
#'
#' @description Get the case classifier of an object of class \code{eventlog}
#'
#' @param eventlog An object of class \code{eventlog}.
#'
#' @seealso \code{\link{eventlog}}, \code{\link{activity_id}},
#' \code{\link{lifecycle_id}}, \code{\link{activity_instance_id}}
#'
#'
#'
#' @export case_id
#'
case_id <- function(eventlog){
if("eventlog" %in% class(eventlog))
return(attr(eventlog, "case_id"))
else
stop("Function only applicable on objects of type 'eventlog'")
}
|
5efcf92d21ba448b7a1f574dd4c0b968d1ffb23a | b63ad7afa41c810687e5d312056a4443cfa42aac | /R/R6UMLR2Base.R | 96506864078e8b63f86f1973d0f9c9d835e14b32 | [] | no_license | Grandez/umlr2 | 0f42ba1499e64785ea95feedef94e086a120140a | 0cbde78918086afcded5fc7cfa8e83e00d53ce11 | refs/heads/master | 2022-12-05T01:31:27.667844 | 2020-08-27T22:06:36 | 2020-08-27T22:06:36 | 283,737,694 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,991 | r | R6UMLR2Base.R | #' @name UMLR2Base
#' @title UMLR2Base
#' @rdname R6UMLR2BASE
#' @docType class
#' @description Clase base del paquete.
UMLR2Base = R6::R6Class("R6UMLR2BASE"
,portable = FALSE
,lock_objects = TRUE
,lock_class = TRUE
,public = list(
#' @description Inicializador
#' @param ... datos para configuracion
initialize = function(...) {
# if (substr(as.character(sys.call(-1))[1], 1, 9) == "UMLR2BASE") msg$err("E900", "UMLR2BASE")
cfg$setConfig(...)
}
#' @description Cambia los datos de configuracion de la instancia
#' @param ... named values para definir la configuracion
#' @return La instancia del objeto
,setConfig = function(...) {
cfg$setConfig(...)
invisible(self)
}
#' @description Obtiene los datos de configuracion
#' @return Una lista con los datos de configuracion
,getConfig = function() { cfg }
#' @description Chequea si la configuracion es correcta
#' @details Esta funcion no verifica que los datos sean reales.
#' Para verificar completamente el sistema usar checkInstallation
#' @param verbose Muestra mensajes informativos
#' @param first Detiene el proceso en el primer error (si lo hay)
#' @return TRUE si lo es
#' FALSE si no
,checkConfiguration = function(verbose=TRUE, first=FALSE) { cfg$checkConfiguration(verbose, first) }
#' @description Chequea si la configuracion y las dependencias son correctas y estan disponibles
#' @param verbose Muestra mensajes informativos
#' @param first Detiene el proceso en el primer error (si lo hay)
#' @return TRUE si lo es
#' FALSE si no
,checkInstallation = function(verbose=TRUE, first=FALSE) { cfg$checkInstallation (verbose, first) }
)
,private = list(S3Class = "S3UMLR2"
,cfg = CONFIG$new()
,msg = UMLR2MSG$new()
)
)
|
94915657aefeb358d8de4ae8a3546cf4a404150d | 632acd6591c71ab7f638092b152794c230f26967 | /siliconvalley/dplyr.R | 57037402292fe0415a81305eae77a395dd737b51 | [] | no_license | kimjh2807/rstudio | 2fdc3dba4dcccf659b2d3a9a3e1cf0331815ef4e | dfdc1831fa65cf82c0cd926967494d573ff45fc5 | refs/heads/master | 2021-06-03T05:50:57.981205 | 2020-04-03T06:54:11 | 2020-04-03T06:54:11 | 111,923,598 | 0 | 6 | null | null | null | null | UTF-8 | R | false | false | 2,580 | r | dplyr.R | # dplyr (p.51~)
library(dplyr)
# tbl_df
iris
i2 <- tbl_df(iris) # tbl_df()
class(i2)
i2
# glimpse
glimpse(i2) # all variable can see with transpose
# %>%
iris %>% head
iris %>% head(10)
# install "gapminder"
install.packages("gapminder")
library(gapminder)
gapminder <- tbl_df(gapminder)
gapminder
glimpse(gapminder)
# filter()
filter(gapminder, country=='Korea, Rep.')
filter(gapminder, year==2007)
filter(gapminder, country=='Korea, Rep.' & year== 2007)
gapminder %>% filter(country == 'Korea, Rep.')
gapminder %>% filter(year == 2007)
gapminder %>% filter(country == 'Korea, Rep.' & year == 2007)
# arrange()
arrange(gapminder, year, country)
gapminder %>% arrange(year, country)
# select()
select(gapminder, pop, gdpPercap)
gapminder %>% select(pop, gdpPercap)
# mutate()
gapminder %>% mutate(total_gdp = pop * gdpPercap,
le_gdp_ratio = lifeExp / gdpPercap,
lgrk = le_gdp_ratio * 100)
# summarize()
gapminder %>%
summarize(n_obs = n(),
n_countries = n_distinct(country),
n_year = n_distinct(year),
med_gdpc = median(gdpPercap),
max_gdppc = max(gdpPercap))
# distinct()
distinct(select(gapminder, country))
distinct(select(gapminder, year))
gapminder %>% select(country) %>% distinct()
gapminder %>% select(year) %>% distinct()
# group_by()
gapminder %>%
filter(year == 2007) %>%
group_by(continent) %>%
summarise(median(lifeExp))
gapminder %>%
filter(year == 2002) %>%
group_by(country) %>%
summarise(lifeExp = median(lifeExp)) %>%
arrange(-lifeExp)
# join
tbl_df(gapminder)
distinct(select(gapminder, country))
filter(gapminder, country == 'Korea, Rep.')
filter(gapminder, year == 2007)
filter(gapminder, country == 'Korea, Rep.' & year == 2007)
gapminder %>% filter(country == 'Korea, Rep.') %>% filter(year == 2007)
gapminder %>% filter(country == 'Korea, Rep.') %>% select(year, lifeExp)
data_yl <- gapminder %>% filter(country == 'Korea, Rep.') %>% select(year, lifeExp)
plot(data_yl)
data_yp <- gapminder %>% filter(country == 'Korea, Rep.') %>% select(year, pop)
plot(data_yp)
data_ygpc <- gapminder %>% filter(country == 'Korea, Rep.') %>% select(year, gdpPercap)
plot(data_ygpc)
gapminder %>% mutate(total_gdp = pop * gdpPercap)
# method 1
d1 = filter(gapminder, year == 2007)
d2 = group_by(d1, continent)
d3 = summarize(d2, lifeExp = median(lifeExp))
arrange(d3, -lifeExp)
arrange(d1, -lifeExp)
# method 2
gapminder %>% filter(year == 2007) %>% group_by(continent) %>%
summarize(lifeExp = median(lifeExp)) %>% arrange(-lifeExp)
|
7b7f9a85728ae8f8fcd5cbba408868c243dc395a | 2c1805e79d915c88faa0f6c258fc41e95937dba5 | /R/Unity/quest_step_position.R | 7e998d751f34c2b35955d200d636e3a8af9203ea | [] | no_license | hejtmy/VR_City_Analysis | b85c14ddc7aad5db8aeeb353ae02462986b20e59 | b149d3f52d76fc8fb0104fa42ec7b38ae7470ba0 | refs/heads/master | 2021-01-18T16:16:53.962471 | 2017-05-21T22:01:26 | 2017-05-21T22:01:34 | 49,779,651 | 0 | 0 | null | 2017-02-18T17:35:16 | 2016-01-16T15:48:50 | R | UTF-8 | R | false | false | 1,073 | r | quest_step_position.R | quest_step_position = function(quest = NULL, step_id){
#parameter validation
if(is.null(quest)){
SmartPrint(c("ERROR:quest_step_position:MissingParameter", "TYPE:quest", "DESCRIPTION:", "parameter not provided"))
return(NULL)
}
if(!is.numeric(step_id)){
SmartPrint(c("ERROR:quest_step_position:WrongParameterType", "TYPE:step_id", "DESCRIPTION:", "Parameter has type ", (class(step_id))," required is numeric"))
return(NULL)
}
step = quest$steps %>% filter(ID == step_id)
if(nrow(step) == 0){
SmartPrint(c("ERROR:quest_step_position:MissingStep", "ID: ", step_id, "DESCRIPTION:", "There is no quest of such ID"))
return(NULL)
}
if(nrow(step) == 0){
SmartPrint(c("ERROR:quest_step_position:NonUnique", "ID: ", step_id, "DESCRIPTION:", "There are more quests with given ID"))
return(NULL)
}
if(step$Transform =="NO transform"){
SmartPrint(c("WARNING:quest_step_position:NOTransform", "ID: ", step_id, "DESCRIPTION:", "Step has no transform"))
return(NULL)
}
return(text_to_vector3(step$Transform)[c(1,3)])
} |
11dbdcadcac15db777b2c15029723b7e54d7acc8 | f5f887250c22676073946936c27306e1d61c48e8 | /test_shiny_app.R | b8340a1fede633f5713ea2e7a08c74e480c8ce47 | [] | no_license | conorotompkins/model_allegheny_house_sales | 8b6d015056c4fc3b55d6e23a37edb2c7ab559f7a | a844e549ab1ab28574847a1ba7215ee9668a223c | refs/heads/main | 2023-03-23T23:35:22.513918 | 2021-03-17T21:17:36 | 2021-03-17T21:17:36 | 320,881,494 | 1 | 0 | null | null | null | null | UTF-8 | R | false | false | 11,072 | r | test_shiny_app.R | #set up
# Load R packages
library(shiny)
library(shinythemes)
library(shinyWidgets)
library(tidyverse)
library(tidymodels)
library(usemodels)
library(hrbrthemes)
library(scales)
library(leaflet)
#https://towardsdatascience.com/build-your-first-shiny-web-app-in-r-72f9538f9868
#https://shiny.rstudio.com/tutorial/
source("scripts/shiny_app/read_ui_input_values.R")
# Define UI
ui <- fluidPage(theme = shinytheme("cerulean"),
title = "Allegheny County Home Sale Price Estimator",
titlePanel(title = "Allegheny County Home Sale Price Estimator"),
fluidRow(
#tabPanel("Inputs", # sidebarPanel
column(3, #column 1
# selectInput(inputId = "school_desc_choice",
# label = "School district",
# choices = pull(school_desc_distinct, school_desc),
# multiple = FALSE,
# selectize = TRUE),
selectInput(inputId = "style_desc_choice",
label = "Style",
choices = pull(style_desc_distinct, style_desc),
selectize = TRUE,
multiple = FALSE),
selectInput(inputId = "grade_desc_choice",
label = "Grade",
choices = pull(grade_desc_distinct, grade_desc),
multiple = FALSE,
selected = "Average"),
selectInput(inputId = "condition_desc_choice",
label = "Condition",
choices = pull(condition_desc_distinct, condition_desc),
multiple = FALSE,
selected = "Average"),
sliderInput(inputId = "lot_area_choice",
label = "Lot Area (sq. ft)",
#min = pull(lot_area_range_min, lot_area),
#max = pull(lot_area_range_max, lot_area),
min = 0,
max = 10000,
value = 2000),
sliderInput(inputId = "finished_living_area_choice",
label = "Finished Living Area (sq. ft)",
#min = pull(finished_living_area_min, finished_living_area),
#max = pull(finished_living_area_max, finished_living_area),
min = 0,
max = 4000,
value = 2000),
sliderInput(inputId = "bedrooms_choice",
label = "Bedrooms",
min = 1,
max = 6,
value = 1),
sliderInput(inputId = "fullbaths_choice",
label = "Full bathrooms",
min = 1,
max = 4,
value = 1),
sliderInput(inputId = "halfbaths_choice",
label = "Half bathrooms",
min = 0,
max = 4,
value = 0),
sliderInput(inputId = "year_blt_choice",
label = "Year house was built",
min = pull(year_blt_min, year_blt),
max = pull(year_blt_max, year_blt),
value = 1948,
sep = ""),
verbatimTextOutput("txtout")
), #column 1
column(9, # column 2
#plotOutput("school_desc_map"),
leafletOutput("school_district_map"),
plotOutput("model_output_graph"),
tableOutput("model_output_table")
) #column 2
) # fluidRow
) # fluidPage
# Define server function
server <- function(input, output) {
#create data to predict on
predict_data_reactive <- reactive({
req(selected_school_desc())
tibble(par_id = "test",
house_age_at_sale = 2020 - input$year_blt_choice,
lot_area = input$lot_area_choice,
finished_living_area = input$finished_living_area_choice,
bedrooms = input$bedrooms_choice,
fullbaths = input$fullbaths_choice,
halfbaths = input$halfbaths_choice,
school_desc = selected_school_desc(),
style_desc = input$style_desc_choice,
grade_desc = input$grade_desc_choice,
condition_desc = input$condition_desc_choice,
longitude = 1,
latitude = 1) %>%
left_join(finished_living_area_summary) %>%
left_join(lot_area_summary) %>%
mutate(finished_living_area_zscore = (finished_living_area - finished_living_area_mean) / finished_living_area_sd,
lot_area_zscore = (lot_area - lot_area_mean) / lot_area_sd) %>%
select(-c(matches("mean$|sd$"), lot_area, finished_living_area))
})
predictions_reactive <- reactive({
#predict on data
model_fit %>%
predict(predict_data_reactive()) %>%
mutate(.pred = 10^.pred) #%>%
# bind_cols(model_fit %>%
# predict(predict_data_reactive(), type = "conf_int") %>%
# mutate(across(matches("^.pred"), ~10^.x)))
})
representative_sample_reactive <- reactive({
full_results %>%
semi_join(predict_data_reactive(), by = c("school_desc", "style_desc"))
})
plot_parameters_reactive <- reactive({
representative_sample_reactive() %>%
pull(sale_price_adj) %>%
hist(breaks = 30) %>%
.$counts %>%
enframe() %>%
summarize(max_count = max(value)) %>%
pull(max_count)
})
output$txtout <- renderText({
list(str_c("School district:", selected_school_desc(), sep = " "),
str_c("Grade:", input$grade_desc_choice, sep = " "),
str_c("Condition:", input$condition_desc_choice, sep = " "),
str_c("Style:", input$style_desc_choice, sep = " "),
str_c("Lot area:", comma(input$lot_area_choice), sep = " "),
str_c("Finished living area:", comma(input$finished_living_area_choice), sep = " "),
str_c("Bedrooms:", input$bedrooms_choice, sep = " "),
str_c("Full Bathrooms:", input$fullbaths_choice, sep = " "),
str_c("Half Bathrooms:", input$halfbaths_choice, sep = " "),
str_c("Year built:", input$year_blt_choice, sep = " ")) %>%
glue::glue_collapse(sep = "\n")
})
output$model_output_table <- renderTable({
predictions_reactive() %>%
mutate(.pred = dollar(.pred)#,
#.pred_upper = dollar(.pred_upper),
#.pred_lower = dollar(.pred_lower)
) %>%
rename(`Average Predicted Price` = .pred#,
#`Upper bound` = .pred_upper,
#`Lower bound` = .pred_lower
) #%>%
#select(`Lower bound`, `Average Predicted Price`, `Upper bound`)
})
output$model_output_graph <- renderPlot({
representative_sample_reactive() %>%
ggplot(aes(x = sale_price_adj)) +
geom_histogram(fill = "grey", color = "black") +
# annotate(geom = "rect",
# xmin = predictions_reactive()$.pred_lower, xmax = predictions_reactive()$.pred_upper,
# ymin = 0, ymax = Inf, fill = "#FCCF02", alpha = .7) +
geom_vline(aes(xintercept = predictions_reactive()$.pred),
color = "#FCCF02",
size = 2) +
scale_x_continuous(labels = scales::dollar_format()) +
scale_y_comma() +
coord_cartesian(ylim = c(0, plot_parameters_reactive() * 1.4)) +
labs(title = str_c(nrow(representative_sample_reactive()) %>% comma(), "sales of",
distinct(representative_sample_reactive())$style_desc, "homes in",
distinct(representative_sample_reactive())$school_desc,
sep = " "),
x = "Sale Price",
y = "Count of similar homes") +
theme_ipsum(base_size = 20) +
theme(panel.background = element_rect(fill = "black"),
axis.title.x = element_text(size = 18),
axis.title.y = element_text(size = 18))
})
# output$school_desc_map <- renderPlot({
#
# #full_results %>%
# school_district_shapes %>%
# semi_join(predict_data_reactive(), by = "school_desc") %>%
# ggplot() +
# geom_sf(data = ac_boundary, fill = "black") +
# geom_sf(data = ac_water, fill = "white") +
# geom_sf(fill = "#FCCF02", color = "#FCCF02", alpha = .7, size = NA) +
# theme_void()
#
# })
output$school_district_map <- renderLeaflet({
school_district_shapes %>%
leaflet("school_district_map") %>%
addProviderTiles(providers$Stamen.TonerLite,
options = providerTileOptions(noWrap = TRUE,
minZoom = 9,
#maxZoom = 8
)) %>%
setView(lng = -80.01181092430839, lat = 40.44170119122286, zoom = 9) %>%
setMaxBounds(lng1 = -79.5, lng2 = -80.5, lat1 = 40.1, lat2 = 40.7) %>%
addPolygons(layerId = ~school_desc,
fillColor = "#FCCF02",
fillOpacity = .7,
stroke = TRUE,
color = "black",
weight = 1)
})
#capture click from leaflet map
selected_school_desc <- reactive({input$school_district_map_shape_click$id})
observe({ #observer
req(selected_school_desc())
# if (length(selected_school_desc()) == 0)
# return()
#
# else {
#filter and map
leafletProxy("school_district_map", data = filter(school_district_shapes, school_desc == input$school_district_map_shape_click$id)) %>%
clearGroup("highlight_shape") %>%
clearGroup("popup") %>%
addPolygons(group = "highlight_shape") %>%
addPopups(popup = ~school_desc,
group = "popup",
lng = ~lng,
lat = ~lat)
#}
}) #observer
}
# Create Shiny object
shinyApp(ui = ui, server = server)
|
5cc7c6d913ff866a08f142818926d595f57634f5 | e0219998a64a696a974e41fc341115704d4a9787 | /source/ini_pr_i.R | eae90295da687b245ac5827cc360cb2472be5348 | [] | no_license | micheledemeo/datacontrol | 81022d65b11d219d11753e7d842c46e102207c8b | c9189a5174976c52d478f682670291785f8b3d49 | refs/heads/master | 2020-06-02T09:16:36.883941 | 2016-08-27T11:23:22 | 2016-08-27T11:23:22 | 27,638,132 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,567 | r | ini_pr_i.R |
# aggiorna i sent in flotta
setkey(flotta, id_battello)
flotta[.( all[sent==1,unique(id_battello)] ) , sent:=1 ]
flotta[is.na(sent) , sent:=0]
# strati con almeno un sent=0 e almeno 2 unità con sent=1
flotta[,remove_to_hv:=0]
setkey(flotta, id_strato,sent)
#str_sent: strati con almeno un campionario inviato. Mettendo sent=0, i battelli campionari di questo strato che hanno sent=0(non inviati), vanno eliminati dal calcolo
str_sent=flotta[,sum(sent), by=id_strato][V1>1,.(id_strato,sent=0)]
setkey(str_sent, id_strato,sent)
# tag per battelli campionari da non considerare nel calcolo hv, essendoci nello strato almeno due unità con sent=1 che possono esser usati come proxy per le mancate risposte (ib_battello>0 & sent=0)
flotta[str_sent, remove_to_hv:=ifelse(id_battello>0,1,0)]
flotta_temp=flotta[,list(id_strato,lft,id_battello=ifelse(remove_to_hv==1,0,id_battello))]
# calcola pr_i e gestisci censimenti####
setkey(flotta_temp, id_strato,lft)
flotta_temp[,eti:=1:nrow(.SD), by=id_strato]
strati_censimento=flotta_temp[,list(.N,n=sum(ifelse(id_battello>0,1,0))),keyby=id_strato][N-n==0,id_strato]
if(length(strati_censimento)>0) {
pr_i=flotta_temp[!id_strato %in% strati_censimento,data.table(.SD[id_battello>0,.(lft,id_battello,eti)],pr_i=diag(hv_pij(lft, n=nrow(.SD[id_battello>0]), eti=.SD[id_battello>0,eti], M=T) ) ) , keyby=id_strato]
pr_i=pr_i[,list(id_battello,pr_i)]
pr_i=rbindlist(list(pr_i,flotta_temp[id_strato %in% strati_censimento,.(id_battello, pr_i=1)]))
} else {
pr_i=flotta_temp[,data.table(.SD[id_battello>0,.(lft,id_battello,eti)],pr_i=diag(hv_pij(lft, n=nrow(.SD[id_battello>0]), eti=.SD[id_battello>0,eti], M=T) ) ) , keyby=id_strato]
pr_i=pr_i[,list(id_battello,pr_i)]
}
# calcolo fattore di correzione
if ( exists("cy") ) {
setkey(cy,id_strato)
ric=all[var=="ricavi", list(id_battello,id_strato,value)]
setkey(ric,id_battello)
setkey(pr_i,id_battello)
#nomatch = 0 => esclude i battelli che vengono rimossi perché non inviati, ma presenti in strati che in cui ci sono almeno 2 inviati
ric=pr_i[ric,nomatch=0][,ric_esp_nisea:=sum(value/pr_i),by=id_strato]
setkey(ric,id_strato)
ric=cy[ric]
ric[is.na(ricavi), ricavi:=ric_esp_nisea] # corr_fact will be 1 for that
ric[,corr_fact:=ricavi/ric_esp_nisea]
setkey(ric,id_battello)
# weight_with_correction = weight * corr_fact = 1/pr * corr_fact -->
# pr_with_correction= 1 / weight_with_correction = 1/(1/pr * corr_fact) = pr / corr_fact
pr_i[ric, pr_i:=pr_i*(1/corr_fact)]
rm(ric)
}
setkey(pr_i,id_battello)
rm(flotta_temp) |
8866763c368c5648482638a048015b4b2a052fbc | ca807743c5b9f9c4e17ee8e5526486a8288e4193 | /RUN_FIRST-create_data_matrix.R | 295d2f7d7f82b99fcf3b3d85bf7671db34649630 | [] | no_license | wsdaniels/COmodeling | bb144f09435d4b193e2e1c66f0cc37a8c626f4d4 | 20785951935226beae72f5a7ced746761e24fbc8 | refs/heads/main | 2023-05-27T17:24:19.414034 | 2023-05-10T20:40:39 | 2023-05-10T20:40:39 | 404,949,625 | 1 | 0 | null | null | null | null | UTF-8 | R | false | false | 6,242 | r | RUN_FIRST-create_data_matrix.R | rm(list = ls())
# Install and load required packages
if ("lubridate" %in% rownames(installed.packages()) == F){
install.packages("lubridate")
}
if ("RAMP" %in% rownames(installed.packages()) == F){
install.packages("RAMP")
}
library(lubridate)
library(RAMP)
# Set base directory
base.dir <- 'https://raw.github.com/wsdaniels/COmodeling/main/'
# GET RESPONSE DATA
response <- read.csv(paste0(base.dir, "MSEA_V8JMOPITT_weeklyanomalies_WEDCEN_nofill.csv"))
response$time <- ymd(response$time)
# Placeholder for missing CO values. This will get used later
missing.val <- -9999
# GET PREDICTOR DATA
nino <- read.csv(paste0(base.dir, "nino34_weekly_avg.csv"))
aao <- read.csv(paste0(base.dir, "aao_weekly_avg.csv"))
tsa <- read.csv(paste0(base.dir, "tsa_weekly_avg.csv"))
dmi <- read.csv(paste0(base.dir, "dmi_weekly_avg.csv"))
olr.msea <- read.csv(paste0(base.dir, "msea_olr.csv"))
# remove partial first entry in olr
olr.msea <- olr.msea[2:nrow(olr.msea),]
# PUT PREDICTOR DATA INTO A LIST
predictors <- list("nino" = nino,
"dmi" = dmi,
"tsa" = tsa,
"aao" = aao,
"olr.msea" = olr.msea)
# Fix column alignment
# NOTE: aao doesn't need the correction for some reason
for (i in 1:3){
this.var <- predictors[[i]]
this.var[,2] <- this.var[,1]
this.var[,1] <- rownames(this.var)
row.names(this.var) <- NULL
predictors[[i]] <- this.var
}
# Convert time variable to lubridate datetime
for (i in 1:length(predictors)){
this.var <- predictors[[i]]
this.var$time <- ymd(this.var$time)
predictors[[i]] <- this.var
}
# Clean up
rm(aao,dmi,nino,tsa,olr.msea,this.var,i)
# ALIGN START OF PREDICTOR TIME SERIES
# Get the latest start date - this will be used to align the starts
start.date <- max(as_date(sapply(predictors, function(X) X$time[1])))
for (i in 1:length(predictors)){
this.var <- predictors[[i]]
this.var <- this.var[!(this.var$time < start.date), ]
predictors[[i]] <- this.var
}
# ALIGN END OF PREDICTOR TIME SERIES
terminal.date <- response$time[length(response$time)]
for (i in 1:length(predictors)){
this.var <- predictors[[i]]
this.var <- this.var[!(this.var$time > terminal.date), ]
predictors[[i]] <- this.var
}
# Clean up
rm(this.var, terminal.date, i)
#------------------------------------------------
# ALL END DATES ARE ALLIGNED AT THIS POINT
# ALL TIME VARIABLES ARE THE SAME AT THIS POINT
#------------------------------------------------
# COMPUTE OFFSETS FROM FINAL RESPONSE OBSERVATION
# This will be used for lag calculations later
for (i in 1:length(predictors)){
predictors[[i]]$offset <- seq(nrow(predictors[[i]])-1, 0)
}
response$offset <- seq(nrow(response)-1, 0)
# COMPUTE SMOOTHED CURVES
# Here we compute gradually smoother gaussian kernels
# Smoothed indices are used for longer lags
x.dist <- response$time[2] - response$time[1]
max.mult <- 8
min.mult <- 1
mult.seq <- seq(min.mult, max.mult, length.out = 8)
for (i in 1:length(predictors)){
for (j in 1:length(mult.seq)){
this.gaussian.kernel <- ksmooth(x = predictors[[i]]$time,
y = predictors[[i]]$anomaly,
kernel = "normal",
bandwidth = mult.seq[j]*x.dist)
predictors[[i]][ length(predictors[[i]]) + 1 ] <- this.gaussian.kernel$y
names(predictors[[i]])[ length(predictors[[i]]) ] <- paste0("gaussian.kernel.",
mult.seq[j])
}
}
rm(this.gaussian.kernel, i, j, x.dist)
# CREATE MASKS FOR THE MONTHS WE WANT TO EXPLAIN THE RESPONSE
#---------------------------------------------------------------------
# Period over which to explain response data (in months)
# This corresponds to fire season in MSEA
study.period <- c(9, 12)
# Create vector of months to keep
if (study.period[1] <= study.period[2]){
months.to.keep <- study.period[1] : study.period[2]
} else {
months.to.keep <- c(study.period[1] : 12, 1 : study.period[2])
}
# Create month mask
month.mask <- month(response$time) %in% months.to.keep
# Apply month mask
response <- response[month.mask,]
rm(month.mask, months.to.keep, study.period)
#---------------------------------------------------------------------
# Remove NA from response
# NOTE: this must be done AFTER setting up the offsets!!!!!!
response <- response[!(response$num_obs == missing.val), ]
# SET LAG LIMITS
min.lag <- 1
max.lag <- 52
lag.vals <- list(nino = min.lag:max.lag,
dmi = min.lag:max.lag,
tsa = min.lag:max.lag,
aao = min.lag:max.lag,
olr.msea = min.lag:max.lag)
# DEFINE SMOOTHING PARAMETERS
start.smoothing <- 4
lag.dividers <- seq(start.smoothing, max.lag, length.out = length(mult.seq)+1)
lag.dividers <- round(lag.dividers)
# COUNT TOTAL NUMBER OF LAGS TO CONSIDER
n.lag <- 0
for (i in 1:length(lag.vals)){
n.lag <- n.lag + length(lag.vals[[i]])
}
# BUILD DATA MATRIX TO BE USED IN RAMP
data.matrix <- data.frame(matrix(NA, ncol = n.lag, nrow = nrow(response)))
# FILL DATA MATRIX
it <- 1
for (i in 1:length(predictors)){
these.lags <- lag.vals[[i]]
if (length(these.lags) > 0){
for (j in 1:length(these.lags)){
var.offsets <- predictors[[i]]$offset
required.offsets <- response$offset + these.lags[j]
to.keep <- var.offsets %in% required.offsets
if (these.lags[j] < start.smoothing){
this.var <- predictors[[i]]$anomaly[to.keep]
} else {
for (k in seq( 1, length(lag.dividers)-1 )){
if (these.lags[j] >= lag.dividers[k] & these.lags[j] < lag.dividers[k+1]){
this.string <- paste0("gaussian.kernel.", mult.seq[k])
this.var <- predictors[[i]][, this.string][to.keep]
break
}
}
}
data.matrix[,it] <- this.var
var.name <- paste0(names(predictors)[i], "_", these.lags[j])
names(data.matrix)[it] <- var.name
it <- it + 1
}
}
}
rm(i, it, j, n.lag, required.offsets, these.lags, this.var, to.keep,
var.name, var.offsets, lag.vals)
write.csv(data.matrix, "data_matrix.csv")
write.csv(response, "response.csv")
|
60d0040694a46d68d8b547b133051395f8e802be | e8bb53f264224f2b72b9b6e2f715080f98914fdf | /04_ExploratoryDataAnalysis/code/Lesson1_LatticePlottingSystem_w2.R | 43a92cc5e4374d55738643f9c19ae9d9007e0e2c | [] | no_license | pritraj90/DataScienceR | 0ef2550590f101bd0886ba7db22c6aa6df755ce0 | 54a67ad080756699138d083bd495da1dfa100d09 | refs/heads/master | 2020-03-21T16:31:22.158157 | 2018-03-07T16:52:26 | 2018-03-07T16:52:26 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,065 | r | Lesson1_LatticePlottingSystem_w2.R | # Week2 : Lesson 1: Lattice Plotting System
library(lattice)
library(datasets)
# Simple scatterplot
xyplot(Ozone~Wind, data = airquality)
# Convert Month to a factor variable
airquality <- transform(airquality, Month = factor(Month))
xyplot(Ozone~Wind | Month, data = airquality, layout = c(5, 1)) # by month
# Lattice behavior
p <- xyplot(Ozone~Wind, data = airquality)
print(p)
## Panel Functions
set.seed(10)
x <- rnorm(100)
f <- rep(0:1, each = 50)
y <- x + f - f * x + rnorm(100, sd = 0.5)
f <- factor(f, labels, c ("Group 1", "Group 2"))
xyplot(y~x | f, layout = c(2,1)) # plot with 2 panels
# Custom panel function
xyplot(y~x | f, panel = function(x, y, ...){
panel.xyplot(x, y, ...) # 1st call the default panel function for xyplot
panel.abline(h = median(y), lty = 2) # add a horizontal line at the median
})
# Custom panel function
xyplot(y~x | f, panel = function(x, y, ...){
panel.xyplot(x, y, ...) # 1st call the default panel function for xyplot
panel.lmline(x, y, col = 2) # overlay a simple linear regression line
})
|
43445899dd3c124dd5a260b25f36a345e7580ddf | 87a10b6ceddd21d6d0195f79648fa2fab473638d | /Food Services by County.R | 33f2cd065815fd47109bc916140e42e7a6b7dc7b | [] | no_license | vineetdcunha/Data_visualization | 8ef1b63a47e6f2082567b3f45367a096ed28ab8b | 83682c55c0d2a6ff25dbc97311296a2ca353071a | refs/heads/main | 2023-01-23T10:05:43.196622 | 2020-11-24T18:43:08 | 2020-11-24T18:43:08 | 313,099,542 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,237 | r | Food Services by County.R | library(tidyverse)
library(geojsonio)
library(RColorBrewer)
library(rgdal)
library(sf)
library(broom)
# Download the Hexagones boundaries at geojson format here: https://team.carto.com/u/andrew/tables/andrew.us_states_hexgrid/public/map.
spdf <- geojson_read("us_states_hexgrid.geojson", what = "sp")
# Bit of reformating
spdf@data = spdf@data %>%
mutate(google_name = gsub(" \\(United States\\)", "", google_name))
spdf@data = spdf@data %>% mutate(google_name = gsub(" \\(United States\\)", "", google_name))
spdf_fortified <- tidy(spdf, region = "google_name")
spdf_fortified_name <- tidy(spdf, region = "google_name")
FoodSrvcByCounty$county = as.character(FoodSrvcByCounty$County)
spdf_fortified$county = toupper(spdf_fortified$id)
spdf_fortified <- spdf_fortified %>%
left_join(. , FoodSrvcByCounty, by = c("county" = "county"))
head(spdf_fortified)
ggplot() +
geom_polygon(data = spdf_fortified,
aes(
x = long,
y = lat,
group = group,
fill = FoodServices.2007
)) +
theme_void() + labs(fill = 'Food Services - 2007',
title = "Food Services by State- 2007")
|
f2ebf5f917934428031c40c49ba1cdc6bc46b6b2 | 2e4afcf0f120a9d36ae9eee3c0d10df688c4cb37 | /js_RcircosPlotting.R | c5ae82f9f7fc29241b1673e5cccc5da4c3f5b937 | [] | no_license | CellFateNucOrg/afterMC-HiCplots | e7b83de0319371baa949696d1f34f1a1f19f3f2f | 8b3af55da3be3f60e3632a2052e46187d6e39a3b | refs/heads/master | 2022-02-11T07:54:11.044287 | 2022-01-31T17:53:41 | 2022-01-31T17:53:41 | 244,604,363 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 4,044 | r | js_RcircosPlotting.R | #' Prepare reads for plotting with Rcircos
#'
#' Input data frame should have column names ReadID, Chr, RefStart and RefEnd
#' @param readData - data frame of fragments detected in MC-HiC
#' @param readsToDraw - vector of read IDs
#' @return data frame with 6 columns for pairs of interacting loci
#' @export
prepareLinkData<-function(readData,readsToDraw) {
firstChr<-c()
firstStart<-c()
firstEnd<-c()
secondChr<-c()
secondStart<-c()
secondEnd<-c()
for (rd in readsToDraw) {
currentRead<-readData[readData$ReadID==rd,]
firstChr <- c(firstChr, currentRead$Chr[1:(dim(currentRead)[1]-1)])
firstStart <- c(firstStart, currentRead$RefStart[1:(dim(currentRead)[1]-1)])
firstEnd <- c(firstEnd, currentRead$RefEnd[1:(dim(currentRead)[1]-1)])
secondChr <- c(secondChr, currentRead$Chr[2:(dim(currentRead)[1])])
secondStart <- c(secondStart, currentRead$RefStart[2:(dim(currentRead)[1])])
secondEnd <- c(secondEnd, currentRead$RefEnd[2:(dim(currentRead)[1])])
}
RCircosLink<- data.frame(firstChr=firstChr,firstStart=firstStart,firstEnd=firstEnd,
secondChr=secondChr,secondStart=secondStart,secondEnd=secondEnd,
stringsAsFactors=F)
return(RCircosLink)
}
#' Prepare the core Rcircos plot for C. elegans data
#'
#' @param base.per.unit - integer for the size of the units that are plotted
#' @param chr.exclude - vector of names of chromosomes to exclude
#' @param track.inside - number of tracks to have inside the circle
#' @param track.outside - number of tracks to have outside the circle
#' @return plots ideogram
#' @export
baseRcircosCE<-function(base.per.unit=3000, chr.exclude=NULL, highlight.width=10, tracks.inside=1, tracks.outside=0){
Chrnames<-c("chrI","chrII","chrIII","chrIV","chrV","chrX","MtDNA") # used to get rid of mtDNA
ce11 <- list( "chrI" = 15072434,
"chrII" = 15279421,
"chrIII" = 13783801,
"chrIV" = 17493829,
"chrV" = 20924180,
"chrX" = 17718942,
"MtDNA" = 13794)
ce11.ideo<-data.frame(Choromsome=Chrnames,ChromStart=0,ChromEnd=unlist(ce11),Band=1,Stain="gvar")
cyto.info <- ce11.ideo
RCircos.Set.Core.Components(cyto.info, chr.exclude,tracks.inside, tracks.outside)
rcircos.params <- RCircos.Get.Plot.Parameters()
rcircos.params$base.per.unit<-base.per.unit
rcircos.params$chrom.width=0 #0.1
rcircos.params$highlight.width=highlight.width #1
RCircos.Reset.Plot.Parameters(rcircos.params)
RCircos.Set.Plot.Area()
par(mai=c(0.25, 0.25, 0.25, 0.25))
plot.window(c(-1.5,1.5), c(-1.5, 1.5))
RCircos.Chromosome.Ideogram.Plot()
}
#' Prepare a list of points of view for 4C
#'
#' Will use chromosome length to find positions at 20%, 50% and 80% of chromosome's
#' length to act as points of view for arms and center
#' @param chrLengthList - a named list with lengths of chromsomes
#' @param winSize - the size of the window around the POV for selecting interactions (must be an even number)
#' @return data.frame with points of view
#' @export
generatePOV<-function(chrLengthList=NULL,winSize=10000){
if (is.null(chrLengthList)){
chrLengthList <- list( "chrI" = 15072434,
"chrII" = 15279421,
"chrIII" = 13783801,
"chrIV" = 17493829,
"chrV" = 20924180,
"chrX" = 17718942)
chrLengthList<-(unlist(chrLengthList))
}
left<-round(0.2*chrLengthList/1000,0)*1000
center<-round(0.5*chrLengthList/1000,0)*1000
right<-round(0.8*chrLengthList/1000,0)*1000
names(left)<-paste(names(left),"left",sep="_")
names(right)<-paste(names(right),"right",sep="_")
names(center)<-paste(names(center),"center",sep="_")
POV<-data.frame(POVname=c(names(left),names(center),names(right)),
POVpos=c(left,center,right),row.names=NULL)
POV$chr<-gsub("_.*","",POV$POVname)
POV$start<-POV$POVpos-winSize/2
POV$end<-POV$POVpos+winSize/2
POV<-POV[order(POV$chr,POV$start),]
return(POV)
}
|
4504238bc35c861acdf4c5858cacd0d7f7fb23e5 | e48a5a75c97b0e8b4d3c3b3f7f8484173baa7a3d | /ui.r | f86dac8e8e4f5b2dc3def2fa10b7a15ca612d861 | [] | no_license | patilv/bb50citiesrank | 08f3e08509c33ee6d409a2c8b81139dee51e1a2f | 1a407ba3cbff264ae676cdb3fdcf51c4fae59fe6 | refs/heads/master | 2021-01-01T19:24:31.875817 | 2013-06-24T10:35:23 | 2013-06-24T10:35:23 | 10,894,910 | 0 | 1 | null | null | null | null | UTF-8 | R | false | false | 3,738 | r | ui.r | library(shiny)
shinyUI(pageWithSidebar(
headerPanel("50 Best US Cities of 2012 - Ranked by Bloomberg Businessweek - Location and Characteristics"), #Application title
# User input for determine characteristic to use for sizing dots on map
sidebarPanel(
selectInput("var1", "Characteristic 1:",
choices =c(
"Number of Bars"="Bars",
"Population"="Population",
"Number of Restaurants"="Restaurants",
"Number of Museums"="Museums",
"Number of Libraries"="Libraries",
"Number of Pro Sports Teams"="Pro.Sports.Teams",
"Parks acres per 1000 residents"="Park.acres.per.1000.residents",
"Number of Colleges"="Colleges",
"Percent with Graduate Degree"="Percent.with.Graduate.Degree",
"Median Household Income"="Median.household.income",
"Percent Unemployed"="Percent.unemployed")
),
HTML("<br><br>"),
h5(textOutput("hits")),# Hit counter Output
HTML("<font size=1>Hit counter courtesy: <a href = 'http://www.econometricsbysimulation.com/2013/06/more-explorations-of-shiny.html' target='_blank'> Francis Smart</a></font>"),
HTML("<br><br>Application code <a href = 'https://github.com/patilv/bb50cities' target='_blank'> is available here.</a>")
),
mainPanel(
HTML("<a href ='http://www.businessweek.com/articles/2012-09-26/san-francisco-is-americas-best-city-in-2012' target='_blank'> Original article </a> <font color='red'>
required one to click on 50 slides to find the best city --- Advertising is important, right? </font><hr> "),
tabsetPanel(
# viewing the map
tabPanel("Geographic Locations",HTML ("<div> <font color='red'>Please be patient for all dots to show. Actually, the 50th dot will be the best city :-) </font> <br>1. Color indicates rank of the city - LOWER VALUE IS BETTER (see the color coding below)
<br>2. Size of dot indicates value of selected Characteristic - LARGER dot is HIGHER value <br>
3. You can hover over the dots to know the city and value of the characteristic</div>"),htmlOutput("gvisgeoplot")
),
# for scatter plot
tabPanel ("Characteristics and Cities", HTML("<div> <font color='red'>This chart can be played around with in the following ways.</font><br>
1. The two small tabs on the top right show either bubble charts or bar graphs, depending on what's selected. <br>
2. The horizontal and vertical axes can be changed to other variables by clicking at existing axis labels (the arrow mark) <br>
3. Size of dot indicates Rank of the city - LARGER size is BETTER ranked. (The rank is 51 minus the displayed value. This shows as variable 'RankReordered' in list in the axes.
See below the plot for explanation.)<br>
4. You can hover over the dots to know the city and value of the characteristics<br>
5. Color is used to identify the city. No other purpose.<br><br></div>"),htmlOutput("scatterplot"),
HTML("<div><font size=1>Since ranking of 1 is better than 50, the default approach would've placed smaller dots for better ranked cities.
So, to have larger bubbles for better cities a reranking was done to have a higher value for better city (51 minus the 'real' rank).
This was done only for this plot.</font></div>")),
tabPanel("Data", htmlOutput("bestcitiesdata")) # viewing data
)))) |
54e5a7076b9ff98e8bbb8c87a838f75b5ddcc12c | e141aebdf1eee3f692848a88e6a4ef1db6b854a9 | /plot2.R | e18d95007a883045442199bd99e8dde1d245aeeb | [] | no_license | bobb72/ExData_Plotting1 | fd9fe3dea393c40cd6928919e9c35b3181d0ab3c | 94eacf8c495e3f6ac7337c047e0ee5da9a391ad2 | refs/heads/master | 2021-01-23T21:03:23.022982 | 2016-06-08T10:26:05 | 2016-06-08T10:26:05 | 60,681,145 | 0 | 0 | null | 2016-06-08T08:20:22 | 2016-06-08T08:20:22 | null | UTF-8 | R | false | false | 1,192 | r | plot2.R | # Here is the data for the project:
# https://d396qusza40orc.cloudfront.net/exdata%2Fdata%2Fhousehold_power_consumption.zip
# Create an R script called plot2.R that reproduces the plot as per course instructions
#setwd("C:/Users/Bob/Desktop/DS Specialization/4_Exploratory_Data_Analysis/w1_assignment/exdata_data_household_power_consumption")
#fileUrl <- "https://d396qusza40orc.cloudfront.net/exdata%2Fdata%2Fhousehold_power_consumption.zip"
#download.file(fileUrl,destfile="../Electric power consumption.zip")
#unzip(zipfile="../Electric power consumption.zip")
# Read file:
house_pwr_cons <- read.table("household_power_consumption.txt", header = TRUE, sep = ";", stringsAsFactors=FALSE, na.strings = "?")
house_pwr_cons_sub <- house_pwr_cons[house_pwr_cons$Date %in% c("1/2/2007","2/2/2007"), ]
rm("house_pwr_cons")
# Create 2nd plot:
globalActivePower <- as.numeric(house_pwr_cons_sub$Global_active_power)
timeStamp <- strptime(paste(house_pwr_cons_sub$Date, house_pwr_cons_sub$Time, sep=" "), "%d/%m/%Y %H:%M:%S")
png("ExData_Plotting1/plot2.png", width=480, height=480)
plot(timeStamp, globalActivePower, type="l", xlab="", ylab="Global Active Power (kilowatts)")
dev.off()
|
2da797878ff65d5df11d5bb9a8c0f1a06bd422b4 | 9cc0308c75c50b5869c783fdd83fb00d36703e98 | /R/Time_Series.R | 8daabdead67b75dfb44f50408048f5711e61f9c6 | [] | no_license | SantiagoGallon/TimeSeries | 2cbf0157cb195bf8b213357f915739cf91d04592 | c40bdd4733796d5ba6f22e0a799afdadd56937da | refs/heads/master | 2020-09-01T01:47:32.635616 | 2019-11-01T22:59:15 | 2019-11-01T22:59:15 | 218,847,481 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 24,672 | r | Time_Series.R | rm(list = ls())
ls()
library(astsa)
library(forecast)
library(fma)
# Import data
# Índice Accionario de Capitalización Bursatil, Bolsa de Valores de Colombia (daily 15/01/2008 - 22/06/2016)
data <- read.table("/Users/Santiago/Dropbox/Teaching/Time Series/data/colcap.txt", sep="\t", header=TRUE)
x <- ts(data[,2], start=c(2008,15,1), end=c(2016,22,6), frequency=256.25)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/colcap.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(1,1), mar=c(2,4,0.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab="Index", xlab ="")
dev.off()
# Índice mensual de actividad económica -IMACO- (monthly 01/1992 - 05/2015)
data <- read.table("/Users/Santiago/Dropbox/Teaching/Time Series/data/imaco.txt", sep="\t", header=TRUE)
x <- ts(data[,3], start=c(1992,1), end=c(2015,5), frequency=12)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/imaco.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(1,1), mar=c(2,4,0.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab="Index", xlab ="")
dev.off()
# Consumption Price Index (monthly jul/54 - may/16)
data <- read.table("/Users/Santiago/Dropbox/Teaching/Time Series/data/ipc.txt", sep="\t", header=TRUE)
x <- ts(data[,2], start=c(1954,7), end=c(2016,5), frequency=12)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/ipc.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(1,1), mar=c(2,4,0.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab="Index", xlab ="")
dev.off()
# Industrial Production Index (monthly 01/1980 - 04/2016)
data <- read.table("/Users/Santiago/Dropbox/Teaching/Time Series/data/ipi.txt", sep="\t", header=TRUE)
x <- ts(data[,2], start=c(1980,1), end=c(2016,4), frequency=12)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/ipi.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(1,1), mar=c(2,4,0.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab="Index", xlab ="")
dev.off()
# Gross Domestic Product (quarterly 2000Q1-2015Q4)
data <- read.table("/Users/Santiago/Dropbox/Teaching/Time Series/data/gdp.txt", sep="\t", header=TRUE)
x <- ts(data[,2]/1000, start=c(2000,1), end=c(2015,4), frequency=4)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/gdp.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(1,1), mar=c(2,4.5,0.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab="Thousands of millions (pesos)", xlab ="")
dev.off()
# Exchange rate (pesos/dollar) (daily 27/11/1991-24/06/2016)
data <- read.table("/Users/Santiago/Dropbox/Teaching/Time Series/data/trm.txt", sep="\t", header=TRUE)
x <- ts(data[,2], start=c(1991,11), end=c(2016,6), frequency=359.24)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/trm.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(1,2), mar=c(2,4,0.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab="", xlab ="")
ts.plot(diff(log(x)), ylab="", xlab ="")
dev.off()
# West Texas Intermediate - WTI - Crude Oil Price (1986 -2015). Source: http://www.eia.gov/
library(xlsx)
data_d <- read.xlsx("/Users/santiagogallon/Dropbox/Teaching/Time Series/data/wti.xls", header=TRUE, sheetIndex = 1)
data_w <- read.xlsx("/Users/santiagogallon/Dropbox/Teaching/Time Series/data/wti.xls", header=TRUE, sheetIndex = 2)
data_m <- read.xlsx("/Users/santiagogallon/Dropbox/Teaching/Time Series/data/wti.xls", header=TRUE, sheetIndex = 3)
data_y <- read.xlsx("/Users/santiagogallon/Dropbox/Teaching/Time Series/data/wti.xls", header=TRUE, sheetIndex = 4)
x_d <- ts(data_d[1:7568,2], start=c(1986,1,2), frequency=365)
x_w <- ts(data_w[1:1565,2], start=c(1986,1), end=c(2015,52), frequency=53)
x_m <- ts(data_m[1:360,2], start=c(1986,1), end=c(2015,12), frequency=12)
x_y <- ts(data_y[,2], start=c(1986), end=2015, frequency=1)
postscript("/Users/santiagogallon/Dropbox/Teaching/Time Series/Slides/wti.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(2,2), mar=c(2,4.2,1.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x_d, main= "Daily", ylab="US dollars per barrel", xlab ="")
ts.plot(x_w, main= "Weekly", ylab="US dollars per barrel", xlab ="")
ts.plot(x_m, main= "Monthly", ylab="US dollars per barrel", xlab ="")
ts.plot(x_y, main= "Yearly", ylab="US dollars per barrel", xlab ="")
dev.off()
# Monthly totals (in thousands) of international airline passengers between 1949 and 1960. Source: Box-Jenkins
postscript("/Users/santiagogallon/Dropbox/Teaching/Time Series/Slides/air.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(1,1), mar = c(2,5.4,0.2,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(AirPassengers, ylab="", xlab="")
title(ylab="No. of passengers (in thousands)", line=4)
dev.off()
# L.A. Pollution Study. The scales are different, mortality, temperature, and emissions (weekly 1970 - 1980)
postscript("/Users/santiagogallon/Dropbox/Teaching/Time Series/Slides/pollu.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(3,1), mar=c(0,4.8,0,0.2), oma=c(4,0,0.2,0), las=1, cex.axis=1.5, cex.lab=1.5, tcl=-.3)
plot(cmort, ylab="No. of Deaths", xaxt="no", type='n')
#grid(lty=1, col=gray(.9))
lines(cmort, col="blue")
#text(1974, 132, 'Bad Year', col=rgb(.5,0,.5), cex=1.25) # just for fun
#arrows(1973.5, 130, 1973, 127, length=0.05, angle=30, col=rgb(.5,0,.5))
plot(tempr, ylab=expression(~Temperature~(degree~F)), xaxt="no", type='n')
#grid(lty=1, col=gray(.9))
lines(tempr, col="red")
plot(part, ylab="Emissions (PPM)")
#grid(lty=1, col=gray(.9))
title(xlab="Time (week)", outer=TRUE)
dev.off()
# Seasonal series
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/seas.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(3,2), mar=c(4,4.3,0.7,0.7), las=0, cex.axis=1.5, cex.lab=1.5, tcl=-.3)
plot(gas, ylab="AU monthly gas production") # Australian monthly gas production: 1956–1995
plot(taylor, ylab="UK half-hourly electricity demand") # Half-hourly electricity demand in England and Wales from Monday 5 June 2000 to Sunday 27 August 2000
plot(wineind, ylab="AU monthly wine sales") # Australian total wine sales by wine makers in bottles <= 1 litre. Jan 1980 – Aug 1994
plot(USAccDeaths, ylab="US monthly accidental deaths") # Accidental Deaths in the US 1973-1978
plot(milk, ylab="Monthly milk production") #Monthly milk production per cow
plot(part, ylab="LA weekly particulate levels") # Particulate levels from the LA pollution study
#plot(birth, ylab="US Monthly live births") # Monthly live births (adjusted) in thousands for the United States, 1948-1979.
#plot(woolyrnq) # Quarterly production of woollen yarn in Australia: tonnes. Mar 1965 – Sep 1994
#plot(ldeaths)
#plot(unemp)# Monthly U.S. Unemployment series 1948-1978
dev.off()
# Simulated AR(1) processes
x_1 <- arima.sim(list(order=c(1,0,0), ar= 0.8), n=250)
x_2 <- arima.sim(list(order=c(1,0,0), ar=-0.8), n=250)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/ar_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(3,2), mar=c(4.2,4.5,1.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x_1, ylab=expression(X[t]), main=(expression(AR(1)~~~phi==0.8)))
Acf(x_1, main="", xlab=expression(k))
Pacf(x_1, main="", ylab="PACF", xlab=expression(k))
#
ts.plot(x_2, ylab=expression(X[t]), main=(expression(AR(1)~~~phi==-0.8)))
Acf(x_2, main="", xlab=expression(k))
Pacf(x_2, main="", ylab="PACF", xlab=expression(k))
#
dev.off()
# Simulated AR(2) processes
x_1 <- arima.sim(list(order=c(2,0,0), ar=c(0.5,-0.8)), n=250)
x_2 <- arima.sim(list(order=c(2,0,0), ar=c(0.6,0.3)), n=250)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/ar_2.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(3,2), mar=c(4.2,4.5,1.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x_1, ylab=expression(X[t]), main=(expression(AR(2)~~~phi[1]==0.5~~phi[2]==-0.8)))
Acf(x_1, main="", xlab=expression(k))
Pacf(x_1, main="", ylab="PACF", xlab=expression(k))
#
ts.plot(x_2, ylab=expression(X[t]), main=(expression(AR(2)~~~phi[1]==0.6~~phi[2]==0.3)))
Acf(x_2, main="", xlab=expression(k))
Pacf(x_2, main="", ylab="PACF", xlab=expression(k))
#
dev.off()
# Stationary regions for an AR(2) process
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/roots.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(1,1), mar=c(4,4.4,0.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
curve(1+x, -2,2, ylim=c(-0.929,0.929), xlab=expression(phi[1]), ylab=expression(phi[2]))
curve(1-x, -2,2, add=TRUE)
curve((-x^2)/4, -2,2, add=TRUE)
text(0,.4,"Real roots", cex=1.5)
text(0,.32,expression(abs(lambda[1])<1~~~abs(lambda[2])<1), cex=1.5)
text(-1.05,.4,expression(phi[2]==1+phi[1]), cex=1.5)
arrows(-0.8,0.4, -0.6,0.4, length=0.1, angle=15)
text(1.05,.4,expression(phi[2]==1-phi[1]), cex=1.5)
arrows(0.8,0.4, 0.6,0.4, length=0.1, angle=15)
text(0, -.5,"Complex roots", cex=1.5)
text(0,-0.59,expression(abs(lambda[1])<1~~~abs(lambda[2])<1), cex=1.5)
text(0,-0.13,expression(phi[1]^2+4*phi[2]==0), cex=1.5)
arrows(0,-0.08, 0,0, length=0.1, angle=15)
abline(0.0015,0, col="gray70", lty=2)
dev.off()
# Simulated MA(1) processes
x_1 <- arima.sim(list(order=c(0,0,1), ma= 0.8), n=250)
x_2 <- arima.sim(list(order=c(0,0,1), ma=-0.8), n=250)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/ma_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(3,2), mar=c(4.2,4.5,1.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x_1, ylab=expression(X[t]), main=(expression(MA(1)~~~theta==0.8)))
Acf(x_1, main="", xlab=expression(k))
Pacf(x_1, main="", ylab="PACF", xlab=expression(k))
#
ts.plot(x_2, ylab=expression(X[t]), main=(expression(MA(1)~~~theta==-0.8)))
Acf(x_2, main="", xlab=expression(k))
Pacf(x_2, main="", ylab="PACF", xlab=expression(k))
#
dev.off()
# Simulated MA(2) processes
x_1 <- arima.sim(list(order=c(0,0,2), ma=c(0.3,0.3)), n=250)
x_2 <- arima.sim(list(order=c(0,0,2), ma=c(0.6,-0.4)), n=250)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/ma_2.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(3,2), mar=c(4.2,4.5,1.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x_1, ylab=expression(X[t]), main=(expression(MA(2)~~~theta[1]==0.3~~theta[2]==0.3)))
Acf(x_1, main="", xlab=expression(k))
Pacf(x_1, main="", ylab="PACF", xlab=expression(k))
#
ts.plot(x_2, ylab=expression(X[t]), main=(expression(MA(2)~~~theta[1]==0.6~~theta[2]==-0.4)))
Acf(x_2, main="", xlab=expression(k))
Pacf(x_2, main="", ylab="PACF", xlab=expression(k))
#
dev.off()
# Simulated ARMA(1,1) processes
x_1 <- arima.sim(list(order=c(1,0,1), ar=0.8, ma=0.6), n=250)
x_2 <- arima.sim(list(order=c(1,0,1), ar=0.9, ma=-0.5), n=250) # ar=0.9, ma=-0.5
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/arma_1_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(3,2), mar=c(4.2,4.5,1.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x_1, ylab=expression(X[t]), main=(expression(ARMA(1,1)~~~phi[1]==0.8~~theta[1]==0.6)))
Acf(x_1, main="", xlab=expression(k))
Pacf(x_1, main="", ylab="PACF", xlab=expression(k))
#
ts.plot(x_2, ylab=expression(X[t]), main=(expression(ARMA(1,1)~~~phi[1]==0.9~~theta[1]==-0.5)))
Acf(x_2, main="", xlab=expression(k))
Pacf(x_2, main="", ylab="PACF", xlab=expression(k))
#
dev.off()
# Stationary AR(1) and Random walk processes with drift alpha
n <- 250
delta <- 0.5
phi <- 0.9
#
x <- y <- z <- numeric(n)
t <- seq(1,n,1)
for(i in 2:n){
x[i] <- delta + x[i-1] + rnorm(1,0,1)
y[i] <- delta + phi*y[i-1] + rnorm(1,0,1)
}
# Deterministic trend
for(i in 1:n){
z[i] <- delta*t[i] + rnorm(1,0,2)
}
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/rw.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(1,1), mar=c(4.2,3.3,0.2,0.5), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab="", main="")
lines(delta*t, lty=2, col ="gray70")
text(190,107,expression(E(Y[t])==delta~t), cex=1.1)
arrows(194,105, 200,101, length=0.1, angle=15)
lines(y, lty=5)
abline(h=mean(y), lty=2, col ="gray70")
text(110,12,expression(E(X[t])==delta/(1-phi[1])), cex=1.1)
arrows(106,10, 110,6, length=0.1, angle=15)
legend("topleft",title="",legend=c("Random Walk with Drift","Stationary AR(1) with Drift"), lty=c(1,5), cex=1.2, bty="n")
dev.off()
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/acf_rw.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(1,2), mar=c(4.2,4.5,3,0.5), las=1, cex.axis=1.5, cex.lab=1.5)
Acf(x, ylab="ACF", xlab=expression(k), main="Random Walk with Drift")
Acf(y, ylab="ACF", xlab=expression(k), main="Stationary AR(1) with Drift", ylim=c(-0.2,1))
dev.off()
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/dt.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(1,2), mar=c(4.2,4.5,1.5,0.3), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(z, ylab=expression(X[t]), main="Trend-stationary")
lines(predict(lm(z~t-1)), col=2)
ts.plot(residuals(lm(z~t-1)), ylab=expression(X[t]-DT[t]), main="Detrending")
dev.off()
summary(lm(z~t-1))
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/rw_ts.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(1,2), mar=c(4.2,4.5,1.5,0.3), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(z, ylab=expression(X[t]), main="Trend-stationary")
ts.plot(x, ylab="", main="Random walk with drift")
dev.off()
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/rw_diff.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(1,2), mar=c(4.2,4.5,1.5,0.5), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, ylab=(expression(X[t])), main="Random Walk with Drift")
ts.plot(diff(x), ylab=(expression(Delta~X[t])), main="First difference")
dev.off()
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/acf_rw_diff.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(1,2), mar=c(4.2,4.5,3,0.5), las=1, cex.axis=1.5, cex.lab=1.5)
Acf(x, ylab="ACF", xlab=expression(k), main="Random Walk with Drift")
Acf(diff(x), ylab="ACF", xlab=expression(k), main="First difference", ylim=c(-0.2,1))
dev.off()
# x <- arima.sim(list(order=c(1,0,0), ar=0.9), mean=0.5, n=250)
# y <- arima.sim(list(order=c(0,1,0)), mean=0.5, n=250)
# y <- 1:101 * b +arima.sim(list(order=c(1,1,0), ar=0), n=250)
# Simulated ARIMA(1,1,1) processes
x_1 <- arima.sim(list(order=c(1,1,0), ar=0.8), n=250) # ARIMA(1,1,0) or ARI(1,1)
x_2 <- arima.sim(list(order=c(0,1,1), ma=0.75), n=250) # ARIMA(0,1,1) or IMA(1,1)
x_3 <- arima.sim(list(order=c(1,1,1), ar=0.9, ma=0.5), n=250) # ARIMA(1,1,1)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/arima_1_1_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(3,3), mar=c(4.2,4.5,1.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x_1, ylab=expression(X[t]), main=(expression(ARIMA(1,1,0)~~~phi[1]==0.8)))
Acf(x_1, main="", xlab=expression(k))
Pacf(x_1, main="", ylab="PACF", xlab=expression(k))
#
ts.plot(x_2, ylab=expression(X[t]), main=(expression(ARIMA(0,1,1)~~~theta[1]==0.75)))
Acf(x_2, main="", xlab=expression(k))
Pacf(x_2, main="", ylab="PACF", xlab=expression(k))
#
ts.plot(x_3, ylab=expression(X[t]), main=(expression(ARIMA(1,1,1)~~~phi[1]==0.9~~theta[1]==0.5)))
Acf(x_3, main="", xlab=expression(k))
Pacf(x_3, main="", ylab="PACF", xlab=expression(k))
#
dev.off()
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/diff_arima_1_1_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfcol=c(3,3), mar=c(4.2,4.5,1.5,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(diff(x_1), ylab=expression(Delta~X[t]), main=(expression(Delta~X[t]~~of~~ARIMA(1,1,0)~~~phi[1]==0.8)))
Acf(diff(x_1), main="", xlab=expression(k))
Pacf(diff(x_1), main="", ylab="PACF", xlab=expression(k))
#
ts.plot(diff(x_2), ylab=expression(Delta~X[t]), main=(expression(Delta~X[t]~~of~~ARIMA(0,1,1)~~~theta[1]==0.75)))
Acf(diff(x_2), main="", xlab=expression(k))
Pacf(diff(x_2), main="", ylab="PACF", xlab=expression(k))
#
ts.plot(diff(x_3), ylab=expression(Delta~X[t]), main=(expression(Delta~X[t]~~of~~ARIMA(1,1,1)~~~phi[1]==0.9~~theta[1]==0.5)))
Acf(diff(x_3), main="", xlab=expression(k))
Pacf(diff(x_3), main="", ylab="PACF", xlab=expression(k))
#
dev.off()
# Seasonal Processes
# Seasonal AR(1) model
Phi <- c(rep(0,11),0.9)
sAR <- arima.sim(list(order=c(12,0,0), ar=Phi), n=37)
sAR <- ts(sAR, freq=12)
dev.off()
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/sar_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
layout(matrix(c(1,2, 1,3), nc=2))
par(mar=c(3,3.5,2,0.2), mgp=c(1.6,.6,0), las=0, cex.axis=1.5, cex.lab=1.5)
plot(sAR, axes=FALSE, main="Seasonal AR(1)", ylab=expression(X[t]), xlab="year", type="c")
months = c("J","F","M","A","M","J","J","A","S","O","N","D")
points(sAR, pch=months, cex=1, font=1, col=1:12)
axis(1, 1:4)
abline(v=1:4, lty=2, col=gray(.6))
axis(2)
box()
ACF <- ARMAacf(ar=Phi, ma=0, 100)[-1]
PACF <- ARMAacf(ar=Phi, ma=0, 100, pacf=TRUE)
plot(ACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
plot(PACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
dev.off()
# Seasonal MA(1) model
Theta <- c(rep(0,11),0.5)
sMA <- arima.sim(list(order=c(0,0,12), ma=Theta), n=37)
sMA <- ts(sMA, freq=12)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/sma_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
layout(matrix(c(1,2, 1,3), nc=2))
par(mar=c(3,3.5,2,0.2), mgp=c(1.6,.6,0), las=0, cex.axis=1.5, cex.lab=1.5)
plot(sMA, axes=FALSE, main="Seasonal MA(1)", ylab=expression(X[t]), xlab="year", type="c")
months = c("J","F","M","A","M","J","J","A","S","O","N","D")
points(sMA, pch=months, cex=1, font=1, col=1:12)
axis(1, 1:4)
abline(v=1:4, lty=2, col=gray(.6))
axis(2)
box()
ACF <- ARMAacf(ar=0, ma=Theta, 100)[-1]
PACF <- ARMAacf(ar=0, ma=Theta, 100, pacf=TRUE)
plot(ACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
plot(PACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
dev.off()
# Seasonal ARMA(1,1) model
Phi <- c(rep(0,11),0.8)
Theta <- c(rep(0,11),-0.5)
sARMA <- arima.sim(list(order=c(12,0,12), ar=Phi, ma=Theta), n=37)
sARMA <- ts(sARMA, freq=12)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/sarma_1_1.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
layout(matrix(c(1,2, 1,3), nc=2))
par(mar=c(3,3.5,2,0.2), mgp=c(1.6,.6,0), las=0, cex.axis=1.5, cex.lab=1.5)
plot(sARMA, axes=FALSE, main="Seasonal ARMA(1,1)", ylab=expression(X[t]), xlab="year", type="c")
months = c("J","F","M","A","M","J","J","A","S","O","N","D")
points(sARMA, pch=months, cex=1, font=1, col=1:12)
axis(1, 1:4)
abline(v=1:4, lty=2, col=gray(.6))
axis(2)
box()
ACF <- ARMAacf(ar=Phi, ma=Theta, 100)[-1]
PACF <- ARMAacf(ar=Phi, ma=Theta, 100, pacf=TRUE)
plot(ACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
plot(PACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
dev.off()
# Seasonal Multiplicative ARMA(0,1)x(1,0)_12
Phi <- c(rep(0,11),0.8)
theta <- -0.5
smARMA <- arima.sim(list(order=c(12,0,1), ar=Phi, ma=theta), n=37)
smARMA <- ts(smARMA, freq=12)
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/sarma_1_1_1_0.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
layout(matrix(c(1,2, 1,3), nc=2))
par(mar=c(3,3.5,2,0.2), mgp=c(1.6,.6,0), las=0, cex.axis=1.5, cex.lab=1.5)
plot(smARMA, axes=FALSE, main="Seasonal Multiplicative ARMA(0,1)x(1,0)_12", ylab=expression(X[t]), xlab="year", type="c")
months = c("J","F","M","A","M","J","J","A","S","O","N","D")
points(smARMA, pch=months, cex=1, font=1, col=1:12)
axis(1, 1:4)
abline(v=1:4, lty=2, col=gray(.6))
axis(2)
box()
ACF <- ARMAacf(ar=Phi, ma=theta, 100)[-1]
PACF <- ARMAacf(ar=Phi, ma=theta, 100, pacf=TRUE)
plot(ACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
plot(PACF, axes=FALSE, type="h", xlab=expression(k), ylim=c(-1,1))
axis(1, seq(0,96,12))
axis(2)
abline(h=0)
box()
dev.off()
# Deterministic Seasonallity
n <- 160
x <- numeric(n)
t <- ts(seq(1,n,1), freq=4)
S <- seasonaldummy(t)
s4 <- 1-rowSums(S)
S <- cbind(S,s4)
for(i in 1:n){
x[i] <- 6*S[i,1] + 8*S[i,2] - 4*S[i,3] + 5*S[i,4] + rnorm(1,0,2)
}
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/ds.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(2,2), mar=c(4.2,4.5,2.7,0.3), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(x, main=expression(X[t]), ylab="")
lines(predict(lm(x~S-1)), col=2)
ts.plot(residuals(lm(x~t+S-1)), main=expression(X[t]-sum(hat(beta)[j]~D[jt], j==1, s)), ylab="")
Acf(x, ylab="ACF", xlab=expression(k), main="", ylim=c(-1,1))
Acf(residuals(lm(x~t+S-1)), ylab="ACF", xlab=expression(k), main="", ylim=c(-1,1))
dev.off()
# Seasonal Multiplicative ARIMA(0,1,1)x(0,1,1)_12
library(forecast)
model <- Arima(ts(rnorm(100),freq=12), order=c(0,1,1), seasonal=c(0,1,1), fixed=c(theta=-0.4, Theta=-0.6))
smARIMA <- simulate(model, nsim=48)
smARIMA <- ts(smARIMA[9:44], freq=12)
dev.off()
layout(matrix(c(1,2, 1,3), nc=2))
par(mar=c(3,3.5,2,0.2), mgp=c(1.6,.6,0), las=0, cex.axis=1.5, cex.lab=1.5)
plot(smARIMA, axes=FALSE, main="Seasonal Multiplicative ARMA(0,1)x(1,0)_12", ylab=expression(X[t]), xlab="year", type="c")
months = c("J","F","M","A","M","J","J","A","S","O","N","D")
points(smARIMA, pch=months, cex=1, font=1, col=1:12)
axis(1, 1:4)
abline(v=1:4, lty=2, col=gray(.6))
axis(2)
box()
Acf(smARIMA, main="", xlab=expression(k), ylim=c(-1,1))
Pacf(smARIMA, main="", ylab="PACF", xlab=expression(k), ylim=c(-1,1))
# Also the "sarima" package can be used
library(sarima)
sAR <-sim_sarima(n=48, model=list(sar=0.9, nseasons=12)) # SAR(1)
sMA <-sim_sarima(n=48, model=list(sma=0.5, nseasons=12)) # SMA(1)
sARMA <-sim_sarima(n=48, model=list(sar=0.8, sma=-0.5, nseasons=12)) # SARMA(1,1)
smARMA <-sim_sarima(n=48, model=list(ma=-0.5, sar=0.8, nseasons=12)) # Seasonal Multiplicative ARMA(0,1)x(1,0)_12
smARIMA <- sim_sarima(n=48, model=list(ma=-0.4, iorder=1, siorder=1, sma=-0.6, nseasons=12)) # Seasonal Multiplicative ARIMA(0,1,1)x(0,1,1)_12
plot(smARIMA[13:48], axes=FALSE, main="Seasonal AR(1)", ylab=expression(X[t]), xlab="time", type="c")
months = c("J","F","M","A","M","J","J","A","S","O","N","D")
points(smARIMA[13:48], pch=months, cex=1, font=1, col=1:12)
axis(1, 0:36)
abline(v=c(12,24,36), lty=2, col=gray(.6))
axis(2)
box()
# Spurious Regression
# Le be two random walk processes
n <- 250
#
x <- y <- numeric(n)
t <- seq(1,n,1)
for(i in 2:n){
x[i] <- x[i-1] + rnorm(1,0,1)
y[i] <- y[i-1] + rnorm(1,0,1)
}
postscript("/Users/Santiago/Dropbox/Teaching/Time Series/Slides/spurious.eps",width = 15.5, height = 8.5, horizontal = TRUE, onefile = FALSE, paper = "a4")
par(mfrow=c(2,2), mar=c(4,4.2,1,0.2), las=1, cex.axis=1.5, cex.lab=1.5)
ts.plot(y, main=expression(y[t]==y[t-1]+u[t]), ylab="", xlab="")
ts.plot(x, main=expression(x[t]==x[t-1]+v[t]), ylab="", xlab="")
plot(x,y, xlab = expression(x[t]), ylab=expression(y[t]))
abline(lm(y~x), col="red")
ts.plot(residuals(lm(y~x)), main=expression(epsilon[t]==y[t]+2.6917-0.5055*x[t]), ylab="", xlab="")
dev.off()
#fit <- Arima(foo, order=c(0,1,1), seasonal=c(0,1,1))
#model <- Arima(ts(rnorm(100),freq=4), order=c(1,1,1), seasonal=c(1,1,1), fixed=c(phi=0.5, theta=-0.4, Phi=0.3, Theta=-0.2))
# http://stackoverflow.com/questions/20273104/simulating-a-basic-sarima-model-in-r
#library(gmwm)
# Specify a SARIMA(2,1,1)(1,1,1)[12]
#mod = SARIMA(ar=c(.3,.5), i=1, ma=.1, sar=.2, si = 1, sma = .4, s = 12, sigma2 = 1.5)
# Generate the data
#xt2 = gen.gts(mod, 1e3)
|
1075e3cff2e78edf30dea16a30c5360b51512a3a | dc3665fa074c42cd25d3eca313b90f4ae4482520 | /vendor_behavior.R | 29c382822735c9e2c2d05a9f9dbadf9b15b404e9 | [] | no_license | andfdiazrod/darkweb_functions | 5f6a350e6902bfbb9a9ce8886425ed62c48dbf3e | b8f20f47c916494103a9f7f2f418ed2a39f80b6d | refs/heads/master | 2022-05-16T02:01:45.786947 | 2019-11-29T16:53:37 | 2019-11-29T16:53:37 | 216,660,996 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,605 | r | vendor_behavior.R | vendor_behavior <- function(df){
columns = c("day_format", 'vendor_name',"in_sample","appears", "consistency",
"relative_consistency")
vendor_consistency <- data.frame(matrix(ncol=length(columns)))
colnames(vendor_consistency) <- columns
vendor_consistency$day_format <- as.Date(vendor_consistency$day_format)
sum_up_to <- function(x){
return(unlist(lapply(1:length(x),function(y) sum(x[1:y]))))
}
for(vn in unique(df$vendor_name)){
vn_day <- data.frame(day_format = unique(as.Date(info_total %>%
filter(vendor_name == vn) %>% pull(day_format))),
appears = 1)
vn_day_range <- read.csv('time_series.csv', stringsAsFactors = F)
vn_day_range$day_format <- as.Date(vn_day_range$day_format)
vn_day_range <- vn_day_range %>%
filter(day_format %in% seq.Date(min(vn_day[,1]),
max(vn_day[,1]),1))
vn_day_range <- na.omit(left_join(vn_day_range, vn_day, by='day_format') %>%
replace_na(list(appears = -1)))
vn_day_range$consistency <- sum_up_to(vn_day_range$appears)
vn_day_range$relative_consistency <- vn_day_range$consistency/(1:nrow(vn_day_range))
vn_day_range$vendor_name <- vn
vendor_consistency <- rbind(vendor_consistency, vn_day_range)
}
return(vendor_consistency)
}
if(FALSE){
a = vendor_consistency %>% group_by(day_format) %>%
summarise(a = median(relative_consistency, na.rm=T))
plot(a, type='l')
b <- left_join(date_range,a,by='day_format')
}
|
69170b9ed1285a26df786dda9db67a76a136ec4b | f8c92559534dba1aaec173f86b22bfd2bff913bc | /Lecture 1/hw1_factor8.R | fbff896114e429129a6ddeff0a094d9cabb093c0 | [] | no_license | Zijie-Xia/GR5206-Introduction-to-Data-Science | 499f8b2999a194431891fb6c019e82542ca111bd | f017e55171a79f1833e8ad4fe0b9697a8634678e | refs/heads/master | 2020-12-07T07:06:56.675558 | 2020-01-08T22:01:39 | 2020-01-08T22:01:39 | 232,179,811 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,002 | r | hw1_factor8.R | # HW1: factor8
#
# 1. Create an ordered factor `f1` consist of letters 'a' to 'z' ordered alphabetically.
# 2. Create an ordered factor `f2` consist of letters 'a' to 'z' in descending alphabetical order.
# 3. Create a 30 elements, ordered factor `f3` consist of letters 'a' to 'z' followed by 4 NA. The order of `f3` is 'a'<...<'z'<NA.
# 4. Delete the element 'c' with the level 'c' and assign it to `f4`.
## Do not modify this line! ## Write your code for 1. after this line! ##
f1<-factor(letters[1:26],order=TRUE)
## Do not modify this line! ## Write your code for 2. after this line! ##
f2<-factor(letters[1:26],order=TRUE,levels=letters[26:1])
## Do not modify this line! ## Write your code for 3. after this line! ##
f3<-factor(c(letters[1:26],NA,NA,NA,NA),order=TRUE,levels=c(letters[1:26],NA),exclude=TRUE)
## Do not modify this line! ## Write your code for 4. after this line! ##
f4<-factor(c("a","b",letters[4:26],NA,NA,NA,NA),order=TRUE,levels=c("a","b",letters[4:26],NA),exclude=TRUE)
|
13b38bd70df1c10c75b0a356f69cc07a7161787a | facce126b08e76ad542ff63258afe1e327e2d563 | /cpp_adv_r_questions.R | 66e1e24427b58632a7d6798a712d4f4c9594f245 | [] | no_license | bweiher/r_cpp_learning | 7f7c1d3f989850e5e3f5deabfd7144269fd83def | f395529814df37e724702df6f14effb59226799f | refs/heads/master | 2020-04-01T04:10:29.847091 | 2018-12-17T02:20:44 | 2018-12-17T02:20:44 | 152,852,572 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,271 | r | cpp_adv_r_questions.R | # rcpp questions
library(Rcpp)
cppFunction("double f1(NumericVector x) {
int n = x.size();
double y = 0;
for(int i = 0; i < n; ++i) {
y += x[i] / n;
}
return y;
}")
x <- 1:10
for(g in seq_along(x)){
y = x[g] / length(x)
print(y)
}
f1(x)
median(x)
f1(c(1,2,3))
# Vector input, vector output
cppFunction('NumericVector pdistC(double x, NumericVector ys) {
int n = ys.size();
NumericVector out(n);
for(int i = 0; i < n; ++i) {
out[i] = sqrt(pow(ys[i] - x, 2.0));
}
return out;
}')
pdistC(0.5, runif(10))
cppFunction('NumericVector rowSumsC(NumericMatrix x) {
int nrow = x.nrow(), ncol = x.ncol();
NumericVector out(nrow);
for (int i = 0; i < nrow; i++) {
double total = 0;
for (int j = 0; j < ncol; j++) {
total += x(i, j);
}
out[i] = total;
}
return out;
}')
set.seed(1014)
x <- matrix(sample(100), 10)
rowSums(x)
# examples -----
# += addition operator
# mean function
cppFunction('double f1(NumericVector x) {
int n = x.size();
double y = 0;
for(int i = 0; i < n; ++i) {
y += x[i] ;
}
y = y / n;
return y;
}')
x <- 1:10
f1(x)
sum(x / length(x))
mean(x)
# cumsum ~
cppFunction('NumericVector f2(NumericVector x) {
int n = x.size();
NumericVector out(n);
out[0] = x[0];
for(int i = 1; i < n; ++i) {
out[i] = out[i - 1] + x[i];
}
return out;
}'
)
f2(x)
cumsum(x)
# does the bool x contain a TRUE value
# bool f3(LogicalVector x) {
# int n = x.size();
#
# for(int i = 0; i < n; ++i) {
# if (x[i]) return true;
# }
# return false;
# }
cppFunction('int f3(LogicalVector x) {
int n = x.size();
for(int i = 0; i < n; ++i) {
if (x[i]) return i;
}
return 99;
}')
f3(c(F,F,F))
f3(c(T,T,F))
f3(c(F,F,F,F,T))
f33(1:3)
cppFunction('int f4(Function pred, List x) {
int n = x.size();
for(int i = 0; i < n; ++i) {
LogicalVector res = pred(x[i]);
if (res[0]) return i + 1;
}
return 0;
}')
# NumericVector f5(NumericVector x, NumericVector y) {
# int n = std::max(x.size(), y.size());
# NumericVector x1 = rep_len(x, n);
# NumericVector y1 = rep_len(y, n);
#
# NumericVector out(n);
#
# for (int i = 0; i < n; ++i) {
# out[i] = std::min(x1[i], y1[i]);
# }
#
# return out;
# } |
d9cce87a0ab01cf0916523ea4b8c368636748c74 | a9c8e9612975e42f68e5a08b0d65fcbb5edb7616 | /plot3.R | 1dcc2f9f1da5febed5dfd9582e13349d76d37dce | [] | no_license | RaghavVacher/ExData_Plotting1 | e41b562dd975e12882f3b55b6eb46ae5543d2c4a | b665bc853f08ca05462953765239928f788caede | refs/heads/master | 2023-02-07T05:25:51.740550 | 2020-12-28T08:46:02 | 2020-12-28T08:46:02 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 846 | r | plot3.R | x <- read.table("C:\\Users\\Hp\\Downloads\\exdata_data_household_power_consumption\\household_power_consumption.txt", skip = 1, sep = ";")
colnames(x) <- c("Date", "Time", "Global_active_power","Global_reactive_power","Voltage","Global_intensity","Sub_metering_1","Sub_metering_2","Sub_metering_3")
sub <- subset(x, x$Date == "1/2/2007" | x$Date == "2/2/2007")
datetime <- strptime(paste(sub$Date, sub$Time), "%d/%m/%Y %H:%M:%S")
sub1 <- as.numeric(sub$Sub_metering_1)
sub2 <- as.numeric(sub$Sub_metering_2)
sub3 <- as.numeric(sub$Sub_metering_3)
plot(datetime, sub1, type = "n", ylab="Energy Submetering", xlab="")
lines(datetime, sub1)
lines(datetime, sub2, col = "red")
lines(datetime, sub3, col = "blue")
legend("topright", c("Submetering 1", "Submetering 2", "Submetering 3"), col = c("black", "red", "blue"), lty = 1)
|
89308d7bff06ed1304981eed274130bbda2cbe6c | 9478cff072f07ea24c94b233a96a3cdb30f27e95 | /basic_script.R | f4d600dd0387463be5f7450481f35a8bc04917d0 | [] | no_license | gozdebudak/r-programming-basics | fce5d091af385c25e4369e6bc1d82e2c42c5beba | 22cfeaf3f3302013d2c872dc694d53de1e1b92a6 | refs/heads/master | 2023-04-22T21:26:28.445300 | 2021-05-09T17:58:49 | 2021-05-09T17:58:49 | 365,807,244 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,158 | r | basic_script.R | data <- read.csv("sample_data.csv") # Reading CSV file and creating dataframe object
print(data) # Printing the data
print(is.data.frame(data)) # Checking if the data object is a dataframe
print(ncol(data)) # The column count of the data dataframe
print(nrow(data)) # The row count of the data dataframe
names(data) # The column names of data dataframe
print(data[1:2,]) # Printing the first 2 rows of the data frame
print(data[152:153,]) # Printing the last 2 rows of the data frame
print(data[47,"Ozone"]) # Printing the value of "Ozone" in the 47th row
# Missing values count in the "Ozone" column of this data frame
sum(is.na(data$Ozone))
# The mean value of "Ozone" column except missing values
mean(data$Ozone[complete.cases(data$Ozone)])
# Mean of the "Solar.R" column in the subset of rows of the data frame where
# "Ozone" values are above 31 and "Temp" values are above 90
mean(subset(x=data, subset= Ozone > 31 & Temp > 90)$Solar.R)
# The mean of "Temp" when "Month" is equal to 6
mean(subset(x=data, subset=Month==6)$Temp)
# The max value of "Ozone" when "Month" is equal to 5
max(subset(x=data, subset=Month==5)$Ozone, na.rm=TRUE)
|
6cf05033fe8aaab69014012d593bb736d2a3070b | 9ef445e42d40f7bedfb6091877a1c1ca8e2cb8d1 | /server.R | 016d1238d59e2f89c70d20134a994290ea56c311 | [] | no_license | marco-vene/datitalia | 03a1d0285ab4f3b6a646639078494acbc29b3060 | e02c4c6aacefee4d2a1b8b62a3f2f66af73d2454 | refs/heads/master | 2020-09-07T15:37:22.771643 | 2019-11-10T18:29:12 | 2019-11-10T18:29:12 | 220,829,641 | 1 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,161 | r | server.R | # Define the server
shinyServer( function(input, output) {
# output$trendPlot <- renderGirafe({
# ggiraph(code = print(trendLine(dati, input$gruppo, input$metrica)))
# })
output$trendPlot <- renderPlotly({
trendLine(dati, input$gruppo, input$metrica, input$periodo)
})
output$description <- renderUI({
if(input$metrica == "Popolazione")
{
includeMarkdown("def_popolazione.md")
}
else if(input$metrica == "Occupati")
{
includeMarkdown("def_occupati.md")
}
else if(input$metrica == "Disoccupati")
{
includeMarkdown("def_disoccupati.md")
}
else if(input$metrica == "Inattivi")
{
includeMarkdown("def_inattivi.md")
}
else if(input$metrica == "Tasso_Occupazione")
{
includeMarkdown("def_tasso_occ.md")
}
else if(input$metrica == "Tasso_Disoccupazione")
{
includeMarkdown("def_tasso_dis.md")
}
else if(input$metrica == "Tasso_Inattivita")
{
includeMarkdown("def_tasso_ina.md")
}
else{
NULL
}
})
})
|
b08f8c6ad9736f9a76e630d88ff43d8d938cfcc3 | ddb120b0aaa38527d4eded97552e63d1cad7fb9a | /Project3/server.R | eab74849a4929780e3558b5542a7a86097df77c7 | [] | no_license | dwatie/project3 | df858e17c541ca9039c154a47665734c8189c589 | 593d6da14ebf6c80ef2ab97a5b6d81fbe265954c | refs/heads/main | 2023-07-01T05:25:48.125878 | 2021-08-03T03:32:39 | 2021-08-03T03:32:39 | 390,863,936 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 5,039 | r | server.R | #
# This is the server logic of a Shiny web application. You can run the
# application by clicking 'Run App' above.
#
# Find out more about building applications with Shiny here:
#
# http://shiny.rstudio.com/
#
library(shiny)
library(shinydashboard)
library(tidyverse)
library(DT)
library(caret)
library(plotly)
library(ggplot2)
library(randomForest)
library(mathjaxr)
# Define server logic required to draw a histogram
shinyServer(function(input, output,session) {
collegeBall <- read.csv("cbb19.csv")
getData <- reactive({
wccB10 <- collegeBall %>% select(TEAM, CONF, G, W, X2P_O, X2P_D, X3P_O, X3P_D, TOR, TORD) %>% filter(CONF == c("WCC", "B10"))
})
output$wccB10Table <- renderDataTable({
getData()
})
output$plots <- renderPlot({
if (input$plotnum == "gmesPlot"){
G = ggplot(wccB10, aes(x = CONF, y = G)) + geom_bar(stat = "identity")
}
if (input$plotnum == "winsPlot"){
G = ggplot(wccB10, aes(x = CONF, y = W))+ geom_bar(position = "dodge",
stat = "identity",
aes(fill = TEAM))
}
if (input$plotnum == "densityPlot"){
G = ggplot(wccB10, aes(x = G)) + geom_density(aes(fill = CONF),
alpha = 0.5,
kernel = "gaussian")
}
if (input$plotnum == "boxPlot"){
G = plot_ly(wccB10, x = ~X3P_O, color = ~CONF, type = "box")
}
G
})
output$confsumms <- renderDataTable({
if (input$datasums == "confWins"){
M = wccB10 %>% group_by(CONF) %>%
summarise(Min = min(W),
Med = median(W),
Avg = mean(W),
Max = max(W),
StDev = sd(W))
}
if (input$datasums == "confTwo"){
M = wccB10 %>% group_by(CONF) %>%
summarise(Min = min(X2P_O),
Med = median(X2P_O),
Avg = mean(X2P_O),
Max = max(X2P_O),
StDev = sd(X2P_O))
}
if (input$datasums == "confThree"){
M = wccB10 %>% group_by(CONF) %>%
summarise(Min = min(X3P_O),
Med = median(X3P_O),
Avg = mean(X3P_O),
Max = max(X3P_O),
StDev = sd(X3P_O))
}
if (input$datasums == "confTor"){
M = wccB10 %>% group_by(CONF) %>%
summarise(Min = min(TOR),
Med = median(TOR),
Avg = mean(TOR),
Max = max(TOR),
StDev = sd(TOR))
}
M
set.seed(1)
train <- sample(1:nrow(wccB10), size = nrow(wccB10)*0.7)
test <- dplyr::setdiff(1:nrow(wccB10), train)
wccB10Train <- wccB10[train, ]
wccB10Test <- wccB10[test, ]
wccB10TrainA <-wccB10Train %>% select(CONF, G, W, X2P_O, X2P_D, X3P_O, X3P_D, TOR, TORD)
wccB10TestA <- wccB10Test %>% select(CONF, G, W, X2P_O, X2P_D, X3P_O, X3P_D, TOR, TORD)
gmesVar <- wccB10TrainA$G
two<- wccB10TrainA
rfFit <- train(W ~ input$gmesVar, data = wccB10TrainA,
method = "rf",
trControl = trainControl(method = "repeatedcv",
repeats = 3,
number = 10),
linout = TRUE,
tuneGrid = data.frame(mtry = 1:10),
data = wccB10TrainA)
bestLm <- lm(W ~ input$gmesVar, data = wccB10TrainA)
ClassFit <- train(W ~ input, data = wccB10TrainA,
method = "rpart",
preProcess = c("center", "scale"),
trControl = trCtrl)
models <- list(c(ranfor, linreg, classtree))
models[[1]] <<- rfFit
models[[2]] <<- bestLm
models[[3]] <-- classfit
output$info <- renderText({
p("The three modeling approaches that will be used are the Multiple Linear Regression Model, Classification Tree Model, and the Random Forest Model. They will be used to find a linear regression equation that is made up of a response variable which in this case will the wins variable, an intercept, and a combination of predictor variables.) Multiple Linear Regression Model pros and cons go here Classification Tree Model pros and cons go here Random Forest Model pros and cons go here")
})
})
})
|
02a84ccdc624056781ec1b970cb8184d48945ba0 | 68562f910349b41cdf4432c0921940f0513ab516 | /tests/testthat/test-style_xaringan.R | 1fc0ebd071fe53304582101422ddc11b8cf70c81 | [
"MIT"
] | permissive | gadenbuie/xaringanthemer | 2990406aff24a458695c6e4793c891dff5feb506 | 85091cd16af5a938b6d927ff5f6b0fe990ee0e63 | refs/heads/main | 2022-09-15T18:32:49.954381 | 2022-08-20T18:03:58 | 2022-08-20T22:47:52 | 129,549,154 | 446 | 28 | NOASSERTION | 2022-08-20T16:58:02 | 2018-04-14T19:44:17 | R | UTF-8 | R | false | false | 483 | r | test-style_xaringan.R |
test_that("style_xaringan() writes to specified outfile", {
tmpfile <- tempfile(fileext = ".css")
expect_equal(style_xaringan(outfile = tmpfile), tmpfile)
expect_true(file.exists(tmpfile))
expect_true(grepl("xaringanthemer", readLines(tmpfile)[3]))
})
test_that("style_xaringan() warns if base_font_size is not absolute", {
tmpfile <- tempfile(fileext = ".css")
expect_warning(
style_xaringan(outfile = tmpfile, base_font_size = "1em"),
"absolute units"
)
})
|
c831e18ce1ad8d2e4000e4d8f190a872bfffcdaf | 4b402d90385a6a291c4761d08adac6d5ce547d18 | /antweb.R | d199f2e25e93344bc0428b6dad21ae05c241ed55 | [] | no_license | karthik/antweb_paper | 273cf0425f3154f0afb742690c067c4c73e22b65 | 07337be82311cde0f16e69aa603649f48b5629e2 | refs/heads/master | 2020-05-27T12:35:37.434962 | 2014-10-23T15:05:48 | 2014-10-23T15:05:48 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,025 | r | antweb.R |
## @knitr counts
library(AntWeb)
genera <- aw_distinct("genus")$count
species <- aw_distinct("species")
species <- species$count
## @knitr how_many_species
madagascar <- aw_data(country = "Madagascar")
total_results <- madagascar$count
offset <- seq(0,ceiling(total_results), by = 1000)
madagascar_all <- lapply(offset, function(x) {
message(sprintf("finishing %s results", x))
output <- aw_data(country = "Madagascar", offset = x, quiet = TRUE)
return(output$data)
})
all_madagascar_data <- data.table::rbindlist(madagascar_all)
saveRDS(all_madagascar_data, file = "data/madagascar.rda", compress = 'xz')
write.csv(all_madagascar_data, file = "all_madagascar_data.csv")
## @knitr elevation_gradient
message("placeholder for ...")
## @knitr latitude_gradient
message("placeholder for ...")
## @knitr across_habitats
message("placeholder for ...")
## @knitr two_localities
message("placeholder for ...")
## @knitr endemism
message("placeholder for ...")
## @knitr accumulation
message("placeholder for ...") |
63e89e861bb9084efd36049f01cb602b882b7065 | 6d96dbaeb9e3985a278e81cacb92eabed0908e1e | /R/create_dsproject.R | 74783a7d4729075d0eb0944f9539d8278edca045 | [
"MIT"
] | permissive | cimentadaj/dsproj | f23368f11ab5dec53b4999ac1159ce31a2669361 | aa99fa025921cf82524064935185b63d5d71a5a8 | refs/heads/master | 2020-04-13T18:42:51.477873 | 2019-02-10T20:27:17 | 2019-02-10T20:27:17 | 163,382,997 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,680 | r | create_dsproject.R | #' Creates a template of folders and files for the 'ideal' data science project
#'
#' @param path A path where to create the project template. Can be relative, absolute and non existent.
#' @param open whether to open the RStudio project or not. Set to
#' FALSE by default
#'
#' @details
#' The function accepts a valid path (either relative or absolute) and applies these steps:
#'
#' @details text describing parameter inputs in more detail.
#' \itemize{
#' \item{"Folders"}{Creates folders code, data, report and misc inside \code{path}}
#' \item{"RProjects"}{Creates an R project in \code{path}}
#' \item{"Git"}{Initializes a Git repository in \code{path}}
#' \item{"Documentation"}{Adds a README.Rmd for project purposes}
#' \item{"Package dependency"}{Installs and loads \code{packrat} for package dependency management}
#' \item{"Fresh start"}{Restarts R and opens the newly created .Rproj with packrat loaded}
#' }
#'
#' @return Nothing, it creates and edits several folders and files in \code{path}
#' @export
#'
#' @examples
#'
#' \dontrun{
#' create_dsproject()
#' }
#'
create_dsproject <- function(path, open = FALSE) {
stopifnot(is.character(path))
path <- normalizePath(path, winslash = .Platform$file.sep, mustWork = FALSE)
dirs_create <- file.path(path, c("code", "data", "report", "misc"))
for (folder in dirs_create) dir.create(folder, recursive = TRUE)
print_styler('Created folder ', dirs_create)
if (!requireNamespace("rmarkdown", quietly = TRUE)) {
print_styler("Installing rmarkdown for reporting")
cat("\n")
utils::install.packages("rmarkdown")
}
usethis::proj_set(path, force = TRUE)
usethis::use_rstudio()
print_styler("Created RStudio project")
cat("\n")
r <- git2r::init(usethis::proj_get())
print_styler("Created git repository")
usethis::use_git_ignore(c(".Rhistory", ".RData", ".Rproj.user"))
# I think this adds .Rbuildignore -- exclude
# Add a custom readme for ds projects
usethis::use_readme_rmd(open = FALSE)
# Add a set of preinstalled packages for every project
print_styler("Installing packrat for package dependency")
cat("\n")
unloadNamespace('packrat')
utils::install.packages("packrat")
write("packrat::on()", file.path(path, ".Rprofile"), append = TRUE)
write("options(repos = c(CRAN = 'https://cran.rstudio.com'))", file.path(path, ".Rprofile"), append = TRUE)
initial_styler(paste0("Set packrat mode on as default in", crayon::blue(" .Rprofile")))
print_styler("Activating packrat project")
cat("\n")
packrat::init(path, infer.dependencies = FALSE, enter = FALSE)
if (open) rstudioapi::openProject(usethis::proj_get())
invisible(TRUE)
}
|
4df02e56965e31ec8a15aeeee2542ac2245c3df0 | d1a87fe12e6f3eba49346d4d89c8f1a931e8715a | /Face_update/face_update.R | 4cfcabecb6da360c0c19cf64d2af04875cb0445d | [] | no_license | danmrc/azul | 9f557876557c046112a9374fcd7fabf612271090 | 87752b17778368b63ff43054a56b83048cc973c4 | refs/heads/master | 2023-08-22T01:13:39.342663 | 2023-08-04T22:10:26 | 2023-08-04T22:10:26 | 141,443,005 | 3 | 1 | null | 2023-07-11T17:31:52 | 2018-07-18T14:04:22 | HTML | UTF-8 | R | false | false | 632 | r | face_update.R | parse_website <- function(url){
require(xml2)
pag <- read_html(url)
fs <- xml_find_all(pag,xpath = "//h3[@class= 'item-title']/a")
ss <- xml_attr(fs,"href")
return(ss)
}
checkBlog <- function(newList,oldList,token){
require(Rfacebook)
teste <- prod(newList == oldList)
if(teste==1){
return("No updates")
} else{
readline(prompt= "New post. Press [enter] to continue")
new_post_url <- newList[newList != oldList][1]
new_post_url <- paste0("https://azul.netlify.com/",new_post_url)
updateStatus("Novo post", token = token, link = new_post_url)
}
}
save(obj, file = "Face_update/Lista.Rdata")
|
1a5a569560105ee5717001f2ef97cc695ecf55e8 | 16cbcd4b55e9df1e91f2d69702790023c9cf6780 | /799435798.r | d3095c5443b480fb5b9b95b7ba5ed7f9e14ddc78 | [] | no_license | erex/MT3607-peer-review | 3f65c9a168f34e947fe0e531e773029384c19314 | bc0750e9a7fb5f2d0a7c7e35b34b3a80213d9fde | refs/heads/master | 2020-06-03T06:12:34.093705 | 2014-10-10T09:49:51 | 2014-10-10T09:49:51 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 6,095 | r | 799435798.r | #I confirm that the attached is my own work, except where clearly indicated in the text.
my.rnorm<-function(n,mean=0,sd=1){
#Purpose:
#Returns n pseudo-random variables from a normal distribution
#Inputs:
#n- number of observations: a numeric scalar,
#mean - mean: a numeric scalar with default 0,
#sd- standard deviation: a numeric scalar with default 1
#Outputs:
#a vector of n pseudo-random values from a normal distribution
#Stops the function and returns an error message if the arguments are invalid
if(is.numeric(n) & is.numeric(mean) & is.numeric(sd) ){
}else{
stop("invalid arguments")
}
if(sd < 0) stop("invalid arguments")
if(n <= 0) stop("invalid arguments")
if(n!=round(n)) stop("invalid arguments")
#Vector to contain the normally distributed deviates
normdev<-c(rep(0,n))
#for loop using the central limit method to find n normally distributed deviates
for(i in 1:n){
U<-runif(16,0,1)
x<-((sum(U)-8)*sqrt(12/16))*sd+mean
normdev[i]<-x
}
return(normdev)
}
my.rchisq<-function(n,df=1){
#Purpose:
#Returns n pseudo-random chi-squared distributed deviates
#Inputs:
#n- number of observations: a numeric scalar
#df-degrees of freedom: a numeric scalar with default 1
#Outputs:
#a vector of n pseudo-random chi-squared distributed deviates
#Stops the function and returns an error message if the arguments are invalid
if(is.numeric(n) & is.numeric(df)){
}else{
stop("invalid arguments")
}
if(df <= 0) stop("invalid arguments")
if(n <=0 ) stop("invalid arguments")
if(n!=round(n)) stop("invalid arguments")
if(df!=round(df)) stop("invalid arguments")
#Create vector for the pseudo-random chi-squared distributed deviates
chisqdev<-c(rep(0,n))
#for loop to calculate the chi-squared distributed deviates by summing
#the squares of normal random variables
for(i in 1:n){
z<-my.rnorm(df)
chi<-sum(z^2)
chisqdev[i]<-chi
}
return(chisqdev)
}
my.rf<-function(n,df1=1,df2=1){
#Purpose:
#Returns n pseudo-random F-distributed deviates
#Inputs:
#n- number of observations: a numeric scalar
#df1- degrees of freedom: a numeric scalar with default 1
#df2- degrees of freedom: a numeric scalar with default 1
#Outputs:
#a vector of n pseudo-random F-distributed deviates
#Stops the function and returns an error message if the arguments are invalid
if(is.numeric(n) & is.numeric(df1) & is.numeric(df2) ){
}else{
stop("invalid arguments")
}
if(df1 <= 0) stop("invalid arguments")
if(df2 <= 0) stop("invalid arguments")
if(n <= 0) stop("invalid arguments")
if(n!=round(n)) stop("invalid arguments")
if(df1!=round(df1)) stop("invalid arguments")
if(df2!=round(df2)) stop("invalid arguments")
#Vector to contain the f distributed deviates
ftestdev<-c(rep(0,n))
#For loop to calcualte the n f-distributed deviates
for(i in 1:n){
u<-my.rchisq(1,df1)
v<-my.rchisq(1,df2)
f<-(u/df1)/(v/df2)
ftestdev[i]<-f
}
return(ftestdev)
}
test.myrnorm<-function(n,mean=0,sd=1){
#Purpose:
#Test to check that my.rnorm recognises invalid inputted arguments
#and consequently doesn't produce a vector of incorrect numeric values
#(checks my.rnorm gives an error when a negative n or sd is inputted and
#when n isn't a whole number).
#Inputs:
#(Same as my.rnorm)
#n- number of observations: a numeric scalar,
#mean - mean: a numeric scalar with default 0,
#sd- standard deviation: a numeric scalar with default 1
#Outputs:
#Pass or Fail
x<-try(my.rnorm(n,mean,sd),silent=TRUE)
if(((sd<0) | (n<=0) | (n!=round(n))) & is.numeric(x)){
cat("Fail")
}else{
cat("Pass")
}
}
test.myrchisq<-function(n,df=1){
#Purpose:
#Test to check that my.rchisq recognises invalid inputted arguments
#and consequently doesn't produce a vector of incorrect values
#(checks my.rchisq gives an error when a negative n or df is inputted and
#when n or df aren't a whole number)
#Inputs:
#(Same as my.rchisq)
##n- number of observations: a numeric scalar
#df-degrees of freedom: a numeric scalar with default 1
#Outputs:
#Pass or Fail
x<-try(my.rchisq(n,df),silent=TRUE)
if(((n!=round(n)) | (n<=0) | (df!=round(df)) | (df<=0)) & is.numeric(x)){
cat("Fail")
}else{
cat("Pass")
}
}
test.myrf<-function(n,df1=1,df2=1){
#Purpose:
#Test to check that my.rf recognises invalid inputted arguments
#and consequently doesn't produce a vector of incorrect values
#(checks my.rf gives an error when a negative n,df1 or df2 is inputted and
#when n, df1 or df2 isn't a whole number)
#Inputs:
#(Same as my.rf)
#n- number of observations: a numeric scalar
#df1- degrees of freedom: a numeric scalar with default 1
#df2- degrees of freedom: a numeric scalar with default 1
#Outputs:
#Pass or Fail
x<-try(my.rf(n,df1,df2),silent=TRUE)
if(((n!=round(n))|(n<=0)|df1!=round(df1)|(df1<=0)|df2!=round(df2)|(df2<=0)) & is.numeric(x)){
cat("Fail")
}else{
cat("Pass")
}
}
outputtest<-function(n,mean=0,sd=1,df=1,df1=1,df2=1,test){
#Purpose:
#Test to check that the output of the function (when it hasn't given an error)
#is the correct amount of values and that is numeric
#Inputs:
#To test the function my.rnorm: test=1
#To test the function my.rchisq: test=2
#To test the function my.rf: test=3
#n- number of observations: a numeric scalar,
#mean - mean: a numeric scalar with default 0 (argument only used when test=1)
#sd- standard deviation: a numeric scalar with default 1 (argument only used when test=1)
#df-degrees of freedom: a numeric scalar with default 1 (argument only used when test=2)
#df1- degrees of freedom: a numeric scalar with default 1 (argument only used when test=3)
#df2- degrees of freedom: a numeric scalar with default 1 (argument only used when test=3)
#Outputs:
#Pass or Fail
switch(as.character(test),
"1"= (x<-my.rnorm(n,mean,sd)),
"2"= (x<-my.rchisq(n,df)),
"3"= (x<-my.rf(n,df1,df2)))
if(length(x)==n & is.numeric(x)){
cat("Pass")
}else{
cat("Fail")
}
}
|
84aa0d7c65b9eaa63ebdeb7fd77a5c6e2dfb4a44 | 39f7d071437cb3489029f0f751600b71ac798962 | /1a 使用R語言進行資料分析/助教課 Week_5/finalExamSol.R | 06e5a392dad0af947345ec51379a186549992ab6 | [] | no_license | evan950608/Evan-R-Programming | 070044950d51e2370d827c41c300160a169f3773 | 586a0f706b71b7e53ea78ad7e1b76be35de0ea81 | refs/heads/master | 2020-04-07T11:49:01.354918 | 2019-01-26T12:05:19 | 2019-01-26T12:05:19 | 158,342,221 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 7,702 | r | finalExamSol.R | #### 注意事項 ####
# 0. 請勿改動 angry_fruit.R,否則將導致此檔案中讀取資料出錯
# 1. 變數名稱請勿改動,若造成判斷錯誤一蓋不負責。
# 2. 請不要用 rm(list=ls()) 之類的東西,我們的 judge 會壞掉。
# 3. ggplot2 的 ggplot() 會回傳東西,第二大題的所有答案都請存到變數中
# ex: gg_exam <- ggplot(data=..., aes(...)) + ...
# 4. 提交答案之前請再次檢查變數存的東西是否符合題目要求。
# 5. 滿分不是一百分
#### 0 ####
# 0.0 (5%)
# 請自行查詢 require() 回傳值
# 請寫出程式碼 "若 沒安裝 rstudioapi 套件,則 安裝 rstudioapi。引入 rstudioapi "
if(!require('rstudioapi')){
install.packages('rstudioapi')
require('rstudioapi')
}
# 0.1 (5%)
# 已知 dirname(rstudioapi::getSourceEditorContext()$path) 會顯示當前.R檔案所在位置
# 請自行查詢 setwd() 和 dirname() 如何使用後,
# 寫出程式碼 "將當前 .R 檔案所在位置設為工作目錄"
# (如果寫不出來一樣請手動設定檔案當前目錄 final2 為 working directory)
setwd(dirname(rstudioapi::getSourceEditorContext()$path))
# 設定檔案當前目錄 final2 為 working directory 後才能 run 下面這行
# 請不要改動 angry_fruit.R 檔案
source('angry_fruit.R')
#### 1 ####
## 俊俊是賣憤怒水果的商人
## 以下是今天他所進貨的憤怒水果的資料
# 水果名稱
fruit_name
# 憤怒程度
anger
# 美味程度
deliciousness
# 1.0 (5%)
# 把三筆資料做成一個 dataframe 存到變數 angry_fruit_na
# column names 分別是 Name, Anger, Deliciousness
angry_fruit_na <- data.frame(Name=fruit_name, Anger=anger, Deliciousness=deliciousness)
angry_fruit_na
# 1.1 (5%)
# 將 anger 或是 delicious 為 NA 值的資料從 angry_fruit_na 移除後
# 將結果存入 angry_fruit
# column names 分別是 Name, Anger, Deliciousness
angry_fruit <- na.omit(angry_fruit_na)
angry_fruit
# 1.2 (5%)
# 水果不能太憤怒,容易過熟影響美味程度
# 俊俊不販賣不美味的憤怒水果
# 理想的憤怒區間為 [0, 100]、而且美味門檻 >= 50
# 請將 angry_fruit 照理想的憤怒區間和美味門檻篩選後
# 將結果存入 ideal_fruit
ideal_fruit <- angry_fruit[angry_fruit$Anger <= 100 & angry_fruit$Anger > 0 & angry_fruit$Deliciousness >= 50,]
# 1.3 (5%)
# 俊俊希望販賣的水果有接近的憤怒值
# 請將憤怒程值超過一個標準差以外的資料從 ideal_fruit 移除
# 並將結果存入 very_ideal_fruit
v <- ideal_fruit$Anger
very_ideal_fruit <- ideal_fruit[(v >= (mean(v) - sd(v))) & (v <= (mean(v) + sd(v))),]
very_ideal_fruit
# 1.4 (5%)
# 冠冠想吃美味的水果
# 請將 very_ideal_fruit 美味值超過 80% 位數(pr80以上)的資料存入 good_fruit
good_fruit <- very_ideal_fruit[very_ideal_fruit$Deliciousness >= quantile(very_ideal_fruit$Deliciousness, 0.8),]
good_fruit
# 1.5 (5%)
# 冠冠有選擇障礙
# 請隨機從 good_fruit 選出一個水果 存入 cm_fruit (存 Name 就好)
# 請用 R 的函數去隨機,不能自己想一個數字
cm_fruit <- as.character(sample(good_fruit$Name, 1))
cm_fruit
# 1.6 (10%)
# 冠冠很滿足,但他不會寫程式
# 請幫他寫一個函式 which_to_eat
# 傳入參數 angry_fruit_na (ex: which_to_eat(angry_fruit_na),輸入保證只有 angry_fruit_na)
# 回傳值為依照 1.1~1.5 步驟篩選後的水果名
# 並將輸出存進 cm_today_fruit
require('dplyr')
which_to_eat <- function(df){
df <- na.omit(df)
df <- df %>%
filter(Anger <= 100, Deliciousness >= 50) %>%
filter(abs(Anger - mean(.$Anger)) <= sd(.$Anger)) %>%
filter(Deliciousness >= quantile(.$Deliciousness, 0.8)) %>%
sample_n(1)
return(as.character(df$Name))
}
cm_today_fruit <- which_to_eat(angry_fruit_na)
cm_today_fruit
# 1.7 (5%)
# 請將 angry_fruit 的 row 依照美味度由大到小排序後存入 angry_fruit_rank
# 若美味度相等,依照憤怒度由大到小排序
# 都一樣的話,依照索引值由小到大排序
angry_fruit_rank <- arrange(angry_fruit, desc(Deliciousness), desc(Anger))
angry_fruit_rank
# 1.8 (10%)
############################################################################
# 已載入 get_50d() (寫在 angry_fruit.R 裡面) #
# 呼叫 get_50d() 會得到一個 list,包含50筆結構如 angry_fruit_na 的 #
# dataframe #
# 注意資料是隨機生成,每次呼叫會不同 #
############################################################################
#
# 俊俊的水果資料每天都會更新
# 俊俊想觀察過去50天水果的資訊,以多進貨冠冠會想吃的水果
# 請寫一個函式 past_50_info()
# 沒有參數
# 過去50天俊俊進貨的水果資訊請用 get_50d() 來產生
#
# 回傳值是一個結構如 angry_fruit_na 的 dataframe
# 第一個 col 為 水果名稱,請使用 fruit_name
# 第二個 col 為「這 50 天中,該水果的平均憤怒程度」,50天皆為 NA 的水果請設為 NaN
# 第三個 col 為「這 50 天中,該水果的平均美味程度」,50天皆為 NA 的水果請設為 NaN
# 請先忽略 NA 後再取平均
# column names 分別為 Name, avg_Anger, avg_Deliciousness
past_50_info <- function(){
past_50_days <- get_50d()
avg_Anger <- sapply(past_50_days, function(df) df$Anger) %>%
rowMeans(na.rm = T)
avg_Deliciousness <- sapply(fifty, function(df) df$Deliciousness) %>%
rowMeans(na.rm = T)
avg_angry_fruit <- data.frame(Name = fruit_name,
avg_Anger = avg_Anger,
avg_Deliciousness = avg_Deliciousness)
return (avg_angry_fruit)
}
#### 2 ####
require('ggplot2')
# 2.1 (10%)
# 利用 ggplot2 畫出內建資料集 airquality 中 Month 為 8 的資料
# Ozone 與 Temp 的 x-y 關係點圖
# 並把圖存到變數 gg1 中
gg1 <- airquality %>%
filter(Month == 8) %>%
ggplot(aes(Ozone, Temp)) + geom_point()
gg1
# 2.2 (10%)
# 利用 ggplot2 畫出內建資料集 mtcars 中 wt, mpg
# 的 x-y 關係點圖,並依據不同的 cyl 分出不同顏色
# 並把圖存到變數 gg2 中
gg2 <- mtcars %>%
ggplot(aes(wt, mpg, color=cyl)) +
geom_point()
gg2
# 2.3 (10%)
# 畫出內建資料集 airquality 中
# x 軸為不同月份 Month,y 軸為該月份 Ozone 的平均值
# 的長條圖,並把圖存到變數 gg3 中
gg3 <- airquality %>%
group_by(Month) %>%
summarise(OzoneMean = mean(Ozone, na.rm=T)) %>%
ggplot(aes(Month, OzoneMean)) +
geom_bar(stat="identity")
gg3
# 2.4 (15%)
# (請依照 0.0 0.1 提示或手動將 .R 檔案當前位置設為 working directory)
# 讀取 "106_student.csv" 檔,存到變數 student 中
# 這份 csv 是 106 學年度全台各地大學專院校的學生人數
# 請先把「等級別」為 "B 學士"、「日間.進修別」為 "D 日" 的資料篩選出來
# 再畫出臺北市大學與非臺北市大學的一年級人數總數的長條圖
# x 軸為是否位於臺北市,y 軸為一年級學生總數 (記得要把男女加起來)
# 並把圖存到變數 gg4 中
student <- read.csv('106_student.csv', sep=",", encoding='utf8')
student
view(student)
gg4 <- student %>%
filter(等級別 == "B 學士", 日間.進修別 == "D 日") %>%
# use gsub() to replace ',' as ''
# ex: '1,024'
transmute(freshman = as.numeric(gsub(",", "", 一年級男生)) + as.numeric(gsub(",", "", 一年級女生)),
atTaipei = (縣市名稱 == "30 臺北市")) %>%
ggplot(aes(atTaipei, freshman)) +
geom_bar(stat="identity")
gg4
|
bd7799bf6f31a58a482690980539bd8ca50838b9 | 7bdee0060e806b64dede482401398149cac7271e | /cointegration/pairs_plot.R | 2768fee098d1c7f5f66f0c8a1ac1e9b063f11473 | [] | no_license | maxim5/stat-arbitrage-r | b775133cdb0aa2e906fde3bcf92dfdf542ab5393 | 5e21c623015f2e8df7bd9ae07435da61bb188669 | refs/heads/master | 2022-11-29T07:58:23.162422 | 2015-08-26T09:10:55 | 2015-08-26T09:10:55 | 287,816,691 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,412 | r | pairs_plot.R | #!/usr/bin/Rscript
suppressMessages(library(ggplot2))
suppressMessages(require(reshape2))
invisible(Sys.setlocale("LC_TIME", "en_US.UTF-8"))
load("pairs.RData")
Plot.Price = function(symbol1, symbol2) {
series1 = all.logs[[symbol1]]
series2 = all.logs[[symbol2]]
dates = as.Date(rownames(all.logs))
data.to.plot = data.frame(exp(series1), exp(series2), dates)
colnames(data.to.plot) = c(symbol1, symbol2, "Date")
data.to.plot = melt(data.to.plot, id="Date")
plot = ggplot(data.to.plot, aes(x=Date, y=value, color=variable)) +
geom_line() +
labs(title=paste0(symbol1, " vs ", symbol2),
x="Time", y="Price") +
scale_colour_discrete(name="Legend")
print(plot)
}
Plot.Logs = function(symbol1, symbol2) {
series1 = all.logs[[symbol1]]
series2 = all.logs[[symbol2]]
dates = as.Date(rownames(all.logs))
data.to.plot = data.frame(series1, series2, dates)
colnames(data.to.plot) = c(symbol1, symbol2, "Date")
data.to.plot = melt(data.to.plot, id="Date")
plot = ggplot(data.to.plot, aes(x=Date, y=value, color=variable)) +
geom_line() +
labs(title=paste0("Log(", symbol1, ") vs Log(", symbol2, ")"),
x="Time", y="Log(Price)") +
scale_colour_discrete(name="Legend")
print(plot)
}
Plot.Spread = function(spread, x=NULL, title="Spread") {
mean = mean(spread)
sd = sd(spread)
stats = boxplot.stats(spread)
if (is.null(x)) {
x = index(spread)
}
plot(x=x, y=spread, type="l", lwd=2, col="darkorchid",
xlab="Time", ylab="Spread", main=title)
Add.HLine = function(level, color) {
abline(h=level, col=color)
text(x[1], level, signif(level, 3), col=color, adj=c(0.5, 0))
}
Add.HLine(mean, "aquamarine3")
Add.HLine(mean+sd, "aquamarine4")
Add.HLine(mean-sd, "aquamarine4")
Add.HLine(stats$stats[1], "firebrick1")
Add.HLine(stats$stats[5], "firebrick1")
}
Plot.Pair = function(symbol1, symbol2) {
Plot.Price(symbol1, symbol2)
Plot.Logs(symbol1, symbol2)
series1 = all.logs[[symbol1]]
series2 = all.logs[[symbol2]]
dates = as.Date(rownames(all.logs))
gamma = as.numeric(cointegrated.pairs[cointegrated.pairs$Symbol1 == symbol1 &
cointegrated.pairs$Symbol2 == symbol2, "Gamma"])
spread = series1 - gamma * series2
Plot.Spread(spread=spread, x=dates,
title=paste0("Log(", symbol1, ") - ", signif(gamma, 3), "*Log(", symbol2, ")"))
}
|
afd471986821ab83c70ad6bd8e3990c4423f09e9 | f73e7dbcc24064028c81f9f778f9892bd55d9066 | /shiny/ui.R | e8f4a2e0534c9642cc3952490b60ed30fff55bf2 | [] | no_license | kchaaa/INFO-498F-Final-Project | 030d235760ad20a62e63cf9afcf9a07174ea61a1 | 63aad059a020e3ef784de29a7cd0c2af1e6a277a | refs/heads/master | 2021-01-10T08:49:05.885601 | 2016-03-11T21:21:33 | 2016-03-11T21:21:33 | 52,401,370 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 232 | r | ui.R | library(shiny)
library(plotly)
library(ggplot2)
shinyUI(fluidPage(
titlePanel("Flint Water Contamination"),
sidebarLayout(
sidebarPanel(
h4("Test Plot")),
mainPanel(
plotOutput("plot1")
)
)
)
) |
3ad426a297e95eeb7c1e5fcbca87562d796a73c9 | 92befee27f82e6637c7ed377890162c9c2070ca9 | /R/summary.lsem.R | f1a4e61558f390df2c6ec42f1e672834d3b21603 | [] | no_license | alexanderrobitzsch/sirt | 38e72ec47c1d93fe60af0587db582e5c4932dafb | deaa69695c8425450fff48f0914224392c15850f | refs/heads/master | 2023-08-31T14:50:52.255747 | 2023-08-29T09:30:54 | 2023-08-29T09:30:54 | 95,306,116 | 23 | 11 | null | 2021-04-22T10:23:19 | 2017-06-24T15:29:20 | R | UTF-8 | R | false | false | 3,312 | r | summary.lsem.R | ## File Name: summary.lsem.R
## File Version: 0.412
#-- summary lsem
summary.lsem <- function( object, file=NULL, digits=3, ... )
{
# open sink for a file
sirt_osink( file=file )
cat('-----------------------------------------------------------------\n')
cat('Local Structural Equation Model \n\n')
#-- print packages
packages <- c('sirt', 'lavaan')
if (object$use_lavaan_survey){
packages <- c(packages, 'lavaan.survey')
}
sirt_summary_print_packages(packages=packages)
#-- print R session
cat('\n')
sirt_summary_print_rsession()
cat(paste0('Function \'sirt::lsem.estimate\', type=\'', object$type,'\''), '\n\n')
#- print call
sirt_summary_print_call(CALL=object$CALL)
#-- print computation time
sirt_summary_print_computation_time_s1(object=object)
# space between equality sign
sp_eq <- paste0( c(' ', '=', ' '), collapse='')
cat( paste0( 'Number of observations in datasets', sp_eq,
round(object$N, digits) ), '\n')
cat( paste0( 'Used observations in analysis', sp_eq,
round(object$nobs, digits) ), '\n')
cat('Used sampling weights:', ! object$no_sampling_weights, '\n')
if ( object$type=='LSEM'){
cat( paste0( 'Bandwidth factor', sp_eq, round(object$h,digits) ), '\n')
cat( paste0( 'Bandwidth', sp_eq, round(object$bw,digits) ), '\n')
cat( paste0( 'Number of focal points for moderator', sp_eq,
length(object$moderator.grid ) ), '\n')
cat('\n')
cat('Used joint estimation:', object$est_joint, '\n')
cat('Used sufficient statistics:', object$sufficient_statistics, '\n')
cat('Used local linear smoothing:', object$loc_linear_smooth, '\n')
cat('Used pseudo weights:', object$use_pseudo_weights, '\n')
cat('Used lavaan package:', TRUE, '\n')
cat('Used lavaan.survey package:', object$use_lavaan_survey, '\n\n')
cat('Mean structure modelled:', object$is_meanstructure, '\n')
if (object$class_boot){
v1 <- paste0('\nStatistical inference based on ', object$R,
' bootstrap samples.')
cat(v1,'\n')
}
}
if ( object$type=='MGM'){
cat( paste0( 'Number of groups for moderator=',
length(object$moderator.grid ) ), '\n')
}
cat('\nlavaan Model\n')
cat(object$lavmodel)
if (object$est_joint){
cat('\n\n')
cat('Global Fit Statistics for Joint Estimation\n\n')
obji <- object$fitstats_joint
sirt_summary_print_objects(obji=obji, digits=digits)
}
cat('\n\n')
cat('Parameter Estimate Summary\n\n')
obji <- object$parameters_summary
sirt_summary_print_objects(obji=obji, digits=digits, from=2)
cat('\n')
cat('Distribution of Moderator: Density and Effective Sample Size\n\n')
cat( paste0('M=', round(object$m.moderator, digits), ' | SD=',
round(object$sd.moderator, digits), '\n\n') )
obji <- object$moderator.density
sirt_summary_print_objects(obji=obji, digits=digits, from=1)
cat('\n')
obji <- object$moderator.stat
sirt_summary_print_objects(obji=obji, digits=digits, from=2)
# close file
sirt_csink(file)
}
|
fd26877562d244d617b56f04f2ec149e1201f122 | 2a0e90441bb5edc22344aff9019dea4d825183bf | /auto_learner_manish_1.r | 1e7be0463b39cc70d1d21f31585d6a32fc6b0a24 | [] | no_license | srijan55/1ml | bb06ac62601beb6a87d4823564bd6fe487fd930f | 802f839e719a6a351c7d1b9b0695e995c4acfaf0 | refs/heads/master | 2021-01-22T06:58:50.134172 | 2015-07-02T02:07:45 | 2015-07-02T02:07:45 | 38,128,894 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 5,677 | r | auto_learner_manish_1.r | library(e1071)
library(Matrix)
library(SparseM)
##########################
## Load raw training data
##########################
rawdata<- read.csv("train_category.dat", sep="\t", nrows = 100 )
################################
## Feature Engineering
#################################
rawdata$UserID <-as.factor(rawdata$UserID) # Convert UserId to factor
rawdata.userid <- rawdata$UserID
###### Convert to a sparse matrix on events
s<- sparse.model.matrix(~0+Event,data=rawdata)
#multiply with count
y <- s*rawdata$Count
rawdata <-cbind( rawdata.userid, sparse.model.matrix(~0+Event+Count,data=rawdata))
rawdata$UserID <-rawdata.userid
###### Multiply count into columns to get the actual sparse matrix
col_num <- ncol(rawdata)
rawdata <-data.frame( UserID=rawdata[,1], (rawdata[,c(-1,-col_num)]*rawdata[,col_num]))
###### Aggregate on UserId's
rawdata <- aggregate(. ~ UserID, FUN=sum, data = rawdata)
#####Add labels from labeldata
labeldata<- read.csv("trainLabel.dat", sep="\t")
rawdata <- merge( rawdata, labeldata, by="UserID", all=FALSE)
rm(labeldata)# get rid of labeldata not needed now
gc()
#####Add weights to non-zero sparse factors
col_num <- ncol(rawdata)
weight <- rawdata[2:(col_num-1)]*5
#rawdata <- data.frame(UserID=rawdata$UserID, weight, Label=rawdata$Label)
rawdata <- data.frame(weight, Label=rawdata$Label)
rawdata$Label <- as.factor(rawdata$Label)
rm(weight)
#####################################
# TODO: Work with demographic data
#####################################
#demodata <- read.csv("../u360_demodata.tsv", sep = "\t", header = FALSE)
###demodata.age <- (demodata$PreGenderProb>0.6)? demodata$PreGender: demodata$RegGender
###names(demodata)<- c("UserID", "RegCountry", "RegBirth", "RegGender", "PreGender", "PreGenderProb", "RegAgeGrp", "PreAgeGrp", "PreAgeGrpProb")
#weighteddata <- weighteddata[,-1]
# augmenteddata<- merge(weighteddata, demodata, by="UserID", all.x=TRUE)
##########################
## train the model
##########################
data.model <- naiveBayes(Label~., data = rawdata)
##########################
## Load raw test data
##########################
rawdata<- read.csv("test_category.dat", sep="\t" )
labeltestdata<- read.csv("testID.dat", sep="\t")
################################
## Feature Engineering on test data
#################################
rawdata$UserID <-as.factor(rawdata$UserID) # Convert UserId to factor
###### Convert to a sparse matrix on events
rawdata<-data.frame( rawdata$UserID, model.matrix(~0+Event+Count,data=rawdata))
###### Multiply count into columns to get the actual sparse matrix
col_num <- ncol(rawdata)
rawdata <-data.frame( UserID=rawdata[,1], rawdata[,c(-1,-col_num)]*rawdata[,col_num])
###### Aggregate on UserId's
rawdata <- aggregate(. ~ UserID, FUN=sum, rawdata)
#####Add weights to non-zero sparse factors
col_num <- ncol(rawdata)
weight <- rawdata[2:(col_num)]*5
#rawdata <- data.frame(UserID=rawdata$UserID, weight, Label=rawdata$Label)
rawdata <- data.frame(weight)
rm(weight)
#########################
# Get the predictions
###########################
data.predictions <- predict(data.model, rawdata, type = "class")
#######################
# Create desired output
#######################
#####Add users from test ID's
labeldata<- read.csv("testID.dat", sep="\t")
labeldata <- data.frame(UserID=as.factor(labeldata$UserID))
rawdata<- data.frame(UserID=labeldata$UserID, Label=data.predictions)
#####Add users and labels from from train_label data
labeldata<- read.csv("trainLabel.dat", sep="\t")
labeldata <- data.frame(UserID=as.factor(labeldata$UserID), Label=as.factor(labeldata$Label))
rawdata <- merge( rawdata, labeldata, by="UserID", all=FALSE)
rawdata<-rbind(rawdata, labeldata)
rm(labeldata)
rm(data.predictions)
gc()
write.csv(file = "auto_output.csv", x=rawdata)
#rawtestdata$UserID <-as.factor(rawtestdata$UserID)
#sparsetestdata<-data.frame( rawtestdata$UserID, model.matrix(~0+Event+Count,data=rawtestdata))
#col_num <- ncol(sparsetestdata)
#multipledtestdata <-data.frame( UserID=sparsetestdata[,1], sparsetestdata[,c(-1,-col_num)]*sparsetestdata[,col_num])
#aggregatetestdata <- aggregate(. ~ UserID, FUN=sum, multipledtestdata)
#labelledtestdata <- merge( aggregatetestdata, labeltedata, by="UserID", all=FALSE)
#col_num <- ncol(aggregatetestdata)
#weight <- aggregatetestdata[2:(col_num-1)]*5
#weightedtestdata <- data.frame(UserID=aggregatetestdata$UserID, weight)
#weighteddata$Label <- as.factor(weighteddata$Label)
#random.rows.train <- sample(1:nrow(weighteddata), 0.5*nrow(weighteddata), replace=F)
#weighteddata.train <- weighteddata[random.rows.train,]
#dim(weighteddata.train)
## select the other 1/2 left as the testing data
#random.rows.test <- setdiff(1:nrow(weighteddata),random.rows.train)
#weighteddata.test <- weighteddata[random.rows.test,]
#dim(weighteddata.test)
## fitting decision model on training set
#weighteddata.model <- naiveBayes(Label~., data = weighteddata)
## MODEL EVALUATION
## make prediction using decision model
#weighteddata.test.predictions <- predict(weighteddata.model, weightedtestdata, type = "class")
## extract out the observations in testing set
#weighteddata.test.observations <- weighteddata.test$Label
## show the confusion matrix
#confusion.matrix <- table(weighteddata.test.predictions, weighteddata.test.observations)
#confusion.matrix
## calculate the accuracy in testing set
#accuracy <- sum(diag(confusion.matrix)) / sum(confusion.matrix)
#accuracy
#predictedata<- data.frame(UserID=weightedtestdata$UserID, Label=weighteddata.test.predictions)
#labeldata <- data.frame(UserID=as.factor(labeldata$UserID), Label=as.factor(labeldata$Label))
#x<-rbind(predictedata, labeldata)
|
74ccce38fd397f404e954cf868cc6c776fb26e74 | 2d88e86736d81b32e957b62bd8b0041e2a9778ad | /R/scores.tables.tweak.R | 1e4cae57350c86dd2118b7c9151872ffa2c8a421 | [] | no_license | cran/amber | c1659595049f230f54db3893704fc67ddb2429ed | e6ef59a25270413a1875c84feac786551bf69315 | refs/heads/master | 2021-07-23T06:25:02.408885 | 2020-08-28T10:20:02 | 2020-08-28T10:20:02 | 212,134,119 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 3,522 | r | scores.tables.tweak.R | ################################################################################
#' Tweak summary table
#' @description This function allows the user to tweak the summary table computed
#' by \link{scores.tables}. Contrary to \link{scores.tables}, this function can be used
#' to create a single summary table that includes the most important metrics only.
#' The user can specify what variables to include and in what order they should appear.
#' @param myVariables An R object with variable names of variables that should be included in table, e.g. c('GPP', 'RECO', 'NEE')
#' @param myCaption A string that is used as table caption, e.g. 'Globally averaged statistical metrics'.
#' @param inputDir A string that gives the input directory, e.g. '/home/project/study'.
#' @param outputDir A string that gives the output directory, e.g. '/home/project/study'. The output will only be written if the user specifies an output directory.
#' @return One table in LaTeX format that shows a subset of statistical metrics
#' @examples
#' library(amber)
#' library(classInt)
#' library(doParallel)
#' library(foreach)
#' library(Hmisc)
#' library(latex2exp)
#' library(ncdf4)
#' library(parallel)
#' library(raster)
#' library(rgdal)
#' library(rgeos)
#' library(scico)
#' library(sp)
#' library(stats)
#' library(utils)
#' library(viridis)
#' library(xtable)
#'
#' myInputDir <- paste(system.file('extdata', package = 'amber'), 'scores', sep = '/')
#' myVariables <- c('GPP', 'LAI', 'ALBS')
#' scores.tables.tweak(myVariables = myVariables, inputDir = myInputDir)
#' @export
scores.tables.tweak <- function(myVariables, myCaption = "Globally averaged statistical metrics", inputDir = getwd(), outputDir = FALSE) {
# summary table with globally averaged inputs for computing scores
my.list <- list.files(path = inputDir, pattern = "scoreinputs_")
my.files <- paste(inputDir, my.list, sep = "/")
data <- lapply(my.files, utils::read.table)
data <- do.call("rbind", data)
colnames(data)
myOrder <- seq(1, length(myVariables), 1)
myOrder <- data.frame(myVariables, myOrder)
colnames(myOrder) <- c("variable.name", "order")
data <- merge(data, myOrder, by = "variable.name")
data <- data[order(data$order, data$ref.id), ]
data <- subset(data, select = -c(order))
colnames(data) <- c("Name", "Variable", "Reference", "Unit", "$v_{mod}$", "$v_{ref}$", "Bias", "Bias (\\%)", "$\\sigma_{ref}$",
"$\\epsilon_{bias}$ (-)", "$S_{bias}$ (-)", "$rmse$", "$crmse$", "$\\sigma_{ref}$", "$\\epsilon_{rmse}$ (-)", "$S_{rmse}$ (-)",
"$max_{cmod}$", "$max_{cref}$", "$\\theta$ (months)", "$S_{phase}$ (-)", "$iav_{mod}$", "$iav_{ref}$", "$\\epsilon_{iav}$ (-)",
"$S_{iav}$", "$\\sigma_{\\overline{v_{mod}}}$", "$\\sigma_{\\overline{v_{ref}}}$", "$\\sigma$ (-)", "$R$ (-)", "$S_{dist}$ (-)")
rownames(data) <- c() # omit rownames
# Make a table that only includes selected reference data and metrics
metricsTable <- data[c(1, 3, 4, 5, 6, 7, 8, 13, 19, 23, 27, 28)] # selection of variables
rownames(metricsTable) <- c() # omit rownames
# convert to LaTeX
metricsTable <- xtable::xtable(metricsTable)
xtable::caption(metricsTable) <- myCaption
if (outputDir != FALSE) {
xtable::print.xtable(metricsTable, include.rownames = FALSE, label = "tab:global_stats", type = "latex", file = "metricsTable.tex",
caption.placement = "top", sanitize.text.function = function(x) {
x
})
}
}
|
977dca5df9467a27cb8d8503ed42572c8bf96028 | a49107cd976c16910f405e92891f81089d46b235 | /16 Dates and Times.R | 8d8f841c0e2b6e5d052bfaca3d7af932a3a8ea5a | [] | no_license | nmoorenz/R4DS | 8fcf397516d95054e24287249267fe58ce6b7d4b | 1227401db2459120b2890ff5f2a391c1daf8b13b | refs/heads/master | 2021-09-18T10:22:10.851213 | 2018-07-12T21:16:20 | 2018-07-12T21:16:20 | 114,934,177 | 0 | 0 | null | 2018-07-12T21:16:21 | 2017-12-20T21:47:41 | R | UTF-8 | R | false | false | 6,694 | r | 16 Dates and Times.R |
############################################
# 16 Dates and Times
library(tidyverse)
library(lubridate)
library(nycflights13)
# 16.2 Creating date/times
today()
now()
# 16.2.1 From strings
# ymd() mdy() dmy()
ymd("2017-01-31")
ymd(20170630)
ymd_hm("2017-12-25 10:00", tz = "NZ")
# 16.2.2 from components
flights %>%
select(year, month, day, hour, minute) %>%
mutate(departure = make_datetime(year, month, day, hour, minute))
# modulus arithmetic for some funky times
make_datetime_100 <- function(year, month, day, time) {
make_datetime(year, month, day, time %/% 100, time %% 100)
}
flights_dt <- flights %>%
filter(!is.na(dep_time), !is.na(arr_time)) %>%
mutate(
dep_time = make_datetime_100(year, month, day, dep_time),
arr_time = make_datetime_100(year, month, day, arr_time),
sched_dep_time = make_datetime_100(year, month, day, sched_dep_time),
sched_arr_time = make_datetime_100(year, month, day, sched_arr_time)
) %>%
select(origin, dest, ends_with("delay"), ends_with("time"))
flights_dt
# visualise departures in a year
flights_dt %>%
ggplot(aes(dep_time)) +
geom_freqpoly(binwidth = 86400)
# or in a single day
flights_dt %>%
filter(dep_time < ymd(20130102)) %>%
ggplot(aes(dep_time)) +
geom_freqpoly(binwidth = 600)
# 16.2.3 From other types
as_datetime(today())
as_date(now())
# Unix Epoch is 1970-01-01
########
# 16.2.4 Exercises
# invalid
ymd(c("20101010", "bananas"))
# tzone
today(tzone = "NZ")
# lubridate functions
d1 <- mdy("January 1, 2010")
d2 <- ymd("2015-Mar-07")
d3 <- dmy("06-Jun-2017")
d4 <- mdy(c("August 19 (2015)", "July 1 (2015)"))
d5 <- mdy("12/30/14") # Dec 30, 2014
######
# 16.3 Date-time components
datetime <- ymd_hms("2016-07-08 12:34:56")
year(datetime)
month(datetime)
mday(datetime) # day of month
yday(datetime) # day of year
wday(datetime) # day of week
month(datetime, label = TRUE) # abbreviated name of month
wday(datetime, label = TRUE, abbr = FALSE) # full name of weekday
flights_dt %>%
mutate(wday = wday(dep_time, label = TRUE)) %>%
ggplot(aes(x = wday)) +
geom_bar()
# on the hour or half hour don't have as much delay
flights_dt %>%
mutate(minute = minute(dep_time)) %>%
group_by(minute) %>%
summarise(
avg_delay = mean(arr_delay, na.rm = TRUE),
n = n()) %>%
ggplot(aes(minute, avg_delay)) +
geom_line()
sched_dep <- flights_dt %>%
mutate(minute = minute(sched_dep_time)) %>%
group_by(minute) %>%
summarise(
avg_delay = mean(arr_delay, na.rm = TRUE),
n = n())
ggplot(sched_dep, aes(minute, avg_delay)) +
geom_line()
ggplot(sched_dep, aes(minute, n)) +
geom_line()
# 16.3.2 Rounding (good for grouping)
flights_dt %>%
count(week = floor_date(dep_time, "week")) %>%
ggplot(aes(week, n)) +
geom_line()
# 16.3.3 setting components
(datetime <- ymd_hms("2016-07-08 12:34:56"))
year(datetime) <- 2010
month(datetime) <- 01
hour(datetime) <- 02
datetime
update(datetime, year = 2020, month = 4, mday = 22, minute = 22)
# large values will roll over, negatives go backwards
ymd("2015-01-01") %>% update(mday = 35)
ymd("2015-03-01") %>% update(hour = -20)
flights_dt %>%
mutate(dep_hour = update(dep_time, yday = 1)) %>%
ggplot(aes(dep_hour)) +
geom_freqpoly(binwidth = 600)
########
# 16.3.4 Exercises
flights_dt %>%
mutate(
dep_hour = update(dep_time, yday = 1),
dep_month = month(dep_time)
) %>%
ggplot(aes(dep_hour, colour = dep_month, group = dep_month)) +
geom_freqpoly(binwidth = 600)
# dep_time, sched_dep_time, dep_delay
flights_dep <- flights %>%
filter(!is.na(dep_time)) %>%
mutate(
dep_time = make_datetime_100(year, month, day, dep_time),
sched_dep_time = make_datetime_100(year, month, day, sched_dep_time),
est_dep_time = sched_dep_time + dep_delay * 60
) %>%
select(origin, dest, ends_with("dep_time"), dep_delay) %>%
filter(dep_time != est_dep_time)
flights_dep
# adjusting timezones for airports and flight times!
# average delay time over day
flights_dt %>%
mutate(dep_hour = update(sched_dep_time, yday = 1)) %>%
group_by(dep_hour) %>%
mutate(delays = mean(dep_delay)) %>%
ggplot(aes(dep_hour, delays)) +
geom_line()
# day of week
flights_dt %>%
mutate(weekday = wday(sched_dep_time)) %>%
group_by(weekday) %>%
mutate(delays = mean(dep_delay)) %>%
ggplot(aes(weekday, delays)) +
geom_line()
# carats and sched_dep_time
flights_dt %>%
mutate(dep_hour = update(sched_dep_time, yday = 1, hour = 1)) %>%
ggplot(aes(dep_hour)) +
geom_freqpoly()
diamonds %>%
ggplot(aes(carat)) +
geom_freqpoly()
######
# 16.4 Time spans
# durations, periods, intervals
# lubridate always measures duration in seconds - vectorised construction
dseconds(20)
dminutes(30)
dhours(36)
ddays(0:7)
dweeks(2) + dyears(2)
today() + ddays(1)
today() + 1
# periods more human like
seconds(15)
minutes(15)
hours(15)
days(15)
months(15)
weeks(15)
years(0:3)
10 * months(10)
today() + hours(50)
# can use these to fix some times in the flights data
flights_dt %>%
filter(arr_time < dep_time)
flights_dt <- flights_dt %>%
mutate(
overnight = arr_time < dep_time,
arr_time = arr_time + days(overnight * 1),
sched_arr_time = sched_arr_time + days(overnight * 1)
)
flights_dt %>%
filter(arr_time < dep_time)
# 16.4.3 Intervals
dyears(1) / ddays(365) # should return one because both are actually in seconds
years(1) / days(1) # warning because not always true (leap year)
# more accurate to use intervals
next_year <- today() + years(1)
(today() %--% next_year) / ddays(1)
(today() %--% next_year) %/% days(1)
########
# 16.4.5 Exercises
# there's no dmonths() because months are all different!
# overnight is boolean, only multiplies if true
# vector of dates first of every month in 2015
ymd("2015-01-01") + months(0:11)
make_date(year(today()), 1, 1) + months(0:11)
# function for age from birthdate
my_age <- function(birthdate) {
(birthdate %--% today()) / dyears(1)
}
my_age(ymd("1984-09-20"))
# wrong?
(today() %--% (today() + years(1))) / months(1)
######
# 16.5 Timezones
Sys.time()
Sys.timezone()
(x1 <- ymd_hms("2015-06-01 12:00:00", tz = "America/New_York"))
(x2 <- ymd_hms("2015-06-01 18:00:00", tz = "Europe/Copenhagen"))
(x3 <- ymd_hms("2015-06-02 04:00:00", tz = "Pacific/Auckland"))
# these are all the same!
x1 - x2
x2 - x3
# can combine these
(x4 <- c(x1, x2, x3))
|
63a7afb8be5cb5181294e32011df820beebdd09f | 09a34862ad70328988389e8a304dcfcfddd2146e | /old_versions/toyData.R | c48825ee51a24a2af3137693b216722c1382bcc1 | [] | no_license | kathiesun/TReC_matnut | 63c33db7ebaee80ffc262975b27e41a3d32ef997 | e8a2f463960a3c59e54d39690a065b0563a957a8 | refs/heads/master | 2023-04-19T11:42:38.918064 | 2021-05-09T19:29:52 | 2021-05-09T19:29:52 | 321,777,614 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 7,684 | r | toyData.R | setwd("~/matnut/src")
library(rstan)
library(tidyverse)
# ---------------------------------
# Generate toy data
# ---------------------------------
set.seed(1)
nmice=2
ngenes=3
nkmer=4
total_counts_per_kmer = pk_matrix = data = list()
pg = rbeta(ngenes, 1, 1)
total_counts_gene = floor(runif(ngenes,0,1) * 10000)
for(j in 1:nmice){
pk_matrix[[j]] = t(sapply(1:ngenes, function(x) rbeta(nkmer,total_counts_gene[x]*pg[x] + 1,(1-pg[x])*total_counts_gene[x]+1)))
total_counts_per_kmer = do.call("rbind", lapply(total_counts_gene, function(x) rpois(nkmer,x/nkmer)))
counts_per_kmer_a1 = sapply(1:length(pk_matrix[[j]]), function(x) rbinom(1, t(total_counts_per_kmer)[x], t(pk_matrix[[j]])[x]))
temp_dat1 = cbind(allele = 1,
count = counts_per_kmer_a1,
kmer = seq(1:nkmer),
gene = rep(1:ngenes, each=nkmer),
mouse = j)
counts_per_kmer_a2 = sapply(1:length(pk_matrix[[j]]), function(x) rbinom(1, t(total_counts_per_kmer)[x], t(1-pk_matrix[[j]])[x]))
temp_dat2 = cbind(allele = 2,
count = counts_per_kmer_a2,
kmer = seq(1:nkmer),
gene = rep(1:ngenes, each=nkmer),
mouse = j)
data[[j]] = rbind(temp_dat1, temp_dat2)
}
data <- do.call("rbind", data)
as.tibble(data) %>%
filter(mouse == 1) %>%
group_by(gene, kmer) %>%
summarise(y1 = count[allele == 1],
sum = (count[allele == 1] + count[allele == 2]),
ratio = count[allele == 1] / (count[allele == 1] + count[allele == 2])) %>%
filter(gene==1) ->
data_model
confInt %>% filter(gene_name == "Reps2", Pup.ID == 379) %>%
group_by(pos) %>% dplyr::slice(1) -> data_model
## Create Stan data
standat <- list(N = nrow(remNanRatios),
K = length(unique(remNanRatios$pos)),
G = length(unique(remNanRatios$gene_name)),
P = 1,
y_gk = remNanRatios$count,
n_gk = remNanRatios$total,
kmer = as.numeric(as.factor(remNanRatios$pos)),
gene = as.numeric(as.factor(remNanRatios$gene_name)))
fileName <- "matnut/logit_binom.stan"
stan_code <- readChar(fileName, file.info(fileName)$size)
cat(stan_code)
chains=2
iter=10000
warmup=round((floor((iter*0.25)/(iter/10))*(iter/10)))
thin=50
resStan <- stan(model_code = stan_code, data = standat,
chains = 2, iter = iter, warmup = warmup, thin = thin)
stanmcmc<- As.mcmc.list(resStan)
summcmc <- summary(stanmcmc)
traceplot(stanmcmc[,1,drop=F])
tset <- sapply(1:(ncol(stanmcmc[[1]])-1), function(x) HPDinterval(stanmcmc[,x,drop=T]), simplify=F)
sigTest <- rep(0, length = length(tset))
sigTest[grep("odds", colnames(stanmcmc[[1]]))] <- 1
sigTest[grep("prob", colnames(stanmcmc[[1]]))] <- 0.5
isSig <- sapply(1:length(tset), function(x) sapply(1:chains, function(y) ifelse(((tset[[x]][[y]][1] < sigTest[x] && tset[[x]][[y]][2] < sigTest[x])
|| (tset[[x]][[y]][1] > sigTest[x] && tset[[x]][[y]][2] > sigTest[x])), T, F)))
trueSig <- intersect(which(isSig[1,] == T), which(isSig[2,] == T))
for(i in 1:10){
ind = trueSig[i]
traceplot(stanmcmc[,ind,drop=F])
}
######################## IGNORE ############################
##########
# STAN #
##########
stan_file <- file.path("../../..","Dropbox", "doubleGLM.stan")
#dissector.sm <- stan_model(file=stan_file)
tsstan
an_fit <- list()
vceg <- list()
for(j in 1:2){
covar <- ifelse(j==1, "Sex", "Diet")
for(i in 1:7){
miss <- which(is.na(compl_phen[[j]][,allvars[[j]][i]]))
mouseID_miss <- compl_phen[[j]]$MouseID[miss]
remove <- which(as.numeric(unlist(strsplit(colnames(compl_kin[[j]]),"[.]"))[c(F,T,F)]) %in% mouseID_miss)
if (length(mouseID_miss) > 0){
temp_phen <- compl_phen[[j]][-which(compl_phen[[j]]$MouseID %in% mouseID_miss),]
temp_kin <- compl_kin[[j]][-remove, -remove]
} else {
temp_phen <- data.frame(compl_phen[[j]])
temp_kin <- compl_kin[[j]]
}
vceg$x <- temp_phen[,covar]
vceg$R <- temp_kin
vceg$Rinv <- solve(vceg$R)
vceg$y <- as.vector(scale(as.numeric(temp_phen[,allvars[[j]][i]])))
stan_fit <- stan(file = stan_file, control = list(adapt_delta = 0.8),
data = list(num_cov = length(unique(vceg$x)),
N = nrow(temp_kin),
phenotype = vceg$y,
cov = vceg$x,
R = temp_kin), chains=3, iter = 5000,
thin=10, warmup = 1000)
}
}
#h2_mcmc <- as.mcmc(stan_fit@sim$samples[[1]]$h2)
#mcmc_trace(regex_pars = 'sigma')
rstan::traceplot(stan_fit, pars = 'h2', inc_warmup=F)
rstan::traceplot(stan_fit, pars = 'grand_sig2', inc_warmup=F)
rstan::traceplot(stan_fit, pars = 'cov_vef', inc_warmup=F)
# ---------------------------------
# Prior parameters
# ---------------------------------
n=10
a <- 1; b <-1;
S <- 20000
tau0 <- 200
sig0 <- 1000
nu0 <- 10
mu0=120
X <- matrix(NA, nrow=n, ncol=S)
sig.1 <- numeric(S)
sig.2 <- numeric(S)
theta.1 <- numeric(S)
theta.2 <- numeric(S)
p <- numeric(S)
theta.min <- numeric(S)
theta.max <- numeric(S)
# ---------------------------------
# Random initialization to groups
# ---------------------------------
p[1] <- rbeta(1, a,b)
X[,1]<- rbinom(n, 1, p[1])
X[,1]<- ifelse(X[,1]==1, 2, 1)
theta.1[1]<- mean(Y)
theta.2[1]<- mean(Y)
sig.1[1]<- var(Y)
sig.2[1]<- var(Y)
# ---------------------------------
# Gibbs sampling algorithm
# ---------------------------------
for (i in 2:S){
# Update for p
n1 <- sum(X[,i-1]==1)
n2 <- sum(X[,i-1]==2)
p[i] <- rbeta(1, a + n1, b + n2)
mean.y1 <- mean(Y[which(X[,i-1] == 1)])
mean.y2 <- mean(Y[which(X[,i-1] == 2)])
var.y1 <- var(Y[which(X[,i-1] == 1)])
var.y2 <- var(Y[which(X[,i-1] == 2)])
# Update parameters
sig.1[i] <- 1/rgamma(1, (n1/2)+(nu0/2), 0.5*((n1-1)*var.y1 + n1*(mean.y1-theta.1[i-1])^2))
sig.2[i] <- 1/rgamma(1, (n2/2)+(nu0/2), 0.5*((n2-1)*var.y2 + n1*(mean.y2-theta.2[i-1])^2))
a1 <-n1*(1/sig.1[i-1])+(1/tau0)
b1 <- n1*(1/sig.1[i-1])*mean.y1+(1/tau0)*mu0
theta.1[i] <- rnorm(1, b1/a1, sqrt(1/a1))
a2 <-n2*(1/sig.2[i-1])+(1/tau0)
b2 <- n2*(1/sig.2[i-1])*mean.y2+(1/tau0)*mu0
theta.2[i] <- rnorm(1, b2/a2, sqrt(1/a2))
# calculate theta statistics
theta.min[i] <- min(theta.1[i], theta.2[i])
theta.max[i] <- max(theta.1[i], theta.2[i])
# Update for X's
for (j in 1:n){
w1 <- p[i-1]*dnorm(Y[j], theta.1[i], sqrt(sig.1[i]))
w0 <- (1-p[i-1])*dnorm(Y[j], theta.2[i], sqrt(sig.2[i]))
w <- w1/(w1+w0)
X[j,i]<- sample(1:0, 1, prob=c(w,1-w))
X[j,i]<- ifelse(X[j,i]==0, 2, 1)
}
}
hist(theta.1, freq = FALSE, xlim = c(90, 190), ylim = c(0, 0.17),
xlab = expression(theta),
main = expression(paste("Approx. of p(",theta,"|y)", sep = "")))
lines(density(theta.1), col = "blue")
lines(density(theta.2), col = "red")
hist(theta.2, freq = FALSE,
add = T)
# ---------------------------------
# Part c
# ---------------------------------
par(mfrow=c(1,2))
plot(acf(theta.min))
plot(acf(theta.max))
print(acf(theta.min, plot=F))
print(acf(theta.max, plot=F))
c(effectiveSize(mcmc(theta.min)),
effectiveSize(mcmc(theta.max)))
|
9d360197f39944a7f69d4bef7fe2e7995d9c7f3b | da1ae08c144c508573a8482a71bd2a2ebe5c21e9 | /R/s3.R | 8beb6fa2b043d7c7eef1d17cf9ca6c065c00367e | [
"MIT"
] | permissive | AmrR101/singleCellHaystack | 48c068eb35319edf7842db17ae6542a715f5420c | 68f41d8cc9cb44b7eff95f318592f33306dae0a4 | refs/heads/master | 2023-02-10T01:12:04.548101 | 2021-01-07T08:44:41 | 2021-01-07T08:44:41 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 11,167 | r | s3.R | #' The main Haystack function
#'
#' @param x a matrix or other object from which coordinates of cells can be extracted.
#' @param dim1 column index or name of matrix for x-axis coordinates.
#' @param dim2 column index or name of matrix for y-axis coordinates.
#' @param assay name of assay data for Seurat method.
#' @param slot name of slot for assay data for Seurat method.
#' @param coord name of coordinates slot for specific methods.
#' @param dims dimensions from coord to use. By default, all.
#' @param cutoff cutoff for detection.
#' @param method choose between highD (default) and 2D haystack.
#' @param detection A logical matrix showing which genes (rows) are detected in which cells (columns)
#' @param use.advanced.sampling If NULL naive sampling is used. If a vector is given (of length = no. of cells) sampling is done according to the values in the vector.
#' @param dir.randomization If NULL, no output is made about the random sampling step. If not NULL, files related to the randomizations are printed to this directory.
#' @param scale Logical (default=TRUE) indicating whether input coordinates in x should be scaled to mean 0 and standard deviation 1.
#' @param grid.points An integer specifying the number of centers (gridpoints) to be used for estimating the density distributions of cells. Default is set to 100.
#' @param grid.method The method to decide grid points for estimating the density in the high-dimensional space. Should be "centroid" (default) or "seeding".
#' @param ... further parameters passed down to methods.
#'
#' @return An object of class "haystack"
#' @export
#'
haystack <- function(x, ...) {
UseMethod("haystack")
}
#' @rdname haystack
#' @export
haystack.matrix <- function(x, dim1 = 1, dim2 = 2, detection, method = "highD", use.advanced.sampling = NULL, dir.randomization = NULL, scale = TRUE, grid.points = 100, grid.method = "centroid", ...) {
method <- match.arg(method, c("highD", "2D"))
switch(method,
"highD" = {
haystack_highD(
x,
detection = detection,
use.advanced.sampling = use.advanced.sampling,
dir.randomization = dir.randomization,
scale = scale,
grid.points = grid.points,
grid.method = grid.method, ...)
},
"2D" = {
haystack_2D(
x[, dim1],
x[, dim2],
detection = detection,
use.advanced.sampling = use.advanced.sampling,
dir.randomization = dir.randomization, ...)
}
)
}
#' @rdname haystack
#' @export
haystack.data.frame <- function(x, dim1 = 1, dim2 = 2, detection, method = "highD", use.advanced.sampling = NULL, dir.randomization = NULL, scale = TRUE, grid.points = 100, grid.method = "centroid", ...) {
haystack(as.matrix(x), dim1 = dim1, dim2 = dim2, detection = detection, method = method, use.advanced.sampling = use.advanced.sampling, dir.randomization = dir.randomization, scale = scale, grid.points = grid.points, grid.method = grid.method, ...)
}
#' @rdname haystack
#' @export
haystack.Seurat <- function(x, assay = "RNA", slot = "data", coord = "pca", dims = NULL, cutoff = 1, method = NULL, use.advanced.sampling = NULL, ...) {
if (!requireNamespace("Seurat", quietly = TRUE)) {
stop("Package \"Seurat\" needed for this function to work. Please install it.", call. = FALSE)
}
y <- Seurat::GetAssayData(x, slot = slot, assay = assay)
z <- Seurat::Embeddings(x, coord)
if (! is.null(dims)) {
z <- z[, dims, drop = FALSE]
}
if(is.null(method)){
if(ncol(z)==2){
method <- "2D"
} else if(ncol(z)>2){
method <- "highD"
}
message("### Input coordinates have ",ncol(z)," dimensions, so method set to \"",method,"\"")
}
y <- y > cutoff
if (use.advanced.sampling) {
use.advanced.sampling = colSums(y)
}
haystack(z, detection = y, method = method, use.advanced.sampling = use.advanced.sampling, ...)
}
#' @rdname haystack
#' @export
haystack.SingleCellExperiment <- function(x, assay = "counts", coord = "TSNE", dims = NULL, cutoff = 1, method = NULL, use.advanced.sampling = NULL, ...) {
if (!requireNamespace("SummarizedExperiment", quietly = TRUE)) {
stop("Package \"SummarizedExperiment\" needed for this function to work. Please install it.", call. = FALSE)
}
if (!requireNamespace("SingleCellExperiment", quietly = TRUE)) {
stop("Package \"SingleCellExperiment\" needed for this function to work. Please install it.", call. = FALSE)
}
y <- SummarizedExperiment::assay(x, assay)
z <- SingleCellExperiment::reducedDim(x, coord)
if(is.null(z)) {
stop("No coordinates named ", coord, " found.")
}
if (! is.null(dims)) {
z <- z[, dims, drop = FALSE]
}
if(is.null(method)){
if(ncol(z)==2){
method <- "2D"
} else if(ncol(z)>2){
method <- "highD"
}
message("### Input coordinates have ",ncol(z)," dimensions, so method set to \"",method,"\"")
}
y <- y > cutoff
if (use.advanced.sampling) {
use.advanced.sampling = colSums(y)
}
haystack(z, detection = y, method = method, use.advanced.sampling = use.advanced.sampling, ...)
}
#' Visualizing the detection/expression of a gene in a 2D plot
#'
#' @param x a matrix or other object from which coordinates of cells can be extracted.
#' @param dim1 column index or name of matrix for x-axis coordinates.
#' @param dim2 column index or name of matrix for y-axis coordinates.
#' @param assay name of assay data for Seurat method.
#' @param slot name of slot for assay data for Seurat method.
#' @param coord name of coordinates slot for specific methods.
#' @param ... further parameters passed to plot_gene_haystack_raw().
#'
#' @export
#'
plot_gene_haystack <- function(x, ...) {
UseMethod("plot_gene_haystack")
}
#' @rdname plot_gene_haystack
#' @export
plot_gene_haystack.matrix <- function(x, dim1 = 1, dim2 = 2, ...) {
plot_gene_haystack_raw(x[, dim1], x[, dim2], ...)
}
#' @rdname plot_gene_haystack
#' @export
plot_gene_haystack.data.frame <- function(x, dim1 = 1, dim2 = 2, ...) {
plot_gene_haystack_raw(x[, dim1], x[, dim2], ...)
}
#' @rdname plot_gene_haystack
#' @export
plot_gene_haystack.SingleCellExperiment <- function(x, dim1 = 1, dim2 = 2, assay = "counts", coord = "TSNE", ...) {
if (!requireNamespace("SummarizedExperiment", quietly = TRUE)) {
stop("Package \"SummarizedExperiment\" needed for this function to work. Please install it.", call. = FALSE)
}
if (!requireNamespace("SingleCellExperiment", quietly = TRUE)) {
stop("Package \"SingleCellExperiment\" needed for this function to work. Please install it.", call. = FALSE)
}
y <- SummarizedExperiment::assay(x, assay)
z <- SingleCellExperiment::reducedDim(x, coord)
plot_gene_haystack_raw(z[, dim1], z[, dim2], expression = y, ...)
}
#' @rdname plot_gene_haystack
#' @export
plot_gene_haystack.Seurat <- function(x, dim1 = 1, dim2 = 2, assay = "RNA", slot = "data", coord = "tsne", ...) {
if (!requireNamespace("Seurat", quietly = TRUE)) {
stop("Package \"Seurat\" needed for this function to work. Please install it.", call. = FALSE)
}
y <- Seurat::GetAssayData(x, slot = slot, assay = assay)
z <- Seurat::Embeddings(x, coord)
plot_gene_haystack_raw(z[, dim1], z[, dim2], expression = y, ...)
}
#' Visualizing the detection/expression of a set of genes in a 2D plot
#'
#' @param x a matrix or other object from which coordinates of cells can be extracted.
#' @param dim1 column index or name of matrix for x-axis coordinates.
#' @param dim2 column index or name of matrix for y-axis coordinates.
#' @param assay name of assay data for Seurat method.
#' @param slot name of slot for assay data for Seurat method.
#' @param coord name of coordinates slot for specific methods.
#' @param ... further parameters passed to plot_gene_haystack_raw().
#'
#' @export
#'
plot_gene_set_haystack <- function(x, ...) {
UseMethod("plot_gene_set_haystack")
}
#' @rdname plot_gene_set_haystack
#' @export
plot_gene_set_haystack.matrix <- function(x, dim1 = 1, dim2 = 2, ...) {
plot_gene_set_haystack_raw(x[, dim1], x[, dim2], ...)
}
#' @rdname plot_gene_set_haystack
#' @export
plot_gene_set_haystack.data.frame <- function(x, dim1 = 1, dim2 = 2, ...) {
plot_gene_set_haystack_raw(x[, dim1], x[, dim2], ...)
}
#' @rdname plot_gene_set_haystack
#' @export
plot_gene_set_haystack.SingleCellExperiment <- function(x, dim1 = 1, dim2 = 2, assay = "counts", coord = "TSNE", ...) {
if (!requireNamespace("SummarizedExperiment", quietly = TRUE)) {
stop("Package \"SummarizedExperiment\" needed for this function to work. Please install it.", call. = FALSE)
}
if (!requireNamespace("SingleCellExperiment", quietly = TRUE)) {
stop("Package \"SingleCellExperiment\" needed for this function to work. Please install it.", call. = FALSE)
}
y <- SummarizedExperiment::assay(x, assay)
z <- SingleCellExperiment::reducedDim(x, coord)
plot_gene_set_haystack_raw(z[, dim1], z[, dim2], detection = y > 1, ...)
}
#' @rdname plot_gene_set_haystack
#' @export
plot_gene_set_haystack.Seurat <- function(x, dim1 = 1, dim2 = 2, assay = "RNA", slot = "data", coord = "tsne", ...) {
if (!requireNamespace("Seurat", quietly = TRUE)) {
stop("Package \"Seurat\" needed for this function to work. Please install it.", call. = FALSE)
}
y <- Seurat::GetAssayData(x, slot = slot, assay = assay)
z <- Seurat::Embeddings(x, coord)
plot_gene_set_haystack_raw(z[, dim1], z[, dim2], detection = y > 1, ...)
}
#' Function for hierarchical clustering of genes according to their expression distribution in 2D or multi-dimensional space
#'
#' @param x a matrix or other object from which coordinates of cells can be extracted.
#' @param dim1 column index or name of matrix for x-axis coordinates.
#' @param dim2 column index or name of matrix for y-axis coordinates.
#' @param ... further parameters passed down to methods.
#'
#' @export
#'
hclust_haystack <- function(x, ...) {
UseMethod("hclust_haystack")
}
#' @rdname hclust_haystack
#' @export
hclust_haystack.matrix <- function(x, dim1 = 1, dim2 = 2, ...) {
hclust_haystack_raw(x[, dim1], x[, dim2], ...)
}
#' @rdname hclust_haystack
#' @export
hclust_haystack.data.frame <- function(x, dim1 = 1, dim2 = 2, ...) {
hclust_haystack_raw(x[, dim1], x[, dim2], ...)
}
#' Function for k-means clustering of genes according to their expression distribution in 2D or multi-dimensional space
#'
#' @param x a matrix or other object from which coordinates of cells can be extracted.
#' @param dim1 column index or name of matrix for x-axis coordinates.
#' @param dim2 column index or name of matrix for y-axis coordinates.
#' @param ... further parameters passed down to methods.
#'
#' @export
#'
kmeans_haystack <- function(x, ...) {
UseMethod("kmeans_haystack")
}
#' @rdname kmeans_haystack
#' @export
kmeans_haystack.matrix <- function(x, dim1 = 1, dim2 = 2, ...) {
kmeans_haystack_raw(x[, dim1], x[, dim2], ...)
}
#' @rdname kmeans_haystack
#' @export
kmeans_haystack.data.frame <- function(x, dim1 = 1, dim2 = 2, ...) {
kmeans_haystack_raw(x[, dim1], x[, dim2], ...)
}
|
a7be91c8531bcdb9ab6f7c76e049369856581dde | 3e374bdfbc0d3bb2933fc248285263dd3e45ec48 | /R/xgboost.R | 52d1017ea94a7a01a9d572afe400d4290cc9970e | [
"Apache-2.0"
] | permissive | RBigData/pbdXGB | c309fc6ce317f5db6c3ac9a838ef71724657f062 | c76b807cfd60aaa61e8ab233873c6e3ea41b3e6c | refs/heads/master | 2020-09-07T05:46:55.322685 | 2020-02-16T23:44:09 | 2020-02-16T23:44:09 | 220,674,364 | 0 | 0 | Apache-2.0 | 2020-02-16T23:44:10 | 2019-11-09T16:59:56 | C++ | UTF-8 | R | false | false | 1,213 | r | xgboost.R | # Simple interface for training an xgboost model that wraps \code{xgb.train}.
# Its documentation is combined with xgb.train.
#
#' @rdname xgb.train
#' @export
xgboost <- function(data = NULL, label = NULL, missing = NA, weight = NULL,
params = list(), nrounds,
verbose = 1, print_every_n = 1L,
early_stopping_rounds = NULL, maximize = NULL,
save_period = NULL, save_name = "xgboost.model",
xgb_model = NULL, callbacks = list(), ...) {
dtrain <- xgb.get.DMatrix(data, label, missing, weight)
watchlist <- list(train = dtrain)
bst <- xgb.train(params, dtrain, nrounds, watchlist, verbose = verbose, print_every_n = print_every_n,
early_stopping_rounds = early_stopping_rounds, maximize = maximize,
save_period = save_period, save_name = save_name,
xgb_model = xgb_model, callbacks = callbacks, ...)
return(bst)
}
# Various imports
#' @importClassesFrom Matrix dgCMatrix
#' @importFrom data.table rbindlist
#' @importFrom stringi stri_split_regex
#' @importFrom stats predict
#'
#' @import methods
#' @import xgboost
#' @import pbdMPI
#'
NULL
|
466692faf81626a468779c27caa9e486cf6312b9 | 01588666e7f7f7c5fbe2e7fea1c1c732851f3f7e | /cachematrix.R | 91c1b97dbcaaf9ca3a288ddc941b7f62bb08d287 | [] | no_license | sanpau/ProgrammingAssignment2 | 38e4537ee82a9ed30a93caa109bdaddec466ac1f | 88626791f6adb10705301618a70ed9ded6e664a7 | refs/heads/master | 2020-07-06T08:47:40.899163 | 2016-11-18T17:29:42 | 2016-11-18T17:29:42 | 74,050,666 | 0 | 0 | null | 2016-11-17T17:25:55 | 2016-11-17T17:25:55 | null | UTF-8 | R | false | false | 1,772 | r | cachematrix.R |
## pair of functions that
## cache the inverse of a matrix.
## `makeCacheMatrix`: This function creates a special "matrix" object
## that can cache its inverse.
makeCacheMatrix <- function(x = matrix()) {
invt <- NULL
# function that sets the value of the matrix IN set
set <- function(y) {
x <<- y
invt <<- NULL
}
#returns matrix stored in the main function
get <- function() x
# store the value of the input in a variable invt into the main function
# makeCacheMatrix (setinvt) and return it (getinvt)
setinvt <- function(inverse) invt <<- inverse
getinvt <- function() invt
list(set = set, get = get, setinvt = setinvt, getinvt = getinvt)
}
## This function computes the inverse of the special
##"matrix" returned by `makeCacheMatrix` above. If the inverse has
##already been calculated (and the matrix has not changed), then
##`cacheSolve` should retrieve the inverse from the cache.
cacheSolve <- function(x, ...) {
## If the inverse has already been calculated, then the cacheSolve should retrieve the inverse from the cache.
invt <- x$getinvt()
if(!is.null(invt)) {
message("getting cached result")
return(invt)
}
## If the inverse has not been calculated, data gets the matrix stored with makeCacheMatrix,
## invt calculates the inverse, and x$setinvt(invt) stores it in the object invt in makeCacheMatrix
data <- x$get()
invt <- solve(data, ...)
x$setinvt(invt)
invt
}
##Output of the above :
## mtrx <- matrix(c(1,1,1,3,4,3,3,3,4),3,3)
##> mtrx
##[,1] [,2] [,3]
##[1,] 1 3 3
##[2,] 1 4 3
##[3,] 1 3 4
##> mtrx1 <- makeCacheMatrix(mtrx)
##> cacheSolve(mtrx1)
##[,1] [,2] [,3]
##[1,] 7 -3 -3
##[2,] -1 1 0
##[3,] -1 0 1
|
e8db9e2229b109ccb9f01002b79904edfe10f166 | 601d899094f8f73a5356e30fd8d721801b6de757 | /R/transformations.R | 707f34bfaea19ea9f8fc62be97b67974c386a40e | [
"MIT"
] | permissive | vmikk/vmik | e8b5b7e90342fad800ac8a139b1590b1becc454e | 4d11200e3928f248050303203e3c7829139817dd | refs/heads/master | 2020-04-07T08:12:12.754247 | 2018-11-19T07:16:02 | 2018-11-19T07:16:02 | 158,204,616 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,111 | r | transformations.R | ## TO DO:
# - IHS - add 'back' flag to perform reverse transformation
# - gelman_scale - see also arm::rescale(, binary.inputs = "full")
# Standardizing by сentering and вividing by 2 standard deviations
gelman_scale <- function(x){
x.obs <- x[!is.na(x)]
mm <- mean(x.obs)
ss <- sd(x.obs)
res <- (x - mm)/(2 * ss)
attr(res, "scaled:center") <- mm
attr(res, "scaled:scale") <- ss
return(res)
}
#' Inverse hyperbolic sine transformation
#'
#' @description Inverse hyperbolic sine transformation. Unlike a log variable, the inverse hyperbolic sine is defined at zero.
#' @param x Numeric vector
#' @return Vector of transformed values.
#' @details It is an alternative to log transformations when some of the variables take on zero or negative values and as an alternative to the Box-Cox when variables are zero or negative.
#' Except for very small values of y, the inverse sine is approximately equal to log(2yi) or log(2)+log(yi), and so it can be interpreted in exactly the same way as a standard logarithmic dependent variable. For example, if the regression coefficient on "urban" is 0.1, that tells us that urbanites have approximately 10 percent higher wealth than non-urban people.
#' @seealso \code{asinh}
#' @references Burbidge J.B., Magee L., Robb A.L. Alternative transformations to handle extreme values of the dependent variable // Journal of the American Statistical Association. 1988. V. 83. № 401. P. 123-127.
#' @examples
#' IHS(seq(-10, 10, 1))
#'
IHS <- function(x){
log(x + sqrt(x^2 + 1))
}
#' Scale numeric vector the specified interval
#'
#' @param x Numeric vector.
#' @param limitMin Lower value of the interval (default = 0).
#' @param limitMax Upper value of the interval (default = 1).
#'
#' @return Transformed values of \code{x} that belongs to the new range.
#' @seealso scales::rescale
#' @examples
#' x <- 1:20
#' scale_to_interval(x)
#' scale_to_interval(x, 2, 5)
#'
scale_to_interval <- function(x, limitMin = 0, limitMax = 1){
res <- (limitMax - limitMin) * (x - min(x))
res <- res / ( (max(x) - min(x)))
res <- res + limitMin
return(res)
}
|
12be28f962e06a5b47e7664009f16b6f5eb80dd8 | dec3db3c118c3aea6f73288b43c5e87c90f60091 | /FigS2_ExtendedData2/FigS2.R | ce60865db5963d9b9f0c0e8468f067480629d387 | [] | no_license | livkosterlitz/Figures-Jordt-et-al-2020 | c3b48d702107a3697d3fecd42410d031ff4a328c | 70e06b4affeddcff15b5c78a8922ee1b8dcd08bf | refs/heads/master | 2021-01-08T10:28:43.307167 | 2020-08-06T16:54:15 | 2020-08-06T16:54:15 | 242,003,143 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 4,229 | r | FigS2.R | library(tidyverse)
library(cowplot)
library(lattice)
library(gridExtra)
library(grid)
library(egg)
#########################
#FigureS2########
#####################
###Ancestor###
dat <- read.csv("FigS2_low.csv")
dat <- dat %>%
#select(-Mixture, -CFUs) %>%
group_by(Host, Antibiotic, Day) %>%
summarise(N = n(),
CFUs = mean(CFUs_sub),
SD = sd(CFUs_sub),
SE = SD/sqrt(N))
levels(dat$Antibiotic)
colors_light <- c('darkgrey', 'firebrick1', 'blue2', 'darkorchid2')
p1 <- ggplot(dat %>% filter(Host=='A'),
aes(x=Day, y=CFUs, color=Antibiotic)) +
theme_cowplot(12)+
geom_line(linetype = "dashed", size = 0.4668623442372146) +
geom_errorbar(aes(ymin=CFUs-SE, ymax=CFUs+SE), width=0, size = 0.4668623442372146) +
geom_point(aes(shape=Antibiotic, color=Antibiotic, fill=Antibiotic, size=Antibiotic)) +
scale_color_manual(values=colors_light) +
scale_fill_manual(values=colors_light)+
scale_size_manual(values=c(2,1,1,2))+
scale_shape_manual(values=c(16,25,24,18)) +
scale_y_log10(expand=c(0, 0),
limits=c(1e0, 1e4),
breaks=10^seq(0, 4, by=1),
labels=seq(0, 4, by=1)) +
scale_x_continuous(expand=c(0, 0), limits = c(0,4.2)) +
theme(legend.position="none") +
theme(axis.title.x=element_blank()) +
theme(axis.title.y=element_blank()) +
theme(axis.line.y = element_line(size = 0.3734899)) +
theme(axis.line.x = element_line(size = 0.3734899)) +
theme(axis.ticks = element_line(size = 0.3734899))+
theme(axis.text.x = element_text(margin=margin(1,0,0,0,"pt")),
axis.text.y = element_text(margin=margin(0,1,0,0,"pt")))+
theme(axis.ticks.length=unit(.025, "in"))+
theme(plot.margin = margin(.2, 0, 0, .05, "in"))+
expand_limits(x = 0, y = 0) +
theme(axis.text = element_text(size = 8))
p1
###Evolved###
dat1 <- read.csv("FigS2_high.csv")
dat1 <- dat1 %>%
#select(-Mixture, -CFUs) %>%
group_by(Host, Antibiotic, Day) %>%
summarise(N = n(),
CFUs = mean(CFUs_sub),
SD = sd(CFUs_sub),
SE = SD/sqrt(N))
colors_dark <- c('darkgrey', 'firebrick1', 'blue2', 'darkorchid2')
p3 <- ggplot(dat1 %>% filter(Host=='B'),
aes(x=Day, y=CFUs, color=Antibiotic)) +
theme_cowplot(12)+
geom_line(size = 0.4668623442372146) +
geom_errorbar(aes(ymin=CFUs-SE, ymax=CFUs+SE), width=0, size = 0.4668623442372146) +
geom_point(aes(shape=Antibiotic, color=Antibiotic, fill=Antibiotic, size=Antibiotic)) +
scale_color_manual(values=colors_light) +
scale_fill_manual(values=colors_light)+
scale_size_manual(values=c(2,1,1,2))+
scale_shape_manual(values=c(16,25,24,18)) +
scale_y_log10(expand=c(0, 0),
limits=c(1e0, 1e4),
breaks=10^seq(0, 4, by=1),
labels=seq(0, 4, by=1)) +
scale_x_continuous(expand=c(0, 0), limits = c(0,4.2)) +
theme(legend.position="none") +
theme(axis.title.x=element_blank()) +
theme(axis.title.y=element_blank()) +
theme(axis.line.y = element_line(size = 0.3734899)) +
theme(axis.line.x = element_line(size = 0.3734899)) +
theme(axis.ticks = element_line(size = 0.3734899))+
theme(axis.text.x = element_text(margin=margin(1,0,0,0,"pt")),
axis.text.y = element_text(margin=margin(0,1,0,0,"pt")))+
theme(axis.ticks.length=unit(.025, "in"))+
theme(plot.margin = margin(.2, 0, 0, .05, "in"))+
expand_limits(x = 0, y = 0) +
theme(axis.text = element_text(size = 8))
p3
Figp1_fixed <- set_panel_size(p1, width = unit(1.35, "in"), height = unit(1.35, "in"))
Figp3_fixed <- set_panel_size(p3, width = unit(1.35, "in"), height = unit(1.35, "in"))
FigureS2_main <- plot_grid(Figp1_fixed, Figp3_fixed,
ncol = 2, align = "v",
labels = c('a','b'), label_size = 10)
FigureS2_main
#create common x and y labels
y.grob <- textGrob("cell density [log ((CFUs/mL)+1)]",
gp=gpar(fontsize=10), rot=90)
x.grob <- textGrob("Transfer",
gp=gpar(fontsize=10))
#add common axis to plot
FigureS2 <- grid.arrange(arrangeGrob(FigureS2_main, left = y.grob, bottom = x.grob))
save_plot("FigureS2.pdf", plot = FigureS2, base_width = 3.45, base_height = 1.85)
|
7cacc52e5e950fa90e77b3274fe74e2a8b05885a | b0a8f8d6984078682450613d8c351664341c47a8 | /Plot1.R | ed4e140738440b7eea40388ab191a374ae7bea02 | [] | no_license | silhouetted/ExData_Plotting1 | c20df765d223f815c1a4f380b8d2c26c32c2cf02 | fb6b06e8149d3a2640eb580e7cbb4169410001ab | refs/heads/master | 2020-04-05T22:11:02.827530 | 2018-11-12T18:35:17 | 2018-11-12T18:35:17 | 157,247,818 | 0 | 0 | null | 2018-11-12T17:04:38 | 2018-11-12T17:04:37 | null | UTF-8 | R | false | false | 2,026 | r | Plot1.R | #### Exploratory data analysis Week 1 Assignment script
### Reading in and subsetting the data to 1st and 2nd Feb 2007
fileUrl <- "https://d396qusza40orc.cloudfront.net/exdata%2Fdata%2Fhousehold_power_consumption.zip"
# download file if it does not exist
if (!file.exists("ElectricalUsageData.zip")) {
download.file(fileUrl, destfile = "ElectricalUsageData.zip", mode = wb)
}
# unzip file if not unzipped
if(!file.exists("EUD")){
unzip("ElectricalUsageData.zip", exdir = "./EUD")
}
# read in the file that was unzipped into the EUD directory marking ? as NA
allElectricData <- read.table("./EUD/household_power_consumption.txt", sep = ";", header = TRUE, colClasses = c("character", "character", "numeric", "numeric", "numeric", "numeric", "numeric", "numeric", "numeric"), na.strings = "?")
# load lubridate to set the date as date rather than character
library(lubridate)
# convert the date to a date format
allElectricData$Date <- dmy(allElectricData$Date) # date in yyyy-mm-dd format
# subset the data from the specific date in Feb we need
febElectricData <- subset(allElectricData, Date == "2007-02-01" | Date == "2007-02-02")
# delete the rest of the data from workspace (taking up nearly 150mb of precious RAM)
rm(allElectricData)
# Draw the first PNG file - a histogram
png("Plot1.png") # set output file
with(febElectricData, hist(Global_active_power, col = "red", main = "Global Active Power", xlab = "Global Active Power (kilowatts)"))
dev.off() # turn off output to png
if(file.exists("Plot1.png")) {
cat("File created successfully")
} else {
warning("File not created. Please try again.")
}
|
d6b4fdaa8c469899a29b7ee5df2dbea7503f947c | 9df25083c9e3b935853cd941e1f6e9dfd4878322 | /01-PH01-SC1-FINAL.R | 4a35d431d9d33a8d49f073252f1790414656252c | [
"MIT"
] | permissive | rintukutum/precisionFDA-BCC-PH01 | 13640c440795f80281ed8835273c32c7923cd10d | e4392a4ea96b266dfdcf3f5e8b952cf73b727190 | refs/heads/master | 2020-12-29T05:26:35.916184 | 2020-02-05T15:57:30 | 2020-02-05T15:57:30 | 238,470,111 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 4,171 | r | 01-PH01-SC1-FINAL.R | #-----------------
# TRAIN, CV & TEST
rm(list=ls())
load('./data/sc1.data.RData')
#-------------------------------
# SPLIT DATA INTO TRAIN AND TEST
set.seed(907)
train.idx <- caret::createDataPartition(
y = sc1.data$outcome$SURVIVAL_STATUS,
times = 1,
p = 0.8
)[[1]]
tr.samp <- sc1.data$outcome$PATIENTID[train.idx]
table(sc1.data$outcome$SURVIVAL_STATUS[train.idx])
#------------------
# 80% dataset
# 0 1
# 40 262
#------------------
sc1.train <- list(
phenotype = sc1.data$phenotype[sc1.data$phenotype$PATIENTID %in% tr.samp,],
y = sc1.data$outcome[sc1.data$outcome$PATIENTID %in% tr.samp,'SURVIVAL_STATUS'],
x = sc1.data$feature[sc1.data$feature$PATIENTID %in% tr.samp,-1]
)
sc1.test <- list(
phenotype = sc1.data$phenotype[!(sc1.data$phenotype$PATIENTID %in% tr.samp),],
y = sc1.data$outcome[!(sc1.data$outcome$PATIENTID %in% tr.samp),'SURVIVAL_STATUS'],
x = sc1.data$feature[!(sc1.data$feature$PATIENTID %in% tr.samp),-1]
)
save(sc1.train,sc1.test,file = './data/PH01-SC1-model-data.RData')
#----------------
# Feature reduction
#----------------
rm(list = ls())
load('./data/PH01-SC1-model-data.RData')
source('func-room.R')
idx.0 <- sc1.train$y == 0
idx.1 <- sc1.train$y == 1
feature.pvals <- apply(sc1.train$x,2,performTEST)
sc1.feature.pvals <- feature.pvals[order(feature.pvals)]
save(sc1.feature.pvals,
file='./data/PH01-SC1-feature-pvals.RData')
#-----------------------
# TRAINING PHASE, 5-FOLD CV
rm(list=ls())
load('./data/PH01-SC1-feature-pvals.RData')
load('./data/PH01-SC1-model-data.RData')
source('func-room.R')
p.cutoffs <- c(1.0e-05,1.0e-04,1.0e-03,1.0e-02)
for(i in 1:length(p.cutoffs)){
message('Threshold pval <= ',p.cutoffs[i])
p.cutoff <- p.cutoffs[i]
sc1.features <- sc1.feature.pvals[sc1.feature.pvals <= p.cutoff]
#---------------
model.feature <- names(sc1.features)
tr.x <- sc1.train$x[,model.feature]
tr.y <- as.factor(sc1.train$y)
tt.x <- sc1.test$x[,model.feature]
tt.y <- as.factor(sc1.test$y)
ph01.sc1.SVM.mods <- performSVM(
tr.x = tr.x,
tr.y = tr.y,
tt.x = tt.x,
tt.y = tt.y,
SEED = 768
)
perf.PH01.SC1 <- getPerfMetrics(x = ph01.sc1.SVM.mods)
dir.create('./data/MODEL-PERF-PH01/',showWarnings = FALSE)
outname <- paste0(
'./data/MODEL-PERF-PH01/perf.PH01.SC1',
p.cutoff,'.RData')
save(perf.PH01.SC1,
file = outname)
p.train <- get.TPR.FPR.plot(x = ph01.sc1.SVM.mods,type = 'train')
p.test <- get.TPR.FPR.plot(x = ph01.sc1.SVM.mods,type = 'test')
dir.create('./figures',showWarnings = FALSE)
outfile <- paste0('./figures/PH01-SC1-SVM-',p.cutoff,'.png')
png(outfile,
width = 1000,height = 1000,
res = 200)
gridExtra::grid.arrange(
p.train + ggtitle(paste('Train | pval <= ',p.cutoff)),
p.test + ggtitle(paste('Test | pval <= ',p.cutoff)))
dev.off()
}
#-----------------------
# FINAL MODEL
rm(list=ls())
load('./data/PH01-SC1-feature-pvals.RData')
load('./data/PH01-SC1-model-data.RData')
sc1.features <- sc1.feature.pvals[sc1.feature.pvals <= 1.0e-05]
#---------------
model.feature <- names(sc1.features)
message('Features used!')
message(paste(model.feature,collapse = ', '))
tr.x <- sc1.train$x[,model.feature]
tr.y <- as.factor(sc1.train$y)
tt.x <- sc1.test$x[,model.feature]
tt.y <- as.factor(sc1.test$y)
sc1.mod.data <- list(
tr.x = tr.x,
tr.y = tr.y,
tt.x = tt.x,
tt.y = tt.y
)
dir.create('./data/MODEL-DATA-PH01',showWarnings = FALSE)
save(sc1.mod.data,
file = './data/MODEL-DATA-PH01/sc1.mod.data.RData')
#-------------------------------------------------------
#-------------------------------------------------------
rm(list=ls())
load('./data/MODEL-DATA-PH01/sc1.mod.data.RData')
source('func-room.R')
ph01.sc1.SVM.mods <- performSVM(
tr.x = sc1.mod.data$tr.x,
tr.y = sc1.mod.data$tr.y,
tt.x = sc1.mod.data$tt.x,
tt.y = sc1.mod.data$tt.y,
SEED = 768
)
perf.PH01.SC1 <- getPerfMetrics(x = ph01.sc1.SVM.mods)
sink('./submission/Summary-Phase1-SC1-TRAIN.txt')
print(perf.PH01.SC1$tr)
sink()
dir.create('./data/MODEL-PERF-FINAL-PH01/',showWarnings = FALSE)
outname <- './data/MODEL-PERF-FINAL-PH01/perf.PH01.SC1.FINAL.RData'
save(perf.PH01.SC1,
file = outname)
|
a96e2c29d4df17c11d0fd581aec8dbf5515a5056 | 7af0de4a6767812f392bd69a2298f45550f8abb5 | /Bagged_Loess_Lattice.R | 00d0ae511c8c7bbe9c2d7af021799907d9afb88f | [] | no_license | SudhakaranP/Statistical_Learning_Basics | 615077494c15c9ae8f28cd3e856eee7b8cd03678 | 40162b9831bdc165da5af926cc2c7ba8a9fe674f | refs/heads/master | 2021-06-14T18:56:40.890625 | 2016-12-14T02:04:34 | 2016-12-14T02:04:34 | 105,226,455 | 0 | 1 | null | 2017-09-29T03:37:24 | 2017-09-29T03:37:23 | null | UTF-8 | R | false | false | 862 | r | Bagged_Loess_Lattice.R | library(ElemStatLearn)
set.seed(105)
ll <- matrix(NA,nrow=100,ncol=155)
for(i in 1:100){
ss <- sample(1:dim(ozone)[1],replace=T)
ozone0 <- ozone[ss,]; ozone0 <- ozone0[order(ozone0$ozone),]
loess0 <- loess(temperature ~ ozone,data=ozone0,span=0.2)
ll[i,] <- predict(loess0,newdata=data.frame(ozone=1:155))
}
xyplot(temperature~ozone, data = ozone
, pch = 19, col = "black"
, bagLines = ll
, prepanel = function(x,y, bagLines) {
list(xlim = c(1, max(x))
, ylim = c(min(bagLines, na.rm = TRUE)
, max(bagLines, na.rm = TRUE)))
}
, panel = function(x,y, ...) {
for (i in 1:100) {
panel.lines(1:155,ll[i,], alpha = 0.1, col = "green")
}
panel.xyplot(x,y, ...)
panel.lines(1:155,apply(ll,2,mean),col="red",lwd=3)
}
)
|
293ff38a2ed703a9fc70b6a468f439e2b5d98b3d | 171398356488ee085e053595b7e9e43d671c1586 | /HW4-Gairola-Abhijit.R | 2cdb3e9de979d1908bfcc973452e18ded7693902 | [] | no_license | dexter11235813/Math185 | 2e83986eba0af687b444bd97d7bfd02a4d0c2bd1 | ae5d60b5c38a2cba38274867657c3e33be380e97 | refs/heads/master | 2020-07-03T23:42:52.900984 | 2016-11-19T16:51:55 | 2016-11-19T16:51:55 | 74,222,485 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 3,222 | r | HW4-Gairola-Abhijit.R | #Q1
head(ToothGrowth)
SST = function(dat,Y...)
{
temp1 = c()
for(i in 1:2)
{
temp1[i] = mean(dat[,i,])
}
return(30*sum((temp1 - Y...)^2))
}
# for B = 1:999, we permute data within each block and calculate SST of that table, and compare it with the SST of the original data.
twowayPermTest = function(ToothGrowth, B = 999)
{
a11 = subset(ToothGrowth,(supp == "VC") & (dose ==0.5))$len
a12 = subset(ToothGrowth,(supp == "VC") & (dose ==1))$len
a13 = subset(ToothGrowth,(supp == "VC") & (dose ==2))$len
a21 = subset(ToothGrowth,(supp == "OJ") & (dose ==0.5))$len
a22 = subset(ToothGrowth,(supp == "OJ") & (dose ==1))$len
a23 = subset(ToothGrowth,(supp == "OJ") & (dose ==2))$len
two.way.table = data.frame(a11,a12,a13,a21,a22,a23)
v = c()
for(i in 1:10)
{
for(j in 1:6)
{
v = c(v,two.way.table[i,j])
}
}
# store the data from ToothGrowth in a 3x2x10 array
two.way.table.3D = array(v,c(3,2,10))
colnames(two.way.table.3D) = c("VC","OJ")
rownames(two.way.table.3D) = c(0.5,1,2)
Y... = mean(two.way.table.3D)
# calculating SST on the original given data
d = SST(two.way.table.3D, Y...)
#
# Permuting across all treatements within each block
d.perm = c()
for(j in 1:B)
{
for(i in 1:3)
{
temp = c(two.way.table.3D[i,1,],two.way.table.3D[i,2,])
temp = sample(temp)
two.way.table.3D[i,1,] = temp[1:10]
two.way.table.3D[i,2,] = temp[11:20]
}
d.perm[j] = SST(two.way.table.3D,Y...)
}
p.val = (length(which(d.perm >= d))+1)/(B+1)
return(p.val)
}
print(twowayPermTest(ToothGrowth,999))
# pvalues range from 0.01 - 0.03
#Q2
load("alon.RData")
alon = as.data.frame(alon)
colnames(alon) = c(seq(1:2000),"y")
p.val.vec = c()
for(i in 1:2000)
{
p.val.vec[i] = as.numeric(t.test(alon[,i] ~ y,data = alon)$p.value)
}
p.val.vec = as.vector(p.val.vec)
#sorting the vector containing p values
temporary = p.val.vec[order(p.val.vec)]
# number of Hypothesis rejected by using the Bonferroni correction
length(which(p.adjust(temporary,"bon") <= 0.05))
# number of Hypothesis rejected by using the Holm correction
length(which(p.adjust(temporary,"holm") <= 0.05))
# number of Hypothesis rejected by using the Hochberg correction
length(which(p.adjust(temporary,"hoch")<=0.05))
# number of Hypothesis rejected by using the FDR correction
length(which(p.adjust(temporary,"BH")<=0.05))
# From the above runs,we see that FDR is the most conservative of all corrections, with 190 rejected hypothesis,whereas
# Bonferroni,Holm and Hochberg all reject 11 out of 2000 hypothesis.
#Q3
B = 50 # low B value used to reduce runtime
t.test.stat = c()
t.test.stat.perm = c()
#vector that holds the p-value for each of the 2000 genes.
p.value = 0
for(i in 1:2000)
{
t.test.stat[i] = t.test(alon[,i] ~ y , data = alon)$statistic
}
for(i in 1:B)
{
# sampling the subjects 'y'
alon$y = sample(alon$y)
for(j in 1:2000)
{
t.test.stat.perm[j] = t.test(alon[,j] ~ y,data = alon)$statistic
}
p.value = p.value + ifelse((abs(t.test.stat.perm) >= abs(t.test.stat)),1,0)/(B+1)
}
print(p.value) |
05dbcc03744c5f696c565a3a23013663a05a62dc | 808f796821392a6ce4dd2a243aac01cce1513e0e | /src/plotFlightLine.R | c13602d3708600dd725d5d1abc7bdc436aeecc47 | [] | no_license | hdugan/WrightValley_AEM | f7c7b9ba72c7fe3a609cd9acade05cb705fc7de3 | 6eb83f7d761ee75d77124808409a4749bf0e4f10 | refs/heads/main | 2023-04-15T03:06:19.803795 | 2022-08-04T15:31:58 | 2022-08-04T15:31:58 | 475,973,442 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 6,740 | r | plotFlightLine.R | #dvdpLocation is UTMX, UTMY, depth. ex) c(416673, 1391082, 85.73)
plotFlightLine <- function(lineNo, aemDF, dvdpLocation = NULL, dvdpLocation2 = NULL) {
# Derive depths of each layer
depths = aemDF |> filter(LINE_NO == lineNo) |>
dplyr::select(DEP_TOP_1:DEP_TOP_30) |>
pivot_longer(cols = DEP_TOP_1:DEP_TOP_30, names_to = "bin", values_to = "depths") |>
mutate(bin = parse_number(bin)) |>
group_by(bin) |>
summarise_all(first)
# Get depth of DOI
doi = aemDF |> filter(LINE_NO == lineNo) |>
dplyr::select(UTMX, UTMY, DOI_CONSERVATIVE)
# Get distance between points
flightDist = aemDF |> filter(LINE_NO == lineNo) |>
dplyr::select(UTMX, UTMY) |>
mutate(row = row_number())
# Make all lines plot west to east
WE = -1
if(flightDist |> slice(1) |> pull(UTMX) > flightDist |> slice(n()) |> pull(UTMX)) {
flightDist = flightDist |> map_df(rev)
WE = 1 # This comes into play later for blackout boxes
}
flightDist = flightDist |>
mutate(across(.fns = lag, .names = '{col}_next')) |>
rowwise() |>
mutate(dist = pointDistance(c(UTMX, UTMY),c(UTMX_next, UTMY_next), method = 'Euclidean', lonlat = F)[1]) |>
mutate(dist = if_else(is.na(dist), 0, dist)) |>
ungroup() |>
mutate(distcum = cumsum(dist)) |>
arrange(row)
# add in thickness and DOIs and flightDist
a = aemDF |> filter(LINE_NO == lineNo) |>
dplyr::select(UTMX, UTMY, ELEVATION, RHO_I_1:RHO_I_30) |>
pivot_longer(cols = RHO_I_1:RHO_I_30, names_to = "bin", values_to = "rho") |>
mutate(bin = parse_number(bin)) |>
left_join(depths) |>
left_join(doi) |>
left_join(flightDist) |>
mutate(depth.elev = round(ELEVATION - depths), doi.elev = round(ELEVATION - DOI_CONSERVATIVE))
# No 30 bin for 2011 flight lines
if (any(is.na(a |> filter(bin ==30) |> pull(rho)))) {
a = a |> filter(bin != 30)
}
mindoi = min(a$depth.elev)
# b = data.frame(new.depths = seq(min(a$depth.elev), max(a$depth.elev), by = 1))
b = expand_grid(distcum = unique(a$distcum), depth.elev = seq(min(a$depth.elev, na.rm = T), max(a$depth.elev, na.rm = T), by = 1)) |>
left_join(a) |>
group_by(distcum) %>%
mutate(rhoInterp = na.approx(rho, na.rm=FALSE, method = 'linear')) |>
mutate(rhoInterp = if_else(rhoInterp < 1, 1, rhoInterp)) |>
mutate(rhoInterp = if_else(rhoInterp > 1e4, 1e4, rhoInterp)) |>
mutate(rhoBackground = if_else(!is.na(rhoInterp), 1e4, NA_real_)) |>
ungroup()
# Get bathymetry
# test extracting depths over line
testline = a |> filter(bin == 1) |> dplyr::select(UTMX,UTMY)
coordinates(testline) <- ~UTMX+UTMY
raster::crs(testline) = raster::crs(DEM3)
testline.depths <- raster::extract(DEM3, # raster layer
testline, # SPDF with centroids for buffer
method = 'simple',
buffer = 5, # buffer size, units depend on CRS
fun = median, # what to value to extract
df = FALSE) # return a dataframe?
testline.df = a |> filter(bin == 1) |>
dplyr::select(UTMX,UTMY, ELEVATION, distcum) |>
mutate(bath.depth = testline.depths) |>
mutate(bathy = ELEVATION - bath.depth)
# Extract DVDP borehole location for plot
if (!is.null(dvdpLocation)) {
dvdpLine = a |> filter(bin == 1) |>
dplyr::select(UTMX,UTMY, ELEVATION, distcum) |>
mutate(UTMXd = dvdpLocation[1], UTMYd = dvdpLocation[2]) |>
mutate(UTMxDIFF = abs(UTMX - UTMXd), UTMyDIFF = abs(UTMY - UTMYd)) |>
filter(UTMxDIFF == min(UTMxDIFF))
}
if (!is.null(dvdpLocation2)) {
dvdpLine2 = a |> filter(bin == 1) |>
dplyr::select(UTMX,UTMY, ELEVATION, distcum) |>
mutate(UTMXd = dvdpLocation2[1], UTMYd = dvdpLocation2[2]) |>
mutate(UTMxDIFF = abs(UTMX - UTMXd), UTMyDIFF = abs(UTMY - UTMYd)) |>
filter(UTMxDIFF == min(UTMxDIFF))
}
# CONTINUOUS SCALE - Interpolated rho
# Get black out zones
rr = b |>
filter(bin == 1) |>
mutate(index = row_number()) |>
mutate(test = dist>100) %>%
filter(test == TRUE) |>
pull(index)
blackout = b |>
filter(bin == 1) |>
mutate(index = row_number()) |>
filter(index %in% rr | index %in% (rr+WE))
# Plot resistivity
options(scipen = 1)
mybreaks = seq(0,4)
mylabels = 10^mybreaks
# Dataframe for DOI polygon
b.ribbon = b |>
filter(!is.na(DOI_CONSERVATIVE)) |>
filter(bin == 30) |>
mutate(doi.elev = if_else(doi.elev < depth.elev, depth.elev, doi.elev))
#Plots using contours
p2 = ggplot(b) +
geom_contour_fill(aes(x = distcum, y = depth.elev, z = log10(rhoBackground))) + # Interpolation can leave gaps... this fills in with max limit of 1e4
geom_contour_fill(bins = 20, aes(x = distcum, y = depth.elev, z = log10(rhoInterp))) +
scale_fill_gradientn(name = 'rho (Ω-m)', colours = rev(met.brewer("Hiroshige", n=100)),
limits = log10(c(1, 1e4)),
breaks = mybreaks, labels = mylabels,
oob = scales::squish) +
geom_ribbon(data = b.ribbon,
aes(x = distcum, ymax = doi.elev, ymin = mindoi), alpha = 0.8) + # DOI ribbon
geom_path(data = testline.df, aes(x = distcum, y = bathy)) + # lake bathymetry
xlab('Distance W → E (m)') + ylab('Elevation (m)') +
theme_minimal(base_size = 8)
if (!is.null(dvdpLocation)) {
p2 = p2 + geom_linerange(data = dvdpLine, aes(x = distcum, ymin = ELEVATION - dvdpLocation[3], ymax = ELEVATION), color = 'black', size = 1) # DVDP borehole
}
if (!is.null(dvdpLocation2)) {
p2 = p2 + geom_linerange(data = dvdpLine2, aes(x = distcum, ymin = ELEVATION - dvdpLocation2[3], ymax = ELEVATION), color = 'black', size = 1) # DVDP borehole
}
if(nrow(blackout) == 2) {
p2 = p2 + geom_rect(data = blackout, aes(xmin = distcum[1], xmax = distcum[2], ymin = mindoi, ymax = ELEVATION[1])) # blackout
}
if(nrow(blackout) > 2) {
p2 = p2 + geom_rect(data = blackout, aes(xmin = distcum[1], xmax = distcum[2], ymin = mindoi, ymax = ELEVATION[1])) # blackout
p2 = p2 + geom_rect(data = blackout, aes(xmin = distcum[3], xmax = distcum[4], ymin = mindoi, ymax = ELEVATION[3])) # blackout
p2 = p2 + geom_rect(data = blackout, aes(xmin = distcum[5], xmax = distcum[6], ymin = mindoi, ymax = ELEVATION[5])) # blackout
p2 = p2 + geom_rect(data = blackout, aes(xmin = distcum[7], xmax = distcum[8], ymin = mindoi, ymax = ELEVATION[7])) # blackout
p2 = p2 + geom_rect(data = blackout, aes(xmin = distcum[9], xmax = distcum[10], ymin = mindoi, ymax = ELEVATION[9])) # blackout
}
return(p2)
}
|
4a24b742e5d454d131d0a57f8f434ff38416b3f1 | 39d7f1d2d81c2a9fc37df54cccc8571f7a1d31e0 | /EM/R_codes/c_em_algorithm_generic.R | 89011d5b58571043b69db3f126b2899df14506c1 | [] | no_license | historical-record-linking/matching-codes | 4a989b65f12791a206c24e8646a1f37881bf44ec | 31aaf4a85056b3f58aa71342611ba2faead66306 | refs/heads/master | 2022-11-21T18:32:15.941226 | 2020-07-08T23:15:26 | 2020-07-08T23:15:26 | 258,260,033 | 16 | 0 | null | null | null | null | UTF-8 | R | false | false | 7,136 | r | c_em_algorithm_generic.R |
rm(list=ls())
dropbox <- "C:/Users/acald/Desktop/test_em_data/"
# set directories
EMdistances <- paste(dropbox,"data/new_codes/em_santi_small2/EMdistances/",sep = "")
EMmatches <- paste(dropbox,"data/new_codes/em_santi_small2/EMmatches/",sep = "")
# Select threshold for Algorithm and maximum number of iterations
stop_at_param<-0.00001
iter<-3000
#Load packages
library(stringdist)
library(doParallel)
library(foreign)
library(readstata13)
library(dplyr)
library(doParallel)
library(plyr)
library(doParallel)
library(foreach)
registerDoParallel(cores=2)
options(scipen=999)
#Import appended data on summary of distances (created by jaro_winkler_names.R)
data<-vector()
data <- read.table(paste(EMdistances,"summary.csv",sep=""), header = T)
data <- data.frame(data)
Dist_FN<-data$Dist_FN
Dist_LN<-data$Dist_LN
Age_Dist<-data$Age_Dist
# *EM Algorithm*
## PRIOR Distributions: Initial value for proportion of matches
n_obs<-sum(data$Count)
p_M_0<-data$N[1]/n_obs
p_U_0<-1-p_M_0
###String distribution: multinomial
###Set number of categorical variables (k) for names and ages
#### 4 groups defined below by the dic_strdist functions
k_N<-4
k_A<-6
i_N<-1:k_N
i_A<-1:k_A
#Priors are set equal to 1/k for the unmatched and to 1/k + [(k-1)/2 - (i-1)]/(k^2) for the matched
###First name priors (on the unmatched assume no further information provided by knowing that unmatched. for the matched impose a decreasing LR):
theta_strdist_FN_U_0<-(aggregate(data$Count, by=list(Dist_FN=data$Dist_FN), FUN=sum)/n_obs)[,2]
theta_strdist_FN_M_0<-(1/k_N+((k_N-1)/2-(i_N-1))/k_N^2)
for (i in c(k_N:2)){
if (theta_strdist_FN_M_0[i-1]/theta_strdist_FN_M_0[i]<theta_strdist_FN_U_0[i-1]/theta_strdist_FN_U_0[i]){
theta_strdist_FN_M_0[i-1]<-theta_strdist_FN_M_0[i]*theta_strdist_FN_U_0[i-1]/theta_strdist_FN_U_0[i]
}
}
theta_strdist_FN_M_0<-theta_strdist_FN_M_0/sum(theta_strdist_FN_M_0)
###Last name priors (my prior is that the Pr(Distance/U) is approx the empirical frequencies of each distance)
theta_strdist_LN_U_0<-(aggregate(data$Count, by=list(Dist_LN=data$Dist_LN), FUN=sum)/n_obs)[,2]
theta_strdist_LN_M_0<-(1/k_N+((k_N-1)/2-(i_N-1))/k_N^2)
for (i in c(k_N:2)){
if (theta_strdist_LN_M_0[i-1]/theta_strdist_LN_M_0[i]<theta_strdist_LN_U_0[i-1]/theta_strdist_LN_U_0[i]){
theta_strdist_LN_M_0[i-1]<-theta_strdist_LN_M_0[i]*theta_strdist_LN_U_0[i-1]/theta_strdist_LN_U_0[i]
}
}
theta_strdist_LN_M_0<-theta_strdist_LN_M_0/sum(theta_strdist_LN_M_0)
###Age distribution: multinomial over 0,1,2,3,4,5 absolute age difference
theta_agedist_U_0<-(aggregate(data$Count, by=list(Age_Dist=data$Age_Dist), FUN=sum)/n_obs)[,2]
theta_agedist_M_0<-(1/k_A+((k_A-1)/2-(i_A-1))/k_A^2)
for (i in c(k_A:2)){
if (theta_agedist_M_0[i-1]/theta_agedist_M_0[i]<theta_agedist_U_0[i-1]/theta_agedist_U_0[i]){
theta_agedist_M_0[i-1]<-theta_agedist_M_0[i]*theta_agedist_U_0[i-1]/theta_agedist_U_0[i]
}
}
theta_agedist_M_0<-theta_agedist_M_0/sum(theta_agedist_M_0)
print(theta_agedist_M_0/theta_agedist_U_0)
print(theta_strdist_FN_M_0/theta_strdist_FN_U_0)
print(theta_strdist_LN_M_0/theta_strdist_LN_U_0)
#Match enriched sample
p_age_M<-theta_agedist_M_0[data[,1]+1]
p_age_U<-theta_agedist_U_0[data[,1]+1]
p_str_FN_M<-theta_strdist_FN_M_0[data[,2]]
p_str_FN_U<-theta_strdist_FN_U_0[data[,2]]
p_str_LN_M<-theta_strdist_LN_M_0[data[,3]]
p_str_LN_U<-theta_strdist_LN_U_0[data[,3]]
m<-(p_str_FN_M*p_str_LN_M*p_age_M*p_M_0)/(p_M_0)
u<-(p_str_FN_U*p_str_LN_U*p_age_U*p_U_0)/(p_U_0)
data<-data[order(-log(m/u)),]
weight<-log(m/u)
weight<-weight[order(-weight)]
data<-cbind(data,c(1:length(data$Count)))
data$Count<-data$Count/(data[,6])
n_obs<-sum(data$Count)
p_M_0<-data$N[1]/n_obs
p_U_0<-1-p_M_0
## Start loop
t<-1
error<-10
error_v<-vector()
while(error>stop_at_param & t<iter){
#Pr(distance/parameters)
p_age_M<-theta_agedist_M_0[data[,1]+1]
p_age_U<-theta_agedist_U_0[data[,1]+1]
p_str_FN_M<-theta_strdist_FN_M_0[data[,2]]
p_str_FN_U<-theta_strdist_FN_U_0[data[,2]]
p_str_LN_M<-theta_strdist_LN_M_0[data[,3]]
p_str_LN_U<-theta_strdist_LN_U_0[data[,3]]
w<-(p_str_FN_M*p_str_LN_M*p_age_M*p_M_0)/((p_str_FN_M*p_str_LN_M*p_age_M*p_M_0)+(p_str_FN_U*p_str_LN_U*p_age_U*p_U_0))
p_M_1<-weighted.mean(w,data[,4])
p_U_1<-1-p_M_1
theta_agedist_M_1<-vector()
theta_agedist_U_1<-vector()
theta_strdist_FN_M_1<-vector()
theta_strdist_FN_U_1<-vector()
theta_strdist_LN_M_1<-vector()
theta_strdist_LN_U_1<-vector()
#Updated values of the parameters
for (i in 1:k_A){
theta_agedist_M_1[i]<-(w[which(data[,1]==i-1)]%*%data[which(data[,1]==i-1),4])/(w%*%data[,4])
theta_agedist_U_1[i]<-((1-w[which(data[,1]==i-1)])%*%data[which(data[,1]==i-1),4])/((1-w)%*%data[,4])
}
for (i in 1:k_N){
theta_strdist_FN_M_1[i]<-(w[which(data[,2]==i)]%*%data[which(data[,2]==i),4])/(w%*%data[,4])
theta_strdist_FN_U_1[i]<-((1-w[which(data[,2]==i)])%*%data[which(data[,2]==i),4])/((1-w)%*%data[,4])
theta_strdist_LN_M_1[i]<-(w[which(data[,3]==i)]%*%data[which(data[,3]==i),4])/(w%*%data[,4])
theta_strdist_LN_U_1[i]<-((1-w[which(data[,3]==i)])%*%data[which(data[,3]==i),4])/((1-w)%*%data[,4])
}
#Check difference in absolute value
t<-t+1
error1<-max(abs(rbind(theta_agedist_M_1,theta_agedist_U_1)-rbind(theta_agedist_M_0,theta_agedist_U_0)))
error2<-max(abs(rbind(theta_strdist_FN_M_1,theta_strdist_FN_U_1,theta_strdist_LN_M_1,theta_strdist_LN_U_1)-rbind(theta_strdist_FN_M_0,theta_strdist_FN_U_0,theta_strdist_LN_M_0,theta_strdist_LN_U_0)))
error3<-abs(p_M_1-p_M_0)
error<-max(error1,error2,error3)
error_v[t]<-error
#Update values
theta_agedist_M_0<-theta_agedist_M_1
theta_agedist_U_0<-theta_agedist_U_1
theta_strdist_FN_M_0<-theta_strdist_FN_M_1
theta_strdist_FN_U_0<-theta_strdist_FN_U_1
theta_strdist_LN_M_0<-theta_strdist_LN_M_1
theta_strdist_LN_U_0<-theta_strdist_LN_U_1
p_M_0<-p_M_1
p_U_0<-p_U_1
print(t)
}
#Create final w:
w_final<-(p_str_FN_M*p_str_LN_M*p_age_M*p_M_0)/((p_str_FN_M*p_str_LN_M*p_age_M*p_M_0)+(p_str_FN_U*p_str_LN_U*p_age_U*p_U_0))
w_final<-round(w_final, 8)
#Save parameters
parameters<-c(theta_agedist_M_0,theta_agedist_U_0,theta_strdist_FN_M_0,theta_strdist_FN_U_0,theta_strdist_LN_M_0,theta_strdist_LN_U_0,p_M_0)
names<-c("AM0","AM1","AM2","AM3","AM4","AM5","AU0","AU1","AU2","AU3","AU4","AU5","FM0","FM1","FM2","FM3","FU0","FU1","FU2","FU3","LM0","LM1","LM2","LM3","LU0","LU1","LU2","LU3","P")
parameters<-cbind(names, parameters)
file1<-paste(EMmatches,"EM_Estimates_parameters.csv",sep="")
write.table(parameters,file1, row.names=F)
#Save estimates of probabilities (w)
probs_export<-cbind(data, w_final)
colnames(probs_export) <- c("Age_Dist", "strdist_FN_index", "strdist_LN_index", "counts", "N", "w_final")
file2<-paste(EMmatches,"EM_Estimates_probabilities.csv",sep="")
write.table(probs_export,file2, row.names=F)
|
0dedd99043c8ea48bf311280366ae359d6bb4841 | bdd8d4b527d36aa0c69d127aa6c071b70dc3fffb | /Admin/testando - fields of study.R | 7f90e47fe53ea76373569052f3ff3a550e94b385 | [] | no_license | antrologos/harmonizeIBGE | 2abc3fa53106c026d29e11f83bae8e642049ab86 | 8c6053b54b434eddab84f6e32f43424f22d06170 | refs/heads/master | 2022-03-13T11:16:06.906636 | 2022-02-28T17:25:21 | 2022-02-28T17:25:21 | 141,569,637 | 2 | 2 | null | null | null | null | UTF-8 | R | false | false | 6,495 | r | testando - fields of study.R | rm(list=ls());gc();Sys.sleep(.5);gc()
options(scipen=999)
library(harmonizeIBGE)
library(Hmisc)
library(descr)
library(fst)
#======================================================================================================
setwd("E:/Dropbox-Ro/Dropbox/Rogerio/Bancos_Dados/Censos")
variaveis <- fread("E:/Google Drive/RCodes/PacotesR/harmonizeIBGE/Admin/variaveis_CENSOS.csv")
anos <- c(1960, 1970, 1980, 1991, 2000, 2010)
themes_to_open <- c("identification", "demographics", "education")
n = 30000000
read_harmonize_and_save = F
if(read_harmonize_and_save == T){
for(i in 1:6){
ano = variaveis$year[i]
print(paste("===================================================================================", ano))
vars_to_open <- harmonizeIBGE:::list_originalVariables_to_drop(ano, themes = themes_to_open) %>%
unlist() %>%
c(., toupper(.), tolower(.)) %>%
unique()
vars_to_drop <- harmonizeIBGE:::list_originalVariables_to_drop(ano, themes = c("identification", "demographics")) %>%
unlist() %>%
c(tolower(.)) %>%
unique()
if(ano == 1970){
vars_to_open <- c(vars_to_open, "CEM005")
vars_to_drop <- c(vars_to_drop, "CEM005")
}
assign(x = paste0("c_",ano),
value = fread(paste0("Censo ", ano, "/", variaveis$file_person[i]),
select = vars_to_open,
nrows = n) %>%
prepare_to_harmonize(type = "census", year = ano, state_var_name = ifelse(ano == 1970, "CEM005", ""))
)
Sys.sleep(.5);gc()
assign(x = paste0("c_",ano),
value = get(paste0("c_",ano)) %>%
harmonize_themes(themes = c("identification", "demographics")) %>%
filter(age >= 17) %>%
select(-vars_to_drop))
gc();Sys.sleep(.5);gc()
assign(x = paste0("c_",ano),
value = get(paste0("c_",ano)) %>%
harmonize_themes(themes = "education") %>%
filter(!is.na(educationAttainment)) %>%
setDT())
gc();Sys.sleep(.5);gc()
}
setwd("e:/censos_tmp")
for(ano in anos){
print(ano)
write_fst(x = get(paste0("c_",ano)), path = paste0("censo_",ano,"_fieldsOfStudy.csv"))
gc()
}
}else{
setwd("e:/censos_tmp")
for(ano in anos){
print(ano)
assign(x = paste0("c_",ano),
value = read_fst(path = paste0("censo_",ano,"_fieldsOfStudy.csv"),as.data.table = T) %>%
select(-idhh, -idperson, -famStatus, -nonrelative, -levelattnd, -literacy) %>%
prepare_to_harmonize(type = "census", year = ano, state_var_name = "CEM005")
)
gc();Sys.sleep(.2);gc()
}
}
c_1991[, wgtperson := wgtperson/(10^8)]
#======================================================================================================
#labels_isced <- readxl::read_xlsx(crosswalk_location, sheet = "Fields_Codes_labels") %>%
# select(isced_code_level3, isced_label_level3_en) %>%
# rename(isced = isced_code_level3,
# label = isced_label_level3_en) %>%
# setDT(key = "isced")
ano = 1960
freq_isced_aggreg = NULL
for(ano in anos){
print(ano)
assign(x = paste0("c_",ano),
value = get(paste0("c_",ano)) %>%
build_education_fieldsOfStudy(aggregated = T)
)
gc();Sys.sleep(.3);gc()
table = get(paste0("c_",ano))[, freq(label_fieldsOfStudy, w = wgtperson)]
freq_isced_aggreg_i = tibble(ano = ano,
isced = attr(table, "dimnames")[[1]],
freq_abs = round(table[,1],digits = 0),
freq_rel = round(table[,3], digits =3)) %>%
filter(complete.cases(.))
freq_isced_aggreg <- bind_rows(freq_isced_aggreg, freq_isced_aggreg_i)
gc();Sys.sleep(.1);gc()
}
freq_isced = NULL
for(ano in c(1980,1991,2000,2010)){
print(ano)
assign(x = paste0("c_",ano),
value = get(paste0("c_",ano)) %>%
build_education_fieldsOfStudy(aggregated = F)
)
gc();Sys.sleep(.3);gc()
table = get(paste0("c_",ano))[, freq(label_fieldsOfStudy, w = wgtperson)]
freq_isced_i = tibble(ano = ano,
isced = attr(table, "dimnames")[[1]],
freq_abs = round(table[,1],digits = 0),
freq_rel = round(table[,3], digits =3)) %>%
filter(complete.cases(.))
freq_isced <- bind_rows(freq_isced, freq_isced_i)
gc();Sys.sleep(.1);gc()
}
#c_1980[v525 == 85, fieldsOfStudy := 999]
#c_1980[, freq(fieldsOfStudy)]
freq_abs_isced_wide_aggreg <- freq_isced_aggreg %>%
select(-freq_rel) %>%
filter(!(isced=="Total")) %>%
spread(key = ano, value = freq_abs) %>%
setDT()
freq_rel_isced_wide_aggreg <- freq_isced_aggreg %>%
select(-freq_abs) %>%
mutate(freq_rel = round(freq_rel, 3)) %>%
filter(!(isced=="Total")) %>%
spread(key = ano, value = freq_rel) %>%
setDT()
freq_abs_isced_wide <- freq_isced %>%
select(-freq_rel) %>%
filter(!(isced=="Total")) %>%
spread(key = ano, value = freq_abs) %>%
setDT()
freq_rel_isced_wide <- freq_isced %>%
select(-freq_abs) %>%
filter(!(isced=="Total")) %>%
mutate(freq_rel = round(freq_rel, 3)) %>%
spread(key = ano, value = freq_rel) %>%
setDT()
|
5ace1e957a32ec7f8ce702bcefd13bc3a46af9c3 | b4a58ba2dbffed266dc06f6dcb32f22797271f2e | /04.heatmap.R | 7ed517177dfd0895e1a617c8823366557784d54d | [] | no_license | SamYangBio/R | 063476d8cb29e1f00d63152a45d4ab78d34067a2 | 7c756d2c74071801f46c33646c7ba19104aef987 | refs/heads/master | 2020-03-24T23:26:01.455474 | 2018-10-14T10:33:13 | 2018-10-14T10:33:13 | 143,134,931 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 761 | r | 04.heatmap.R | args <- commandArgs(TRUE)
input <- args[1]
out <- args[2]
library('gplots')
a = read.table(input, sep = "\t", header = T, check.names = F)
lie = ncol(a)
hang = nrow(a)
high_pre = hang/10
high = round(high_pre, digits = 0)
high = high + 10
high = ifelse(high < 7, 7, ifelse(high>200, 200, high))
x = a[,2:lie]
y = as.matrix(x)
y = log10(y+1)
rownames(y) = a[,1]
scale = ifelse(ncol(y)>2, "row", "none")
pdf(file = out,height=high)
par(oma = c(3,3,3,5))
heatmap.2(y,
Colv=NA,dendrogram=('row'), # make x axis clusterting
col=colorRampPalette(c("navy","white","firebrick3")),
trace = "none",
cexCol=1,
srtCol=45,
scale = scale,
cexRow = 0.6,
lhei=c(10,100)
)#,labRow=F)
|
4c1eb42eb0fac1178462f410ad6b45d2d65a9bed | 67222f69dd1a5b5ced1d28df833a303924dbde35 | /2. Algorithms on Datasets/Supervised Machine Learning Techniques/Decision Tree/Company_Data/Company_Decision+Tree.R | dfd991d189ae2a2d3b632e7d4a1b72a2f44cb5f0 | [] | no_license | mandarmakhi/DataScience-R-code | 4f75906507e303fb9b438b99a5eab0a74bcc77f6 | 8c1728b306e53668b1814283da9936503e0554b9 | refs/heads/master | 2023-01-19T04:55:11.171455 | 2020-11-28T07:59:55 | 2020-11-28T07:59:55 | 263,417,867 | 1 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,263 | r | Company_Decision+Tree.R | #Decision Tree
#A cloth manufacturing company is interested to know about the segment or attributes causes high sale.
#install.packages("caret")
#install.packages("C50")
library(C50)
library(factoextra)
library(caret)
library(gmodels)
#Lets Import the Dataset
company <- read.csv("C:/Users/Mandar/Desktop/data/assignments/decision tree/Company_Data/Company_Data.csv")
View(company)
attach(company)
head(company)
summary(company) #Gives the Summary of the Dataset
str(company) #Gives the Structure of the Dataset
ShelveLoc <- as.factor(ShelveLoc)
Urban <- as.factor(Urban)
US <- as.factor(US)
High <- ifelse(Sales > 8,"Yes","No")
High <- as.factor(High)
#Lets Combine it With the Main Dataset
company_new <- cbind(company[,-1], High) #Here we Exclude the Sales Column as we have Derived a responsive Variable High using it
#Lets Create the Training and Testing sets
indatapartition <- createDataPartition(company_new$High, p=.70, list = F) #This will Hold 70% of the whole dataset
training <- company_new[indatapartition,]
testing <- company_new[-indatapartition,]
#Lets Fit the Model for Entire Data Now
tree_model <- C5.0(High~. , data =company_new)
summary(tree_model) #Here we can See there was an Error of 6.5%
tree_model
#Lets Plot the Tree
plot(tree_model)
#Lets Build the Model for Training Set and Then Predict he Values for Testing set
t_model <- C5.0(High~. , data = training, method ="class")
summary(t_model) #Here we can see that there is an Error rate of 8.5%
t_model
pred <- predict.C5.0(t_model, newdata = testing)
pred
a<- table(testing$High, pred)
Accuracy <- sum(diag(a))/sum(a)
Accuracy
CrossTable(testing$High, pred)
#So here we can see that there are Misclassifications, we can improve the models using Bagging and Boosting techniques
#We use For loop for bagging in order to make multiple models
acc<- c()
for (i in 1:50) #This will create 500 different models
{
print(i)
#Build a Model
fittree <- C5.0(High~. , data = training,method = "class", trials = 10) #Trials is a Boosting Parameter
pred2<- predict.C5.0(fittree,testing[,-11])
ab<- table(testing$High, pred2)
#To save the Accuracy of the models
acc<- c(acc,sum(diag(ab))/sum(ab))
}
summary(acc)
summary(fittree)
plot(fittree)
|
2d2fac878ca51a454eb0d917e7b9081b962bd167 | c8a8ad16a9633de5bf1bd2917ae63702ae58d4ac | /R/allele_divtables-class.R | 1829b19b86b06fb84fecfae1b02ae9168f84828b | [] | no_license | douglasgscofield/dispersalDiversity | 5487b5cc8ce31e9fc64e5398afa71db9a12a6ac6 | 3c826115e6fca06b801346452e2659ec5142a715 | refs/heads/main | 2023-04-14T08:17:59.717926 | 2021-03-23T16:44:16 | 2021-03-23T16:44:16 | 4,307,834 | 1 | 0 | null | null | null | null | UTF-8 | R | false | false | 6,852 | r | allele_divtables-class.R | #' @include divtable-class.R
NULL
#' List of divtables holding allele diversity data
#'
#' An object of class \code{allele_divtables} is a list of
#' \code{\link{divtable}} objects, each representing sites-by-allele counts
#' data for a single genetic locus. Row and column names are
#' the site names and individual alleles, respectively.
#' This is the basic data object for analysis of genetic data using the
#' \code{\link{dispersalDiversity}} package. It is accepted by the
#' function \code{\link{diversity}}, which generates descriptive statistics,
#' and the functions \code{\link{alphaDiversityTest}},
#' \code{\link{gammaContrastTest}} and others that test for differences
#' in the structure of genetic diversity.
#'
#' Assembling an \code{allele_divtables} by hand is labourious. More
#' typically, genotype data will be in a class \code{\link{genalex}} object
#' and then converted to \code{\link{allele_divtables}} using
#' \code{\link{createAlleleTables}}. The \code{\link{createAlleleTables}}
#' and \code{\link{as.allele_divtables}} functions will attempt to convert
#' non-\code{\link{genalex}} objects using to \code{\link{genalex}} format
#' using \code{\link[readGenalex]{as.genalex}}, but this may not be successful.
#'
#' @examples
#'
#' ## One possible way to plot an \code{allele_divtables} object, plotting
#' ## each locus as a separate divtable in a two-column format
#'
#' data(Qagr_pericarp_genotypes)
#' pal <- createAlleleTables(Qagr_pericarp_genotypes)
#' par(mfrow = c(round(length(pal) / 2), 2))
#' lapply(names(pal), function(n) plot(pal[[n]], main = n, l2 = NULL, las = 2))
#'
#' @name allele_divtables-class
#'
#' @aliases allele_divtables
#'
NULL
#' Generate an allele_divtables object from a class genalex object
#'
#' S3 method to convert an object of class \code{genalex} to an object of
#' class \code{\link{allele_divtables}}. a list of \code{\link{divtable}}
#' objects representing sites-by-allele counts. This is an S3 generic
#' so that other methods might be written to convert other genetic
#' formats.
#'
#' If \code{x} is not of class \code{genalex}, an attempt is made to convert
#' it to class \code{genalex} using \code{\link[readGenalex]{as.genalex}}
#' from the
#' \href{http://cran.r-project.org/web/packages/readGenalex/index.html}{readGenalex}
#' package. An error will be produced if \code{x} is of a class or format
#' that cannot be converted to class \code{genalex}.
#'
#' Another option for converting genotypes to \code{\link{allele_divtables}}
#' objects is to convert to one of the formats recognised by
#' \code{\link[readGenalex]{as.genalex}}.
#'
#' Although missing alleles may be common in genotypic data, there is no
#' provision in \code{\link{diversity}} and other functions in this package
#' for missing data. Missing alleles are recognised and excluded if they
#' match one of the values in \code{exclude}. The numbers of missing alleles
#' recognised is reported if \code{quiet = FALSE}.
#'
#' @note \code{as.allele_divtables} is a synonym, unless \code{x} is of class
#' \code{list}. If so, if the class of each element of \code{x} is of class
#' \code{\link{divtable}}, then the class of \code{x} is changed to
#' \code{\link{allele_divtable}}. If \code{x} is of class
#' \code{allele_divtables} it is returned unchanged.
#'
#' @param x Object of class \code{genalex} holding genotypes to be converted,
#' or of a class and format that can be converted to \code{genalex} using
#' \code{\link[readGenalex]{as.genalex}}
#'
#' @param exclude Values in \code{x} that indicate missing alleles, these are
#' excluded from the \code{divtable} entries for each locus
#'
#' @param quiet If \code{TRUE}, report the number of missing alleles excluded
#'
#' @return Object of class \code{\link{allele_divtable}}
#'
#' @examples
#'
#' ## Use genotype data from readGenalex package, already loaded
#' data(Qagr_pericarp_genotypes)
#' pal <- createAlleleTables(Qagr_pericarp_genotypes)
#' str(pal)
#'
#' ## The divtable for the first locus
#' pal[[1]]
#'
#' ## allele_divtables removes and can report missing data
#' data(Qagr_adult_genotypes)
#' aal <- createAlleleTables(Qagr_adult_genotypes, quiet = FALSE)
#'
#' @export createAlleleTables as.allele_divtables as.allele_divtables.default as.allele_divtables.genalex as.allele_divtables.list as.allele_divtables.allele_divtables
#'
#' @aliases createAlleleTables as.allele_divtables as.allele_divtables.default as.allele_divtables.genalex as.allele_divtables.list as.allele_divtables.allele_divtables
#'
#' @name createAlleleTables
#'
NULL
createAlleleTables <- function(x, ...) UseMethod("createAlleleTables")
#' @rdname createAlleleTables
#'
#' @export
#'
createAlleleTables.default <- function(x, ...)
{
if (inherits(x, 'data.frame') || inherits(x, 'loci')) {
x <- as.genalex(x)
return(createAlleleTables.genalex(x))
} else {
stop("Cannot convert to class 'allele_divtables', perhaps ",
deparse(substitute(x)), " can be converted to class 'genalex'?",
" See '?readGenalex::as.genalex'.")
}
}
#' @rdname createAlleleTables
#'
#' @export
#'
createAlleleTables.allele_divtables <- function(x, ...)
x
#' @rdname createAlleleTables
#'
#' @export
#'
createAlleleTables.genalex <- function(x, exclude = c(NA, "0"),
quiet = TRUE, ...)
{
ploidy <- attr(x, "ploidy")
lc <- attr(x, "locus.columns")
ln <- attr(x, "locus.names")
population <- attr(x, "pop.title")
ans <- list()
ex <- list()
for (il in 1:length(lc)) {
alleles <- as.vector(unlist(x[, lc[il]:(lc[il] + ploidy - 1)]))
ex[[ ln[il] ]] <- sum(alleles %in% exclude)
pop <- rep(x[[population]], ploidy)
ans[[ ln[il] ]] <- as.divtable(table(pop, alleles, exclude = exclude))
}
if (sum(unlist(ex)) && ! quiet)
cat(sprintf("Excluding %d entries based on 'exclude = c(%s)'\n",
sum(unlist(ex)), paste(collapse = ", ", exclude)))
structure(ans, class = c('allele_divtables', 'list'))
}
#----------------------------------
#
# synonyms, documented and exported in createAlleleTables above
as.allele_divtables <- function(x, ...) UseMethod("as.allele_divtables")
as.allele_divtables.default <- function(x, ...)
createAlleleTables.default(x, ...)
as.allele_divtables.genalex <- function(x, ...)
createAlleleTables.genalex(x, ...)
# These are not synonyms but are still documented in createAlleleTables above
#
as.allele_divtables.list <- function(x, ...)
{
if (all(sapply(x, inherits, 'divtable')))
structure(x, class('allele_divtables', 'list'))
else stop(deparse(substitute(x)),
" cannot be converted to class allele_divtables,",
" all members must be class 'divtable'")
}
as.allele_divtables.allele_divtables <- function(x, ...)
x
|
443d6f2f7ee000a470d256dcb826509db41a19de | 13cdc54d90f2ba332f558eaebd799de1edfe4a17 | /scripts/visualizations.R | 48dde87aa02f04c7fb3b92fd8302e6d007970ea1 | [
"CC0-1.0"
] | permissive | UACC-renedherrera/UAZCC_deliverables | adecdaf55e811ea6ef1cfda86b4b879c19104202 | dcadbb69d121776c7538c7217ecd8876f5098bc0 | refs/heads/master | 2023-03-08T21:51:32.340103 | 2021-02-27T06:55:16 | 2021-02-27T06:55:16 | 282,557,115 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 15,298 | r | visualizations.R | #### set up ####
# packages
library(here)
library(tidyverse)
library(ggthemes)
# color palette
blues_8 <- c("#f7fbff",
"#deebf7",
"#c6dbef",
"#9ecae1",
"#6baed6",
"#4292c6",
"#2171b5",
"#084594")
blues_3 <- c("#deebf7",
"#9ecae1",
"#3182bd")
mixed_8 <- c("#deebf7",
"#9ecae1",
"#3182bd",
"#edf8e9",
"#bae4b3",
"#74c476",
"#31a354",
"#006d2c")
#### read data ####
df <- read_rds("data/tidy/data_for_visualizations.rds")
distinct(df, category)
str(df)
#### race ####
table_race <- df %>%
filter(category == "Race") %>%
group_by(area)
table_race$attribute <- as.factor(table_race$attribute)
table_race$attribute <- ordered(table_race$attribute, levels = c("Non-Hispanic White", "Hispanic", "American Indian", "Non-Hispanic Black"))
table_race %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value, fill = attribute)) +
geom_bar(stat = "identity", position = "stack", color = "black", alpha = 0.666) +
labs(
title = "Proportion of each Race and Ethnicity Category in Catchment Geographies",
subtitle = "",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("demographics_race.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
#### demographics ####
# age ----
table_demographics <- df %>%
filter(category == "Demographics",
area != "Catchment")
unique(table_demographics$area)
table_demographics$area <- ordered(table_demographics$area, levels = c("USA", "AZ", "Cochise", "Pima", "Pinal", "Santa Cruz", "Yuma"))
unique(table_demographics$attribute)
table_demographics_age <- table_demographics %>%
filter(attribute == "Median Age")
unique(table_demographics_age$attribute)
table_demographics_age$attribute <- as.factor(table_demographics_age$attribute)
table_demographics_age %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value)) +
geom_bar(stat = "identity", fill = "#0C234B") +
geom_label(aes(label = value)) +
labs(
title = "Median Age in Catchment Geographies",
subtitle = "",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("demographics_age.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# sex ----
table_demographics <- df %>%
filter(category == "Demographics")
unique(table_demographics$attribute)
table_demographics_sex <- table_demographics %>%
filter(attribute == "Female")
unique(table_demographics_sex$attribute)
table_demographics_sex$attribute <- as.factor(table_demographics_sex$attribute)
table_demographics_sex %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value)) +
geom_bar(stat = "identity", fill = "#0C234B") +
geom_label(aes(label = round(value, digits = 2))) +
labs(
title = "Female Proportion in Catchment Geographies",
subtitle = "",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("demographics_sex.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# education attainment ----
table_demographics <- df %>%
filter(category == "Demographics")
unique(table_demographics$attribute)
table_demographics_edu <- table_demographics %>%
filter(attribute == "High School Graduation" | attribute == "Some College" | attribute == "College Graduate")
unique(table_demographics_edu$attribute)
table_demographics_edu$attribute <- as.factor(table_demographics_edu$attribute)
table_demographics_edu$attribute <- ordered(table_demographics_edu$attribute, levels = c("High School Graduation", "Some College", "College Graduate"))
table_demographics_edu %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value, fill = attribute)) +
geom_bar(stat = "identity", position = "dodge", color = "black", alpha = 0.666) +
labs(
title = "Proportion of Educational Attainment in Catchment Geographies",
subtitle = "Proportion of population with high school diploma or college",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("demographics_edu.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# income ----
table_demographics <- df %>%
filter(category == "Demographics")
table_demographics$area <- ordered(table_demographics$area, levels = c("USA", "AZ", "Catchment", "Cochise", "Pima", "Pinal", "Santa Cruz", "Yuma"))
unique(table_demographics$area)
unique(table_demographics$attribute)
table_demographics_income <- table_demographics %>%
filter(attribute == "Median Family income" | attribute == "Mean family income")
unique(table_demographics_income$attribute)
table_demographics_income$attribute <- as.factor(table_demographics_income$attribute)
table_demographics_income %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value)) +
geom_bar(stat = "identity", position = "dodge", color = "black", fill = "#0C234B") +
geom_label(aes(label = value)) +
facet_wrap("attribute") +
labs(
title = "Mean and Median Income in Catchment Geographies",
subtitle = "in US dollars",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("demographics_income.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# poverty ----
table_demographics <- df %>%
filter(category == "Demographics")
unique(table_demographics$attribute)
table_demographics_poverty <- table_demographics %>%
filter(attribute == "Food Insecurity" | attribute == "Households Below Poverty Level" | attribute == "Unemployment" | attribute == "Uninsured")
unique(table_demographics_poverty$attribute)
table_demographics_poverty$attribute <- as.factor(table_demographics_poverty$attribute)
table_demographics_poverty %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value, fill = attribute)) +
geom_bar(stat = "identity", position = "dodge", color = "black", alpha = .666) +
labs(
title = "Indicators of Poverty in Catchment Geographies",
subtitle = "",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates;
U.S. Bureau of Labor Statistics May 2020;
Map the Meal Gap 2020"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("demographics_poverty.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
#### Health Behavior Risk Factors ####
# physical activity ----
table_risk <- df %>%
filter(category == "Health Behavior Risk Factor",
area != "Catchment")
unique(table_risk$area)
table_risk$area <- ordered(table_risk$area, levels = c("USA", "AZ", "Cochise", "Pima", "Pinal", "Santa Cruz", "Yuma"))
unique(table_risk$attribute)
table_risk_pa <- table_risk %>%
filter(attribute == "No Leisure-Time Physical Activity" | attribute == "Adult Obesity" | attribute == "Diabetes")
unique(table_risk_pa$attribute)
table_risk_pa$attribute <- as.factor(table_risk_pa$attribute)
table_risk_pa$attribute <- ordered(table_risk_pa$attribute,
levels = c("No Leisure-Time Physical Activity", "Adult Obesity", "Diabetes"))
table_risk_pa %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value, fill = attribute)) +
geom_bar(stat = "identity", position = "dodge", color = "black", alpha = .666) +
geom_label(aes(label = value)) +
labs(
title = "Health Behavior and Risk Factors in Catchment Geographies",
subtitle = "Rates of Physical Activity, Obesity, and Diabetes",
y = "",
x = "",
caption = "Source: 2018 BRFSS Survey Data;
US Diabetes Surveillance System"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("risk_pa.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# alcohol and smoking ----
table_risk <- df %>%
filter(category == "Health Behavior Risk Factor",
area != "Catchment")
unique(table_risk$area)
table_risk$area <- ordered(table_risk$area, levels = c("USA", "AZ", "Cochise", "Pima", "Pinal", "Santa Cruz", "Yuma"))
unique(table_risk$attribute)
table_risk_behavior <- table_risk %>%
filter(attribute == "Excessive drinking (BRFSS)" | attribute == "Adult Smoking")
unique(table_risk_behavior$attribute)
table_risk_behavior$attribute <- as.factor(table_risk_behavior$attribute)
table_risk_behavior %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value, fill = attribute)) +
geom_bar(stat = "identity", position = "dodge", color = "black", alpha = .666) +
labs(
title = "Health Behavior and Risk Factors in Catchment Geographies",
subtitle = "Rates of Alcohol and Smoking Use",
y = "",
x = "",
caption = "Source: 2017 & 2018 Behavioral Risk Factor Surveillance System"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("risk_smoking.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# vaccination ----
table_risk <- df %>%
filter(category == "Health Behavior Risk Factor",
area != "Catchment")
unique(table_risk$area)
table_risk$area <- ordered(table_risk$area, levels = c("USA", "AZ", "Cochise", "Pima", "Pinal", "Santa Cruz", "Yuma"))
unique(table_risk$attribute)
table_risk_vac <- table_risk %>%
filter(attribute == "HPV vaccination (age range 13-17) 3+ doses")
unique(table_risk_vac$attribute)
table_risk_vac %>%
filter(race == "All") %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value, fill = attribute)) +
geom_bar(stat = "identity", position = "dodge", color = "black", alpha = .666) +
labs(
title = "Health Behavior and Risk Factors in Catchment Geographies",
subtitle = "HPV Vaccination",
y = "",
x = "",
caption = "Source: 2018 National Immunization Survey;
2017 Immunizations for Adolescents Completion Rates, AHCCCS
"
) +
theme_clean() +
theme(legend.position = "") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("risk_vaccination.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
#### Screening ####
table_screening <- df %>%
filter(category == "Screening",
area != "Catchment",
attribute != "Cervical Cancer Screening") %>%
group_by(area)
unique(table_screening$area)
table_screening$area <- ordered(table_screening$area, levels = c("USA", "AZ", "Cochise", "Pima", "Pinal", "Santa Cruz", "Yuma"))
unique(table_screening$attribute)
table_screening$attribute <- as.factor(table_screening$attribute)
table_screening %>%
arrange(desc(attribute)) %>%
ggplot(mapping = aes(x = area, y = value, fill = attribute)) +
geom_bar(stat = "identity", position = "dodge", color = "black", alpha = 0.666) +
labs(
title = "Cancer Screening Rates in Catchment Geographies",
subtitle = "All races and sexes combined",
y = "",
x = "",
caption = "Source: Directly Estimated 2018 BRFSS Data;
2008-2010 County Level Modeled Estimate Combining BRFSS & NHIS"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_fill_brewer(palette = "Accent")
# save plot to file
ggsave("risk_screening.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# demographic disparities ----
unique(df$attribute)
# all 8 geographic areas
df %>%
filter(attribute == "College Graduate" |
attribute == "Rural" |
attribute == "Hispanic" |
attribute == "Households Below Poverty Level" |
attribute == "Food Insecurity" |
attribute == "Unemployment",
race == "All") %>%
ggplot(mapping = aes(x = attribute, y = value, fill = area)) +
geom_bar(color = "black", stat = "identity", position = "dodge", alpha = .5) +
labs(
title = "",
subtitle = "",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates;
Map the Meal Gap 2020"
) +
theme_clean() +
theme(legend.position = "bottom") +
scale_y_continuous(labels = scales::percent_format()) +
scale_fill_manual(values = mixed_8)
# save plot to file
ggsave("demographic_disparities_01_complete.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
# only USA, AZ, Catchment
df %>%
filter(attribute == "College Graduate" |
attribute == "Rural" |
attribute == "Hispanic" |
attribute == "Households Below Poverty Level" |
attribute == "Food Insecurity" |
attribute == "Unemployment",
race == "All",
area == "USA" |
area == "AZ" |
area == "Catchment") %>%
ggplot(mapping = aes(x = attribute, y = value, fill = area)) +
geom_bar(color = "black", stat = "identity", position = "dodge", alpha = .5) +
labs(
title = "",
subtitle = "",
y = "",
x = "",
caption = "Source: U.S. Census Bureau, 2014-2018 American Community Survey 5-Year Estimates;
Map the Meal Gap 2020;
May 2020 U.S. Bureau of Labor Statistics") +
theme_clean() +
theme(legend.position = "bottom") +
scale_y_continuous(labels = scales::percent_format()) +
scale_fill_manual(values = blues_3)
# save plot to file
ggsave("demographic_disparities_01_us_az_catchment.svg",
width = 20,
height = 11.25,
device = svg,
path = "figures/graphics/",
scale = .5
)
|
7af4f28ec4f0c862ced4c64ee8524db6ee098332 | eca503411624bec763cace42856c2c9fdf7b26d5 | /tests/testthat/test-02_flow_data_if.R | 82d4e64794ca0d2c45863dd636d432593c995c1b | [] | no_license | yuewangpanda/flow | d1eaf3da9ab32bbde89deff85ccff39f37a26e9f | e6a812392ea9d4ea91e70c5aa6fb799e14ba32fe | refs/heads/master | 2022-12-24T18:07:49.606674 | 2020-09-25T16:49:25 | 2020-09-25T16:49:25 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 5,888 | r | test-02_flow_data_if.R |
#### IF ####
# simple if call without else and empty body
test_that("flow_data works with simple if and empty body",{
fun <- function(x) {
if(x) {}
}
data <- flow_data(fun)
# flow_data(fun)
# dput2(data$nodes[1:4])
# dput2(data$edges)
expect_equal(
data$nodes[1:4],
data.frame(
id = c(0, 1, 2, -1, 3),
block_type = c("header", "if", "standard", "end", "return"),
code_str = c("fun(x)", "if (x)", "", "", ""),
label = c("", "", "", "", ""),
stringsAsFactors = FALSE))
expect_equal(
data$edges,
data.frame(
from = c(0, 1, 2, 1, -1),
to = c(1, 2, -1, -1, 3),
edge_label = c("", "y", "", "n", ""),
arrow = c("->", "->", "->", "->", "->"),
stringsAsFactors = FALSE))
})
# simple if call without else and a symbol in body
test_that("flow_data works with simple if",{
fun <- function(x) {
if(x) foo
}
data <- flow_data(fun)
# flow_data(fun)
# dput2(data$nodes[1:4])
# dput2(data$edges)
expect_equal(
data$nodes[1:4],
data.frame(
id = c(0, 1, 2, -1, 3),
block_type = c("header", "if", "standard", "end", "return"),
code_str = c("fun(x)", "if (x)", "foo", "", ""),
label = c("", "", "", "", ""),
stringsAsFactors = FALSE))
expect_equal(
data$edges,
data.frame(
from = c(0, 1, 2, 1, -1),
to = c(1, 2, -1, -1, 3),
edge_label = c("", "y", "", "n", ""),
arrow = c("->", "->", "->", "->", "->"),
stringsAsFactors = FALSE))
})
# simple if else call
test_that("flow_data works with simple if else",{
fun <- function(x) {
if(x) foo else bar
}
data <- flow_data(fun)
# flow_data(fun)
# dput2(data$nodes[1:4])
# dput2(data$edges)
expect_equal(
data$nodes[1:4],
data.frame(
id = c(0, 1, 2, 3, -1, 4),
block_type = c("header", "if", "standard", "standard", "end", "return"),
code_str = c("fun(x)", "if (x)", "foo", "bar", "", ""),
label = c("", "", "", "", "", ""),
stringsAsFactors = FALSE))
expect_equal(
data$edges,
data.frame(
from = c(0, 1, 2, 1, 3, -1),
to = c(1, 2, -1, 3, -1, 4),
edge_label = c("", "y", "", "n", "", ""),
arrow = c("->", "->", "->", "->", "->", "->"),
stringsAsFactors = FALSE))
})
# simple if else call without else and a symbol in body
# simple if else call without else and a call in body
# simple if else call without else and 2 calls in body
# if else call returning on the left
test_that("flow_data works returning on the yes branch",{
fun <- function(x) {
if(x) return(foo) else bar
}
data <- flow_data(fun)
# flow_data(fun)
# dput2(data$nodes[1:4])
# dput2(data$edges)
expect_equal(
data$nodes[1:4],
data.frame(
id = c(0, 1, 2, -2, 3, -1, 4),
block_type = c("header", "if", "standard", "return", "standard", "end", "return"),
code_str = c("fun(x)", "if (x)", "return(foo)", "", "bar", "", ""),
label = c("", "", "", "", "", "", ""),
stringsAsFactors = FALSE))
expect_equal(
data$edges,
data.frame(
from = c(0, 1, 2, 1, 3, -1),
to = c(1, 2, -2, 3, -1, 4),
edge_label = c("", "y", "", "n", "", ""),
arrow = c("->", "->", "->", "->", "->", "->"),
stringsAsFactors = FALSE))
})
# if else call stopping on the right
test_that("flow_data works stopping on the no branch",{
fun <- function(x) {
if(x) foo else stop(bar)
}
data <- flow_data(fun)
# flow_data(fun)
# dput2(data$nodes[1:4])
# dput2(data$edges)
expect_equal(
data$nodes[1:4],
data.frame(
id = c(0, 1, 2, 3, -3, -1, 4),
block_type = c("header", "if", "standard", "standard", "stop", "end", "return"),
code_str = c("fun(x)", "if (x)", "foo", "stop(bar)", "", "", ""),
label = c("", "", "", "", "", "", ""),
stringsAsFactors = FALSE))
expect_equal(
data$edges,
data.frame(
from = c(0, 1, 2, 1, 3, -1),
to = c(1, 2, -1, 3, -3, 4),
edge_label = c("", "y", "", "n", "", ""),
arrow = c("->", "->", "->", "->", "->", "->"),
stringsAsFactors = FALSE))
})
# if else call stopping on the left AND returning on the right
test_that("flow_data works stopping on the yes branch and returning on the right branch",{
fun <- function(x) {
if(x) stop(foo) else return(bar)
}
data <- flow_data(fun)
# flow_data(fun)
# dput2(data$nodes[1:4])
# dput2(data$edges)
expect_equal(
data$nodes[1:4],
data.frame(
id = c(0, 1, 2, -2, 3, -3, 4),
block_type = c("header", "if", "standard", "stop", "standard", "return", "return"),
code_str = c("fun(x)", "if (x)", "stop(foo)", "", "return(bar)", "", ""),
label = c("", "", "", "", "", "", ""),
stringsAsFactors = FALSE))
expect_equal(
data$edges,
data.frame(
from = c(0, 1, 2, 1, 3),
to = c(1, 2, -2, 3, -3),
edge_label = c("", "y", "", "n", ""),
arrow = c("->", "->", "->", "->", "->"),
stringsAsFactors = FALSE))
})
# simple if call with a nested if else call
test_that("flow_data works with nested if calls",{
fun <- function(x) {
if(x) if(y) foo else bar
}
data <- flow_data(fun)
# flow_data(fun)
# dput2(data$nodes[1:4])
# dput2(data$edges)
expect_equal(
data$nodes[1:4],
data.frame(
id = c(0, 1, 2, 3, 4, -2, -1, 5),
block_type = c("header", "if", "if", "standard", "standard", "end", "end", "return"
),
code_str = c("fun(x)", "if (x)", "if (y)", "foo", "bar", "", "", ""),
label = c("", "", "", "", "", "", "", ""),
stringsAsFactors = FALSE))
expect_equal(
data$edges,
data.frame(
from = c(0, 1, 2, 3, 2, 4, -2, 1, -1),
to = c(1, 2, 3, -2, 4, -2, -1, -1, 5),
edge_label = c("", "y", "y", "", "n", "", "", "n", ""),
arrow = c("->", "->", "->", "->", "->", "->", "->", "->", "->"),
stringsAsFactors = FALSE))
})
|
a70c095b72aa6332e1e7daa48f9a51c4759d6f59 | 473cf48d1e85a74718b80a7e7aaeb86731f9ad98 | /R/ratio.plot.ade.R | c1633fba8adc1d0781246fcc169389d5f266429c | [] | no_license | cran/epade | f3584067b7b925b1420103559d9292dd7f702de1 | 779623aac20dd6cf53dcefa524d7a59fba1aecb6 | refs/heads/master | 2022-11-15T03:02:32.786022 | 2022-10-27T14:35:16 | 2022-10-27T14:35:16 | 17,695,818 | 0 | 1 | null | null | null | null | UTF-8 | R | false | false | 25,545 | r | ratio.plot.ade.R | ratio.plot.ade <-
function( M, vnames=NULL, sectext=NULL, main=NULL,xlab=NULL, ylab=NULL, legenlab=NULL, rlab=NULL, col=NULL, tcol=NULL, bgcol=NULL, lcol=NULL, r=NULL, v=c(0,1), lty=c(1,2), xticks=18, hlines=TRUE, legends=TRUE, logaxe=FALSE, wall=0){
if(any(par('mfg')!=c(1,1,1,1)) & any(par('mai') < c(1.02, 0.82, 0.82, 0.42))){
maidiff<-rep(0, 4)
norm<-c(1.02, 0.82, 0.82, 0.42)
maidiff[par('mai')<norm]<- norm[par('mai')<norm] - par('mai')[par('mai')<norm]
par(mai=par('mai')+maidiff)
}
oldpar<-par(no.readonly =TRUE)
oldpar<-oldpar[-which(names(oldpar)%in%c('usr', 'plt', 'pin', 'fin', 'fig', 'mfg', 'mfcol', 'mfrow', 'omd', 'omi', 'oma'))]
on.exit(par(oldpar))
if(!is.list(M)){
ML<-NULL
ML<-as.list(ML)
ML[[1]]<-M
M<-ML
}
if(is.null(vnames)) legends=FALSE
n <- dim(as.matrix(M[[1]]))[1]
if(!is.null(vnames)) vnames<-as.character(as.vector(vnames[1:n]))
if(is.null(vnames)) vnames<-''
N<-length(M)
if(length(legenlab)<N) legenlab<- c(legenlab, rep('?', N-length(legenlab)))
################################################################################
# Colors
if(length(col)<N) col<-NULL
if(is.null(tcol) & wall==0) tcol<-1
if(is.null(tcol) & wall!=0) tcol<-rgb(0.1,0.1,0.25)
if(is.null(bgcol) & wall==0) bgcol<-1
if(is.null(bgcol) & wall!=0) bgcol<-'#DBE0E8'
if(is.null(lcol) & (wall==0 | wall==2| wall==5)) lcol<-bgcol
if(is.null(lcol) & (wall==1 | wall==4)) lcol<-rgb(1,1,1)
if(is.null(lcol) & (wall==3)) lcol<-a.coladd.ade(bgcol, -50)
if(is.null(col) & N==1) col <- tcol
if(is.null(col) & N>1) col <- a.getcol.ade(N)
rcol=col
col2<-a.coladd.ade(col, 175)
fcol= a.coladd.ade(bgcol, -35)
fcol2=a.coladd.ade(bgcol, -100)
fcol3=bgcol
bgcol2<-a.coladd.ade(bgcol, -50)
#
################################################################################
if(is.list(M)){
for(k in 1:length(M)) {
M[[k]]<- apply(as.matrix(M[[k]]), c(1,2), as.numeric)
if(logaxe) M[[k]]<-log(M[[k]])
}
xmin <- min(unlist(M), na.rm=TRUE)
xmax <- max(unlist(M), na.rm=TRUE)
}
if(is.matrix(M)){
xmin <- min(M, na.rm=TRUE)
xmax <- max(M, na.rm=TRUE)
}
if(is.data.frame(M)){
M<-as.matrix(M)
xmin <- min(M, na.rm=TRUE)
xmax <- max(M, na.rm=TRUE)
}
if(logaxe) v=log(v[v!=0])
xrange <- xmax - xmin
if(!is.null(r)){
r<-(r*2.5)
print(xmin)
print(xmax)
xlimw <- c(xmin , (xmax+(xrange*r)))
}
if(is.null(r)){
vlength <-max(nchar(c(vnames, rlab, sectext)))
if(wall==4) r= vlength*0.025 + (sqrt((vlength^4))*0.0004)+0.18
if(wall!=4) r= vlength*0.025 + (sqrt((vlength^4))*0.0002)+0.18
xlimw <- c(xmin , (xmax+(xrange*r)))
}
schift <- (xrange)/6
if(N==1) legends<-F
if(legends) ylimw <- c(0.5, (n+(sqrt(n)/2.25)))
if(!legends & is.null(rlab)) ylimw <- c(0.5, (n+0.5))
if(!legends & !is.null(rlab)) ylimw <- c(0.5, (n+0.75))
par(lend='square')
tud <- (diff(ylimw)/20)/n^0.5
################################################################################
################################################################################
#Walltype 0
if(wall==0){
par(col.axis=tcol)
# Plot
plot(0, 0, type='p', pch='', bg=col, main=main, cex=1, xlim=xlimw, ylim=ylimw, ylab='', xlab='', axes = FALSE, col.main=tcol, col=rgb(1,1,1,0))
if(length(xticks)==1) ticksade<-pretty(c(xmin, xmax), n = xticks, min.n = n %/% 3)
if(length(xticks)>1 ) ticksade<-xticks
onetick<-ticksade[length(ticksade)]-ticksade[length(ticksade)-1]
lastline<-xmax+(0.15*xrange)
abline(v=lastline, col=bgcol, lwd=1)
ys<-seq(n,1)
yz<-seq(0.5, -0.5, length.out=(N+2))
if(N==1) yz<-c(0, 0)
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = bgcol, lty = 1, lwd = 1)
if(hlines & N>1) abline(h=(0:n)+0.5, col=bgcol, lwd=1)
if(legends) lagendram<-legend("topleft", legenlab, pch=c(22,22), col=col, pt.bg=col, horiz=TRUE, bg=rgb(1,1,1, 0), box.col=bgcol, text.col=tcol)
if(!legends) abline(v=v, lty=lty, col=lcol)
if(legends) segments(v, par('usr')[3] , v, par('usr')[4]-lagendram$rect$h, lty=lty, col=lcol)
for(k in 1:N) {
segments(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], col = col[k], lty = 1, lwd = 3)
points(M[[k]][ ,1], ys+yz[k+1], type='p', pch=22, cex=1.1, col=col[k], bg=col[k])
}
if(is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1), labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)+tud, labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)-tud, labels =sectext, pos=4 , col=tcol)
if(logaxe & length(xticks)==1) logis<- ticksade<-pretty(c(exp(xmin), exp(xmax)), n = xticks, min.n = n %/% 3)
if(logaxe & length(xticks)> 1) logis<- xticks
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=bgcol, lwd.ticks=1)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=bgcol)
text( lastline+diff(par('usr')[c(1, 2)])/75 , n+0.5+tud*1.5, labels = rlab, pos=4 , font=2, cex = 1.1, col=tcol)
mtext(xlab, side = 1, at = lastline, adj = 0, padj=1.5, font=1, col=tcol, cex=par('cex.lab'))
mtext(ylab, line=1.5, side = 2, font=1, col=tcol, cex=par('cex.lab'))
box(col=bgcol)
}
################################################################################
################################################################################
################################################################################
################################################################################
#Walltype 1
if(wall==1){
par(col.axis=tcol)
# Plot
plot(0, 0, type='p', pch='', bg=col, main=main, cex=1, xlim=xlimw, ylim=ylimw, ylab='', xlab='', axes = FALSE, col.main=tcol, col=rgb(1,1,1,0))
if(length(xticks)==1) ticksade<-pretty(c(xmin, xmax), n = xticks, min.n = n %/% 3)
if(length(xticks)>1 ) ticksade<-xticks
onetick<-ticksade[length(ticksade)]-ticksade[length(ticksade)-1]
lastline<-xmax+(0.15*xrange)
polygon( c(par('usr')[c(1,1)], lastline, lastline), par('usr')[c(3,4,4,3)], col=bgcol, border=FALSE)
abline(v=v, lty=lty, col=lcol)
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = rgb(1,1,1), lty = 1, lwd = 1)
if(hlines & N>1) segments(par('usr')[1], (0:n)+0.5 ,lastline, (0:n)+0.5, col = rgb(1,1,1), lty = 1, lwd = 1)
if(hlines & N>1) segments(lastline, (0:n)+0.5 ,par('usr')[2], (0:n)+0.5, col = bgcol, lty = 1, lwd = 1)
ys<-seq(n,1)
yz<-seq(0.5, -0.5, length.out=(N+2))
if(N==1) yz<-c(0, 0)
for(k in 1:N) {
segments(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], col = a.coladd.ade(col[k], -75), lty = 1, lwd = 3)
segments(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], col = col[k], lty = 1, lwd = 1)
points(M[[k]][ ,1], ys+yz[k+1], type='p', pch=22, cex=1.1, col=a.coladd.ade(col[k], -75), bg=col[k])
}
if(legends) legend("topleft", legenlab, fill=col, border=a.coladd.ade(col, -75), horiz=TRUE, bg=bgcol, box.col=rgb(1,1,1), box.lwd=2, text.col=tcol)
if(is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1), labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)+tud, labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)-tud, labels =sectext, pos=4 , col=tcol)
if(logaxe & length(xticks)==1) logis<- ticksade<-pretty(c(exp(xmin), exp(xmax)), n = xticks, min.n = n %/% 3)
if(logaxe & length(xticks)> 1) logis<- xticks
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=tcol, lwd.ticks=1)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=tcol)
text( lastline+diff(par('usr')[c(1, 2)])/75 , n+0.5+tud*1.5, labels = rlab, pos=4 , font=2, cex = 1.1, col=tcol)
mtext(xlab, side = 1, at = lastline, adj = 0, padj=1.5, font=1, col=tcol, cex=par('cex.lab'))
mtext(ylab, line=1.5, side = 2, font=1, col=tcol, cex=par('cex.lab'))
box(col=rgb(1,1,1), lwd=1)
}
################################################################################
################################################################################
################################################################################
################################################################################
#Walltype 2
if(wall==2){
par(col.axis=tcol)
# Plot
plot(0, 0, type='p', pch='', bg=col, main=main, cex=1, xlim=xlimw, ylim=ylimw, ylab='', xlab='', axes = FALSE, col.main=tcol, col=rgb(1,1,1,0))
if(length(xticks)==1) ticksade<-pretty(c(xmin, xmax), n = xticks, min.n = n %/% 3)
if(length(xticks)>1 ) ticksade<-xticks
onetick<-ticksade[length(ticksade)]-ticksade[length(ticksade)-1]
lastline<-xmax+(0.15*xrange)
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = bgcol, lty = 1, lwd = 1)
if(hlines & N>1) segments(par('usr')[1], (0:n)+0.5 ,lastline, (0:n)+0.5, col = bgcol, lty = 1, lwd = 1)
if(hlines & N>1) segments(lastline, (0:n)+0.5 ,par('usr')[2], (0:n)+0.5, col = bgcol, lty = 1, lwd = 1)
abline(v=v, lty=lty, col=lcol)
ys<-seq(n,1)
yz<-seq(0.5, -0.5, length.out=(N+2))
if(N==1) yz<-c(0, 0)
for(k in 1:N) {
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = a.coladd.ade(col[k], -75), lty = 1, lwd = 3)
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = col[k], lty = 1, lwd = 1)
points(M[[k]][ ,1], ys+yz[k+1], type='p', pch=22, cex=1.1, col=a.coladd.ade(col[k], -75), bg=col[k])
}
abline(v=lastline, col=a.coladd.ade(bgcol, -75), lwd=1)
if(legends) legend("topleft", legenlab, fill=col, border=a.coladd.ade(col, -75), horiz=TRUE, bg=rgb(1,1,1), box.col=a.coladd.ade(bgcol, -75), box.lwd=1, text.col=tcol, text.width=max(strwidth(legenlab,font = 2)))
if(is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1), labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)+tud, labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)-tud, labels =sectext, pos=4 , col=tcol)
if(logaxe & length(xticks)==1) logis<- ticksade<-pretty(c(exp(xmin), exp(xmax)), n = xticks, min.n = n %/% 3)
if(logaxe & length(xticks)> 1) logis<- xticks
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=a.coladd.ade(bgcol, -75), lwd.ticks=1)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=a.coladd.ade(bgcol, -75))
text( lastline+diff(par('usr')[c(1, 2)])/75 , n+0.5+tud*1.5, labels = rlab, pos=4 , font=2, cex = 1.1, col=tcol)
mtext(xlab, side = 1, at = lastline, adj = 0, padj=1.5, font=1, col=tcol, cex=par('cex.lab'))
mtext(ylab, line=1.5, side = 2, font=1, col=tcol, cex=par('cex.lab'))
box(col=a.coladd.ade(bgcol, -75))
}
################################################################################
################################################################################
################################################################################
################################################################################
#Walltype 3
if(wall==3){
par(col.axis=tcol)
# Plot
plot(0, 0, type='p', pch='', bg=col, main=main, cex=1, xlim=xlimw, ylim=ylimw, ylab='', xlab='', axes = FALSE, col.main=tcol, col=rgb(1,1,1,0))
if(length(xticks)==1) ticksade<-pretty(c(xmin, xmax), n = xticks, min.n = n %/% 3)
if(length(xticks)>1 ) ticksade<-xticks
onetick<-ticksade[length(ticksade)]-ticksade[length(ticksade)-1]
lastline<-xmax+(0.15*xrange)
polygon( c(par('usr')[c(1,1)], lastline, lastline), par('usr')[c(3,4,4,3)], col=bgcol, border=FALSE)
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = rgb(1,1,1), lty = 1, lwd = 1)
if(hlines & N>1) segments(par('usr')[1], (0:n)+0.5 ,lastline, (0:n)+0.5, col = a.coladd.ade(bgcol, -50), lty = 1, lwd = 1)
if(hlines & N>1) segments(lastline, (0:n)+0.5 ,par('usr')[2], (0:n)+0.5, col = a.coladd.ade(bgcol, -50), lty = 1, lwd = 1)
abline(v=v, lty=lty, col=lcol)
ys<-seq(n,1)
yz<-seq(0.5, -0.5, length.out=(N+2))
if(N==1) yz<-c(0, 0)
for(k in 1:N) {
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = a.coladd.ade(col[k], -75), lty = 1, lwd = 3)
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = col[k], lty = 1, lwd = 1)
points(M[[k]][ ,1], ys+yz[k+1], type='p', pch=22, cex=1.1, col=a.coladd.ade(col[k], -75), bg=col[k])
}
abline(v=lastline, col=a.coladd.ade(bgcol, -75), lwd=1)
if(legends) legend("topleft", legenlab, fill=col, border=a.coladd.ade(col, -75), horiz=TRUE, bg=rgb(1,1,1), box.col=a.coladd.ade(bgcol, -75), box.lwd=1, text.col=tcol)
if(is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1), labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)+tud, labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)-tud, labels =sectext, pos=4 , col=tcol)
if(logaxe & length(xticks)==1) logis<- ticksade<-pretty(c(exp(xmin), exp(xmax)), n = xticks, min.n = n %/% 3)
if(logaxe & length(xticks)> 1) logis<- xticks
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=a.coladd.ade(bgcol, -75), lwd.ticks=1)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=a.coladd.ade(bgcol, -75))
text( lastline+diff(par('usr')[c(1, 2)])/75 , n+0.5+tud*1.5, labels = rlab, pos=4 , font=2, cex = 1.1, col=tcol)
mtext(xlab, side = 1, at = lastline, adj = 0, padj=1.5, font=1, col=tcol, cex=par('cex.lab'))
mtext(ylab, line=1.5, side = 2, font=1, col=tcol, cex=par('cex.lab'))
box(col=a.coladd.ade(bgcol, -75))
}
################################################################################
################################################################################
################################################################################
################################################################################
#Walltype 4
if(wall==4){
par(col.axis=tcol)
par(font=2)
# Plot
plot(0, 0, type='p', pch='', bg=col, main=NULL, cex=1, xlim=xlimw, ylim=ylimw, ylab='', xlab='', axes = FALSE, col.main=tcol, col=rgb(1,1,1,0))
if(length(xticks)==1) ticksade<-pretty(c(xmin, xmax), n = xticks, min.n = n %/% 3)
if(length(xticks)>1 ) ticksade<-xticks
onetick<-ticksade[length(ticksade)]-ticksade[length(ticksade)-1]
lastline<-xmax+(0.15*xrange)
polygon( c(par('usr')[c(1,1)], lastline, lastline), par('usr')[c(3,4,4,3)], col=bgcol, border=FALSE)
polygon( c(lastline, lastline, par('usr')[c(2,2)]), par('usr')[c(3,4,4,3)], col=tcol, border=rgb(1,1,1))
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = rgb(1,1,1), lty = 1, lwd = 1)
if(hlines & N>1) segments(par('usr')[1], (0:n)+0.5 ,lastline, (0:n)+0.5, col = rgb(1,1,1), lty = 1, lwd = 1)
if(hlines & N>1) segments(lastline, (0:n)+0.5 ,par('usr')[2], (0:n)+0.5, col = rgb(1,1,1), lty = 1, lwd = 1)
par(xpd=TRUE)
dx<-7/par('din')[1]
dy<-7/par('din')[2]
xr<-(diff(par('usr')[1:2])/11)*dx
yr<-(diff(par('usr')[3:4])/10)*dy
polygon(a.glc(side=c(2,2,4,4), line=c(0,0,0,0)), a.glc(side=3, line=c(0, 2.75, 2.75, 0)), col=tcol, border=rgb(1,1,1))
if(ylab!='' & ylab!=' ') polygon( a.glc(side=2, line=c(2, 2, 0, 0)), a.glc(side=c(1, 3, 3, 1), line=0), col=bgcol, border=rgb(1,1,1))
text(a.glc(side=0), a.glc(side=3, line=1), labels=main, cex = 1.25, font=2, col=rgb(1,1,1), adj=c(0.5,0))
text(a.glc(side=2, line=0.75), a.glc(side=5), labels=ylab, cex = 1.1, font=2, col=tcol, adj=c(0.5,0), srt=90)
par(xpd=FALSE)
abline(v=v, lty=lty, col=lcol)
ys<-seq(n,1)
yz<-seq(0.5, -0.5, length.out=(N+2))
if(N==1) yz<-c(0, 0)
for(k in 1:N) {
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = a.coladd.ade(col[k], -75), lty = 1, lwd = 3)
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = col[k], lty = 1, lwd = 1)
points(M[[k]][ ,1], ys+yz[k+1], type='p', pch=22, cex=1.1, col=a.coladd.ade(col[k], -75), bg=col[k])
}
if(legends) legend("topleft", legenlab, fill=col, border=rgb(1,1,1), horiz=TRUE, bg=tcol, box.col=rgb(1,1,1), box.lwd=1, text.col=rgb(1,1,1), text.width=max(strwidth(legenlab,font = 2)))
if(is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1), labels =vnames, pos=4 , col=rgb(1,1,1))
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)+tud, labels =vnames, pos=4 , col=rgb(1,1,1))
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)-tud, labels =sectext, pos=4 , col=rgb(1,1,1))
if(logaxe & length(xticks)==1) logis<- ticksade<-pretty(c(exp(xmin), exp(xmax)), n = xticks, min.n = n %/% 3)
if(logaxe & length(xticks)> 1) logis<- xticks
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=tcol, lwd.ticks=1)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=tcol)
text( lastline+diff(par('usr')[c(1, 2)])/75 , n+0.5+tud*1.5, labels = rlab, pos=4 , font=2, col=rgb(1,1,1))
mtext(xlab, side = 1, at = lastline, adj = 0, padj=1.5, font=1, col=tcol)
box(col=rgb(1,1,1))
}
################################################################################
################################################################################
################################################################################
################################################################################
#Walltype 5
if(wall==5){
par(col.axis=tcol)
par(col.lab=tcol)
par(col.main=tcol)
newmai<-rep(0, 4)
oldmai<-par('mai')
if(oldmai[2]>0.80 & oldmai[2]<=0.82) newmai[2]<- 0.8-oldmai[3]
if(oldmai[3]>0.75 & oldmai[3]<=0.82) newmai[3]<- 0.75-oldmai[3]
if(oldmai[4]>0.25 & oldmai[4]<=0.42) newmai[4]<- 0.25-oldmai[4]
par(mai=(oldmai+newmai))
# Plot
plot(0, 0, type='p', pch='', bg=col, main=NULL, cex=1, xlim=xlimw, ylim=ylimw, ylab='', xlab='', axes = FALSE, col.main=tcol, col=rgb(1,1,1,0))
if(length(xticks)==1) ticksade<-pretty(c(xmin, xmax), n = xticks, min.n = n %/% 3)
if(length(xticks)>1 ) ticksade<-xticks
onetick<-ticksade[length(ticksade)]-ticksade[length(ticksade)-1]
lastline<-xmax+(0.15*xrange)
par(xpd=TRUE)
dx<-7/par('din')[1]
dy<-7/par('din')[2]
xr<-(diff(par('usr')[1:2])/10)*dx
yr<-(diff(par('usr')[3:4])/10)*dy
polygon(a.glc(side=2, line=c(3.25, 3.25, 0, 0)), a.glc(side=3, line=c(0.6, 3, 3, 0.6)), col=bgcol, border=tcol)
polygon(a.glc(side=c(2,2,4,4), line=c(0,0,0,0)), a.glc(side=3, line=c(0.6, 3, 3, 0.6)), col=rgb(1,1,1,0), border=tcol)
polygon(a.glc(side=4, line=c(0, 0 ,0.6, 0.6)), a.glc(side=3, line=c(0.6, 3, 3, 0.6)), col=bgcol, border=tcol)
polygon(a.glc(side=2, line=c(3.25, 3.25 ,2.65, 2.65)), a.glc(side=c(1,3,3,1), line=c(2.6, 0.6, 0.6, 2.6)), col=bgcol, border=tcol)
polygon(a.glc(side=4, line=c(0, 0 ,0.6, 0.6)), a.glc(side=c(1, 3, 3, 1), line=0), col=bgcol, border=tcol)
polygon(a.glc(side=2, line=c(3.25, 3.25, 0, 0)), a.glc(side=1, line=c(2.6, 4.5, 4.5, 2.6)), col=bgcol, border=tcol)
polygon(a.glc(side=c(2, 2, 4, 4), line=0), a.glc(side=1, line=c(2.6, 4.5, 4.5, 2.6)), col=rgb(1,1,1,0), border=tcol)
polygon(a.glc(side=4, line=c(0, 0, 0.6, 0.6)), a.glc(side=1, line=c(2.6, 4.5, 4.5, 2.6)), col=bgcol, border=tcol)
text(a.glc(side=0), a.glc(side=3, line=1.5), labels=main, cex = 1.25, font=2, col=tcol, adj=c(0.5,0))
text(a.glc(side=0), a.glc(side=1, line=3.75), labels=xlab, cex = 1.1, font=2, col=tcol, adj=c(0.5,0))
text(a.glc(side=2, line=1.5), a.glc(side=5), labels=ylab, cex = 1.1, font=2, col=tcol, adj=c(0.5,0), srt=90)
par(xpd=FALSE)
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = bgcol, lty = 1, lwd = 1)
if(hlines & N>1) segments(par('usr')[1], (0:n)+0.5 ,lastline, (0:n)+0.5, col = bgcol, lty = 1, lwd = 1)
if(hlines & N>1) segments(lastline, (0:n)+0.5 ,par('usr')[2], (0:n)+0.5, col = bgcol, lty = 1, lwd = 1)
abline(v=v, lty=lty, col=lcol)
ys<-seq(n,1)
yz<-seq(0.5, -0.5, length.out=(N+2))
if(N==1) yz<-c(0, 0)
for(k in 1:N) {
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = a.coladd.ade(col[k], -75), lty = 1, lwd = 3)
arrows(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], length = 0.0275, angle = 90, code = 3, col = col[k], lty = 1, lwd = 1)
points(M[[k]][ ,1], ys+yz[k+1], type='p', pch=22, cex=1.1, col=a.coladd.ade(col[k], -75), bg=col[k])
}
abline(v=lastline, col=tcol, lwd=1)
if(legends) legend("topleft", legenlab, fill=col, border=a.coladd.ade(col, -75), horiz=TRUE, bg=rgb(1,1,1), box.col=tcol, box.lwd=1, text.col=tcol)
if(is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1), labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)+tud, labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)-tud, labels =sectext, pos=4 , col=tcol)
if(logaxe & length(xticks)==1) logis<- ticksade<-pretty(c(exp(xmin), exp(xmax)), n = xticks, min.n = n %/% 3)
if(logaxe & length(xticks)> 1) logis<- xticks
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=tcol, lwd.ticks=1)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=tcol)
text( lastline+diff(par('usr')[c(1, 2)])/75 , n+0.5+tud*1.5, labels = rlab, pos=4 , font=2, cex = 1.1, col=tcol)
box(col=tcol)
}
################################################################################
################################################################################
################################################################################
################################################################################
#Walltype 6
if(wall==6){
par(col.axis=tcol)
# Plot
plot(0, 0, type='p', pch='', bg=col, main=main, cex=1, xlim=xlimw, ylim=ylimw, ylab='', xlab='', axes = FALSE, col.main=tcol, col=rgb(1,1,1,0))
if(length(xticks)==1) ticksade<-pretty(c(xmin, xmax), n = xticks, min.n = n %/% 3)
if(length(xticks)>1 ) ticksade<-xticks
onetick<-ticksade[length(ticksade)]-ticksade[length(ticksade)-1]
lastline<-xmax+(0.15*xrange)
polygon( c(par('usr')[c(1,1)], lastline, lastline), par('usr')[c(3,4,4,3)], col=bgcol, border=NA)
abline(v=v, lty=lty, col=lcol)
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = a.coladd.ade(bgcol, -35), lty = 1, lwd = 3)
if(hlines & N==1) segments(par('usr')[1], 1:n ,lastline, 1:n, col = rgb(1,1,1), lty = 1, lwd = 1)
if(hlines & N>1) segments(par('usr')[1], (0:n)+0.5 ,par('usr')[2], (0:n)+0.5, col = a.coladd.ade(bgcol, -35), lty = 1, lwd = 3)
if(hlines & N>1) segments(par('usr')[1], (0:n)+0.5 ,par('usr')[2], (0:n)+0.5, col = rgb(1,1,1), lty = 1, lwd = 1)
ys<-seq(n,1)
yz<-seq(0.5, -0.5, length.out=(N+2))
if(N==1) yz<-c(0, 0)
for(k in 1:N) {
segments(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], col = a.coladd.ade(col[k], -75), lty = 1, lwd = 3)
segments(M[[k]][ ,2], ys+yz[k+1] ,M[[k]][ ,3] , ys+yz[k+1], col = col[k], lty = 1, lwd = 1)
points(M[[k]][ ,1], ys+yz[k+1], type='p', pch=22, cex=1.1, col=a.coladd.ade(col[k], -75), bg=col[k])
}
if(legends) legend("topleft", legenlab, fill=col, border=a.coladd.ade(col, -75), horiz=TRUE, bg=bgcol, box.col=rgb(1,1,1), box.lwd=3, text.col=tcol)
if(legends) legend("topleft", legenlab, fill=col, border=a.coladd.ade(col, -75), horiz=TRUE, bg=rgb(1,1,1,0), box.col=a.coladd.ade(bgcol, -35), box.lwd=1, text.col=tcol)
if(is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1), labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)+tud, labels =vnames, pos=4 , col=tcol)
if(!is.null(sectext)) text( rep(lastline+diff(par('usr')[c(1, 2)])/75 , n) ,seq(n, 1)-tud, labels =sectext, pos=4 , col=tcol)
if(logaxe & length(xticks)==1) logis<- ticksade<-pretty(c(exp(xmin), exp(xmax)), n = xticks, min.n = n %/% 3)
if(logaxe & length(xticks)> 1) logis<- xticks
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=a.coladd.ade(bgcol, -35), lwd.ticks=3)
if(logaxe) axis(1, at=log(logis), labels=logis, col=bgcol, col.ticks=rgb(1,1,1), lwd.ticks=1)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=a.coladd.ade(bgcol, -35), lwd.ticks=3)
if(!logaxe) axis(1, at=ticksade, col=bgcol, col.ticks=rgb(1,1,1), lwd.ticks=1)
text( lastline+diff(par('usr')[c(1, 2)])/75 , n+0.5+tud*1.5, labels = rlab, pos=4 , font=2, cex = 1.1, col=tcol)
mtext(xlab, side = 1, at = lastline, adj = 0, padj=1.5, font=1, col=tcol, cex=par('cex.lab'))
mtext(ylab, line=1.5, side = 2, font=1, col=tcol, cex=par('cex.lab'))
box(col=rgb(1,1,1), lwd=3)
box(col=a.coladd.ade(bgcol, -35), lwd=1)
}
################################################################################
################################################################################
}
|
c26aeb3644fc3d7d0f476c8c3a5de1357c27cca7 | 7eef8780fd24ebc5deab2b2f68bf5209c65cd056 | /plot1.R | 15fffe909a9de12a8b6dc0ee4c60d2c741c8031a | [] | no_license | agnecede/ExData_Plotting1 | fd091f10fdc8e66cb195bfe19a4fd8dcbbc09d78 | 3781872f3904fcdbb5c1373735d84202abfa51ce | refs/heads/master | 2020-04-04T00:28:07.144654 | 2014-12-06T13:48:09 | 2014-12-06T13:48:09 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 980 | r | plot1.R | household_power_consumption_dates <- read.csv("~/Coursera/Exploratory Data Analysis/CourseProject1/household_power_consumption.txt",
sep=";",na.strings = "?",skip=66636,nrows=2880)
names(household_power_consumption_dates) <- c("Date","Time","Global_active_power","Global_reactive_power","Voltage",
"Global_intensity","Sub_metering_1","Sub_metering_2","Sub_metering_3")
household_power_consumption_dates$Date<-as.Date(household_power_consumption_dates$Date,format="%d/%m/%Y")
dt<-paste(household_power_consumption_dates$Date,household_power_consumption_dates$Time)
household_power_consumption_dates$datetime<-strptime(dt,format="%Y-%m-%d %H:%M:%S")
png("~/Coursera/Exploratory Data Analysis/CourseProject1/plot1.png",480,480)
hist(household_power_consumption_dates$Global_active_power,main="Global Active Power",
xlab="Global Active Power (kilowatts)",col="red",breaks=12)
dev.off() |
fe19acf0eca73551cca203f1aeee677c65ea649c | 7804575cd506b4c42defb796dee66fe083061ff9 | /dynamic_ex.R | 88bbea7a9aeb1179f99846325780d35a4fae3972 | [] | no_license | tmastny/reactor | 62b9bf5bacfc46316d82792bdaa817ac2bf8ca77 | ff49938baf0d5e8d403eb39ce7257f6e3efb177e | refs/heads/master | 2020-03-25T00:29:07.550128 | 2019-11-30T16:39:24 | 2019-11-30T16:39:24 | 143,188,063 | 2 | 0 | null | null | null | null | UTF-8 | R | false | false | 957 | r | dynamic_ex.R | library(shiny)
ui <- fluidPage(
#includeScript('www/enter_bind.js'),
singleton(tags$head(tags$script(src = "enter_bind.js"))),
tags$input(type='command', id='command1', class='reactnb-command',
autocomplete='off', autocorrect='off'
),
tags$br(),
# tags$input(type='command', id='command2', class='reactnb-command',
# autocomplete='off', autocorrect='off'
# ),
verbatimTextOutput("o1"),
verbatimTextOutput("o2"),
verbatimTextOutput("o3")
)
server <- function(input, output) {
output$o1 <- renderPrint({
input$command1
})
output$o2 <- renderPrint({
input$command2
})
output$o3 <- renderPrint({
print(reactiveValuesToList(input))
})
}
runApp(shinyApp(ui, server), launch.browser = TRUE)
# it looks like `input$...` works on element id. However,
# in the example app, the input is of type `shiny-bound-input`.
# I believe this happens with a call to `Shiny.inputBindings.register(...)`
|
90f138e8372fdee05511793ea45e331c03d3de44 | 21ac23387cf8bb8f0dba10571ed89b1300fa6d17 | /funcoes-em-R.R | 87dcc6dea01faa21227a5de305182c6f27851343 | [] | no_license | carlosafs/curso-universidade-mexico-unam | eb4e1b2e3fd38b5f0078f75dfae6166b361ed514 | 0674b4b8f91aa16ffe907e1c3563e4c2cafa8681 | refs/heads/master | 2020-05-22T18:28:47.476574 | 2019-06-07T14:52:37 | 2019-06-07T14:52:37 | 186,471,981 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 577 | r | funcoes-em-R.R |
#funcoes
#exemplo 1:
minha_funcao <- function(x, y, operacao = "soma"){
if (operacao == "soma"){
return(x + y)
}
if (operacao == "subtracao"){
return(x - y)
}
if (operacao == "multiplicacao"){
return(x * y)
}
if (operacao == "divisao"){
return(x / y)
}
}
minha_funcao(2, 2)
minha_funcao(2, 2, "soma")
minha_funcao(2, 2, "subtracao")
minha_funcao(2, 2, "multiplicacao")
minha_funcao(2, 2, "divisao")
#exemplo 2:
maiores <- function(x, limite){
#indices eh um vetor logico
indices <- x > limite
x[indices]
}
a <- 1:50
maiores(a, 25)
|
997f08dad4c2c296f6797b046d4d5becf6912eb0 | 107535e88a0314595086a2502d00a766518bbbe3 | /ui/ui_trans.R | b8bee1b14eaec4079836d24654bc0b0518268173 | [] | no_license | aravindhebbali/exploriment | 1d8b7cc4fe518fa2ade6b22e6fe3cdb4f287c506 | 8cf4e86086751853bd7df7c3dc67699c3ab2b765 | refs/heads/master | 2021-03-16T10:13:51.198920 | 2017-07-10T12:23:43 | 2017-07-10T12:23:43 | null | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 219 | r | ui_trans.R | tabPanel('Trans', value = 'tab_trans', icon = icon('database'),
navlistPanel(id = 'navlist_trans',
well = FALSE,
widths = c(2, 10),
source('ui/ui_transform2.R', local = TRUE)[[1]]
)
)
|
d78770038ebcc673391f979a632c3bcb820ad496 | 0312ccadd2937b536aaf655a0b35dd8a551b07b9 | /Plot6.R | e7008af599679a3d826543c750e739dd31fa108e | [] | no_license | daiane1989/ASSIGMENT | 78121840fb28c24d55ddd7f3ffaca46891866aaa | 17a56da5ecd0ccaff165b608449ae76e2f21a9f5 | refs/heads/master | 2020-03-27T21:16:21.531262 | 2018-09-03T00:08:25 | 2018-09-03T00:08:25 | 147,130,284 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 1,445 | r | Plot6.R | setwd("C:/Users/Daiane/Desktop/COURSERA/Exploratory data analysis/semana4")
## This first line will likely take a few seconds. Be patient!
NEI <- readRDS("summarySCC_PM25.rds")
SCC <- readRDS("Source_Classification_Code.rds")
library(dplyr)
#Compare emissions from motor vehicle sources in Baltimore City with emissions from motor vehicle sources in Los Angeles County, California (
#fips == "06037"\color{red}{\verb|fips == "06037"|}
#fips=="06037"). Which city has seen greater changes over time in motor vehicle emissions?
baltimore_NEI = NEI[NEI$fips=="24510" & NEI$type=="ON-ROAD", ]
la_NEI = NEI[NEI$fips=="06037" & NEI$type=="ON-ROAD", ]
emissionBaltimore = aggregate(Emissions ~ year, baltimore_NEI, sum)
emissionLa = aggregate(Emissions ~ year, la_NEI, sum)
png("plot6.png")
rng = range(emissionBaltimore$Emissions, emissionLa$Emissions)
plot(x = emissionBaltimore$year , y = emissionBaltimore$Emissions,
type = "p", pch = 16, col = "blue",
ylab = "PM2.5 Emission", xlab = "Year", ylim = rng,
main = "Motor vehicle PM2.5 Emission in LA & Baltimore from 1999 to 2008")
lines(x =emissionBaltimore$year, y = emissionBaltimore$Emissions, col = "blue")
points(x = emissionLa$year, y = emissionLa$Emissions, pch = 16, col = "red")
lines(x =emissionLa$year, y = emissionLa$Emission, col = "red")
legend("right", legend = c("LA", "Baltimore"), pch = 20, lty=1, col = c("red", "blue"), title = "City")
dev.off() |
e23ba59343c41efe57f93a2d22dcb0d64403c3de | 00bc0fb8ba893f6cf3a0dcb9f73f6816e267c96e | /CleanHousePrices.R | a22e6769badd615eda2bd2926e5a497a1bfd8ac7 | [] | no_license | oliengist/HousePrices | 3e72c78c3f56d0503e4dfd518bdb1dbf4053c2ec | 4369ca26203ccaa20da848ea1af19f3cda2e70a6 | refs/heads/master | 2021-01-19T21:21:22.753896 | 2017-02-20T04:47:54 | 2017-02-20T04:47:54 | 82,503,946 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 2,757 | r | CleanHousePrices.R |
# Author: Oliver Engist
# Corresponding dataset: https://www.kaggle.com/c/house-prices-advanced-regression-techniques
# Description: Run this script to produce a cleaned up testing and training dataset of the original
# House Price datasets.
#--------------------------------------------------------------------------------------------------
# Read in the data:
H.train <- read.csv('~/Dropbox/PORTFOLIO/Regression/train.csv',sep=',',header = T)
H.test <- read.csv('~/Dropbox/PORTFOLIO/Regression/test.csv',sep=',',header = T)
#Function to take a data vector and replace all NA with another factor ("None" in our case).
#If the vector is numeric, a message appears and the original vector is returned.
replaceNAWithNewFactor <- function(dataVector,newFactor){
if(class(dataVector)!="numeric"){
data <- as.character(dataVector)
na.idx <- is.na(data)
data[na.idx] <- newFactor
return(as.factor(data))
}else{
print("numeric variable, no NAs replaced")
return(dataVector)
}
}
# Create a vector with all the variable names that should be cleaned up:
columnsToClean <- data.frame("Fence","MiscFeature","FireplaceQu","PoolQC","^Bsmt","^Garage","^Mas")
# The ^ symbol is a wildcard, so all vectors with the pattern "Bsmt" or "Garage" are considered.
# Function to apply the grep function to the vector of column names.
applyGrepToHousing <- function(pattern){
#returns a vector of indices of the columns that have this pattern in the title.
return(grep(pattern,colnames(H.train)))
}
# Apply the new function on the entire vector of patterns we want to consider
# Creates a vector of all column indices that have NAs and need to be cleaned.
col.idx <- unlist(apply(columnsToClean,2,applyGrepToHousing))
# Take this vector of indices and use the replaceNAWithFactor function.
# Testing and training dataset separately.
H.train[,col.idx] <- mapply(replaceNAWithNewFactor,H.train[,col.idx],"none")
H.test[,col.idx] <- mapply(replaceNAWithNewFactor,H.test[,col.idx],"none")
# Replace all missing LotFrontage NAs with the square root of the LotArea
LotFront.na <- is.na(H.train$LotFrontage)
H.train$LotFrontage[LotFront.na] <- sqrt(H.train$LotArea[LotFront.na])
#Replace the single missing Electricity entry with the most frequent one:
electr.na <- is.na(H.train$Electrical)
Mode <- function(x) {
ux <- unique(x)
ux[which.max(tabulate(match(x, ux)))]
}
H.train$Electrical[electr.na] <- Mode(H.train$Electrical)
#Resulting numbers of NA per variable:
apply(apply(H.train,2,is.na),2,sum)
# Write data to a new file:
write.table(H.train, file = "~/Dropbox/PORTFOLIO/Regression/train_clean.csv", row.names = FALSE,sep=',')
write.table(H.test, file="~/Dropbox/PORTFOLIO/Regression/test_clean.csv", row.names=FALSE,sep=",")
|
3bfcdfc654f9d9c42054ebc6d6adc766a90e0249 | c9e02a75abbd1d5048446a65aa23b10f79492b2f | /scripts/cheaters.R | a97b4dc2876923ad576fc00fbe2bc7fa24960645 | [] | no_license | somasushma/R-code | f8290d3ecd8ea87ef778b1deb0b7222e84b811be | 2e1f0e05ae56ebe87354caeb374aebb19bf00080 | refs/heads/master | 2021-10-27T14:42:26.847193 | 2021-10-25T22:02:51 | 2021-10-25T22:02:51 | 137,162,116 | 0 | 1 | null | null | null | null | UTF-8 | R | false | false | 809 | r | cheaters.R | te=c(27, 20, 7, 4, 1,1)
ye=c("European", "Indian", "Chinese", "West Asian", "Korean", "Japanese" )
df=data.frame(cbind(ye,te))
colnames(df)=c("origin", "count")
df$count=as.numeric(as.character(df$count))
df$percentage= round(df$count/sum(te)*100,1)
df$pop=c(197.3, 3.18, 3.79, 10.5, 1.7, 1.3 )
totalpop=sum(df$pop[c(2,3,5,6)])
df$biomed.pop=c(.58,.34*1.8/totalpop,.34*5.2/totalpop, .025, .34*1.67/totalpop, .34*1.3/totalpop)*69000
df$cheaters.ppm=df$count/df$pop
df$cheaters.p1000=df$count/df$biomed.pop*1000
#tabulate
knitr::kable(df[,c(1,2,3,5,7)],format = "pandoc")
xtable::xtable(df[,c(1,2,3,5,7)])
#probability of Indian fraud
pbinom(q=20,size = 4236, prob = 60/69000,lower.tail = F)
#how many times more likely are Indians to commit fraud
4.7219854/((60-20)/(69000-4236)*1000)
|
c41a1f398e58e5b2c367b7ea3115818ef265cd8f | d82a996f50f6b553f645af24a6dd1600b19084cf | /lascar_data_analysis/Lascar Duplicate Identification.R | 5d59254c5213b22978d271c522b8ad721061c964 | [] | no_license | ashlinn/GRAPHS_exposure_data | 99f3035d2746b318f42113b3759a33543e83d91a | 9f5923734d00f5a63f66fbc30d318537bd235da9 | refs/heads/master | 2021-01-21T11:18:34.561236 | 2018-04-06T20:03:51 | 2018-04-06T20:03:51 | 91,735,083 | 0 | 0 | null | 2017-05-18T20:30:11 | 2017-05-18T20:30:11 | null | UTF-8 | R | false | false | 3,530 | r | Lascar Duplicate Identification.R | # Lascar file duplicate/problem identification by session
# set path
path <- "~/Dropbox/Ghana_exposure_data_SHARED (1)/Main_study_exposure_assessment"
################# Run script from here to the end
## get Lascar files
files<-list.files(path,recursive=T,pattern="^(CU_CO|CU_C0|CO_USB|COL_USB|CU-CO|CU-C0|CO-USB|COL-USB)", full.names=F)
length(files) #4894
# make a data frame of the files
Lascar_data1 <- data.frame(file = files)
# grab mother and child id info
id_pattern <- "BM....."
Lascar_data1$mstudyid <- regmatches(Lascar_data1$file, regexpr(id_pattern, Lascar_data1$file))
child_pattern <- "BM....C"
Lascar_data1$cstudyid <- regexpr(child_pattern, Lascar_data1$file)
Lascar_data1$cstudyid <- ifelse(Lascar_data1$cstudyid == -1, NA, substr(x = Lascar_data1$file, start = Lascar_data1$cstudyid, stop = Lascar_data1$cstudyid + 6))
# identify those files with ids matching neither mother nor child
Lascar_data1$problem_id <- regmatches(Lascar_data1$file, gregexpr(id_pattern, Lascar_data1$file))
Lascar_data1$problem_id <- ifelse(sapply(Lascar_data1$problem_id, unique, "[[") == Lascar_data1$mstudyid, NA, sapply(Lascar_data1$problem_id, unique, "[["))
Lascar_data1$problem_id <- ifelse(substr(Lascar_data1$problem_id, 15, 21) %in% Lascar_data1$cstudyid & nchar(Lascar_data1$problem_id) < 24, NA, Lascar_data1$problem_id)
# can't parse further
# grab session info
session_pattern <- "s_.."
Lascar_data1$session <- regmatches(Lascar_data1$file, regexpr(session_pattern, Lascar_data1$file))
Lascar_data1$session2 <- regmatches(Lascar_data1$file, gregexpr(session_pattern, Lascar_data1$file))
Lascar_data1$session2 <- ifelse(sapply(Lascar_data1$session2, unique, "[[") == Lascar_data1$session, NA, sapply(Lascar_data1$session2, unique, "[["))
# identify those paths that contain more than 1 session and there is no child
Lascar_data1$problem_session <- ifelse(nchar(as.character(Lascar_data1$session2) > 2) & is.na(Lascar_data1$cstudyid), Lascar_data1$session2, NA)
# get rid of files that are actual duplicates (where 2 monitors were deployed simultaneously)
dupfiles_Lascar <- Lascar_data1[grep("dup", Lascar_data1$file),1] #24
Lascar_data1 <- Lascar_data1[!Lascar_data1$file %in% dupfiles_Lascar,] # row numbers get added here, not sure how to avoid
nrow(Lascar_data1) # 4888
# search for duplicated files
# separate the file name from the full path
Lascar_data1$filename<-basename(as.character(Lascar_data1$file))
# id the duplicates and send to new data frame
dups <- Lascar_data1$filename[which(duplicated(Lascar_data1$filename))]
Lascar_data1$duplicated <- ifelse(Lascar_data1$filename %in% dups, TRUE, FALSE)
sum(Lascar_data1$duplicated) #78
# Lascar_duplicates <- Lascar_data1[Lascar_data1$filename %in% dups,]
# Lascar_duplicates <- Lascar_duplicates[order(Lascar_duplicates$filename), c("mstudyid", "cstudyid", "problem_id", "problem_session", "filename", "file")]
Naming_problems <- Lascar_data1[!is.na(Lascar_data1$problem_id) | !is.na(Lascar_data1$problem_session) | Lascar_data1$duplicated == TRUE, c("mstudyid", "cstudyid", "problem_id", "problem_session", "duplicated", "filename", "file")]
Naming_problems$problem_id <- as.character(Naming_problems$problem_id)
Naming_problems$problem_session <- as.character(Naming_problems$problem_session)
Naming_problems <- Naming_problems[order(Naming_problems$mstudyid),]
# save as .csv
write.csv(Naming_problems, file = paste0("Naming_problems_", format(Sys.Date(), format = "%Y%b%d"), ".csv"), row.names = FALSE)
############################
|
e837e01db2f501783c6604d0ae61ec50edf29835 | 6b451cbe2d5ce230262a6fcbdc22c33561bd3fb6 | /R/R_Functions.r | 897843be81ac5d6c266cb6dbf691fe79001d2e50 | [] | no_license | phattdoan/Code-Base | 9423d794e391439ecde7a500027dda6b3523bffa | f0a988c5cc765d25e5a9c208c02ed82a8b9464bf | refs/heads/master | 2021-03-27T18:23:21.970097 | 2017-09-25T14:45:13 | 2017-09-25T14:45:13 | 102,536,140 | 0 | 0 | null | null | null | null | UTF-8 | R | false | false | 3,294 | r | R_Functions.r | -----------------------------------------------------------------------------------------------
## read CSV
####################################
raws_score = read.csv(file = "Data/RAF2016_v2.csv", header=TRUE, sep=",")
head(raws_score)
-----------------------------------------------------------------------------------------------
## joining table
####################################
members.15.over65.transformed4 = left_join(members.15.over65.transformed3,
members.15.over65, "EMPI")
-----------------------------------------------------------------------------------------------
## Function: bin and aggregate
####################################
fn_bin_vector = function(vec, bin){
vec = sort(vec, decreasing = TRUE)
#str(vec)
bin.sum = rep(0, bin)
idx = 1
for (b in 1:bin){
cumsum = 0
#flag = FALSE
for (i in idx:length(vec)){
cumsum = cumsum + vec[i]
#print(idx)
#print(b)
if (i/length(vec) > (b/bin)){
break
}
}
idx = i+1
bin.sum[b] = cumsum
}
return(bin.sum)
}
-----------------------------------------------------------------------------------------------
## Fucntion to calculate distance between 2 geocodes
####################################
```{r}
get_geo_distance = function(lon1, lat1, lon2, lat2, units = "km") {
distance_Haversine = distm(c(lon1, lat1), c(lon2, lat2), fun = distHaversine)
if (units == "km") {
distance = distance_Haversine / 1000.0
}
else if (units == "miles") {
distance = distance_Haversine / 1609.344
}
#else if (units == "meters"){
# distance = distance_Haversine / 1000000.0
#}
else {
distance = distance_Haversine
# This will return in meter as same way as distHaversine function.
}
distance
}
#distance between one grid
grid_length = get_geo_distance(-113, 42, -112.5, 41.5, "km")
print("Grid length in km: ")
print(grid_length[1])
#QA Script to test the function
#test = get_geo_distance(members$Longitude[1], members$Latitude[1],
# centers$lon[1], centers$lat[1], 'meters')
#test[1]
```
## Calculate distance of each member to each clinic
```{r}
for (i in 1:length(centers$name.cleaned)){
name = centers$name.cleaned[i]
members.15.over65 <- cbind(members.15.over65, name)
members.15.over65$name = 0
print(centers$Name[i])
for (j in 1:length(members.15.over65$Latitude)){
temp = get_geo_distance(centers$lon[i], centers$lat[i],
members.15.over65$Longitude[j],
members.15.over65$Latitude[j],
"meters")
members.15.over65$name[j] = temp[1]
}
colnames(members.15.over65)[(names(members.15.over65)) == "name"] = paste(centers$name.cleaned[i],
"distance",sep=".")
}
colnames(members.15.over65)
# QA Script
#centers$lon[1]
#centers$lat[1]
#members.15$Latitude[1]
#members.15$Longitude[1]
#print(get_geo_distance(centers$lon[1], centers$lat[1],
# members.15$Longitude[1],
# members.15$Latitude[1],
# "meters"))
```
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.