## ----setup, include=FALSE----------------------------------------------------- knitr::opts_chunk$set(dpi = 300) knitr::opts_chunk$set(cache=FALSE) ## ----message=FALSE------------------------------------------------------------ require(flowAI) ## ----collapse = TRUE---------------------------------------------------------- data(Bcells) Bcells ## ----eval=FALSE--------------------------------------------------------------- # setwd(...) # fcsfiles <- dir(".", pattern="*fcs$") ## ----eval=FALSE--------------------------------------------------------------- # resQC <- flow_auto_qc(Bcells) # using a flowSet # resQC <- flow_auto_qc(Bcells[[1]]) # using a flowFrame # resQC <- flow_auto_qc(fcsfiles) # using a character vector ## ----eval=FALSE--------------------------------------------------------------- # GbLimit <- 2 # decide the limit in gigabyte for your batches of FCS files # size_fcs <- file.size(fcsfiles)/1024/1024/1024 # it calculates the size in gigabytes for each FCS file # groups <- ceiling(sum(size_fcs)/GbLimit) # cums <- cumsum(size_fcs) # batches <- cut(cums, groups) ## ----eval=FALSE--------------------------------------------------------------- # for(i in 1:groups){ # flow_auto_qc(fcsfiles[which(batches == levels(batches)[i])], output = 0) # }