## ----setup, include=FALSE------------------------------------------------
knitr::opts_chunk$set(dpi = 300)
knitr::opts_chunk$set(cache=FALSE)

## ---- message=FALSE------------------------------------------------------
require(flowAI)

## ---- collapse = TRUE----------------------------------------------------
data(Bcells)
Bcells

## ---- eval=FALSE---------------------------------------------------------
## setwd(...)
## fcsfiles <- dir(".", pattern="*fcs$")

## ---- eval=FALSE---------------------------------------------------------
## resQC <- flow_auto_qc(Bcells)  # using a flowSet
## resQC <- flow_auto_qc(Bcells[[1]]) # using a flowFrame
## resQC <- flow_auto_qc(fcsfiles) # using a character vector

## ---- eval=FALSE---------------------------------------------------------
## GbLimit <- 2    # decide the limit in gigabyte for your batches of FCS files
## size_fcs <- file.size(fcsfiles)/1024/1024/1024    # it calculates the size in gigabytes for each FCS file
## groups <- ceiling(sum(size_fcs)/GbLimit)
## cums <- cumsum(size_fcs)
## batches <- cut(cums, groups)

## ---- eval=FALSE---------------------------------------------------------
## for(i in 1:groups){
##     flow_auto_qc(fcsfiles[which(batches == levels(batches)[i])], output = 0)
## }