The hardware and bandwidth for this mirror is donated by dogado GmbH, the Webhosting and Full Service-Cloud Provider. Check out our Wordpress Tutorial.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]dogado.de.
Modern scientific methods can be automated and produce disparate samples sizes. In many cases, it is desirable to retain identical or pre-defined sample sizes between replicates or groups. However, choosing which subset of originally acquired data that best matches the entirety of the data set without introducing bias is not trivial.
The fairsubset() function automates the choice of subsets. Choices which alter the definition of the “best” subset choice are built into the function. These choices fit many types of scientific data.
For subset_setting = “mean” or “median” : The fairsubset$best_subset will have the closest average and standard deviation (equally weighted) to the original data.
For subset_setting = “ks” : The fairsubset$best_subset will have the maximal P-value given by ks.test, relative to the original data.
In this vignette, we discuss how to appropriately use fairsubset using these choices.
##Normally distributed data If you expect your data to have standard Gaussian curve distribution, subset_setting = “mean” or subset_setting = “median” makes sense to use.
#Create normally distributed data
input_list <- list(a= stats::rnorm(100, mean = 3, sd = 2),
b = stats::rnorm(50, mean = 5, sd = 5),
c= stats::rnorm(75, mean = 2, sd = 0.5))
#Run fairsubset, using "mean" as the averaging choice
output <- fairSubset(input_list, subset_setting = "mean", manual_N = 10, random_subsets = 1000)
#Print the report, which describes best and worst subset features
output$report
#> a b c
#> Mean of original data 2.8536849 4.471270 2.0129401
#> Mean of best subset of data 2.8372313 4.455207 2.0112059
#> Mean of worst subset of data 0.9727307 8.327654 1.6767161
#> Standard deviation of original data 2.0008876 4.633997 0.5348014
#> Standard deviation of best subset of data 1.9807153 4.690388 0.5390717
#> Standard deviation of worst subset of data 1.3335198 2.632652 0.8419289
##Bimodal or skewed data In cases which may have non-normal distributions, such as bimodal or skewed data, subset_setting = “ks” makes sense to use. This can also be used for normally distributed data, or when the distribution is unknown. Unknown distributions are most often the case for automated methods of data acquisition, so this setting is most broadly useful.
#Create bimodal distributed data
input_list <- list(a= c( stats::rnorm(100, mean = 3, sd = 2)
,stats::rnorm(100, mean = 5, sd = 4) ),
b = c( stats::rnorm(50, mean = 5, sd = 5)
,stats::rnorm(50, mean = 2, sd = 1)),
#Create skewed data
c= stats::rnbinom(75, 10, .5))
#Run fairsubset, using "ks" to enable ks.test metrics to decide the best subset
output <- fairSubset(input_list, subset_setting = "ks", manual_N = 10, random_subsets = 1000)
#Print the report, which describes best and worst subset features
output$report
#> a b c
#> Median of original data 3.516929 2.260110 9.000000
#> Median of best subset of data 3.499177 2.284833 9.000000
#> Median of worst subset of data 7.837717 8.917156 14.500000
#> Standard deviation of original data 3.617471 4.336588 4.957550
#> Standard deviation of best subset of data 3.633917 4.306124 4.971027
#> Standard deviation of worst subset of data 4.766373 5.275105 6.795423
##Allow an automatic determination of the largest sample size Since automated data acquisition will produce an unknown sample size, it may be desireable to always choose the largest subset possible. This will be the number equivalent to the smallest sample from input columns.
#Run fairsubset, using manual_N = NULL to default to largest possible subsets consistent between all samples
output <- fairSubset(input_list, subset_setting = "ks", manual_N = NULL, random_subsets = 1000)
#Print the size of the largest possible consistent N subset
nrow(output$best_subset)
#> [1] 75
#Evidently, sample c the lowest number of observed events, so fairsubset chose N = 75 for all subsets. Note that the "best_subset" using this setting retains the original set of data for the smallest sample (ie, output$best_subset$c is the same as input_list$c, just shuffled)
Fairsubset allows these random choices to best fit the original data. The worst subset demonstrates the worst-case scenario which could happen from a simple single random pick of data.
For further examples using real-life scientific examples, please refer to the original publication: https://pubmed.ncbi.nlm.nih.gov/31583263/ FairSubset: A tool to choose representative subsets of data for use with replicates or groups of different sample sizes
Please cite fairsubset when you have used it. We appreciate your time to do so!
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.
Health stats visible at Monitor.