MSclassifR
is a R package dedicated for automated
classification of mass-spectra with machine learning methods.It was
developed with the aim of identifying very similar species or phenotypes
from mass spectra obtained by Matrix Assisted Laser Desorption
Ionisation - Time Of Flight Mass Spectrometry (MALDI-TOF MS). However,
the different functions of this package can also be used to classify
other categories associated to mass spectra; or from mass spectra
obtained with other mass spectrometry techniques. It includes
easy-to-use functions for pre-processing mass spectra, functions to
determine discriminant mass-over-charge values (m/z) from a library of
mass spectra corresponding to different categories, and functions to
predict the category (species, phenotypes, etc.) associated to a mass
spectrum from a list of selected mass-over-charge values.
In this vignette, we search to discriminate five species of the genus
Ecrobia by MALDI-TOF mass spectrometry with a data set of 55
mass spectra (Lasch et al.
(2016)) using MSclassifR
package. The raw data
concerning from a research paper (Wilke et
al., 2020) were downloaded from this link.
The installation of the MSclassifR
package requires the
installation of packages from Bioconductor
, so you you
might have to install the latest version of the BiocManager
package. The MSclassifR
package imports the other necessary
packages from the CRAN. In addition, it is recommended to install the
latest version of R.
MSclassifR
using the following code:## install BiocManager if not installed
if (!require("BiocManager", quietly = TRUE))
install.packages("BiocManager")
## Install the mixOmics and multtest packages from Bioconductor
BiocManager::install(c("multtest","mixOmics", "limma", "qvalue"))
## Install MSclassifR package
install.packages("MSclassifR")
MSclassifR
package:## For spectral easy spectral treatment and machine learning
require(MSclassifR)
## Le chargement a nécessité le package : MSclassifR
MSclassifR
package is installed. Also we need
the MALDIquantForeign
, MALDIquant
and
caret
packages from the CRAN. The installation is possible
with the command install.packages()
.## time measurement
start_time <- Sys.time()
## For signal processing and machine learning
require(MSclassifR)
## For additional signal processingt functions
### install MALDIquantForeign and MALDIquant if not installed:
#install.packages("MALDIquantForeign")
#install.packages("MALDIquant")
require(MALDIquantForeign)
## Le chargement a nécessité le package : MALDIquantForeign
## Le chargement a nécessité le package : MALDIquant
##
## This is MALDIquant version 1.21
## Quantitative Analysis of Mass Spectrometry Data
## See '?MALDIquant' for more information about this package.
require(MALDIquant)
## For additional machine learning functions
### install caret if not installed:
#install.packages("caret")
require(caret)
## Le chargement a nécessité le package : caret
## Le chargement a nécessité le package : ggplot2
## Warning: le package 'ggplot2' a été compilé avec la version R 4.2.1
## Le chargement a nécessité le package : lattice
## url for load the MS data
MSdataE <- "https://agodmer.github.io/MSData/Ecrobia/MassSpectra_Ecrobia.Rdata"
MSMetadataE <- "https://agodmer.github.io/MSData/Ecrobia/metaData_Ecrobia.Rdata"
## Load mass spectra
### Mass spectra
load(url(MSdataE))
### Metadata
load(url(MSMetadataE))
importBrukerFlex
from
MALDIquantForeign package
. Of note, you will have to create
your own corresponding metadata as a data frame in this case.## for data .flex format import
#MyBrukerFlexspectra <- MALDIquantForeign::importBrukerFlex("YourPathway")
## Visual quality control
MSclassifR::PlotSpectra(SpectralData=MassSpectra_Ecrobia[[1]])
We need to perform the signal processing (including intensity
transformation, smoothing, removing baseline) and peaks detection
(including peaks detection and peaks binning) steps with the two
SignalProcessing
and PeakDetection
functions.
Of note, the optimal workflow can be determined by changing the default
parameters of this two functions, see ?SignalProcessing
and
?PeakDetection
for more details.
## Signal processing
spectra <- MSclassifR::SignalProcessing(MassSpectra_Ecrobia)
## Peaks detection
peaks <- MSclassifR::PeakDetection(spectra, labels = metaData_Ecrobia$Number_strain)
You can plot the spectral processing performed and the peaks selected (black crosses).
## Inspect spectra after signal processing
MSclassifR::PlotSpectra(SpectralData=spectra[[1]],Peaks=peaks[[1]],col_spec="blue",col_peak="black")
At this step, we create an intensity matrix with m/z values in
columns and sample names in rows data with the
intensityMatrix
function in MALDIquant
package
(Gibb and Strimmer (2012)).
## Perfom an Intensity Matrix with spectra data
IntMat <- MALDIquant::intensityMatrix(peaks)
## Rows are named according to selected metadata
rownames(IntMat) <- paste(metaData_Ecrobia$Strain_name_spot, metaData_Ecrobia$Number_strain)
## Remove "NA" in the intensityMatrix
IntMat[is.na(IntMat)] <- 0
## Normalise peaks according to the maximum intensity
## Create basic function
norma<-function(x) x/(max(x))
## Apply this function
IntMat <- apply(IntMat,1,norma)
## Transpose Matrix for statistical analysis and named X
X <- t(IntMat)
Y <- factor(metaData_Ecrobia$Species)
The dataset is splitted it into training (70%) and test (30%)
datasets using caret
package Kuhn
(2008) and the createDataPartition
function for
machine learning step.
set.seed(90) # For reproductibility
## Split data to tune the model according train set
## Index creation with p = 0,7
Index_skfold <- caret::createDataPartition(Y, p = 0.7, list=FALSE)
## Train set creation
X_train <- X[Index_skfold,]
Y_train <- Y[Index_skfold]
## Test set creation
X_test <- X[-Index_skfold,]
Y_test <- Y[-Index_skfold]
## Format Y_test as a factor for analysis
Y_testb <- gsub(" ", ".",Y_test)
The selection of discriminant mass-over-charge values can be
performed with several methods proposed in the SelectionVar
function of our MSclassifR
package. The RFERF
method uses the Recursive Feature Elimination (RFE) algorithm coupled
with Random Forests (RF); the RFEGlmnet
method uses RFE
coupled with Logistic regression; the sPLSDA
method uses
sparse partial least squares discriminant analysis (sPLS-DA) (Rohart F et al. (2017)). The sPLSDA
method selects variables from the ones kept in latent components of the
model using an automatic choice of the number of components; the
VSURF
method uses RF with the VSURF
function
and package (Genuer et al.
(2015)).
For this example, we used a k-fold cross validation (k = 2) with 5 to
70 variables. The VSURF
method performs three steps
variables selection : thresholding step, interpretation step and
prediction step (Genuer et al.
(2015)).
## a. using RFE and Random forest model
set.seed(123) # for reproducibility
a <- SelectionVar(X_train,
Y_train,
MethodSelection = c("RFERF"),
MethodValidation = c("cv"),
Metric = "Accuracy",
NumberCV = 2,
PreProcessing = c("center","scale","nzv","corr"),
Sizes = c(5:70))
## b. using RFE and Random forest model
set.seed(123) # for reproducibility
b <- SelectionVar(X_train,
Y_train,
MethodSelection = c("RFEGlmnet") ,
MethodValidation = c("cv"),
Metric = "Accuracy",
NumberCV = 2,
PreProcessing = c("center","scale","nzv","corr"),
Sizes = c(5:70))
## c. using sPLDA method
set.seed(123) # for reproducibility
c <- SelectionVar(X_train,
Y_train,
MethodSelection = c("sPLSDA"),
MethodValidation = c("cv"),
NumberCV = 2,
PreProcessing = c("scale","nzv"),
Sizes = c(5:70))
## d. using VSURF
set.seed(123) # for reproducibility
d <- SelectionVar(X_train,
Y_train,
MethodSelection = c("VSURF"))
## Keep selected variables in a list
list_moz <- list("RFERF" = a$sel_moz,
"RFEglmnet" = b$sel_moz,
"sPLSDA" = c$sel_moz,
"VSURF" = d$sel_moz)
RFERF
with
the red dotted lines.PlotSpectra(SpectralData=spectra[[1]],Peaks=peaks[[1]],
Peaks2=a$sel_moz,col_spec="blue",col_peak="black")
RFEglmnet
method with the red dotted lines.PlotSpectra(SpectralData=spectra[[1]],Peaks=peaks[[1]],
Peaks2=b$sel_moz,col_spec="blue",col_peak="black")
sPLSDA
method with the red dotted lines.MSclassifR::PlotSpectra(SpectralData=spectra[[1]],Peaks=peaks[[1]],
Peaks2=c$sel_moz,col_spec="blue",col_peak="black")
VSURF
method
with the red dotted lines.MSclassifR::PlotSpectra(SpectralData=spectra[[1]],Peaks=peaks[[1]],
Peaks2=d$sel_moz,col_spec="blue",col_peak="black")
With the selected discriminant masses, we can create a prediction
model. The LogReg
function allows estimating a prediction
model with k-fold cross validation in order to predict the category to
which a mass spectrum belongs. Two main kinds of models can be
estimated: linear or nonlinear (with neural networks (nnet
method), random forests (rf
method), support vector
machines with linear kernel (svm
method), or eXtreme
Gradient Boosting (xgb
method)). Hyperparameters are
randomly searched, except for the eXtreme Gradient Boosting where a grid
search is performed.
In this part, we will create different prediction models with discriminant mass-over-charge values previously selected ones using the 4 methods described above.
Each prediction models were estimated with a repeated k-fold cross validation (k = 2, and 2 repeats)in the train data set. Next, the accuracies of estimated models are evaluated on the test data set.
set.seed(123) # for reproducibility
## Select variables found with RFERF
sel_moz=a$sel_moz
## linear multinomial regression
model_lm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2)
#Estimated model:
#model_lm
## nonlinear multinomial regression using neural networks
model_nn=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="nnet")
#Estimated model:
#model_nn
## nonlinear multinomial regression using random forests
model_rf=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="rf")
#Estimated model:
#model_rf
## nonlinear multinomial regression using xgboost
model_xgb=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="xgb")
#Estimated model:
#model_xgb
## nonlinear multinomial regression using svm
model_svm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="svm")
#Estimated model:
#model_svm
Let’s look at the performances of different prediction models on test dataset.
## Keep models in a list
list_model_A <- list("lm" = model_lm$train_mod,
"nnet" = model_nn$train_mod,
"rf" = model_rf$train_mod,
"Xgboost" = model_xgb$train_mod,
"svm" = model_svm$train_mod)
## Plot performances of prediction model
model_A <- caret::resamples(list_model_A)
bwplot(model_A)
## Probabilities of belonging to each category for the mass spectra on test set
prob_cat=MSclassifR::PredictLogReg(peaks = peaks[-Index_skfold],
model = list_model_A,
moz = sel_moz,
Reference = Y_test)
## Split table according method used
ResultatsModelA <- split(prob_cat$Prob.results, prob_cat$Prob.results$method)
## Calcul accuracy for each method
DF_A <- data.frame(lapply(ResultatsModelA, function(x)(sum(Y_testb == x[,8])/length(Y_testb))), row.names = "RFERF accuracy")
t(DF_A)
## RFERF accuracy
## comb_fisher 1.0000000
## max_vote 1.0000000
## multinom 1.0000000
## nnet 0.8571429
## rf 1.0000000
## svmLinear2 0.9285714
## xgbTree 1.0000000
set.seed(123) # for reproducibility
## Select variables found with RFGlmnet
sel_moz=b$sel_moz
## linear multinomial regression
model_lm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2)
#Estimated model:
#model_lm
## nonlinear multinomial regression using neural networks
model_nn=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="nnet")
#Estimated model:
#model_nn
## nonlinear multinomial regression using random forests
model_rf=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="rf")
#Estimated model:
#model_rf
## nonlinear multinomial regression using xgboost
model_xgb=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="xgb")
#Estimated model:
#model_xgb
## nonlinear multinomial regression using svm
model_svm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="svm")
#Estimated model:
#model_svm
## Keep models in a list
list_model_B <- list("lm" = model_lm$train_mod,
"nnet" = model_nn$train_mod,
"rf" = model_rf$train_mod,
"Xgboost" = model_xgb$train_mod,
"svm" = model_svm$train_mod)
## Plot performances of prediction model
model_B <- caret::resamples(list_model_B)
bwplot(model_B)
## Probabilities of belonging to each category for the mass spectra on test set
prob_cat=MSclassifR::PredictLogReg(peaks = peaks[-Index_skfold],
model = list_model_B,
moz = sel_moz,
Reference = Y_test)
## Split table according method used
ResultatsModelB <- split(prob_cat$Prob.results, prob_cat$Prob.results$method)
## Calcul accuracy for each method
DF_B <- data.frame(lapply(ResultatsModelB, function(x)(sum(Y_testb == x[,8])/length(Y_testb))), row.names = "RFEGlmnet accuracy")
t(DF_B)
## RFEGlmnet accuracy
## comb_fisher 1.0000000
## max_vote 1.0000000
## multinom 0.9285714
## nnet 0.9285714
## rf 0.8571429
## svmLinear2 1.0000000
## xgbTree 0.8571429
set.seed(123) # for reproducibility
## Select variables found with RFGlmnet
sel_moz=c$sel_moz
## linear multinomial regression
model_lm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2)
#Estimated model:
#model_lm
## nonlinear multinomial regression using neural networks
model_nn=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="nnet")
#Estimated model:
#model_nn
## nonlinear multinomial regression using random forests
model_rf=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="rf")
#Estimated model:
#model_rf
## nonlinear multinomial regression using xgboost
model_xgb=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="xgb")
#Estimated model:
#model_xgb
## nonlinear multinomial regression using svm
model_svm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="svm")
#Estimated model:
#model_svm
## Keep models in a list
list_model_C <- list("lm" = model_lm$train_mod,
"nnet" = model_nn$train_mod,
"rf" = model_rf$train_mod,
"Xgboost" = model_xgb$train_mod,
"svm" = model_svm$train_mod)
## Plot performances of prediction model
model_C <- caret::resamples(list_model_C)
bwplot(model_C)
## Probabilities of belonging to each category for the mass spectra on test set
prob_cat=MSclassifR::PredictLogReg(peaks = peaks[-Index_skfold],
model = list_model_C,
moz = sel_moz,
Reference = Y_test)
## Split table according method used
ResultatsModelC <- split(prob_cat$Prob.results, prob_cat$Prob.results$method)
## Calcul accuracy for each method
DF_C <- data.frame(lapply(ResultatsModelC, function(x)(sum(Y_testb == x[,8])/length(Y_testb))), row.names = "sPLSDA accuracy")
t(DF_C)
## sPLSDA accuracy
## comb_fisher 1
## max_vote 1
## multinom 1
## nnet 1
## rf 1
## svmLinear2 1
## xgbTree 1
set.seed(123) # for reproducibility
## Select variables found with RFGlmnet
sel_moz=d$sel_moz
## linear multinomial regression
model_lm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2)
#Estimated model:
#model_lm
## nonlinear multinomial regression using neural networks
model_nn=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="nnet")
#Estimated model:
#model_nn
## nonlinear multinomial regression using random forests
model_rf=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="rf")
#Estimated model:
#model_rf
## nonlinear multinomial regression using xgboost
model_xgb=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="xgb")
#Estimated model:
#model_xgb
## nonlinear multinomial regression using svm
model_svm=MSclassifR::LogReg(X=X_train, moz=sel_moz, Y=factor(Y_train), Metric = "Accuracy", number=2, repeats=2, kind="svm")
#Estimated model:
#model_svm
## Keep models in a list
list_model_D <- list("lm" = model_lm$train_mod,
"nnet" = model_nn$train_mod,
"rf" = model_rf$train_mod,
"Xgboost" = model_xgb$train_mod,
"svm" = model_svm$train_mod)
## Plot performances of prediction model
model_D <- caret::resamples(list_model_D)
bwplot(model_D)
## Probabilities of belonging to each category for the mass spectra on test set
prob_cat=MSclassifR::PredictLogReg(peaks = peaks[-Index_skfold],
model = list_model_D,
moz = sel_moz,
Reference = Y_test)
## Split table according method used
ResultatsModelD <- split(prob_cat$Prob.results, prob_cat$Prob.results$method)
## Calcul accuracy for each method
DF_D <- data.frame(lapply(ResultatsModelD, function(x)(sum(Y_testb == x[,8])/length(Y_testb))), row.names = "VSURF accuracy")
t(DF_D)
## VSURF accuracy
## comb_fisher 1.0000000
## max_vote 1.0000000
## multinom 1.0000000
## nnet 1.0000000
## rf 0.9285714
## svmLinear2 0.9285714
## xgbTree 0.9285714
the accuracies of all the used method in the test dataset can be summarized:
## Summarize results on test set
DFf <- rbind(DF_A, DF_B, DF_C, DF_D)
## Print results
t(DFf)
## RFERF accuracy RFEGlmnet accuracy sPLSDA accuracy VSURF accuracy
## comb_fisher 1.0000000 1.0000000 1 1.0000000
## max_vote 1.0000000 1.0000000 1 1.0000000
## multinom 1.0000000 0.9285714 1 1.0000000
## nnet 0.8571429 0.9285714 1 1.0000000
## rf 1.0000000 0.8571429 1 0.9285714
## svmLinear2 0.9285714 1.0000000 1 0.9285714
## xgbTree 1.0000000 0.8571429 1 0.9285714
This vignette illustrates how MSclassifR
allows
estimating many different machine learning-based models to identify five
species of the genus Ecrobia from MALDI-TOF mass spectra, and
how to use these models to predict the species of a new coming mass
spectrum. In the example we used, 19 models of the 28 models have
accuracy equal to 1. For instance, perfect accuracies were obtained on
the test data set with variables selection by RFE-RF
,
sPLSDA
and coupled with lm
and rf
methods.
## time measurement
end_time <- Sys.time()
end_time - start_time
## Time difference of 1.216329 hours
sessionInfo()
## R version 4.2.0 (2022-04-22 ucrt)
## Platform: x86_64-w64-mingw32/x64 (64-bit)
## Running under: Windows 10 x64 (build 19044)
##
## Matrix products: default
##
## locale:
## [1] LC_COLLATE=French_France.utf8 LC_CTYPE=French_France.utf8
## [3] LC_MONETARY=French_France.utf8 LC_NUMERIC=C
## [5] LC_TIME=French_France.utf8
##
## attached base packages:
## [1] stats graphics grDevices utils datasets methods base
##
## other attached packages:
## [1] caret_6.0-92 lattice_0.20-45 ggplot2_3.3.6
## [4] MALDIquantForeign_0.13 MALDIquant_1.21 MSclassifR_0.3.0
##
## loaded via a namespace (and not attached):
## [1] utf8_1.2.2 ggstance_0.3.5
## [3] tidyselect_1.1.2 grid_4.2.0
## [5] ranger_0.14.1 BiocParallel_1.30.3
## [7] maptools_1.1-4 pROC_1.18.0
## [9] munsell_0.5.0 codetools_0.2-18
## [11] mutoss_0.1-12 performanceEstimation_1.1.0
## [13] xgboost_1.6.0.1 future_1.27.0
## [15] withr_2.5.0 colorspace_2.0-3
## [17] Biobase_2.56.0 highr_0.9
## [19] knitr_1.39 rstudioapi_0.13
## [21] stats4_4.2.0 robustbase_0.95-0
## [23] listenv_0.8.0 Rdpack_2.4
## [25] labeling_0.4.2 mnormt_2.1.0
## [27] polyclip_1.10-0 farver_2.1.1
## [29] parallelly_1.32.1 vctrs_0.4.1
## [31] generics_0.1.3 TH.data_1.1-1
## [33] cp4p_0.3.6 ipred_0.9-13
## [35] xfun_0.31 geepack_1.3.4
## [37] randomForest_4.7-1.1 R6_2.5.1
## [39] doParallel_1.0.17 cachem_1.0.6
## [41] reshape_0.8.9 scales_1.2.0
## [43] multcomp_1.4-19 nnet_7.3-17
## [45] gtable_0.3.0 globals_0.15.1
## [47] sandwich_3.0-2 timeDate_4021.104
## [49] rlang_1.0.4 Rborist_0.2-3
## [51] splines_4.2.0 ModelMetrics_1.2.2.2
## [53] broom_1.0.0 mosaicCore_0.9.0
## [55] yaml_2.3.5 reshape2_1.4.4
## [57] abind_1.4-5 backports_1.4.1
## [59] qvalue_2.28.0 tools_4.2.0
## [61] lava_1.6.10 gstat_2.0-9
## [63] ellipsis_0.3.2 readBrukerFlexData_1.9.0
## [65] jquerylib_0.1.4 RColorBrewer_1.1-3
## [67] proxy_0.4-27 geeM_0.10.1
## [69] BiocGenerics_0.42.0 ggformula_0.10.1
## [71] ggridges_0.5.3 TFisher_0.2.0
## [73] Rcpp_1.0.9 plyr_1.8.7
## [75] base64enc_0.1-3 waveslim_1.8.3
## [77] purrr_0.3.4 rpart_4.1.16
## [79] MBA_0.0-9 zoo_1.8-10
## [81] haven_2.5.0 ggrepel_0.9.1
## [83] magrittr_2.0.3 data.table_1.14.2
## [85] RSpectra_0.16-1 spacetime_1.2-8
## [87] mvtnorm_1.1-3 matrixStats_0.62.0
## [89] hms_1.1.1 evaluate_0.15
## [91] XML_3.99-0.10 mclust_5.4.10
## [93] gridExtra_2.3 shape_1.4.6
## [95] compiler_4.2.0 ellipse_0.4.3
## [97] tibble_3.1.8 crayon_1.5.1
## [99] htmltools_0.5.3 corpcor_1.6.10
## [101] tidyr_1.2.0 lubridate_1.8.0
## [103] tweenr_1.0.2 MASS_7.3-56
## [105] Matrix_1.4-1 car_3.1-0
## [107] cli_3.3.0 rbibutils_2.2.8
## [109] parallel_4.2.0 metap_1.8
## [111] gower_1.0.0 qqconf_1.2.3
## [113] igraph_1.3.4 forcats_0.5.1
## [115] pkgconfig_2.0.3 sn_2.0.2
## [117] numDeriv_2016.8-1.1 foreign_0.8-82
## [119] signal_0.7-7 sp_1.5-0
## [121] readMzXmlData_2.8.1 recipes_1.0.1
## [123] foreach_1.5.2 rARPACK_0.11-0
## [125] bslib_0.4.0 hardhat_1.2.0
## [127] multtest_2.52.0 VSURF_1.1.0
## [129] prodlim_2019.11.13 stringr_1.4.0
## [131] digest_0.6.29 rmarkdown_2.14
## [133] intervals_0.15.2 MESS_0.5.9
## [135] MALDIrppa_1.1.0-1 lifecycle_1.0.1
## [137] nlme_3.1-157 mltools_0.3.5
## [139] jsonlite_1.8.0 carData_3.0-5
## [141] mixOmics_6.20.0 UBL_0.0.7
## [143] limma_3.52.2 fansi_1.0.3
## [145] labelled_2.9.1 pillar_1.8.0
## [147] fuzzyjoin_0.1.6 fastmap_1.1.0
## [149] plotrix_3.8-2 DEoptimR_1.0-11
## [151] survival_3.3-1 glue_1.6.2
## [153] xts_0.12.1 FNN_1.1.3.1
## [155] iterators_1.0.14 glmnet_4.1-4
## [157] ggforce_0.3.3 class_7.3-20
## [159] stringi_1.7.8 sass_0.4.2
## [161] automap_1.0-16 mathjaxr_1.6-0
## [163] dplyr_1.0.9 e1071_1.7-11
## [165] future.apply_1.9.0