* using log directory ‘/data/gannet/ripley/R/packages/tests-LENGTH1/mlr3proba.Rcheck’ * using R Under development (unstable) (2022-04-25 r82252) * using platform: x86_64-pc-linux-gnu (64-bit) * using session charset: UTF-8 * using option ‘--no-stop-on-test-error’ * checking for file ‘mlr3proba/DESCRIPTION’ ... OK * this is package ‘mlr3proba’ version ‘0.4.9’ * package encoding: UTF-8 * checking package namespace information ... OK * checking package dependencies ... OK * checking if this is a source package ... OK * checking if there is a namespace ... OK * checking for executable files ... OK * checking for hidden files and directories ... OK * checking for portable file names ... OK * checking for sufficient/correct file permissions ... OK * checking whether package ‘mlr3proba’ can be installed ... [70s/365s] OK * checking package directory ... OK * checking DESCRIPTION meta-information ... OK * checking top-level files ... OK * checking for left-over files ... OK * checking index information ... OK * checking package subdirectories ... OK * checking R files for non-ASCII characters ... OK * checking R files for syntax errors ... OK * checking whether the package can be loaded ... OK * checking whether the package can be loaded with stated dependencies ... OK * checking whether the package can be unloaded cleanly ... OK * checking whether the namespace can be loaded with stated dependencies ... OK * checking whether the namespace can be unloaded cleanly ... OK * checking loading without being on the library search path ... OK * checking use of S3 registration ... OK * checking dependencies in R code ... OK * checking S3 generic/method consistency ... OK * checking replacement functions ... OK * checking foreign function calls ... OK * checking R code for possible problems ... [92s/445s] OK * checking Rd files ... [3s/12s] OK * checking Rd metadata ... OK * checking Rd line widths ... OK * checking Rd cross-references ... OK * checking for missing documentation entries ... OK * checking for code/documentation mismatches ... OK * checking Rd \usage sections ... OK * checking Rd contents ... OK * checking for unstated dependencies in examples ... OK * checking contents of ‘data’ directory ... OK * checking data for non-ASCII characters ... OK * checking LazyData ... OK * checking data for ASCII and uncompressed saves ... OK * checking line endings in C/C++/Fortran sources/headers ... OK * checking pragmas in C/C++ headers and code ... OK * checking compilation flags used ... OK * checking compiled code ... OK * checking examples ... [8s/27s] OK * checking for unstated dependencies in ‘tests’ ... OK * checking tests ... Running ‘testthat.R’ [96s/303s] [96s/305s] ERROR Running the tests in ‘tests/testthat.R’ failed. Complete output: > if (requireNamespace("testthat", quietly = TRUE)) { + library("testthat") + library("checkmate") # for more expect_*() functions + library("mlr3proba") + test_check("mlr3proba") + } Loading required package: mlr3 ----------- FAILURE REPORT -------------- --- failure: length > 1 in coercion to logical --- --- srcref --- : --- package (from environment) --- distr6 --- call from context --- .merge_matpdf_cols(lapply(mds, gprm, "pdf")) --- call from argument --- nc == 1 && length(cnms) == nc --- R stacktrace --- where 1: .merge_matpdf_cols(lapply(mds, gprm, "pdf")) where 2: distr6::mixMatrix(distr) where 3: .__PipeOpSurvAvg__weighted_avg_predictions(self = self, private = private, super = super, inputs = inputs, weights = weights, row_ids = row_ids, truth = truth) where 4: private$weighted_avg_predictions(inputs, weights, row_ids, truth) where 5: .__PipeOpEnsemble__.predict(self = self, private = private, super = super, inputs = inputs) where 6: private$.predict(input) where 7: withCallingHandlers({ output = private$.predict(input) }, error = function(e) { e$message = sprintf("%s\nThis happened PipeOp %s's $predict()", e$message, self$id) stop(e) }, warning = function(w) { w$message = sprintf("%s\nThis happened PipeOp %s's $predict()", w$message, self$id) warning(w) invokeRestart("muffleWarning") }) where 8: .__PipeOp__predict(self = self, private = private, super = super, input = input) where 9: op[[fun]](input) where 10: graph_reduce(self, input, "predict", single_input) where 11: .__Graph__predict(self = self, private = private, super = super, input = input, single_input = single_input) where 12: self$graph$predict(task) where 13: .__GraphLearner__.predict(self = self, private = private, super = super, task = task) where 14: get_private(learner)$.predict(task) where 15: .f(task = , learner = ) where 16: eval(expr, p) where 17: eval(expr, p) where 18: eval.parent(expr, n = 1L) where 19: invoke(.f, .args = .args, .opts = .opts, .seed = .seed, .timeout = .timeout) where 20: encapsulate(learner$encapsulate["predict"], .f = predict_wrapper, .args = list(task = task, learner = learner), .pkgs = learner$packages, .seed = NA_integer_, .timeout = learner$timeout["predict"]) where 21: learner_predict(self, task, row_ids) where 22: .__Learner__predict(self = self, private = private, super = super, task = task, row_ids = row_ids) where 23 at test_pipelines.R#46: pipe$predict(task) where 24: eval(code, test_env) where 25: eval(code, test_env) where 26: withCallingHandlers({ eval(code, test_env) if (!handled && !is.null(test)) { skip_empty() } }, expectation = handle_expectation, skip = handle_skip, warning = handle_warning, message = handle_message, error = handle_error) where 27: doTryCatch(return(expr), name, parentenv, handler) where 28: tryCatchOne(expr, names, parentenv, handlers[[1L]]) where 29: tryCatchList(expr, names[-nh], parentenv, handlers[-nh]) where 30: doTryCatch(return(expr), name, parentenv, handler) where 31: tryCatchOne(tryCatchList(expr, names[-nh], parentenv, handlers[-nh]), names[nh], parentenv, handlers[[nh]]) where 32: tryCatchList(expr, classes, parentenv, handlers) where 33: tryCatch(withCallingHandlers({ eval(code, test_env) if (!handled && !is.null(test)) { skip_empty() } }, expectation = handle_expectation, skip = handle_skip, warning = handle_warning, message = handle_message, error = handle_error), error = handle_fatal, skip = function(e) { }) where 34: test_code(desc, code, env = parent.frame(), reporter = reporter) where 35 at test_pipelines.R#40: test_that("survbagging", { pipe = mlr3pipelines::ppl("survbagging", lrn("surv.kaplan"), iterations = 2) expect_class(pipe, "Graph") pipe = mlr3pipelines::ppl("survbagging", lrn("surv.kaplan"), iterations = 2, graph_learner = TRUE) expect_class(pipe, "GraphLearner") pipe$train(task) p = pipe$predict(task) expect_prediction_surv(p) }) where 36: eval(code, test_env) where 37: eval(code, test_env) where 38: withCallingHandlers({ eval(code, test_env) if (!handled && !is.null(test)) { skip_empty() } }, expectation = handle_expectation, skip = handle_skip, warning = handle_warning, message = handle_message, error = handle_error) where 39: doTryCatch(return(expr), name, parentenv, handler) where 40: tryCatchOne(expr, names, parentenv, handlers[[1L]]) where 41: tryCatchList(expr, names[-nh], parentenv, handlers[-nh]) where 42: doTryCatch(return(expr), name, parentenv, handler) where 43: tryCatchOne(tryCatchList(expr, names[-nh], parentenv, handlers[-nh]), names[nh], parentenv, handlers[[nh]]) where 44: tryCatchList(expr, classes, parentenv, handlers) where 45: tryCatch(withCallingHandlers({ eval(code, test_env) if (!handled && !is.null(test)) { skip_empty() } }, expectation = handle_expectation, skip = handle_skip, warning = handle_warning, message = handle_message, error = handle_error), error = handle_fatal, skip = function(e) { }) where 46: test_code(NULL, exprs, env) where 47: source_file(path, child_env(env), wrap = wrap) where 48: FUN(X[[i]], ...) where 49: lapply(test_paths, test_one_file, env = env, wrap = wrap) where 50: doTryCatch(return(expr), name, parentenv, handler) where 51: tryCatchOne(expr, names, parentenv, handlers[[1L]]) where 52: tryCatchList(expr, classes, parentenv, handlers) where 53: tryCatch(code, testthat_abort_reporter = function(cnd) { cat(conditionMessage(cnd), "\n") NULL }) where 54: with_reporter(reporters$multi, lapply(test_paths, test_one_file, env = env, wrap = wrap)) where 55: test_files(test_dir = test_dir, test_package = test_package, test_paths = test_paths, load_helpers = load_helpers, reporter = reporter, env = env, stop_on_failure = stop_on_failure, stop_on_warning = stop_on_warning, wrap = wrap, load_package = load_package) where 56: test_files(test_dir = path, test_paths = test_paths, test_package = package, reporter = reporter, load_helpers = load_helpers, env = env, stop_on_failure = stop_on_failure, stop_on_warning = stop_on_warning, wrap = wrap, load_package = load_package, parallel = parallel) where 57: test_dir("testthat", package = package, reporter = reporter, ..., load_package = "installed") where 58: test_check("mlr3proba") --- value of length: 2 type: logical --- [1] FALSE FALSE --- function from context --- function (pdfs) { nc <- unique(viapply(pdfs, ncol)) cnms <- sort(unique(as.numeric(unlist(lapply(pdfs, colnames))))) if (nc == 1 && length(cnms) == nc) { return(pdfs) } nc <- length(cnms) lapply(pdfs, function(.x) { out <- matrix(0, nrow(.x), nc, FALSE, list(NULL, cnms)) out[, match(as.numeric(colnames(.x)), cnms)] <- .x out }) } --- function search by body --- Function .merge_matpdf_cols in namespace distr6 has this body. ----------- END OF FAILURE REPORT -------------- Fatal error: length > 1 in coercion to logical * checking PDF version of manual ... OK * checking for non-standard things in the check directory ... OK * checking for detritus in the temp directory ... OK * DONE Status: 1 ERROR See ‘/data/gannet/ripley/R/packages/tests-LENGTH1/mlr3proba.Rcheck/00check.log’ for details. Command exited with non-zero status 1 Time 28:51.73, 377.02 + 24.36