Last updated on 2025-03-09 22:53:04 CET.
Flavor | Version | Tinstall | Tcheck | Ttotal | Status | Flags |
---|---|---|---|---|---|---|
r-devel-linux-x86_64-debian-clang | 0.9.0 | 5.33 | 97.15 | 102.48 | ERROR | |
r-devel-linux-x86_64-debian-gcc | 0.9.0 | 4.46 | 83.36 | 87.82 | ERROR | |
r-devel-linux-x86_64-fedora-clang | 0.9.0 | 213.53 | ERROR | |||
r-devel-linux-x86_64-fedora-gcc | 0.9.0 | 206.36 | ERROR | |||
r-devel-macos-arm64 | 0.9.0 | 45.00 | OK | |||
r-devel-macos-x86_64 | 0.9.0 | 119.00 | OK | |||
r-devel-windows-x86_64 | 0.9.0 | 6.00 | 111.00 | 117.00 | ERROR | |
r-patched-linux-x86_64 | 0.9.0 | 6.37 | 109.64 | 116.01 | ERROR | |
r-release-linux-x86_64 | 0.9.0 | 5.42 | 108.20 | 113.62 | ERROR | |
r-release-macos-arm64 | 0.9.0 | 40.00 | OK | |||
r-release-macos-x86_64 | 0.9.0 | 141.00 | OK | |||
r-release-windows-x86_64 | 0.9.0 | 6.00 | 110.00 | 116.00 | ERROR | |
r-oldrel-macos-arm64 | 0.9.0 | 54.00 | OK | |||
r-oldrel-macos-x86_64 | 0.9.0 | 103.00 | OK | |||
r-oldrel-windows-x86_64 | 0.9.0 | 10.00 | 50.00 | 60.00 | OK | --no-examples --no-tests --no-vignettes |
Version: 0.9.0
Check: tests
Result: ERROR
Running ‘testthat.R’ [41s/22s]
Running the tests in ‘tests/testthat.R’ failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 1 | WARN 0 | SKIP 21 | PASS 141 ]
══ Skipped tests (21) ══════════════════════════════════════════════════════════
• On CRAN (13): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-predict-dpar.R:1:1', 'test-vcov.R:1:1'
• On Linux (3): 'test-plot-facet.R:1:1', 'test-plot.R:1:1', 'test-print.R:1:1'
• {rstanarm} is not installed (5): 'test-estimate_predicted.R:3:3',
'test-estimate_predicted.R:27:3', 'test-estimate_predicted.R:67:3',
'test-estimate_predicted.R:127:3', 'test-estimate_predicted.R:163:3'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
[ FAIL 1 | WARN 0 | SKIP 21 | PASS 141 ]
Error: Test failures
Execution halted
Flavor: r-devel-linux-x86_64-debian-clang
Version: 0.9.0
Check: tests
Result: ERROR
Running ‘testthat.R’ [35s/19s]
Running the tests in ‘tests/testthat.R’ failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 3 | WARN 0 | SKIP 17 | PASS 166 ]
══ Skipped tests (17) ══════════════════════════════════════════════════════════
• .Platform$OS.type == "windows" is not TRUE (1):
'test-estimate_predicted.R:56:3'
• On CRAN (13): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-predict-dpar.R:1:1', 'test-vcov.R:1:1'
• On Linux (3): 'test-plot-facet.R:1:1', 'test-plot.R:1:1', 'test-print.R:1:1'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
── Failure ('test-estimate_predicted.R:149:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:155:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
[ FAIL 3 | WARN 0 | SKIP 17 | PASS 166 ]
Error: Test failures
Execution halted
Flavor: r-devel-linux-x86_64-debian-gcc
Version: 0.9.0
Check: tests
Result: ERROR
Running ‘testthat.R’ [92s/131s]
Running the tests in ‘tests/testthat.R’ failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 4 | WARN 22 | SKIP 17 | PASS 165 ]
══ Skipped tests (17) ══════════════════════════════════════════════════════════
• .Platform$OS.type == "windows" is not TRUE (1):
'test-estimate_predicted.R:56:3'
• On CRAN (13): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-predict-dpar.R:1:1', 'test-vcov.R:1:1'
• On Linux (3): 'test-plot-facet.R:1:1', 'test-plot.R:1:1', 'test-print.R:1:1'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
── Failure ('test-estimate_predicted.R:149:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:155:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:204:3'): estimate_expectation - predicting RE works ──
out$Predicted (`actual`) not equal to c(...) (`expected`).
`actual`: 12.2617 12.0693 11.1560 11.6318 11.1657 10.3811 11.1074 11.0749
`expected`: 12.2064 12.0631 11.2071 11.6286 11.2327 10.5839 11.2085 11.1229
[ FAIL 4 | WARN 22 | SKIP 17 | PASS 165 ]
Error: Test failures
Execution halted
Flavor: r-devel-linux-x86_64-fedora-clang
Version: 0.9.0
Check: tests
Result: ERROR
Running ‘testthat.R’ [80s/121s]
Running the tests in ‘tests/testthat.R’ failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 4 | WARN 22 | SKIP 17 | PASS 165 ]
══ Skipped tests (17) ══════════════════════════════════════════════════════════
• .Platform$OS.type == "windows" is not TRUE (1):
'test-estimate_predicted.R:56:3'
• On CRAN (13): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-predict-dpar.R:1:1', 'test-vcov.R:1:1'
• On Linux (3): 'test-plot-facet.R:1:1', 'test-plot.R:1:1', 'test-print.R:1:1'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
── Failure ('test-estimate_predicted.R:149:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:155:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:204:3'): estimate_expectation - predicting RE works ──
out$Predicted (`actual`) not equal to c(...) (`expected`).
`actual`: 12.2617 12.0693 11.1560 11.6318 11.1657 10.3811 11.1074 11.0749
`expected`: 12.2064 12.0631 11.2071 11.6286 11.2327 10.5839 11.2085 11.1229
[ FAIL 4 | WARN 22 | SKIP 17 | PASS 165 ]
Error: Test failures
Execution halted
Flavor: r-devel-linux-x86_64-fedora-gcc
Version: 0.9.0
Check: tests
Result: ERROR
Running 'testthat.R' [28s]
Running the tests in 'tests/testthat.R' failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 3 | WARN 0 | SKIP 23 | PASS 168 ]
══ Skipped tests (23) ══════════════════════════════════════════════════════════
• On CRAN (23): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-plot-facet.R:7:1', 'test-plot.R:7:1', 'test-predict-dpar.R:1:1',
'test-print.R:14:3', 'test-print.R:26:3', 'test-print.R:37:3',
'test-print.R:50:3', 'test-print.R:65:3', 'test-print.R:78:5',
'test-print.R:92:3', 'test-print.R:106:3', 'test-vcov.R:1:1'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
── Failure ('test-estimate_predicted.R:149:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:155:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
[ FAIL 3 | WARN 0 | SKIP 23 | PASS 168 ]
Error: Test failures
Execution halted
Flavor: r-devel-windows-x86_64
Version: 0.9.0
Check: tests
Result: ERROR
Running ‘testthat.R’ [49s/26s]
Running the tests in ‘tests/testthat.R’ failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 3 | WARN 0 | SKIP 17 | PASS 166 ]
══ Skipped tests (17) ══════════════════════════════════════════════════════════
• .Platform$OS.type == "windows" is not TRUE (1):
'test-estimate_predicted.R:56:3'
• On CRAN (13): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-predict-dpar.R:1:1', 'test-vcov.R:1:1'
• On Linux (3): 'test-plot-facet.R:1:1', 'test-plot.R:1:1', 'test-print.R:1:1'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
── Failure ('test-estimate_predicted.R:149:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:155:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
[ FAIL 3 | WARN 0 | SKIP 17 | PASS 166 ]
Error: Test failures
Execution halted
Flavor: r-patched-linux-x86_64
Version: 0.9.0
Check: tests
Result: ERROR
Running ‘testthat.R’ [47s/25s]
Running the tests in ‘tests/testthat.R’ failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 3 | WARN 0 | SKIP 17 | PASS 166 ]
══ Skipped tests (17) ══════════════════════════════════════════════════════════
• .Platform$OS.type == "windows" is not TRUE (1):
'test-estimate_predicted.R:56:3'
• On CRAN (13): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-predict-dpar.R:1:1', 'test-vcov.R:1:1'
• On Linux (3): 'test-plot-facet.R:1:1', 'test-plot.R:1:1', 'test-print.R:1:1'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
── Failure ('test-estimate_predicted.R:149:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:155:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
[ FAIL 3 | WARN 0 | SKIP 17 | PASS 166 ]
Error: Test failures
Execution halted
Flavor: r-release-linux-x86_64
Version: 0.9.0
Check: tests
Result: ERROR
Running 'testthat.R' [30s]
Running the tests in 'tests/testthat.R' failed.
Complete output:
> # This file is part of the standard setup for testthat.
> # It is recommended that you do not modify it.
> #
> # Where should you do additional test configuration?
> #
> # * https://r-pkgs.org/tests.html
> # * https://testthat.r-lib.org/reference/test_package.html#special-files
> library(testthat)
> library(modelbased)
>
> test_check("modelbased")
Starting 2 test processes
[ FAIL 3 | WARN 0 | SKIP 23 | PASS 168 ]
══ Skipped tests (23) ══════════════════════════════════════════════════════════
• On CRAN (23): 'test-brms-marginaleffects.R:1:1', 'test-brms.R:1:1',
'test-estimate_contrasts.R:1:1', 'test-estimate_contrasts_methods.R:1:1',
'test-estimate_means.R:1:1', 'test-estimate_means_counterfactuals.R:1:1',
'test-estimate_means_mixed.R:1:1', 'test-g_computation.R:1:1',
'test-get_marginaltrends.R:1:1', 'test-glmmTMB.R:1:1', 'test-ordinal.R:1:1',
'test-plot-facet.R:7:1', 'test-plot.R:7:1', 'test-predict-dpar.R:1:1',
'test-print.R:14:3', 'test-print.R:26:3', 'test-print.R:37:3',
'test-print.R:50:3', 'test-print.R:65:3', 'test-print.R:78:5',
'test-print.R:92:3', 'test-print.R:106:3', 'test-vcov.R:1:1'
══ Failed tests ════════════════════════════════════════════════════════════════
── Failure ('test-estimate_expectation.R:49:3'): estimate_expectation - data-grid ──
dim(estim) (`actual`) not identical to c(10L, 5L) (`expected`).
`actual`: 3 5
`expected`: 10 5
── Failure ('test-estimate_predicted.R:149:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
── Failure ('test-estimate_predicted.R:155:3'): estimate_expectation - Frequentist ──
dim(estim) (`actual`) not equal to c(10, 6) (`expected`).
`actual`: 3.0 6.0
`expected`: 10.0 6.0
[ FAIL 3 | WARN 0 | SKIP 23 | PASS 168 ]
Error: Test failures
Execution halted
Flavor: r-release-windows-x86_64
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.
Health stats visible at Monitor.