* using log directory ‘/data/gannet/ripley/R/packages/tests-clang/performance.Rcheck’
* using R Under development (unstable) (2025-03-01 r87860)
* using platform: x86_64-pc-linux-gnu
* R was compiled by
    clang version 20.1.0-rc3
    flang version 20.1.0-rc3
* running under: Fedora Linux 40 (Workstation Edition)
* using session charset: UTF-8
* using option ‘--no-stop-on-test-error’
* checking for file ‘performance/DESCRIPTION’ ... OK
* checking extension type ... Package
* this is package ‘performance’ version ‘0.13.0’
* package encoding: UTF-8
* checking package namespace information ... OK
* checking package dependencies ... OK
* checking if this is a source package ... OK
* checking if there is a namespace ... OK
* checking for executable files ... OK
* checking for hidden files and directories ... OK
* checking for portable file names ... OK
* checking for sufficient/correct file permissions ... OK
* checking whether package ‘performance’ can be installed ... [38s/106s] OK
See 'https://www.r-project.org/nosvn/R.check/r-devel-linux-x86_64-fedora-clang/performance-00install.html' for details.
* checking installed package size ... OK
* checking package directory ... OK
* checking DESCRIPTION meta-information ... OK
* checking top-level files ... OK
* checking for left-over files ... OK
* checking index information ... OK
* checking package subdirectories ... OK
* checking code files for non-ASCII characters ... OK
* checking R files for syntax errors ... OK
* checking whether the package can be loaded ... OK
* checking whether the package can be loaded with stated dependencies ... OK
* checking whether the package can be unloaded cleanly ... OK
* checking whether the namespace can be loaded with stated dependencies ... OK
* checking whether the namespace can be unloaded cleanly ... OK
* checking loading without being on the library search path ... OK
* checking use of S3 registration ... OK
* checking dependencies in R code ... OK
* checking S3 generic/method consistency ... OK
* checking replacement functions ... OK
* checking foreign function calls ... OK
* checking R code for possible problems ... [101s/278s] OK
* checking Rd files ... OK
* checking Rd metadata ... OK
* checking Rd line widths ... OK
* checking Rd cross-references ... INFO
Undeclared package ‘ggplot2’ in Rd xrefs
* checking for missing documentation entries ... OK
* checking for code/documentation mismatches ... OK
* checking Rd \usage sections ... OK
* checking Rd contents ... OK
* checking for unstated dependencies in examples ... OK
* checking R/sysdata.rda ... OK
* checking examples ... [63s/167s] OK
* checking for unstated dependencies in ‘tests’ ... OK
* checking tests ... [139s/191s] ERROR
  Running ‘testthat.R’ [138s/190s]
Running the tests in ‘tests/testthat.R’ failed.
Complete output:
  > library(testthat)
  > library(performance)
  > 
  > test_check("performance")
  Starting 2 test processes
  [ FAIL 5 | WARN 18 | SKIP 34 | PASS 375 ]
  
  ══ Skipped tests (34) ══════════════════════════════════════════════════════════
  • On CRAN (29): 'test-bootstrapped_icc_ci.R:2:3',
    'test-bootstrapped_icc_ci.R:44:3', 'test-binned_residuals.R:137:3',
    'test-binned_residuals.R:164:3', 'test-check_dag.R:1:1',
    'test-check_collinearity.R:181:3', 'test-check_collinearity.R:218:3',
    'test-check_distribution.R:35:3', 'test-check_itemscale.R:28:3',
    'test-check_model.R:1:1', 'test-check_predictions.R:2:1',
    'test-check_residuals.R:2:3', 'test-check_singularity.R:2:3',
    'test-check_singularity.R:23:3', 'test-check_zeroinflation.R:73:3',
    'test-check_zeroinflation.R:112:3', 'test-compare_performance.R:21:3',
    'test-helpers.R:1:1', 'test-icc.R:2:1', 'test-mclogit.R:53:3',
    'test-check_outliers.R:110:3', 'test-model_performance.bayesian.R:1:1',
    'test-model_performance.merMod.R:2:3',
    'test-model_performance.merMod.R:22:3', 'test-model_performance.rma.R:33:3',
    'test-pkg-ivreg.R:7:3', 'test-r2_nakagawa.R:20:1', 'test-rmse.R:35:3',
    'test-test_likelihoodratio.R:55:1'
  • On Linux (3): 'test-nestedLogit.R:1:1', 'test-r2_bayes.R:1:1',
    'test-test_wald.R:1:1'
  • getRversion() > "4.4.0" is TRUE (1): 'test-check_outliers.R:258:3'
  • packageVersion("glmmTMB") > "1.1.10" is not TRUE (1): 'test-r2.R:82:5'
  
  ══ Failed tests ════════════════════════════════════════════════════════════════
  ── Failure ('test-check_convergence.R:27:3'): check_convergence ────────────────
  check_convergence(model) is not TRUE
  
  `actual`:   FALSE
  `expected`: TRUE 
  ── Failure ('test-check_distribution.R:18:3'): check_distribution ──────────────
  out$p_Residuals (`actual`) not equal to c(...) (`expected`).
  
       actual | expected               
   [2] 0.000  | 0.000    [2]           
   [3] 0.000  | 0.000    [3]           
   [4] 0.000  | 0.000    [4]           
   [5] 0.938  - 0.906    [5]           
   [6] 0.000  | 0.000    [6]           
   [7] 0.000  | 0.000    [7]           
   [8] 0.000  | 0.000    [8]           
   [9] 0.031  - 0.062    [9]           
  [10] 0.000  | 0.000    [10]          
  [11] 0.000  | 0.000    [11]          
   ... ...      ...      and 1 more ...
  ── Failure ('test-performance_aic.R:29:3'): performance_aic lme4 default ───────
  performance_aic(m1, estimator = "ML") (`actual`) not equal to 125.0043 (`expected`).
  
    `actual`: 128.2
  `expected`: 125.0
  ── Failure ('test-performance_aic.R:31:3'): performance_aic lme4 default ───────
  performance_aic(m2, estimator = "REML") (`actual`) not equal to 128.0054 (`expected`).
  
    `actual`: 132.9
  `expected`: 128.0
  ── Failure ('test-r2_nakagawa.R:7:3'): r2_nakagawa ─────────────────────────────
  r2_nakagawa(model) (`actual`) not equal to structure(...) (`expected`).
  
    `actual$R2_conditional`: 0.950
  `expected$R2_conditional`: 0.969
  
    `actual$R2_marginal`: 0.87
  `expected$R2_marginal`: 0.66
  
  [ FAIL 5 | WARN 18 | SKIP 34 | PASS 375 ]
  Error: Test failures
  Execution halted
* checking PDF version of manual ... [15s/37s] OK
* checking HTML version of manual ... [9s/22s] OK
* checking for non-standard things in the check directory ... OK
* checking for detritus in the temp directory ... OK
* DONE
Status: 1 ERROR