Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Performance optimizations #7

@jemus42

Description

@jemus42

Main bottleneck is the $predict() and $predict_newdata() steps, so minimizing how often they are called will be helpful.

  • One approach is to create the datasets for e.g. PFI individually, rbind them together while keeping track in which chunk of data which feature is permuted, and then only calling $predict once on the combined data.
  • Analogously, this can be done across multiple iterations (iters_perm) of the permutations for PFI and SAGE.
  • Since resamplings are independent and have different trained models, parallelizing across those would be an option as well.
  • Chunking the data beforehand

Sub-issues

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementExtends package features in some way

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions