[Dev][TL] Hardware Aware Tuning Examples with TL #201
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This pull request includes several changes to improve the scheduling and tuning capabilities in the
bitblas
module, along with some code refactoring and cleanup. The most important changes include updating theThreadPoolExecutor
usage, adding hardware-aware configuration methods, introducing a new fine-grained matrix multiplication scheduler, and making various code style improvements.Enhancements to Scheduling and Tuning:
bitblas/base/utils.py
: Changed theThreadPoolExecutor
to use a dynamic number of workers based on themax_workers
parameter.bitblas/ops/base_scheduler.py
: Added a method to get hardware-aware configurations for matrix multiplication schedulers.bitblas/ops/general_matmul/tilelang/dense/matmul_tensorcore.py
: Added methods to get hardware-aware configurations for CUDA architectures.New Scheduler Introduction:
bitblas/ops/general_matmul/tilelang/dense/matmul_simt.py
: IntroducedMatmulFineGrainSIMTScheduler
, a new fine-grained matrix multiplication scheduler.Code Refactoring and Cleanup:
bitblas/ops/operator.py
: Refactored multiple methods for better readability and maintainability, includingapply_fast_tuning
,hardware_aware_finetune
, and_build_default_module
. [1] [2] [3]bitblas/ops/general_matmul/tilelang/dense/matmul.py
tomatmul_tensorcore.py
: Renamed the file for better clarity and organization.