Skip to content

Fix bug causing random initialization of bias when using GPTQ quantization with models without bias #4211

Fix bug causing random initialization of bias when using GPTQ quantization with models without bias

Fix bug causing random initialization of bias when using GPTQ quantization with models without bias #4211

name: BetterTransformer / Python - Test
on:
push:
branches: [ main ]
pull_request:
branches: [ main ]
concurrency:
group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }}
cancel-in-progress: true
jobs:
build:
strategy:
fail-fast: false
matrix:
python-version: [3.8, 3.9]
os: [ubuntu-20.04, macos-latest]
runs-on: ${{ matrix.os }}
steps:
- uses: actions/checkout@v2
- name: Setup Python ${{ matrix.python-version }}
uses: actions/setup-python@v2
with:
python-version: ${{ matrix.python-version }}
- name: Install dependencies
run: |
pip install .[tests]
pip install --no-cache-dir --upgrade torch torchvision torchaudio
pip install accelerate
- name: Test on pytorch stable
working-directory: tests
run: |
pytest bettertransformer/test_*.py -s -vvvvv
- name: Install dependencies 2
run: |
pip uninstall -y torch torchvision torchaudio
pip install --no-cache-dir --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/cpu
- name: Test on pytorch nightly
working-directory: tests
run: |
pytest bettertransformer/test_*.py -s -vvvvv