Skip to content

Commit 2ce3ed9

Browse files
Installation doc updates (#1529)
1 parent e4fe1cf commit 2ce3ed9

File tree

3 files changed

+7
-17
lines changed

3 files changed

+7
-17
lines changed

docs/source/faqs.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# FAQs
22

3-
Please submit your questions in [this Github Discussion thread](https://github.com/TimDettmers/bitsandbytes/discussions/1013) if you feel that they will likely affect a lot of other users and that they haven't been sufficiently covered in the documentation.
3+
Please submit your questions in [this Github Discussion thread](https://github.com/bitsandbytes-foundation/bitsandbytes/discussions/1013) if you feel that they will likely affect a lot of other users and that they haven't been sufficiently covered in the documentation.
44

55
We'll pick the most generally applicable ones and post the QAs here or integrate them into the general documentation (also feel free to submit doc PRs, please).
66

docs/source/fsdp_qlora.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ FSDP-QLoRA combines data parallelism (FSDP enables sharding model parameters, op
55
This guide provides a brief guide on how bitsandbytes supports storing quantized weights to enable FSDP-QLoRA, and how to run training with the Hugging Face libraries.
66

77
> [!TIP]
8-
> Other changes required for bitsandbytes to support FSDP-QLoRA, such as reconstructing the weights from the quantization metadata and preventing quantizing already quantized weights when they're moved from a CPU to GPU, are documented in this [Pull Request](https://github.com/TimDettmers/bitsandbytes/pull/970) and described in the [Enabling 70B Finetuning on Consumer GPUs](https://www.answer.ai/posts/2024-03-14-fsdp-qlora-deep-dive) blog post. We highly recommend reading these resources for a better understanding of FSDP-QLoRA!
8+
> Other changes required for bitsandbytes to support FSDP-QLoRA, such as reconstructing the weights from the quantization metadata and preventing quantizing already quantized weights when they're moved from a CPU to GPU, are documented in this [Pull Request](https://github.com/bitsandbytes-foundation/bitsandbytes/pull/970) and described in the [Enabling 70B Finetuning on Consumer GPUs](https://www.answer.ai/posts/2024-03-14-fsdp-qlora-deep-dive) blog post. We highly recommend reading these resources for a better understanding of FSDP-QLoRA!
99
1010
## Quantized data storage
1111

docs/source/installation.mdx

Lines changed: 5 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ Welcome to the installation guide for the `bitsandbytes` library! This document
1919

2020
## CUDA[[cuda]]
2121

22-
`bitsandbytes` is currently only supported on CUDA GPUs for CUDA versions **11.0 - 12.6**. However, there's an ongoing multi-backend effort under development, which is currently in alpha. If you're interested in providing feedback or testing, check out [the multi-backend section below](#multi-backend).
22+
`bitsandbytes` is currently only supported on CUDA GPUs for CUDA versions **11.0 - 12.8**. However, there's an ongoing multi-backend effort under development, which is currently in alpha. If you're interested in providing feedback or testing, check out [the multi-backend section below](#multi-backend).
2323

2424
### Supported CUDA Configurations[[cuda-pip]]
2525

@@ -28,10 +28,8 @@ The latest version of the distributed `bitsandbytes` package is built with the f
2828
| **OS** | **CUDA Toolkit** | **Host Compiler** |
2929
|-------------|------------------|----------------------|
3030
| **Linux** | 11.7 - 12.3 | GCC 11.4 |
31-
| | 12.4 - 12.6 | GCC 13.2 |
32-
| **Windows** | 11.7 - 12.6 | MSVC 19.42+ (VS2022) |
33-
| | 12.4+ | GCC 13.2 |
34-
| **Windows** | 11.7 - 12.6 | MSVC 19.38+ (VS2022) |
31+
| | 12.4 - 12.8 | GCC 13.2 |
32+
| **Windows** | 11.7 - 12.8 | MSVC 19.42+ (VS2022) |
3533

3634
For CUDA systems, ensure your hardware meets the following requirements:
3735

@@ -104,7 +102,6 @@ Now to install the bitsandbytes package from source, run the following commands:
104102

105103
```bash
106104
git clone https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
107-
pip install -r requirements-dev.txt
108105
cmake -DCOMPUTE_BACKEND=cuda -S .
109106
make
110107
pip install -e . # `-e` for "editable" install, when developing BNB (otherwise leave that out)
@@ -152,7 +149,7 @@ Then locally install the CUDA version you need with this script from bitsandbyte
152149
```bash
153150
wget https://raw.githubusercontent.com/bitsandbytes-foundation/bitsandbytes/main/install_cuda.sh
154151
# Syntax cuda_install CUDA_VERSION INSTALL_PREFIX EXPORT_TO_BASH
155-
# CUDA_VERSION in {110, 111, 112, 113, 114, 115, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126}
152+
# CUDA_VERSION in {110, 111, 112, 113, 114, 115, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126, 128}
156153
# EXPORT_TO_BASH in {0, 1} with 0=False and 1=True
157154

158155
# For example, the following installs CUDA 12.6 to ~/local/cuda-12.6 and exports the path to your .bashrc
@@ -228,7 +225,7 @@ pip install "transformers>=4.45.1"
228225
<hfoption id="AMD ROCm">
229226

230227
> [!WARNING]
231-
> Pre-compiled binaries are only built for ROCm versions `6.1.0`/`6.1.1`/`6.1.2`/`6.2.0` and `gfx90a`, `gfx942`, `gfx1100` GPU architectures. [Find the pip install instructions here](#multi-backend-pip).
228+
> Pre-compiled binaries are only built for ROCm versions `6.1.2`/`6.2.4`/`6.3.2` and `gfx90a`, `gfx942`, `gfx1100` GPU architectures. [Find the pip install instructions here](#multi-backend-pip).
232229
>
233230
> Other supported versions that don't come with pre-compiled binaries [can be compiled for with these instructions](#multi-backend-compile).
234231
>
@@ -320,9 +317,6 @@ bitsandbytes is fully supported from ROCm 6.1 onwards (currently in alpha releas
320317
# Clone bitsandbytes repo, ROCm backend is currently enabled on multi-backend-refactor branch
321318
git clone -b multi-backend-refactor https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
322319

323-
# Install dependencies
324-
pip install -r requirements-dev.txt
325-
326320
# Compile & install
327321
apt-get install -y build-essential cmake # install build tools dependencies, unless present
328322
cmake -DCOMPUTE_BACKEND=hip -S . # Use -DBNB_ROCM_ARCH="gfx90a;gfx942" to target specific gpu arch
@@ -345,7 +339,6 @@ The below commands are for Linux. For installing on Windows, please adapt the be
345339
```
346340
git clone --depth 1 -b multi-backend-refactor https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
347341
pip install intel_extension_for_pytorch
348-
pip install -r requirements-dev.txt
349342
cmake -DCOMPUTE_BACKEND=cpu -S .
350343
make
351344
pip install -e . # `-e` for "editable" install, when developing BNB (otherwise leave that out)
@@ -365,9 +358,6 @@ pip install -e . # `-e` for "editable" install, when developing BNB (otherwise
365358
# Clone bitsandbytes repo, Ascend NPU backend is currently enabled on multi-backend-refactor branch
366359
git clone -b multi-backend-refactor https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
367360
368-
# Install dependencies
369-
pip install -r requirements-dev.txt
370-
371361
# Compile & install
372362
apt-get install -y build-essential cmake # install build tools dependencies, unless present
373363
cmake -DCOMPUTE_BACKEND=npu -S .

0 commit comments

Comments
 (0)