You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+10-1Lines changed: 10 additions & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -21,7 +21,7 @@ Inference of Stable Diffusion and Flux in pure C/C++
21
21
- Accelerated memory-efficient CPU inference
22
22
- Only requires ~2.3GB when using txt2img with fp16 precision to generate a 512x512 image, enabling Flash Attention just requires ~1.8GB.
23
23
- AVX, AVX2 and AVX512 support for x86 architectures
24
-
- Full CUDA, Metal and SYCL backend for GPU acceleration.
24
+
- Full CUDA, Metal, Vulkan and SYCL backend for GPU acceleration.
25
25
- Can load ckpt, safetensors and diffusers models/checkpoints. Standalone VAEs models
26
26
- No need to convert to `.ggml` or `.gguf` anymore!
27
27
- Flash Attention for memory usage optimization (only cpu for now)
@@ -142,6 +142,15 @@ cmake .. -DSD_METAL=ON
142
142
cmake --build . --config Release
143
143
```
144
144
145
+
##### Using Vulkan
146
+
147
+
Install Vulkan SDK from https://www.lunarg.com/vulkan-sdk/.
148
+
149
+
```
150
+
cmake .. -DSD_VULKAN=ON
151
+
cmake --build . --config Release
152
+
```
153
+
145
154
##### Using SYCL
146
155
147
156
Using SYCL makes the computation run on the Intel GPU. Please make sure you have installed the related driver and [Intel® oneAPI Base toolkit](https://www.intel.com/content/www/us/en/developer/tools/oneapi/base-toolkit.html) before start. More details and steps can refer to [llama.cpp SYCL backend](https://github.com/ggerganov/llama.cpp/blob/master/docs/backend/SYCL.md#linux).
0 commit comments