Possibly:
John6666
2
Related topics
| Topic | Replies | Views | Activity | |
|---|---|---|---|---|
| How to generate with a single gpu when a model is loaded onto multiple gpus? | 0 | 917 | February 9, 2024 | |
| Getting error when running inference in multiple GPUs | 0 | 695 | October 13, 2023 | |
| [SOLVED] What's the right way to do GPU paralellism for inference (not training) on AutoModelForCausalLM? | 1 | 295 | August 26, 2024 | |
| If I use llama 70b and 7b for speculative decoding, how should I put them on my multiple gpus in the code | 0 | 71 | October 11, 2024 | |
| Using 3 GPUs for training with Trainer() of transformers | 2 | 2420 | October 18, 2023 |