Added the ability to choose the model size, quantization, and compute device on the fly. Beware, no guardrails yet. So if you choose "cuda" and you don't have an Nvidia GPU I can't guarantee how it'll behave.
To check which quantizations your cpu and gpu support, use my other tool [HERE](https://github.com/BBC-Esq/ctranslate2-compatibility-checker).
Create an issue if you would like see or work on additional features or find a bug.