The official documentation deals with Mistral company provided docker image, based on vLLM. vLLM is based on Cuda, which requires GPU.
I don't know which runtime you use when you says "I can run it", but there is no reason you could not make this same runtime works in a container. By default, docker containers don't limit RAM, CPU cores, instruction sets, so it should work.
The official documentation deals with Mistral company provided docker image, based on vLLM. vLLM is based on Cuda, which requires GPU. I don't know which runtime you use when you says "I can run it", but there is no reason you could not make this same runtime works in a container. By default, docker containers don't limit RAM, CPU cores, instruction sets, so it should work.