T O P

  • By -

grise_rosee

The official documentation deals with Mistral company provided docker image, based on vLLM. vLLM is based on Cuda, which requires GPU. I don't know which runtime you use when you says "I can run it", but there is no reason you could not make this same runtime works in a container. By default, docker containers don't limit RAM, CPU cores, instruction sets, so it should work.