Does unity support Distributed-Model-Infering,such as Multi-GPUs?

with the popular of LLMs,NLP models is becoming more and more big,although can be quantized, its still hard to deploy one LLM model in one GPU Ram. running LLMs on multi-Host-multi-GPUs may be the usage solution right now.

I’m wondering, if we can deploy large model on multiply GPUs with tvm-unity?

based on my previous understading of TVM,if one want to run large model with tvm,he/she can:

  1. split model to multipy sub-graph,compile one-by-one,and infer thems with tvm-pipeline, which different sub-graph can be put on different GPUs based on current hardware resources
  2. maybe use the Heterogeneous Execution for relax with the VDevice. [RFC][Unity][Relax] Heterogeneous Execution for Relax - Development / unity - Apache TVM Discuss

Thanks for chiming in, this is indeed one area we should push for and some of the items you listed are on the right direction