Is it possible to split the 34G model into several pieces?

#20
by kirp - opened

The 34G model is too large for my cpu(16G) to load.
So I wander whether we can set max_shard_size or something to split the model.
Then, we can the load the model to gpu piece by piece.

Sign up or log in to comment