How can I run this with tensor parallelism?

#6
by quilicam - opened

I notice the inference.py script does not provide native functionality for running this model tensor parallel? Is there any plan to update this script to allow for such?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment