Skip to content

Try model partitioning for quantized inference #56

@lantiga

Description

@lantiga

https://twitter.com/_cartick/status/1640903057994285056?s=20

Does it support model partitioning for quantized inference? I have 4x8gb cards so want to see if I can try larger models.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions