Skip to content

seetacloud/codewithgpu

Repository files navigation

CodeWithGPU is a community that focuses on the reproducible AI algorithms. It has close links with Github by leveraging the managed code, and distributes corresponding docker images, models and logs for friendly reproduction.

This repository provides a novel data loading solution that maps data between Python object and serialized bytes automatically. This solution encourages developers to build a hierarchical data loading pipeline, which decouples the reading, transforming and batching. Similar solution, such as NVIDIA DALI, is widely deployed in many HPC systems and ML benchmarks.

Besides, it considers a modular and asynchronous design for the inference of AI models. Developers can easily serve their models on distributed devices by creating a many-to-many "Producer-Consumer" dataflow, and the flow control is dealt by the synchronous queues. By this way, model serving resembles training and can also get great benefit from the efficient data loader.

Also, it develops the benchmarks of modern AI models on diverse accelerators, including the newest NVIDIA GPUs and Apple Silicon processors. It will help users to match their demand on picking the best suitable devices. “The more reasonable GPUs you buy, the more money you save.”

Installation

Install from PyPI:

pip install codewithgpu

Or, clone this repository to local disk and install:

cd codewithgpu && pip instsall .

You can also install from the remote repository:

pip install git+ssh://[email protected]/seetacloud/codewithgpu.git

Quick Start

Deploy Image Inference Application

See Example: Image Inference.

Use Record Dataset To Accelerate Data Loading

See Example: Record Dataset.

Model Benchmarks

See Doc: Model Benchmarks.

License

Apache License 2.0