Skip to content

Releases: cli99/llm-analysis

Bug fixes

13 Nov 04:45
Compare
Choose a tag to compare
v0.2.2

bump version

v0.2.1

02 Nov 17:37
Compare
Choose a tag to compare
fix act checkpointing and add local model config file support

Bug fixes and MoE training analysis support

31 Oct 07:55
Compare
Choose a tag to compare

This release fixes a few bugs when calculating memory usage (e.g. activation, optimizer states), and adds support to analysis MoE training.

Bug fixes and Llama 2 inference support

18 Aug 06:30
Compare
Choose a tag to compare

This release:

  • adds group query attention (GQA) support
  • changes the activation memory calculation in inference to assume maximum tensor buffer
  • fixes the kv cache size calculation
  • adds a gpu cost analysis in the inference
  • adds llama2 inference case study

v0.1.0

02 May 17:19
Compare
Choose a tag to compare
release