Skip to content

AICrossSim/NewComputeBench

Repository files navigation

README

NewComputeBench is a project to develop a benchmark suite for the new compute paradigm (Spiking neural networks, Optical computation, In-Memory computation, etc). The project is divided into three main components:

  • Model Training
  • Model Behavior-Level Simulation
  • Hardware-Performance Simulation

🔖 For tutorials and examples, please refer to this site.

Model Training

LLMs

We adopt Llama-3 architecture and aim to support the following features:

  • Pretraining
  • Generation (inference)
  • 🚧 TODO: Parameter-efficient fine-tuning;
  • 🚧 TODO 🐌 LowPriority: Supervised-fine-tuning
  • Evaluation

PreTraining

The LLM pretraining is built on top of torchtitan.

Generation

We recommend using the HuggingFace Transformers library for generation tasks. We provide a script to convert the torchtitan checkpoint to a HuggingFace checkpoint (See this file).

Parameter-Efficient Fine-tuning

  • 🚧 TODO: For models larger than 1.1B, we fine-tune pretrained checkpoints.
    • LoRA fine-tuning data
    • LoRA fine-tuning scripts

Model Behavior Simulation

  • Random bitflip
    • Post-training bitflip transform
    • Bitflip-aware pretraining
  • Optical compute 🚧 TODO
  • Spiking neural networks 🚧 TODO
  • In-memory compute 🚧 TODO

Hardware-Performance Simulation

🚧 TODO

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •