Skip to content

Cerebras/inference-examples

Repository files navigation

Cerebras Inference API Demos

Welcome to the Cerebras Inference API demo repository! This repository contains various examples showcasing the power of the Cerebras Wafer-Scale Engines and CS-3 systems for AI model inference.

🚀 Introduction

The Cerebras Inference API offers developers a low-latency solution for AI model inference powered by Cerebras Wafer-Scale Engines and CS-3 systems. We invite developers to explore the new possibilities that our high-speed inferencing solution unlocks.

The Cerebras Inference API provides access to models such as OpenAI's GPT-OSS, Meta's Llama family of models, and Alibaba's Qwen models. For the full details of supported models, see the supported models documentation.

📚 Resources

📁 Projects Overview

This repository contains multiple example projects, each demonstrating different capabilities of the Cerebras Inference API. Each project is located in its own folder and contains a detailed README.

Open Val Town Template

🔗 Example Projects


🌟 Getting Started

To explore each project, simply navigate to the corresponding folder and follow the instructions in the README. Happy coding!

🛠️ Requirements

  • Python 3.7+
  • Docker (for RAG examples)
  • Streamlit (for Cerebras + Streamlit example)
  • Other dependencies as noted in each project’s README.

📝 License

This project is licensed under the MIT License - see the LICENSE file for details.

👥 Contributors

We welcome contributions! Feel free to submit a pull request or open an issue.


© 2024 Cerebras Systems

About

Inference examples

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 8