Skip to content

Kuzay3t/Resource-Efficient-ML-Inference-API-for-Low-Compute-Devices

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Resource-Efficient ML Inference API

A lightweight ML deployment system for compressed image classification.

Features

  • MobileNetV2 fine-tuned on CIFAR-10
  • ONNX export for deployment
  • INT8 quantization for size reduction

Project Structure

MLH_PROJECT/
├── training/          # Training and export scripts
├── models/            # ONNX models
├── app/              # API (coming soon)
└── tests/            # Tests (coming soon)

Setup

pip install -r requirements.txt
cd training
python train.py
python export_onnx.py

MLH Fellowship Application

This project demonstrates production ML deployment for the MLH Fellowship.

About

I built a lightweight, production-style ML inference API that demonstrates how compressed models can be deployed efficiently on low-resource systems

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages