Skip to content

enthropy7/YSCV

Repository files navigation

yscv

A complete computer vision and deep learning framework in pure Rust. One cargo add yscv gives you image processing (178 ops, faster than OpenCV), neural network training (39 layer types, 5 optimizers), ONNX inference (128+ operators, INT8 quantization), real-time detection + tracking + recognition (67µs per frame), H.264 video decoding, and GPU compute via Vulkan/Metal/DX12 — all in a single statically-linked binary with zero Python or C++ dependencies.

We built this because deploying ML in production shouldn't require Docker containers with PyTorch, CUDA drivers, and a prayer. YSCV compiles to one binary that runs on a Raspberry Pi, a cloud VM, or a factory floor computer. Every hot path has hand-tuned SIMD for ARM and x86 — 295 functions with runtime dispatch. It's faster than NumPy, PyTorch, and OpenCV on every operation we benchmarked (76 wins, 0 losses).

Quick Start

[dependencies]
yscv = "0.1"

Load an image, process it, save the result — three lines:

use yscv::prelude::*;

let img = imread("photo.jpg")?;
let gray = rgb_to_grayscale(&img)?;
imwrite("gray.png", &gray)?;

What can you build with this?

The short answer: anything you'd normally need Python + OpenCV + PyTorch for.

YSCV covers the full pipeline — from reading pixels off a camera to training a neural network to deploying an optimized model. Here are some real examples:

Video surveillance and security. Hook up a camera, detect people, track them across frames, recognize faces — all in 67 microseconds per frame. That's 15,000 FPS on a single CPU core. No GPU needed. Deploy on any ARM or x86 device as a static binary.

Factory quality control. Train a defect detection model on your laptop, export to ONNX, quantize to INT8, deploy on cheap edge hardware on the production line. The whole thing runs without internet, without Python, without Docker.

Retail and traffic analytics. Count people, track movement paths, measure dwell time. One machine can handle dozens of camera streams because the processing is that fast.

Medical imaging. Process X-rays and DICOM images at scale: resize, normalize, detect edges, extract features. Memory-safe by construction — no segfaults on malformed input, ever.

Robotics and drones. Cross-compile to ARM, get a 5MB binary with feature detection (ORB, SIFT), optical flow, stereo matching, and homography estimation. No Python runtime on your drone.

Training on the edge. Fine-tune a pretrained ResNet or ViT directly on device with Adam optimizer, learning rate scheduling, and gradient clipping. You don't always need a GPU cluster.

Training a model

Build a CNN, train it, done:

use yscv::prelude::*;

let mut graph = Graph::new();
let mut model = SequentialModel::new(&graph);
model.add_conv2d_zero(3, 16, 3, 3, 1, 1, true)?;
model.add_relu();
model.add_flatten();
model.add_linear_zero(&mut graph, 16 * 30 * 30, 10)?;

let config = TrainerConfig {
    optimizer: OptimizerKind::Adam { lr: 0.001 },
    loss: LossKind::CrossEntropy,
    epochs: 50,
    batch_size: 32,
    validation_split: Some(0.2),
};

let result = Trainer::new(config).fit(&mut model, &mut graph, &inputs, &targets)?;
println!("Final loss: {:.4}", result.final_loss);

13 pretrained architectures available out of the box — ResNet, VGG, MobileNet, EfficientNet, ViT, DeiT. Weights download automatically:

let hub = ModelHub::new(); // caches in ~/.yscv/models/
let weights = hub.load_weights("resnet50")?;

Object detection

YOLOv8 detection with tracking and recognition:

use yscv::detect::{detect_yolov8_from_rgb, yolov8_coco_config, non_max_suppression};

let img = imread("scene.jpg")?;
let detections = detect_yolov8_from_rgb(&img, &model, &yolov8_coco_config())?;
let filtered = non_max_suppression(&detections, 0.5);

The detect → track → recognize pipeline runs in 67µs per frame end-to-end. DeepSORT and ByteTrack are built in. VP-tree ANN for recognition.

Performance

We benchmark every hot path against NumPy, PyTorch, OpenCV, onnxruntime, and CoreML. Current score: 77 wins, ~4 ties, 0 losses. Metal GPU inference beats Apple CoreML (which uses dedicated Neural Engine hardware).

Every operation has hand-tuned SIMD on all platforms — NEON on ARM, AVX/SSE on x86, with optional Intel MKL and ARM Performance Libraries for the last few percent.

vs PyTorch (CPU, 1 thread) yscv PyTorch Speedup
sigmoid 921K 0.217ms 1.296ms 6.0×
softmax 512×256 0.098ms 0.216ms 2.2×
layer_norm 512×256 0.065ms 0.117ms 1.8×
batch_norm 64² 0.028ms 0.045ms 1.6×
relu 921K 0.069ms 0.105ms 1.5×
vs OpenCV (u8, 640×480) yscv OpenCV Speedup
resize nearest 0.048ms 0.157ms 3.3×
resize bilinear 0.068ms 0.201ms 3.0×
sobel 3×3 0.074ms 0.169ms 2.3×
dilate 3×3 0.031ms 0.047ms 1.5×
ONNX Inference (YOLOv8n 640×640) yscv Competitor Result
CPU 31.9ms onnxruntime 95.4ms 3× faster
Metal GPU 11.8ms CoreML 13.4ms 14% faster (CoreML uses ANE hw)
YOLO11n CPU 46.0ms all competitors FAIL WIN
YOLO11n Metal 12.4ms all competitors FAIL WIN

Full benchmark results in docs/performance-benchmarks.md.

What's inside

The framework is split into 14 crates, each doing one thing well:

Crate Purpose
yscv-tensor N-dimensional tensor with 115 ops, f32/f16/bf16, SIMD-accelerated
yscv-kernels CPU + GPU compute backends, 295 SIMD functions, 20 GPU shaders
yscv-autograd Reverse-mode autodiff with 40+ backward ops
yscv-optim SGD, Adam, AdamW, LAMB, RAdam — all with SIMD. 11 LR schedulers
yscv-model 39 layer types, Trainer API, model zoo, LoRA, distributed training
yscv-imgproc 178 image processing ops (blur, edges, morphology, features, color)
yscv-video H.264/HEVC decoder, camera I/O, MP4 parsing
yscv-detect YOLOv8/v11 pipeline, NMS, heatmap decoding
yscv-track DeepSORT, ByteTrack, Kalman filter
yscv-recognize Cosine matching, VP-Tree ANN indexing
yscv-eval 41 metrics (mAP, MOTA, HOTA, PSNR, etc.), 11 dataset formats
yscv-onnx 128+ op ONNX runtime, INT8 quantization, graph optimizer, Metal GPU

Building

cargo build --workspace --release
cargo test --workspace --release      # 1,678 tests
cargo run --example train_cnn         # train a CNN on synthetic data
cargo run --example train_linear      # linear regression
cargo run --example image_processing  # image pipeline demo
cargo run --example yolo_detect -- model.onnx photo.jpg  # YOLOv8/v11 detection
cargo run --example yolo_finetune     # fine-tune a detection head

GPU / Metal (Apple Silicon)

# wgpu backend (Vulkan / Metal / DX12 — cross-platform)
cargo run --release --example bench_yolo --features gpu

# Metal-native backend (macOS only — fastest on Apple Silicon)
cargo run --release --example bench_metal_yolo --features metal-backend
cargo run --release --example bench_metal_conv --features metal-backend
cargo run --release --example bench_mps_gemm  --features metal-backend

# Both features can be combined
cargo test --workspace --features metal-backend
cargo clippy --workspace --features metal-backend

The gpu feature uses wgpu and works on any platform with Vulkan, Metal, or DX12. The metal-backend feature talks to Metal directly via metal-rs and is where the fastest Apple Silicon path lives (Winograd F(4×4,3×3), f16 inter-op pipeline, NEON input upload, fused command buffers). On macOS, metal-backend is what you want.

System dependencies

YSCV builds with zero required system dependencies. Optional:

  • OpenBLAS (Linux/Windows) — faster matmul. apt install libopenblas-dev or brew install openblas. macOS uses Accelerate automatically.
  • protoc — for ONNX proto generation. Without it, a built-in fallback is used. apt install protobuf-compiler or brew install protobuf.

Feature flags

Flag What it does Platforms
gpu GPU acceleration via wgpu (Vulkan / Metal / DX12) All
metal-backend Metal-native GPU pipeline — fastest on Apple Silicon macOS only
native-camera Real camera capture (V4L2 / AVFoundation / MediaFoundation) All
blas Hardware BLAS — Accelerate on macOS, OpenBLAS on Linux/Windows All (default)
mkl Intel MKL for vectorized math on x86 x86/x86_64
armpl ARM Performance Libraries on ARM Linux (Graviton, Ampere) aarch64 Linux

Multi-architecture SIMD

All hot paths have hand-tuned SIMD for three architectures with runtime CPU detection:

macOS (Apple Silicon) Linux/Windows (x86_64) Linux (ARM64)
f32 tensor ops NEON 4× unroll AVX 4× unroll → SSE fallback NEON 4× unroll
u8 image ops NEON 16B/iter AVX2 32B/iter → SSE2/SSSE3 16B/iter NEON 16B/iter
Activations NEON 3-term poly AVX/SSE poly NEON 3-term poly
MatMul / Conv Accelerate BLAS OpenBLAS (or MKL with --features mkl) OpenBLAS (or ARMPL with --features armpl)
Vectorized math vDSP (Accelerate) MKL VML (opt-in) ARMPL (opt-in)
Threading GCD dispatch_apply (~0.3µs) std::thread::scope (~1µs) std::thread::scope (~1µs)
GPU inference Metal-native (16.6ms YOLOv8n) wgpu/Vulkan wgpu/Vulkan
Softmax Fused NEON Fused AVX/SSE Fused NEON
Allocator mimalloc mimalloc mimalloc

SIMD dispatch is automatic at runtime — no need for -C target-cpu flags (though they help: -C target-cpu=apple-m1 or -C target-cpu=native for best codegen). The framework uses std::is_x86_feature_detected!("avx") on x86 and compile-time #[cfg(target_arch = "aarch64")] on ARM. 295 SIMD functions total, all with scalar fallback for WASM/RISC-V/Miri.

Recommended release profile

For best performance, the workspace already includes an optimized release profile in Cargo.toml:

[profile.release]
lto = "thin"
codegen-units = 1

For Apple Silicon specifically, add to .cargo/config.toml:

[target.aarch64-apple-darwin]
rustflags = ["-C", "target-cpu=apple-m1"]

Architecture

yscv (prelude)
    ↓
┌─────────────┐ ┌──────────┐ ┌───────────┐
│ yscv-model   │ │ yscv-    │ │ yscv-     │
│ (39 layers,  │ │ imgproc  │ │ detect/   │
│  trainer,    │ │ (178 ops)│ │ track/    │
│  zoo, LoRA)  │ │          │ │ recognize │
└──────┬───────┘ └────┬─────┘ └─────┬─────┘
       ↓              ↓             ↓
┌──────────────┐ ┌──────────┐ ┌──────────┐
│ yscv-autograd│ │ yscv-    │ │ yscv-    │
│ (40+ backward│ │ kernels  │ │ video    │
│  ops)        │ │ (SIMD+   │ │ (H.264,  │
└──────┬───────┘ │ GPU)     │ │ HEVC,    │
       ↓         └────┬─────┘ │ camera)  │
┌──────────────┐      ↓       └──────────┘
│ yscv-tensor  │←─────┘
│ (115 ops,    │
│  f32/f16/bf16│
│  SIMD)       │
└──────────────┘

Cookbook

See docs/cookbook.md for practical recipes: image processing, model training, ONNX inference (CPU/GPU/Metal), YOLO detection, tracking, preprocessing pipelines, fine-tuning, video, cross-compilation, and benchmarking.

When to use YSCV (and when not to)

Use YSCV when you need to deploy a trained model as a single binary without Python. When you're building a real-time CV pipeline on edge hardware. When you want to train a CNN or ViT on moderate-scale data without setting up a GPU cluster. When you need image processing that's faster than OpenCV and memory-safe.

Don't use YSCV when you need the Python ML ecosystem — Hugging Face model hub, thousands of community architectures, Jupyter notebook prototyping, dynamic graph debugging with breakpoints. When you're training foundation models across thousands of GPUs with NCCL. For that, use PyTorch.

YSCV is faster than PyTorch on individual operations (76 benchmark wins), but PyTorch has a decade-old ecosystem with millions of pretrained models and research tooling. Different tools for different jobs. Train in PyTorch, export to ONNX, deploy with YSCV — or train directly in YSCV if your model fits within our 39 layer types and you don't need the Python ecosystem.

License

Licensed under either of Apache 2.0 or MIT, at your option.

About

No description, website, or topics provided.

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors