1 Commits

Author SHA1 Message Date
Augustin
e528b10a0a Add ONNX inference with tokenization support
Implement complete ONNX inference pipeline with NPU acceleration:
- Add OnnxClassifier for text classification via ONNX Runtime
- Integrate HuggingFace tokenizers for text preprocessing
- Support tokenization with padding/truncation
- Implement classification with probabilities (softmax)
- Add distilbert_tokenizer() model config for download

Features:
- Tokenize text input to input_ids and attention_mask
- Run NPU-accelerated inference via DirectML
- Extract logits and convert to probabilities
- RefCell pattern for session management

Note: Current blocker is Windows linker C runtime mismatch between
esaxx-rs (static MT) and ONNX Runtime (dynamic MD). Code compiles
but linking fails. Resolution in progress.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-16 19:16:51 +02:00