-
Notifications
You must be signed in to change notification settings - Fork 29
Open
Description
What can we do for the default backend?
- Model: Let's start with llama 3.2 1B, preferably quantized ones
- Runtime:
- MediaPipe and AI-Edge-Torch: our current default backend is tflite based, so if we can continue using TFLite-based solution, it would be good.
-
- new API in LiteRT (https://ai.google.dev/edge/litert)
- ExecuTorch: https://github.com/pytorch/executorch, https://github.com/pytorch/executorch/blob/main/examples/models/llama/README.md
- llama.cpp: https://github.com/ggerganov/llama.cpp
- onnx runtime: https://github.com/microsoft/onnxruntime, let's check if it works on Android platforms.
- onnx runtime genai: https://github.com/microsoft/onnxruntime-genai
Metadata
Metadata
Assignees
Labels
No labels