-
Notifications
You must be signed in to change notification settings - Fork 1
-
Action Items
- Convert more Intel Model and push to EmbeddedLLM Huggingface Model Space to speed up auto-download of model. Users do not need to convert on the fly. (On-going)
- Separate the model into another process.
- Support auto relaunch of ONNX DirectML Engine when kernel crashes.
- Intel NPU LLM model support
v0.3
- OpenVINO LLM Support
- Support loading and unloading of OpenVINO and IPEX-LLM model using ModelUI.
- Support OpenVINO Vision Model. Phi-3.5-Vision
- AMD iGPU ellm_api_server
- Intel OpenVINO ellm_api_server
- Intel IPEX-LLM ellm_api_server
Beta Was this translation helpful? Give feedback.
All reactions
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment