Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Roadmap (Sept 2024) #23

tjtanaa started this conversation in Ideas
Discussion options

Action Items

  1. Convert more Intel Model and push to EmbeddedLLM Huggingface Model Space to speed up auto-download of model. Users do not need to convert on the fly. (On-going)
  2. Separate the model into another process.
  3. Support auto relaunch of ONNX DirectML Engine when kernel crashes.
  4. Intel NPU LLM model support

v0.3

  1. OpenVINO LLM Support
  2. Support loading and unloading of OpenVINO and IPEX-LLM model using ModelUI.
  3. Support OpenVINO Vision Model. Phi-3.5-Vision
  4. AMD iGPU ellm_api_server
  5. Intel OpenVINO ellm_api_server
  6. Intel IPEX-LLM ellm_api_server
You must be logged in to vote

Replies: 0 comments

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Ideas
Labels
type: discussion / question Further information is requested type: enhancement / feature New feature or request
1 participant

AltStyle によって変換されたページ (->オリジナル) /