-
Notifications
You must be signed in to change notification settings - Fork 153
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
epic: Semantic naming for Cortex Engines? #1168
Comments
IMO, maybe keep it as |
@tikikun @nguyenhoangthuan99 Can I check: if we are using "engine names", should it be:
|
Lack of decision causing issues here: #1283 Gentle bump @nguyenhoangthuan99 @vansangpfiev to come to an agreement on this. Option 1:
Option 2:
Current (inconsistent): ❯ cortex-nightly engines list
+---+--------------+-------------------+---------+--------------+
| # | Name | Supported Formats | Version | Status |
+---+--------------+-------------------+---------+--------------+
| 1 | ONNXRuntime | ONNX | 0.0.1 | Incompatible |
+---+--------------+-------------------+---------+--------------+
| 2 | llama.cpp | GGUF | 0.0.1 | Ready |
+---+--------------+-------------------+---------+--------------+
| 3 | TensorRT-LLM | TensorRT Engines | 0.0.1 | Incompatible |
+---+--------------+-------------------+---------+--------------+ |
@nguyenhoangthuan99 @vansangpfiev adding this as a priority to decide for consistency before the release |
Related issue: #1283
Analysis:
I prefer Option 1 to update our engine naming convention to llama-cpp, onnx-runtime, and tensorrt-llm. This change brings consistency across our engine names while maintaining their familiarity. The new format improves readability and aligns with common practices in the open-source community, enhancing both user experience and code maintainability. This standardized approach will also help us smoothly integrate any future engines or technologies. The output of terminal when list engines:
cc @vansangpfiev @namchuai @dan-homebrew @0xSage |
@nguyenhoangthuan99 @vansangpfiev @namchuai I think we should go with how the upstream authors name it, i.e.
So essentially how @nguyenhoangthuan99 names them, with just a change on onnxruntime.
For the Repos, I recommend we add an
|
Reopen this epic, we have not rename upstream engines repo yet |
Goal / Summary of discussion
Out of scope - renaming repos?
For the Repos, I recommend we add an
*-engine
suffix to differentiate us from the upstream library:janhq/cortex.tensorrt-llm
janhq/tensorrt-llm-engine
janhq/cortex.llamacpp
janhq/llamacpp-engine
janhq/cortex.onnx
janhq/onnxruntime-engine
Tasklist
The text was updated successfully, but these errors were encountered: