-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Updates to the Readme that include mission #1
Conversation
README.md
Outdated
|
||
Because ML practitioners use so many different types of tools, it is a key goal that you can customize the stack to whatever your requirements (within reason), and let the system take care of the "boring stuff." While we have started with a narrow set of technologies, we are working with many different projects to include additional tooling. | ||
|
||
Ultimately, we want to have a single deployable artifact (a manifest) that gives you an easy to use ML stack _anywhere_ Kubernetes is already running and can self configure based on the cluster it deploys into. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
a set of deployable manifests? Because we're doing a directory of them, with the application step being:
kubectl apply -f dir/ -R
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
great point, fixed!
README.md
Outdated
|
||
The Kubeflow project is a new open source Github repo dedicated to reducing friction in doing Machine Learning on Kubernetes. Contained in this repository are: | ||
The Kubeflow project is dedicated to making Machine Learning on Kubernetes easy, portable and scalable. Contained in this repository are: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The elements listed below aren't included in the repository; they are included in the Kubeflow distribution. Should we change the text to something like
Contained in the Kubeflow distribution? or
Contained in the Kubeflow stack?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
+1.
README.md
Outdated
Our goal is to help folks use ML more easily, by letting Kubernetes to do what it's great at: | ||
- Easy, repeatable, portable deployments | ||
- Deploying and managing loosely-coupled microservices | ||
- Scaling based on demand |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
add: managing diverse infrastructure
README.md
Outdated
|
||
### Google Kubernetes Engine |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do not delete this section. It is the best way currently to try out GPUs k8s clusters.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That's fine, but it shouldn't be here - or if it is, it should have other solutions as well. As it stands, it makes it seem that this is the only way to create a production cluster.
@aronchick PTAL. Needs resolving conflict. Good to merge after that |
Squashing and merging this now. |
Initial repository setup.
* rename `getEnvDef` to `getEnvDefault` * Add a comment to describe how the STOP_ANNOTATION gets used Signed-off-by: Kimonas Sotirchos <[email protected]>
* rename `getEnvDef` to `getEnvDefault` * Add a comment to describe how the STOP_ANNOTATION gets used Signed-off-by: Kimonas Sotirchos <[email protected]>
* Create a culler as a package Helper functions for culling resources. Takes for granted that ISTIO is installed to the system and queries Prometheus to get metrics. Specifically, requests/{configurable time}. If the resource should be culled, then it should be done by setting an annotation. This way the UIs can also show that the Resource is stopping and also easily stop a resource by making a PATCH request. Signed-off-by: Kimonas Sotirchos <[email protected]> * Culling logic enhancements Add necessary ENV Vars. Culling won't happen by default. To enable it the user will need to set the ENABLE_CULLING=true Signed-off-by: Kimonas Sotirchos <[email protected]> * Misc fixes in logging and comment cleanup Signed-off-by: Kimonas Sotirchos <[email protected]> * Fix typo Signed-off-by: Kimonas Sotirchos <[email protected]> * Add Notebooks specific culling Query the /api/status endpoint of each Server Signed-off-by: Kimonas Sotirchos <[email protected]> * Remove the generic culling logic We need to discuss if it would make sense to have this logic as a go library, or use knative. Signed-off-by: Kimonas Sotirchos <[email protected]> * Add unit tests Signed-off-by: Kimonas Sotirchos <[email protected]> * Remove unused code Signed-off-by: Kimonas Sotirchos <[email protected]> * Review changes #1 * rename `getEnvDef` to `getEnvDefault` * Add a comment to describe how the STOP_ANNOTATION gets used Signed-off-by: Kimonas Sotirchos <[email protected]> * Make cluster domain configurable Signed-off-by: Kimonas Sotirchos <[email protected]>
* adds ambassador and argo components to manifests * removes ksonnet labels * removed unused stubs for overlays
No description provided.