-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
inception server not working #621
Comments
What is your networking configuration? In general you won't be able to reach the internal pod IPs of your cluster from your local machine. Are you running python label.py from within your cluster or from your local machine? |
@jlewi thanks for replying! actually for running for this one, the following is my network configuration
|
The error Do you look at the logs of TFServer, do they indicate any errors? |
@jlewi yeah it shows errors!!!
|
Are you running on GCP? The log messages indicate its trying to get a GCP credential and failing. The first error message indicates there is no private key file and the HTTP errors are most likely trying to contact the metadata server to get a token. |
@jlewi yes! I used the model path from user_guide. It's thanks a lot |
i'm not running on GCP. But I used |
I think that model is public but it looks like the GCS libraries in TFServing may not be smart enough to detect that. My suggestion would be to copy that model to a directory that you can serve from. You can use the CLI |
As @jlewi said, Issue is with the debian package which is not able to download the model(although it is publicly available on the google cloud storage and should work with google's default auth). I have verified the issue by downloading and running the debian package directly on the ubuntu machine(without any kubeflow) and it threw the same error. One possible fix could be to download the model explicitly in the dockerfile and then run the binary with |
thanks @jlewi @tremblerz the model is public but if u logout google, you can not download that model after that i'll see if inception run well |
@lluunn We should document this in the user guide. |
This is related to kserve/kserve#135
i follow the step of
Serve a model using TensorFlow Serving
in user_guidekubectl get svc -n kubeflow inception
kubectl get po -n kubeflow
it seems running!
also i did
kubectl describe pod/inception-7bc4df4546-6b4gj -n kubeflow
, and it shows:then i use
kubectl get pod/inception-7bc4df4546-6b4gj -n kubeflow -o yaml
to get portIP is192.168.175.202
(it succeed forjupyter
), so i think192.168.175.202:9000
is correct.then i deployed inception-client, after that:
python label.py -s 192.168.175.202 -p 9000 images/sleeping-pepper.jpg
, but the result is not niceanyone can help me?
--
ks version
--
kubectl version
The text was updated successfully, but these errors were encountered: