input
stringlengths
1
18.7k
output
stringlengths
1
18.7k
Thanks so much Jonathon, that's super helpful feedback. I'd be happy to jump on a call. Although I'm also very early on in evaluating Flyte and haven't given it a good test drive yet.
Oliver Mannion / Jonathon Belotti as said earlier, I would like to start a conversation with you guys. :slightly_smiling_face:
Oliver Mannion / Jonathon Belotti as said earlier, I would like to start a conversation with you guys. :slightly_smiling_face:
No worries, if we can agree on a time that crosses between our timezones (i’m in Sydney, Australia) I can get on a call.
No worries, if we can agree on a time that crosses between our timezones (i’m in Sydney, Australia) I can get on a call.
I am in PST (Seattle USA)
I am in PST (Seattle USA)
I'm in Melbourne, Australia same timezone as Sydney :slightly_smiling_face: Jonathon Belotti just wondering, do you ever evaluate airflow?
I'm in Melbourne, Australia same timezone as Sydney :slightly_smiling_face: Jonathon Belotti just wondering, do you ever evaluate airflow?
the team did, just prior to my arrival
the team did, just prior to my arrival
Oliver Mannion Jonathon Belotti where are you guys at. How can I help decide
Hello Guys, Im here triying to make my own module to work with flyte, but always that i want to register the new flow it answers me this : `ModuleNotFoundError`. Im doing something wrong?
Can you share what command are you trying to run?
Can you share what command are you trying to run?
and can you describe what you’re trying to do. and can you please `echo $PYTHONPATH`?
and can you describe what you’re trying to do. and can you please `echo $PYTHONPATH`?
I might know what your issue is You have "workflow_packages" set to `flyte`. This means the system expects your workflow to live in the `flyte` module. You need a directory `flyte` (which I see you have one). It needs to have an `__init__.py` , any chance that is missing?
I might know what your issue is You have "workflow_packages" set to `flyte`. This means the system expects your workflow to live in the `flyte` module. You need a directory `flyte` (which I see you have one). It needs to have an `__init__.py` , any chance that is missing?
In fact you're right it was a problem with the path, i already solve it but now every time i ran my task with something it answers this: ´Retries [0/1], task failed, TaskFailedUnknownError: Container/Pod failed. No message received from kubernetes. Could be permissions?´ I try to run this short example but it does not allow it :'(
In fact you're right it was a problem with the path, i already solve it but now every time i ran my task with something it answers this: ´Retries [0/1], task failed, TaskFailedUnknownError: Container/Pod failed. No message received from kubernetes. Could be permissions?´ I try to run this short example but it does not allow it :'(
Sorry about that confusing error. can you try to `kubectl get pods -n {yourproject}-development` ? Then `kubectl logs -n {yourproject}-deevelopment {yourpod}`
Sorry about that confusing error. can you try to `kubectl get pods -n {yourproject}-development` ? Then `kubectl logs -n {yourproject}-deevelopment {yourpod}`
Hello Johnny first i want to thanks all you're always here to help :smiley:. Im running it locally with minikube, when I put kubectl get pods -n flyte it shows me this: But when i try to put the logs it does not found the pod =(
Hello Johnny first i want to thanks all you're always here to help :smiley:. Im running it locally with minikube, when I put kubectl get pods -n flyte it shows me this: But when i try to put the logs it does not found the pod =(
Eduardo Giraldo `flyteadmin` isn't the pod you're looking for I think. You want the pod which ran your task. your pod ran in the namespace `yourproject-yourdomain` yourdomain is probably "development" alternatively, you can just do `kubectl get pods --all-namespaces` to see every pod in every namespace (your pod should be included)
Eduardo Giraldo `flyteadmin` isn't the pod you're looking for I think. You want the pod which ran your task. your pod ran in the namespace `yourproject-yourdomain` yourdomain is probably "development" alternatively, you can just do `kubectl get pods --all-namespaces` to see every pod in every namespace (your pod should be included)
With the second command i can see the logs, Im going to start to debug. Thanks a lot Johnny. You rock :smiley:
I will move the code there
If the `lyft/protocgenerator` code is open-source that'd be really cool I'm beginning to using the docker image a lot xD
If the `lyft/protocgenerator` code is open-source that'd be really cool I'm beginning to using the docker image a lot xD
Adhita Selvaraj that’s the plan, I am working on the replacement
Hi, can I get a +1 for this PR? <https://github.com/lyft/flyte/pull/184> Thanks.
Do you mind running `make kustomize` to update the generated files?
Do you mind running `make kustomize` to update the generated files?
Thanks for minding, fixed it. wdyt of <https://github.com/lyft/flyte/pull/200> I got an approval already from Yee Lightening fast!
Thanks for minding, fixed it. wdyt of <https://github.com/lyft/flyte/pull/200> I got an approval already from Yee Lightening fast!
yeah i like these github workflow things. trying to use them more across all our repos in place of travis.
yeah i like these github workflow things. trying to use them more across all our repos in place of travis.
seems production ready. one thing as I just commented, i didn’t know force push would remove previous workflow executions. _something worth knowing i think_
seems production ready. one thing as I just commented, i didn’t know force push would remove previous workflow executions. _something worth knowing i think_
just force push on the topic branch right?
just force push on the topic branch right?
yeah
yeah
yeah, all good. as long as it’s not master we don’t care
yeah, all good. as long as it’s not master we don’t care
scary that should be disabled.
scary that should be disabled.
i’ll check in a bit.
Hello guys, I have a couple of questions about config: 1. Is there a way to setup the user and password to download a private docker image or use only a local generated Image on Local Sandbox? 2. How can i setup my own s3 bucket on Local Sandbox?
So for #1, you'd usually use kubernetes imagepullsecrets <https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry/> For #2, you say "setting up a custom bucket". Do you mean creating a custom bucket? Or configuring your Flyte workflow to use a custom bucket that has already been created?
So for #1, you'd usually use kubernetes imagepullsecrets <https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry/> For #2, you say "setting up a custom bucket". Do you mean creating a custom bucket? Or configuring your Flyte workflow to use a custom bucket that has already been created?
Hello Johnny Burns, I want to work with a S3 bucket in my AWS, i already create it but I don't figure it out who to config the sandbox to use my credentials and connect into the bucket
Hello Johnny Burns, I want to work with a S3 bucket in my AWS, i already create it but I don't figure it out who to config the sandbox to use my credentials and connect into the bucket
Ah, yeah, great question. So you can configure which s3 bucket to use in the SDK configs: <https://github.com/lyft/flytesnacks/blob/master/python/sandbox.config#L13> As for permissions to upload/download from that bucket, there are several ways to do that. The first way is using AWS "roles". Since you're running locally, you probably want username/password though, I'm guessing. You can configure that in the SDK configs here: <https://github.com/lyft/flytesnacks/blob/a6ec170713efa76aa392e5d7648c5765012d010c/python/sandbox.config#L15-L17>
Ah, yeah, great question. So you can configure which s3 bucket to use in the SDK configs: <https://github.com/lyft/flytesnacks/blob/master/python/sandbox.config#L13> As for permissions to upload/download from that bucket, there are several ways to do that. The first way is using AWS "roles". Since you're running locally, you probably want username/password though, I'm guessing. You can configure that in the SDK configs here: <https://github.com/lyft/flytesnacks/blob/a6ec170713efa76aa392e5d7648c5765012d010c/python/sandbox.config#L15-L17>
ok Johnny I check it out. Thanks for the info, as always you rock :smiley:
ok Johnny I check it out. Thanks for the info, as always you rock :smiley:
tyty
Hello Everyone! I have a question about running `SPARK tasks`. Is there way to run their executors on the nodes with `GPU`? Yes, we’ve managed to do this for python Sidecar tasks. As I know for sparkoperator we should specify GPU request in `<http://sparkapplications.sparkoperator.k8s.io|sparkapplications.sparkoperator.k8s.io>` CRD. And if is it possible could you please advice how to specify this from python code? Thank you in advance! <https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/user-guide.md#requesting-gpu-resources>
Spark 2.4.x doesn't natively support GPUs and GPU support in the operator needs a web-hook to be enabled. Spark 3.0 will support GPUs natively which will simplify some of this set-up. Coming to Flyte, we again don't currently support this : <https://github.com/lyft/flyteplugins/blob/master/go/tasks/plugins/k8s/spark/spark.go#L153> and would need some changes in flytekit/flyteplugins to pass gpu config through to the spark-operator. Let me cut an issue so that we can prioritize/track this. Filed <https://github.com/lyft/flyte/issues/224> to track this. Feel free to add any details I missed Ruslan Stanevich Ketan Umare
Spark 2.4.x doesn't natively support GPUs and GPU support in the operator needs a web-hook to be enabled. Spark 3.0 will support GPUs natively which will simplify some of this set-up. Coming to Flyte, we again don't currently support this : <https://github.com/lyft/flyteplugins/blob/master/go/tasks/plugins/k8s/spark/spark.go#L153> and would need some changes in flytekit/flyteplugins to pass gpu config through to the spark-operator. Let me cut an issue so that we can prioritize/track this. Filed <https://github.com/lyft/flyte/issues/224> to track this. Feel free to add any details I missed Ruslan Stanevich Ketan Umare
Ruslan Stanevich want to help us launch this in open source? contributions welcome
Ruslan Stanevich want to help us launch this in open source? contributions welcome
Oh great! thank you! We also would like to manage K8s scheduler to `not schedule` other pods on expensive `GPU` nodes. So our GPU nodes should have `taints` and Pods for GPU nodes should have `tolerations`. Moreover, in the future we’d like to set `nodeSelector` to choose different type of nodes for some tasks (default/computed optized/GPU/etc) And we looked at `LaunchPlan Annotations` in python code and how we can use them for managing Spark tasks. For now we implemented quick-n-dirty solution. We wrote simple `k8s webhook controller` which modify `SparkApplication CRD` based on Annotations values. Example of annotations: ```... annotations: <http://jsonPatchSpark.lyft-minsk.com/patches|jsonPatchSpark.lyft-minsk.com/patches>: | - op: add path: /spec/executor/gpu value: name: <http://nvidia.com/gpu|nvidia.com/gpu> quantity: 1 - op: add path: /spec/executor/tolerations value: - key: flyte/gpu value: dedicated operator: Equal effect: NoSchedule ...``` This applies necessary patches for `SparkApplication` to run task on GPU node. Yes, it is too low-level definition with jsonPatches... but It allowed us to make fine tuning for requested resources. For today we got good results on DEV cluster.
Oh great! thank you! We also would like to manage K8s scheduler to `not schedule` other pods on expensive `GPU` nodes. So our GPU nodes should have `taints` and Pods for GPU nodes should have `tolerations`. Moreover, in the future we’d like to set `nodeSelector` to choose different type of nodes for some tasks (default/computed optized/GPU/etc) And we looked at `LaunchPlan Annotations` in python code and how we can use them for managing Spark tasks. For now we implemented quick-n-dirty solution. We wrote simple `k8s webhook controller` which modify `SparkApplication CRD` based on Annotations values. Example of annotations: ```... annotations: <http://jsonPatchSpark.lyft-minsk.com/patches|jsonPatchSpark.lyft-minsk.com/patches>: | - op: add path: /spec/executor/gpu value: name: <http://nvidia.com/gpu|nvidia.com/gpu> quantity: 1 - op: add path: /spec/executor/tolerations value: - key: flyte/gpu value: dedicated operator: Equal effect: NoSchedule ...``` This applies necessary patches for `SparkApplication` to run task on GPU node. Yes, it is too low-level definition with jsonPatches... but It allowed us to make fine tuning for requested resources. For today we got good results on DEV cluster.
this is awesome
this is awesome
And yes important note. It was some time ago but we had to activate webhooks for sparkoperator (which not enabled in flyte default) too in eks cluster with Flyte. Some teams wanted using usual Scheduled SparkApplications for data syncs.
And yes important note. It was some time ago but we had to activate webhooks for sparkoperator (which not enabled in flyte default) too in eks cluster with Flyte. Some teams wanted using usual Scheduled SparkApplications for data syncs.
hmm why not use scheduled jobs through flyte? Ruslan Stanevich, would love to understand whats happening on your side. Can we sync up sometime?
hmm why not use scheduled jobs through flyte? Ruslan Stanevich, would love to understand whats happening on your side. Can we sync up sometime?
Sure, we prefer using scheduled launch plans in Flyte for all python jobs! It works great! But just there are some DWH Team’s jobs written in Scala and they are running as scheduledSparkApplication. So, we use native Flyte features where possible :+1:
Sure, we prefer using scheduled launch plans in Flyte for all python jobs! It works great! But just there are some DWH Team’s jobs written in Scala and they are running as scheduledSparkApplication. So, we use native Flyte features where possible :+1:
Ruslan Stanevich actually we just added support for scala spark jobs in python
Ruslan Stanevich actually we just added support for scala spark jobs in python
Sorry maybe I was confusing. I mean the code is written in Scala language. So, If I understood correctly we cannot run Scala code in flyte. If about spark_jobs (written in python) - yes they are scheduled also via native Flyte approach (Cloudwatch/SQS)
Sorry maybe I was confusing. I mean the code is written in Scala language. So, If I understood correctly we cannot run Scala code in flyte. If about spark_jobs (written in python) - yes they are scheduled also via native Flyte approach (Cloudwatch/SQS)
You can run any language in Flyte, Spotify is actually writing a Java sdk. But scala spark job support is already available in flytekit python The only thing is you use a jar
You can run any language in Flyte, Spotify is actually writing a Java sdk. But scala spark job support is already available in flytekit python The only thing is you use a jar
This is exciting! But I'm not sure that I understand now how run it. Let me check. Can I find some examples of running other languages in flytekit repo?
This is exciting! But I'm not sure that I understand now how run it. Let me check. Can I find some examples of running other languages in flytekit repo?
Hmm let me try and share, Anmol Khurana do you have this example checked in?
Hmm let me try and share, Anmol Khurana do you have this example checked in?
I don’t have that right now. I am planning to work on it this week. Will share once its ready. Ruslan Stanevich this was recently checked in and is available now in beta: <https://github.com/lyft/flytekit/releases/tag/v0.7.0b4> will add docs but <https://github.com/lyft/flytekit/blob/master/tests/flytekit/common/workflows/scala_spark.py#L10> is an example of how a scala spark job can be integrated as a task in Flyte
I don’t have that right now. I am planning to work on it this week. Will share once its ready. Ruslan Stanevich this was recently checked in and is available now in beta: <https://github.com/lyft/flytekit/releases/tag/v0.7.0b4> will add docs but <https://github.com/lyft/flytekit/blob/master/tests/flytekit/common/workflows/scala_spark.py#L10> is an example of how a scala spark job can be integrated as a task in Flyte
Anmol Khurana can you close the issue too?
Anmol Khurana can you close the issue too?
done
done
Anmol Khurana it sounds great! thank you! will share with team
Johnny Burns can you please help Jordan Bramble This is the most common problem everyone has I think i should install minikube to really answer it he is having a problem connecting to admin
Hey Johnny, having some issues building flyte on MacOS with minikube. minikube has had a regression that has broken minikube tunnel. I was able to work around this using: `minikube service --alsologtostderr -v=3 contour -n heptio-contour` which has given me a mapping that I can access the console with. However when trying to run a docker container for myflyteproject example, I am running into networking errors with gprc.
Hey Johnny, having some issues building flyte on MacOS with minikube. minikube has had a regression that has broken minikube tunnel. I was able to work around this using: `minikube service --alsologtostderr -v=3 contour -n heptio-contour` which has given me a mapping that I can access the console with. However when trying to run a docker container for myflyteproject example, I am running into networking errors with gprc.
Hey Jordan. Glad to help :wave: There is a lot to unpack with your question. I don't think you should need minikube tunnel, but let's talk about your problem before I say that for sure. When you say that you're trying to run a docker container, you mean that you've built a container for a flyte workflow? Are you having trouble registering that? or running it?
Hey Jordan. Glad to help :wave: There is a lot to unpack with your question. I don't think you should need minikube tunnel, but let's talk about your problem before I say that for sure. When you say that you're trying to run a docker container, you mean that you've built a container for a flyte workflow? Are you having trouble registering that? or running it?
yes, so I went through the install and set up which implied i needed minikube tunnel. I was able to use service instead to get to the console. Now I am attempting to run a workflow, based on this <https://lyft.github.io/flyte/user/getting_started/create_first.html> I am able to build the container. now I am attempting run it. using the command at the bottom, alongside of the URL and port i received after running `minikube service --alsologtostderr -v=3 contour -n heptio-contour`
yes, so I went through the install and set up which implied i needed minikube tunnel. I was able to use service instead to get to the console. Now I am attempting to run a workflow, based on this <https://lyft.github.io/flyte/user/getting_started/create_first.html> I am able to build the container. now I am attempting run it. using the command at the bottom, alongside of the URL and port i received after running `minikube service --alsologtostderr -v=3 contour -n heptio-contour`
Ah. Yeah. I'm guessing you built that container with a `docker` command on your mac?
Ah. Yeah. I'm guessing you built that container with a `docker` command on your mac?
yes.
yes.
Minikube is a VM, so we can think of it kind of like a remote machine. When you built your docker container locally, that image lives on your mac. Flyte is running inside your Minikube VM, which is a totally separate machine. So when Flyte launches your task, it's looking for this docker image, but that docker image does not exist on the machine (the minikube VM). Make sense? If so, I can propose a few solutions.
Minikube is a VM, so we can think of it kind of like a remote machine. When you built your docker container locally, that image lives on your mac. Flyte is running inside your Minikube VM, which is a totally separate machine. So when Flyte launches your task, it's looking for this docker image, but that docker image does not exist on the machine (the minikube VM). Make sense? If so, I can propose a few solutions.
yes that makes sense.
yes that makes sense.
Cool. So, here are 3 ways you can solve this issue: 1. Build the docker container _inside_ the minikube VM. By that, I mean use `minikube ssh` and treat that as your development environment. Run your same `docker build` command there. Then, your VM will have the image. 2. Push your image to a remote datastore like dockerhub. Your VM will be able to pull the remote image from dockerhub. 3. Use docker-for-mac instead of minikube. This will run K8s on your mac, so Flyte will be running on the same machine you built the image on.
Cool. So, here are 3 ways you can solve this issue: 1. Build the docker container _inside_ the minikube VM. By that, I mean use `minikube ssh` and treat that as your development environment. Run your same `docker build` command there. Then, your VM will have the image. 2. Push your image to a remote datastore like dockerhub. Your VM will be able to pull the remote image from dockerhub. 3. Use docker-for-mac instead of minikube. This will run K8s on your mac, so Flyte will be running on the same machine you built the image on.
what do you recommend here? my goal is to get a working environment up ASAP, so I can build some test workflows locally. They won't require processing large datasets at this time.
what do you recommend here? my goal is to get a working environment up ASAP, so I can build some test workflows locally. They won't require processing large datasets at this time.
I would do #1, tbh Are you using the `flytesnacks` repo to register your first image?
I would do #1, tbh Are you using the `flytesnacks` repo to register your first image?
yes, I am running `make docker_build` from inside of the repo and then attempting to run the container
yes, I am running `make docker_build` from inside of the repo and then attempting to run the container
Cool. So I would git pull that from within your minikube There is one other small snag you're going to hit You'll need to change this line: <https://github.com/lyft/flytesnacks/blob/master/python/Dockerfile#L33> I think the image name is incorrect. When you're done building the image, do `docker images` and verify the name of the docker image. I think you probably just need to remove `<http://docker.io|docker.io>` from that line. If so, LMK because I'm going to put in a PR to fix it in the flytesnacks repo.
Cool. So I would git pull that from within your minikube There is one other small snag you're going to hit You'll need to change this line: <https://github.com/lyft/flytesnacks/blob/master/python/Dockerfile#L33> I think the image name is incorrect. When you're done building the image, do `docker images` and verify the name of the docker image. I think you probably just need to remove `<http://docker.io|docker.io>` from that line. If so, LMK because I'm going to put in a PR to fix it in the flytesnacks repo.
Thanks, I've got this built inside of minikube and registered the xgboost tasks. However, now when I look at my flyte console which I have a tunnel to from my local browser, I don't see any tasks under flytesnacks.
Thanks, I've got this built inside of minikube and registered the xgboost tasks. However, now when I look at my flyte console which I have a tunnel to from my local browser, I don't see any tasks under flytesnacks.
Did your `register` work successfully? Are you registering under the `flytesnacks` project? if not, change the `-p` in your registration command to `flytesnacks`
Did your `register` work successfully? Are you registering under the `flytesnacks` project? if not, change the `-p` in your registration command to `flytesnacks`
yes to both.
yes to both.
no tasks or workflows?
no tasks or workflows?
oh, I got it. my bad I was looking under flytesnacks instead of the project name I created :slightly_smiling_face: thanks all!
Hello Everyone! could you please advise what is the minimal required `Flytekit` version for new `Flyte release`? I think I use old vwersion I got this error when registering workflow built with `flytekit==0.5.3` ``` File "/usr/local/lib/python3.7/dist-packages/grpc/_<http://channel.py|channel.py>", line 729, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: &lt;_InactiveRpcError of RPC that terminated with: status = StatusCode.UNKNOWN details = "failed database operation with column "type" of relation "tasks" does not exist" debug_error_string = "{"created":"@1586327555.355034200","description":"Error received from peer ipv4:10.200.63.62:80","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"failed database operation with column "type" of relation "tasks" does not exist","grpc_status":2}"```
0.5.3 should be fine this honestly sounds like a mismatch internally in admin. assuming this is a local installation, could you kill the admin pod and let a new one come up and then try again?
0.5.3 should be fine this honestly sounds like a mismatch internally in admin. assuming this is a local installation, could you kill the admin pod and let a new one come up and then try again?
hmm, when we come back to the previous flyte release - all worked fine … it is in AWS EKS
hmm, when we come back to the previous flyte release - all worked fine … it is in AWS EKS
which version worked?
which version worked?
<http://docker.io/lyft/flyteadmin:v0.2.1|docker.io/lyft/flyteadmin:v0.2.1> works <http://docker.io/lyft/flyteadmin:v0.2.4|docker.io/lyft/flyteadmin:v0.2.4> failed for us
<http://docker.io/lyft/flyteadmin:v0.2.1|docker.io/lyft/flyteadmin:v0.2.1> works <http://docker.io/lyft/flyteadmin:v0.2.4|docker.io/lyft/flyteadmin:v0.2.4> failed for us
can you run v0.2.4 again? and kill the pod and let a new one come up? i feel like one of the migrations didn’t run for some reason. the migrations should run as part of the init containers. also, can you confirm that the admin version for all the init containers also match that v0.2.4?
can you run v0.2.4 again? and kill the pod and let a new one come up? i feel like one of the migrations didn’t run for some reason. the migrations should run as part of the init containers. also, can you confirm that the admin version for all the init containers also match that v0.2.4?
Yes, it is 0.2.1 in init! (due it places in our overlays) Let me check again!
Yes, it is 0.2.1 in init! (due it places in our overlays) Let me check again!
cool, let me know
cool, let me know
Yes, it was my bad Everything works :+1: Thank you very much!
Yes, it was my bad Everything works :+1: Thank you very much!
of course!
Hello :hand: Could you please help to figure out with my possible misconfiguration for new `Datacatalog`? When starting `v0.2.1` - the one new log record appeared ```{"json":{"src":"stowstore.go:76"},"level":"warning","msg":"stow configuration section missing, defaulting to legacy s3/minio connection config","ts":"2020-04-09T18:48:07Z"}``` but there is no this log when running datacatalog `v0.1.1` Thank you in advice! # Storage config: ```storage: type: s3 container: {{ AWS_S3_BUCKET }} connection: access-key: {{ AWS_ACCESS_KEY }} auth-type: accesskey secret-key: {{ AWS_SECRET_KEY }} region: us-east-1``` # Start logs with Datacatalog 0.1.1 ```Using config file: [/etc/datacatalog/config/datacatalog_config.yaml] time="2020-04-09T18:50:16Z" level=info msg="Using config file: [/etc/datacatalog/config/datacatalog_config.yaml]" time="2020-04-09T18:50:16Z" level=info msg="Config section [logger] updated. Firing updated event." src="viper.go:317" {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [database] updated. No update handler registered.","ts":"2020-04-09T18:50:16Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [datacatalog] updated. No update handler registered.","ts":"2020-04-09T18:50:16Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [application] updated. No update handler registered.","ts":"2020-04-09T18:50:16Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [storage] updated. No update handler registered.","ts":"2020-04-09T18:50:16Z"} {"json":{"src":"serve.go:67"},"level":"info","msg":"Serving DataCatalog http on port :0","ts":"2020-04-09T18:50:16Z"} {"json":{"app_name":"datacatalog","src":"service.go:91"},"level":"info","msg":"Created data storage.","ts":"2020-04-09T18:50:17Z"} {"json":{"app_name":"datacatalog","src":"service.go:110"},"level":"info","msg":"Created DB connection.","ts":"2020-04-09T18:50:17Z"} {"json":{"src":"server.go:95"},"level":"info","msg":"Starting profiling server on port [10254]","ts":"2020-04-09T18:50:17Z"} {"json":{"src":"serve.go:48"},"level":"info","msg":"Serving DataCatalog Insecure on port :8089","ts":"2020-04-09T18:50:17Z"}```
Can we look in a bit
Can we look in a bit
# Start Logs with Datacatalog 0.2.1 ```Warn: No metricsProvider set for the workqueue Using config file: [/etc/datacatalog/config-app/datacatalog_config.yaml] time="2020-04-09T18:48:07Z" level=info msg="Using config file: [/etc/datacatalog/config-app/datacatalog_config.yaml]" time="2020-04-09T18:48:07Z" level=info msg="Config section [database] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [datacatalog] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [application] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [storage] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [logger] updated. Firing updated event." src="viper.go:317" {"json":{"src":"serve.go:71"},"level":"info","msg":"Serving DataCatalog http on port :0","ts":"2020-04-09T18:48:07Z"} {"json":{"src":"stowstore.go:76"},"level":"warning","msg":"stow configuration section missing, defaulting to legacy s3/minio connection config","ts":"2020-04-09T18:48:07Z"} {"json":{"app_name":"datacatalog","src":"service.go:79"},"level":"info","msg":"Created data storage.","ts":"2020-04-09T18:48:07Z"} {"json":{"app_name":"datacatalog","src":"service.go:98"},"level":"info","msg":"Created DB connection.","ts":"2020-04-09T18:48:07Z"} {"json":{"src":"server.go:96"},"level":"info","msg":"Starting profiling server on port [10254]","ts":"2020-04-09T18:48:07Z"} {"json":{"src":"serve.go:49"},"level":"info","msg":"Serving DataCatalog Insecure on port :8089","ts":"2020-04-09T18:48:07Z"}```
# Start Logs with Datacatalog 0.2.1 ```Warn: No metricsProvider set for the workqueue Using config file: [/etc/datacatalog/config-app/datacatalog_config.yaml] time="2020-04-09T18:48:07Z" level=info msg="Using config file: [/etc/datacatalog/config-app/datacatalog_config.yaml]" time="2020-04-09T18:48:07Z" level=info msg="Config section [database] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [datacatalog] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [application] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [storage] updated. No update handler registered." src="viper.go:315" time="2020-04-09T18:48:07Z" level=info msg="Config section [logger] updated. Firing updated event." src="viper.go:317" {"json":{"src":"serve.go:71"},"level":"info","msg":"Serving DataCatalog http on port :0","ts":"2020-04-09T18:48:07Z"} {"json":{"src":"stowstore.go:76"},"level":"warning","msg":"stow configuration section missing, defaulting to legacy s3/minio connection config","ts":"2020-04-09T18:48:07Z"} {"json":{"app_name":"datacatalog","src":"service.go:79"},"level":"info","msg":"Created data storage.","ts":"2020-04-09T18:48:07Z"} {"json":{"app_name":"datacatalog","src":"service.go:98"},"level":"info","msg":"Created DB connection.","ts":"2020-04-09T18:48:07Z"} {"json":{"src":"server.go:96"},"level":"info","msg":"Starting profiling server on port [10254]","ts":"2020-04-09T18:48:07Z"} {"json":{"src":"serve.go:49"},"level":"info","msg":"Serving DataCatalog Insecure on port :8089","ts":"2020-04-09T18:48:07Z"}```
remind me again what the deployment is? is this on eks?
remind me again what the deployment is? is this on eks?
hi Yee! yes this is eks
hi Yee! yes this is eks
this is the configuration section that we have for data catalog ```storage: type: s3 connection: auth-type: iam region: us-east-1 cache: max_size_mbs: 1024 target_gc_percent: 70 container: "your-s3-bucket-name"``` do you have that somewhere? usually when something like this happens, it’s either a misconfiguration issue, or the config files aren’t even being loaded.
this is the configuration section that we have for data catalog ```storage: type: s3 connection: auth-type: iam region: us-east-1 cache: max_size_mbs: 1024 target_gc_percent: 70 container: "your-s3-bucket-name"``` do you have that somewhere? usually when something like this happens, it’s either a misconfiguration issue, or the config files aren’t even being loaded.
we have this one: ```storage: type: s3 container: {{ AWS_S3_BUCKET }} connection: access-key: {{ AWS_ACCESS_KEY }} auth-type: accesskey secret-key: {{ AWS_SECRET_KEY }} region: us-east-1```
we have this one: ```storage: type: s3 container: {{ AWS_S3_BUCKET }} connection: access-key: {{ AWS_ACCESS_KEY }} auth-type: accesskey secret-key: {{ AWS_SECRET_KEY }} region: us-east-1```
near the top of the logs for catalog, when the process starts after the pod has been brought up, it should say all the log files it’s loaded. can you look for something like that? this is from my local sandbox deployment but something along these lines ```time="2020-03-30T16:14:51Z" level=info msg="Using config file: [/etc/datacatalog/config/datacatalog_config.yaml]" time="2020-03-30T16:14:51Z" level=info msg="Config section [logger] updated. Firing updated event." src="viper.go:317" {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [database] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [datacatalog] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [application] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [storage] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"serve.go:67"},"level":"info","msg":"Serving DataCatalog http on port :0","ts":"2020-03-30T16:14:51Z"}```
near the top of the logs for catalog, when the process starts after the pod has been brought up, it should say all the log files it’s loaded. can you look for something like that? this is from my local sandbox deployment but something along these lines ```time="2020-03-30T16:14:51Z" level=info msg="Using config file: [/etc/datacatalog/config/datacatalog_config.yaml]" time="2020-03-30T16:14:51Z" level=info msg="Config section [logger] updated. Firing updated event." src="viper.go:317" {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [database] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [datacatalog] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [application] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"viper.go:315"},"level":"info","msg":"Config section [storage] updated. No update handler registered.","ts":"2020-03-30T16:14:51Z"} {"json":{"src":"serve.go:67"},"level":"info","msg":"Serving DataCatalog http on port :0","ts":"2020-03-30T16:14:51Z"}```
sure, It says `using config file: [ … ]` (line 2) And we got `warning` (line 11) This is logs for starting datacatalog `0.2.1` but yes, using version `0.1.2` I got the same logs like you’ve shown me
sure, It says `using config file: [ … ]` (line 2) And we got `warning` (line 11) This is logs for starting datacatalog `0.2.1` but yes, using version `0.1.2` I got the same logs like you’ve shown me
Andrew Chan any idea what this might be? Ruslan Stanevich stupid question, but you’re obviously not running minio as part of your eks deployment… so does data catalog work? I’m wondering if the log line itself is erroneous.
Andrew Chan any idea what this might be? Ruslan Stanevich stupid question, but you’re obviously not running minio as part of your eks deployment… so does data catalog work? I’m wondering if the log line itself is erroneous.
sure, we don’t use minio as part of deployment! about datacatalog I’m not sure … let me check hmm honestly it works fine ```grpcurl \ -authority="flyte.datacatalog.grpc" \ <http://istio-ingress.domain.com:82|istio-ingress.domain.com:82> \ datacatalog.DataCatalog/ListDatasets``` ```{ "datasets": [ { "id": { ....```
sure, we don’t use minio as part of deployment! about datacatalog I’m not sure … let me check hmm honestly it works fine ```grpcurl \ -authority="flyte.datacatalog.grpc" \ <http://istio-ingress.domain.com:82|istio-ingress.domain.com:82> \ datacatalog.DataCatalog/ListDatasets``` ```{ "datasets": [ { "id": { ....```
would you mind filing an issue in that case on us? in the flyte repo. and we’ll look more into it when we have a bit more time.
would you mind filing an issue in that case on us? in the flyte repo. and we’ll look more into it when we have a bit more time.
I can take a look as well on the GH issue. To confirm you haven’t changed your config map between v0.1.1 and v0.2? And also Admin is up and running fine too? Datacatalog accesses s3 the same way Admin does
I can take a look as well on the GH issue. To confirm you haven’t changed your config map between v0.1.1 and v0.2? And also Admin is up and running fine too? Datacatalog accesses s3 the same way Admin does
yes, no changes for configmap and sure, all components work fine. Usual, scheduled workflows and notifications are running.
yes, no changes for configmap and sure, all components work fine. Usual, scheduled workflows and notifications are running.
And you’re observing that aside from that one `warn` DataCatalog is up and running and works fine?
And you’re observing that aside from that one `warn` DataCatalog is up and running and works fine?
but just got some feedback that team couldn’t use datacatalog when we updated… so maybe will discuss about their compatibility with new datacatalog (if it makes sense) due to time gap communication is slower :joy: Yes, for us I see no issues …
but just got some feedback that team couldn’t use datacatalog when we updated… so maybe will discuss about their compatibility with new datacatalog (if it makes sense) due to time gap communication is slower :joy: Yes, for us I see no issues …
uh, can you elaborate on “team couldn’t use datacatalog when we updated”? why? that sounds bad
uh, can you elaborate on “team couldn’t use datacatalog when we updated”? why? that sounds bad
&gt; but just got some feedback that team couldn’t use datacatalog when we updated… Do you mean trying to use it outside of Flyte? ie., directly communicating with it and not via flyteplugins?
&gt; but just got some feedback that team couldn’t use datacatalog when we updated… Do you mean trying to use it outside of Flyte? ie., directly communicating with it and not via flyteplugins?
I am waiting for details about this we exposed Datacatalog via aws VPC endpoint for another team purposes and this warn log was doubt for me that datacatalog is configured another way Thank you for your help :pray:
I am waiting for details about this we exposed Datacatalog via aws VPC endpoint for another team purposes and this warn log was doubt for me that datacatalog is configured another way Thank you for your help :pray:
Ruslan Stanevich can we get on a hangouts call sometime? I would like to connect and discuss how you are using Flyte etc and see how we could improve our working relationship
Ruslan Stanevich can we get on a hangouts call sometime? I would like to connect and discuss how you are using Flyte etc and see how we could improve our working relationship
I think this is a great idea! And I would like to inform other teams that use Flyte. And we will try to collect questions/interested people and arrange a call in your morning (our evening) So, I think we can arrange a meeting next week. let you know more precisely after the weekend thank you Ketan Umare:pray:
I think this is a great idea! And I would like to inform other teams that use Flyte. And we will try to collect questions/interested people and arrange a call in your morning (our evening) So, I think we can arrange a meeting next week. let you know more precisely after the weekend thank you Ketan Umare:pray:
Ruslan Stanevich absolutely, looking forward to it
Ruslan Stanevich absolutely, looking forward to it
Ketan Umare , do you have a chance to call in morning, ~9..10 am in your timezone? Will Thursday suit you?
Ketan Umare , do you have a chance to call in morning, ~9..10 am in your timezone? Will Thursday suit you?
ya morning works, let me check on Thursday I can do next wednesday? or monday i can also do 10:00 am tomorrow
ya morning works, let me check on Thursday I can do next wednesday? or monday i can also do 10:00 am tomorrow
Hi! on Wednesday at 10am (our 8pm) sounds good tomorrow so can we create hangouts meet at 10am and sent invitation to you?
Hi! on Wednesday at 10am (our 8pm) sounds good tomorrow so can we create hangouts meet at 10am and sent invitation to you?
add me too?
add me too?
great! please advise if we can invite someone else :pray: Ketan Umare Yee Hello! we are in hangouts room
great! please advise if we can invite someone else :pray: Ketan Umare Yee Hello! we are in hangouts room
logging in now
logging in now
Let me know if you need me in the meeting, if there is datacatalog specific topics