input
stringlengths 1
18.7k
| output
stringlengths 1
18.7k
|
---|---|
thanks!
| no problem
|
jburns good job with the article. if you want i can tweet it
| Absolutely. Tyty!
|
Absolutely. Tyty!
| Done
|
Are Flyte workflows appropriate for long-running tasks? Example: if you have security camera video streams and want to sample X frames a second, do object detection, then log results.
• Should each frame be an execution in this case? You have something outside of Flyte sampling frames and creating an execution for each?
• Or should it be a long-running execution per camera with the input being the video stream? (first task would decode/sample frames)
| Hey Zach :wave:
While Flyte supports fairly long running tasks, (we have some Flyte tasks run for multiple days), the system does assume the amount of work is finite (it expects the task to finish at some point).
While there isn't anything stopping you from creating one execution per-frame, I would lean toward submitting frames in batches, since there is some overhead involved with each execution.
You might have a cron that runs on your machine once per minute, and submits a batch of frames to Flyte.
Your Flyte workflow can output a list of objects found in each frame.
|
Hey Zach :wave:
While Flyte supports fairly long running tasks, (we have some Flyte tasks run for multiple days), the system does assume the amount of work is finite (it expects the task to finish at some point).
While there isn't anything stopping you from creating one execution per-frame, I would lean toward submitting frames in batches, since there is some overhead involved with each execution.
You might have a cron that runs on your machine once per minute, and submits a batch of frames to Flyte.
Your Flyte workflow can output a list of objects found in each frame.
| Hi Zach Hobbs welcome to Flyte. Johnny has explained a solution. Let me provide some background. So the workflows execute tasks. Tasks are execute for some finite time. The duration can be very small or very large, but there are caveats and overheads. For example very short duration like milliseconds is not recommended as to launch a container it takes longer (even if cached). But this is very interesting to us an we would love to come up with a solution
|
Hi Zach Hobbs welcome to Flyte. Johnny has explained a solution. Let me provide some background. So the workflows execute tasks. Tasks are execute for some finite time. The duration can be very small or very large, but there are caveats and overheads. For example very short duration like milliseconds is not recommended as to launch a container it takes longer (even if cached). But this is very interesting to us an we would love to come up with a solution
| Thanks for the info Ketan and Johnny! As I think about it more, since a given task can only emit output once during an execution it makes sense that I can't treat it as a pipeline to continuously stream data into.
I'll keep digging into the architecture to see if it would make sense to extend Flyte to support this. Looking for a solution to do async inference on content streams (video stream, new photos/video/audio in S3, etc).
BTW, do you guys use Flyte for inference much?
|
Thanks for the info Ketan and Johnny! As I think about it more, since a given task can only emit output once during an execution it makes sense that I can't treat it as a pipeline to continuously stream data into.
I'll keep digging into the architecture to see if it would make sense to extend Flyte to support this. Looking for a solution to do async inference on content streams (video stream, new photos/video/audio in S3, etc).
BTW, do you guys use Flyte for inference much?
| Hey Zach Hobbs, this is definitely an interesting use-case. If it's desired to process each frame as they come, it seems more appropriate to consider a streaming solution (Flink maybe?) <https://github.com/lyft/flinkk8soperator>
We are looking for ways to integrate Flyte with Flink, as you learn more about the architecture, please keep that in mind.
And as always, contributions are most welcomed!
|
Hello, could someone direct me to the right channel to discuss setting up Flyte locally with Minikube on MacOS? Thanks!
| i think <#CP2HDHKE1|onboarding> could be a good place to chat. Johnny Burns and Yee are pretty experienced with this and might be able to help
|
i think <#CP2HDHKE1|onboarding> could be a good place to chat. Johnny Burns and Yee are pretty experienced with this and might be able to help
| <#CP2HDHKE1|onboarding>
|
<#CP2HDHKE1|onboarding>
| jinx
|
jinx
| oh you beat me to it :slightly_smiling_face:
|
oh you beat me to it :slightly_smiling_face:
| thanks :+1:
|
Hello
| Hello Sergio! thank you for joining
|
Hello Sergio! thank you for joining
| Hi Sergio
|
Hello everyone, I just started playing with Flyte and I'm very excited, it looks like a very promising project
| glad to hear it! welcome and please don’t hesitate to ask any questions!!
|
glad to hear it! welcome and please don’t hesitate to ask any questions!!
| Thank you Matt, as a matter of fact I do have a few…what would be the best channel to discuss them?
|
Thank you Matt, as a matter of fact I do have a few…what would be the best channel to discuss them?
| I think you can ask here and if it’s better to be in another channel, we can help re-route it for you Though i’m about to head off to bed so someone will have to pick it up tomorrow morning.
|
ok I got a couple of questions:
1)
I followed the examples in the flytesnacks repository and created a simple workflow with a few tasks, now I am trying to schedule it to run every 5 minutes by using this code that I found in the docs (<https://lyft.github.io/flyte/user/features/lanuchplans.html>) which I added after the workflow definition
```my_fixed_rate_launch_plan = myworkflow.create_launch_plan(
default_inputs={'string_in': Input(Types.String, default="aaaa")},
schedule=schedules.FixedRate(datetime.timedelta(minutes=5)),
)```
when I rebuild the container and re-register the workflow i see that the launchplan gets registered
```Flyte Admin URL 127.0.0.1:30081
Running task, workflow, and launch plan registration for flytedemo, development, ['fk_tasks'] with version 62ffc8b36883d18f4cf424f8d09510a5dd3db46d
Registering Task: fk_tasks.workflows.download_dataset
Registering Task: fk_tasks.workflows.read_pickle
Registering Workflow: fk_tasks.workflows.myworkflow
Registering Launch Plan: fk_tasks.workflows.my_fixed_rate_launch_plan
Registering Launch Plan: fk_tasks.workflows.myworkflow
Registering Task: fk_tasks.workflows.reverse_task
Registering Task: fk_tasks.workflows.uppercase_task```
unfortunately in the gui I get a “This workflow has no schedules.”
What am I missing?
| Hi Giordano sorry for the late reply. So in the local sandbox cluster mode we do not have a scheduler today. When you deploy to a cloud environment we use either AWS cloud watch schedules or on GCP cloud scheduler. This is an area of active interest and we prefer to write a simple one using k8s operator or better option work with our partner Spotify to use Styx scheduler to run the cron schedules
So to understand, we just don’t have the cron runner, the runner just triggers Flyte control plane and we do the rest. - let me know if you would love to contribute or have any other questions
Let me know if you need help with anything else
|
Hi Giordano sorry for the late reply. So in the local sandbox cluster mode we do not have a scheduler today. When you deploy to a cloud environment we use either AWS cloud watch schedules or on GCP cloud scheduler. This is an area of active interest and we prefer to write a simple one using k8s operator or better option work with our partner Spotify to use Styx scheduler to run the cron schedules
So to understand, we just don’t have the cron runner, the runner just triggers Flyte control plane and we do the rest. - let me know if you would love to contribute or have any other questions
Let me know if you need help with anything else
| Hi Ketan Umare thanks for the info, i was looking at flyte to replace airflow running on a barematal k8s cluster so I guess I’ll have to manage without the scheduling part
|
Hi Ketan Umare thanks for the info, i was looking at flyte to replace airflow running on a barematal k8s cluster so I guess I’ll have to manage without the scheduling part
| Giordano we would love to work with you on a replacement for the cloud scheduler
Try out the other features and e See is you like everything else
The reason for not packaging a scheduler is reinvention of wheel And building a scheduler for our scale will take sometime 100ks execution per day
|
2) are there any examples on how to connect to flyteadmin and launch/manage and query workflow results via the flytesdk?
| There is documentation for this, will share a link. It’s as simple as importing the workflow and there are bunch of API’s on them
Also you can use Flyte-cli
|
There is documentation for this, will share a link. It’s as simple as importing the workflow and there are bunch of API’s on them
Also you can use Flyte-cli
| great thanks looking forward to the link.
|
great thanks looking forward to the link.
| <https://lyft.github.io/flyte/user/features/flytecli.html|https://lyft.github.io/flyte/user/features/flytecli.html> Flyte cli
<https://lyft.github.io/flyte/flytekit/flytekit.common.html#module-flytekit.common.workflow_execution|https://lyft.github.io/flyte/flytekit/flytekit.common.html#module-flytekit.common.workflow_execution>
We don’t have better docs on the interactive part yet
|
ok one more question, I was trying to create a workflow that would call some tasks aready defined in another workflow using an example that I found on the kubecon 19 presentation
```from flytekit.models.launch_plan import LaunchPlanState
from flytekit.common import utils, schedules, tasks
from flytekit.sdk.tasks import python_task, outputs, inputs
from flytekit.sdk.types import Types
from flytekit.sdk.workflow import workflow_class, Output, Input
from flytekit.common.tasks.task import SdkTask
@workflow_class
class myworkflow(object):
string_in = Input(Types.String, required=True, help="input string")
dataset = Input(Types.CSV, default=Types.CSV.create_at_known_location(
"<http://172.16.140.171:8000/label_summary.csv>"),
help="A CSV File")
return_dataset = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset","470cae526836f7a05d41ef81faa07a3b275b9de9")(dataset=dataset)
return_pickle = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","470cae526836f7a05d41ef81faa07a3b275b9de9")(dataset=return_dataset.outputs.out)
myoutput = Output(return_pickle.outputs.csv_head, sdk_type=Types.String)```
unfortunately when I register I get an error
```flytekit.common.exceptions.user.FlyteAssertion: An entity was not found in modules accessible from the workflow packages configuration. Please ensure that entities in 'fk_workflow.workflow' are moved to a configured packaged, or adjust the configuration.```
Does my code look ok? am I missing something?
What I am trying to achieve is to de-couple the tasks from the workflow so that I can dynamically create a workflow from pre-existing tasks. my use case would be some kind of application where a user can choose different pre-built tasks, build a workflow definition through some kind of gui and then run a script that would dynamically create the workflow with those tasks. I understand that the tasks need to be created inside a container, but from what I understand a workflow does not need to be related to a specific container build, is my assumption correct?
| is your code pushed anywhere we can take a look?
|
is your code pushed anywhere we can take a look?
| No but i can create a quick repo
|
No but i can create a quick repo
| sure, if you don’t mind.
always helpful
|
sure, if you don’t mind.
always helpful
| hmm yeh, i’m not seeing where `fk_workflow.workflow` is referenced
(the code you copied originally looks great!)
|
hmm yeh, i’m not seeing where `fk_workflow.workflow` is referenced
(the code you copied originally looks great!)
| hi Matt thanks, i just pushed to the repo my example
<https://github.com/giordyb/flyte_demo>
the workflow inside the fk_tasks folder works just fine (it’s 4 tasks and 1 workflow)
in the workflow inside the fk_workflow folder I’m trying fetch the tasks created with the previous workflow but doesn’t work
|
hi Matt thanks, i just pushed to the repo my example
<https://github.com/giordyb/flyte_demo>
the workflow inside the fk_tasks folder works just fine (it’s 4 tasks and 1 workflow)
in the workflow inside the fk_workflow folder I’m trying fetch the tasks created with the previous workflow but doesn’t work
| what happens if you run `flyte-cli -h <your-url> list-task-names -p flytedemo -d development`
|
what happens if you run `flyte-cli -h <your-url> list-task-names -p flytedemo -d development`
| Welcome to Flyte CLI! Version: 0.4.4
Task Names Found in flytedemo:development
fk_tasks.tasks_and_workflow.download_dataset
fk_tasks.tasks_and_workflow.read_pickle
fk_tasks.tasks_and_workflow.reverse_task
fk_tasks.tasks_and_workflow.uppercase_task
btw I also tried fetch_latest without the version number but i get a
```AttributeError: type object 'SdkTask' has no attribute 'fetch_latest'```
|
Welcome to Flyte CLI! Version: 0.4.4
Task Names Found in flytedemo:development
fk_tasks.tasks_and_workflow.download_dataset
fk_tasks.tasks_and_workflow.read_pickle
fk_tasks.tasks_and_workflow.reverse_task
fk_tasks.tasks_and_workflow.uppercase_task
btw I also tried fetch_latest without the version number but i get a
```AttributeError: type object 'SdkTask' has no attribute 'fetch_latest'```
| cool and also: `flyte-cli -h <your-url> list-task-versions -p flytedemo -d development -n fk_tasks.tasks_and_workflow.download_dataset`
|
cool and also: `flyte-cli -h <your-url> list-task-versions -p flytedemo -d development -n fk_tasks.tasks_and_workflow.download_dataset`
| Welcome to Flyte CLI! Version: 0.4.4
Task Versions Found for flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset
Version Urn
1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216 tsk:flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset:1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216
|
Welcome to Flyte CLI! Version: 0.4.4
Task Versions Found for flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset
Version Urn
1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216 tsk:flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset:1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216
| ok cool, the SHA `470cae526836f7a05d41ef81faa07a3b275b9de9` doesn’t look like it was ever registered
if you swapped that with `1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216`, it should work
|
ok cool, the SHA `470cae526836f7a05d41ef81faa07a3b275b9de9` doesn’t look like it was ever registered
if you swapped that with `1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216`, it should work
| sorry i had rebuilt the deployment in the meantime
just tried with that version and it still throws
```flytekit.common.exceptions.user.FlyteAssertion: An entity was not found in modules accessible from the workflow packages configuration. Please ensure that entities in 'fk_workflow.workflow' are moved to a configured packaged, or adjust the configuration.```
|
sorry i had rebuilt the deployment in the meantime
just tried with that version and it still throws
```flytekit.common.exceptions.user.FlyteAssertion: An entity was not found in modules accessible from the workflow packages configuration. Please ensure that entities in 'fk_workflow.workflow' are moved to a configured packaged, or adjust the configuration.```
| ```read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
download_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
@workflow_class
class myworkflow(object):
string_in = Input(Types.String, required=True, help="input string")
dataset = Input(Types.CSV, default=Types.CSV.create_at_known_location(
"<http://172.16.140.171:8000/label_summary.csv>"),
help="A CSV File")
return_dataset = download_task(dataset=dataset)
return_pickle = read_task(dataset=return_dataset.outputs.out)
myoutput = Output(return_pickle.outputs.csv_head, sdk_type=Types.String)```
hmm that might be a bug in the registration tool.
…sorry still typing
can you try above? i think there is a bug which incorrectly triggers a sanity-check mechanism
anyway, this is actually easier in a ‘pure client’ case. The case where you have a service that is dynamically generating workflows, you can use the client to find task versions, weave them into a workflow, and then simply call register in a pure script like:
```t = SdkTask.fetch(...)
wf = workflow(nodes={'n1': t()}, inputs=...)
lp = wf.create_launch_plan()
wf.register(..)
lp.register(..)```
|
```read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
download_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
@workflow_class
class myworkflow(object):
string_in = Input(Types.String, required=True, help="input string")
dataset = Input(Types.CSV, default=Types.CSV.create_at_known_location(
"<http://172.16.140.171:8000/label_summary.csv>"),
help="A CSV File")
return_dataset = download_task(dataset=dataset)
return_pickle = read_task(dataset=return_dataset.outputs.out)
myoutput = Output(return_pickle.outputs.csv_head, sdk_type=Types.String)```
hmm that might be a bug in the registration tool.
…sorry still typing
can you try above? i think there is a bug which incorrectly triggers a sanity-check mechanism
anyway, this is actually easier in a ‘pure client’ case. The case where you have a service that is dynamically generating workflows, you can use the client to find task versions, weave them into a workflow, and then simply call register in a pure script like:
```t = SdkTask.fetch(...)
wf = workflow(nodes={'n1': t()}, inputs=...)
lp = wf.create_launch_plan()
wf.register(..)
lp.register(..)```
| ```Running task, workflow, and launch plan registration for flytedemo, development, ['fk_workflow'] with version 1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216
Traceback (most recent call last):
File "/app/venv/bin/pyflyte", line 11, in <module>
sys.exit(main())
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__
return self.main(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main
rv = self.invoke(ctx)
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 97, in workflows
register_all(project, domain, pkgs, test, version)
File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 21, in register_all
for m, k, o in iterate_registerable_entities_in_order(pkgs):
File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 112, in iterate_registerable_entities_in_order
for m in iterate_modules(pkgs):
File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 16, in iterate_modules
yield importlib.import_module(name)
File "/usr/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 994, in _gcd_import
File "<frozen importlib._bootstrap>", line 971, in _find_and_load
File "<frozen importlib._bootstrap>", line 955, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 665, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/app/fk_workflow/workflow.py", line 9, in <module>
read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point
return wrapped(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/flytekit/common/tasks/task.py", line 159, in fetch
admin_task = _engine_loader.get_engine().fetch_task(task_id=task_id)
File "/app/venv/lib/python3.6/site-packages/flytekit/engines/flyte/engine.py", line 108, in fetch_task
).client.get_task(task_id)
File "/app/venv/lib/python3.6/site-packages/flytekit/clients/friendly.py", line 162, in get_task
id=id.to_flyte_idl()
File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 12, in handler
return fn(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 136, in get_task
return self._stub.GetTask(get_object_request)
File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 824, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 726, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.NOT_FOUND
details = "entry not found"
debug_error_string = "{"created":"@1579804960.139722800","description":"Error received from peer ipv4:127.0.0.1:30081","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"entry not found","grpc_status":5}"
>
make: *** [register_workflow] Error 1```
nope still throws an error
the pure client solution would be even better i think
there is something that I haven’t quite understood yet, maybe you could shed some light
I understand that in order to register a task i need to do it from a container (or at least set an env variable that points to the container that will run it)
do I still need to reference a container when creating a workflow?
|
```Running task, workflow, and launch plan registration for flytedemo, development, ['fk_workflow'] with version 1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216
Traceback (most recent call last):
File "/app/venv/bin/pyflyte", line 11, in <module>
sys.exit(main())
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__
return self.main(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main
rv = self.invoke(ctx)
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke
return callback(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 97, in workflows
register_all(project, domain, pkgs, test, version)
File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 21, in register_all
for m, k, o in iterate_registerable_entities_in_order(pkgs):
File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 112, in iterate_registerable_entities_in_order
for m in iterate_modules(pkgs):
File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 16, in iterate_modules
yield importlib.import_module(name)
File "/usr/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 994, in _gcd_import
File "<frozen importlib._bootstrap>", line 971, in _find_and_load
File "<frozen importlib._bootstrap>", line 955, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 665, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/app/fk_workflow/workflow.py", line 9, in <module>
read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point
return wrapped(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/flytekit/common/tasks/task.py", line 159, in fetch
admin_task = _engine_loader.get_engine().fetch_task(task_id=task_id)
File "/app/venv/lib/python3.6/site-packages/flytekit/engines/flyte/engine.py", line 108, in fetch_task
).client.get_task(task_id)
File "/app/venv/lib/python3.6/site-packages/flytekit/clients/friendly.py", line 162, in get_task
id=id.to_flyte_idl()
File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 12, in handler
return fn(*args, **kwargs)
File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 136, in get_task
return self._stub.GetTask(get_object_request)
File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 824, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 726, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.NOT_FOUND
details = "entry not found"
debug_error_string = "{"created":"@1579804960.139722800","description":"Error received from peer ipv4:127.0.0.1:30081","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"entry not found","grpc_status":5}"
>
make: *** [register_workflow] Error 1```
nope still throws an error
the pure client solution would be even better i think
there is something that I haven’t quite understood yet, maybe you could shed some light
I understand that in order to register a task i need to do it from a container (or at least set an env variable that points to the container that will run it)
do I still need to reference a container when creating a workflow?
| no, definitely not for a workflow…and you don’t actually technically need to be in a container for a task either. Being in a container that is configured in a certain way just makes it easy to auto-fill in information when creating tasks.
|
no, definitely not for a workflow…and you don’t actually technically need to be in a container for a task either. Being in a container that is configured in a certain way just makes it easy to auto-fill in information when creating tasks.
| right that’s what I figured
if you could provide some simple example of a workflow created with the pure client I would appreciate it
another thing that’s not very clear from the flytekit documentation is how to connect to the admin service
i know that with flyte-cli I can just pass the host parameters but I couldn’t find the equivalent for the flytekit
i think it’s looking for some kind of configuration file because I get an error similar to this
```FlyteAssertion: No configuration set for [platform] url. This is a required configuration.```
|
right that’s what I figured
if you could provide some simple example of a workflow created with the pure client I would appreciate it
another thing that’s not very clear from the flytekit documentation is how to connect to the admin service
i know that with flyte-cli I can just pass the host parameters but I couldn’t find the equivalent for the flytekit
i think it’s looking for some kind of configuration file because I get an error similar to this
```FlyteAssertion: No configuration set for [platform] url. This is a required configuration.```
| do you get this error when using the pyflyte tool? or just when writing a script yourself?
if the latter, take a look at this method: <https://github.com/lyft/flytekit/blob/master/flytekit/configuration/__init__.py#L11>
|
do you get this error when using the pyflyte tool? or just when writing a script yourself?
if the latter, take a look at this method: <https://github.com/lyft/flytekit/blob/master/flytekit/configuration/__init__.py#L11>
| when creating a script myself
|
when creating a script myself
| alternatively, you can set env vars that follow the format: `FLYTE_{SECTION}_{KEY}` . so in this case FLYTE_PLATFORM_URL, but i recommend creating a `flytekit.conf` file
and i will find an example of a pure client use case
|
alternatively, you can set env vars that follow the format: `FLYTE_{SECTION}_{KEY}` . so in this case FLYTE_PLATFORM_URL, but i recommend creating a `flytekit.conf` file
and i will find an example of a pure client use case
| is the flytekit.config supposed to be created when running flyte-cli setup-config?
|
is the flytekit.config supposed to be created when running flyte-cli setup-config?
| it is the same config file format, so it can be.
|
it is the same config file format, so it can be.
| because I tried running
```flyte-cli setup-config -h localhost:30081 -i```
but I got a jsondecode error…should I open an issue about that?
|
because I tried running
```flyte-cli setup-config -h localhost:30081 -i```
but I got a jsondecode error…should I open an issue about that?
| yes please. cc: Yee
|
yes please. cc: Yee
| :+1:
|
:+1:
| there should be a barebones config in the demo repository you can copy
|
there should be a barebones config in the demo repository you can copy
| ok thanks I’ll take a look
|
ok thanks I’ll take a look
| here is an example script for a pure client approach. note: i copy-pasted different elements so there _might_ be a copy-paste error (and the interface definitions for your tasks/workflows will be different):
```my_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset", "1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
input_dict = {
'input_1': Input(Types.Integer, default=10, help='Not required input.'),
'input_2': Input(Types.Integer, help='Required.')
}
nodes = {
'a': my_task(a=input_dict['input_1']),
'b': my_task(a=input_dict['input_2']),
'c': my_task(a=1000)
}
outputs = {
'wf_out': Output(
[
nodes['a'].outputs.b,
nodes['b'].outputs.b,
nodes['c'].outputs.b,
],
sdk_type=[Types.Integer]
)
}
w = workflow(inputs=input_dict, outputs=outputs, nodes=nodes)
w.register('flytedemo', 'development', 'simple_functional', version)))
lp = w.create_launch_plan(
fixed_inputs={'input_2': 100},
schedule=schedules.CronSchedule("0/15 * * * ? *")
lp.register('flytedemo', 'development', 'simple_functional', version)))
ex = lp.execute('flytedemo', 'development', inputs={'input_1': 500})
print(("Execution URN: {}".format(ex.id)))
# wait for execution to complete, then check out the status, inputs, and outputs.
ex.wait_for_completion(timeout=datetime.timedelta(minutes=10))
assert ex.closure.phase == _core_execution.WorkflowExecutionPhase.SUCCEEDED
assert ex.error is None
assert len(ex.inputs) == 2
assert len(ex.outputs) == 1
assert ex.inputs['input_1'] == 500
assert ex.inputs['input_2'] == 100
assert ex.outputs['wf_out'] == [501, 101, 1001]
for k, ne in six.iteritems(ex.node_executions):
if k in {'start-node', 'end-node'}:
continue
ne.sync()
assert ne.closure.phase == _core_execution.NodeExecutionPhase.SUCCEEDED
assert len(ne.inputs) == 1
assert len(ne.outputs) == 1
assert len(ne.executions) == 1
assert len(ne.task_executions) == 1
assert ne.error is None
ne.executions[0].sync()
assert ne.executions[0].error is None
assert len(ne.executions[0].inputs) == 1
assert len(ne.executions[0].outputs) == 1
assert ne.task_executions[0].closure.phase == _core_execution.TaskExecutionPhase.SUCCEEDED
assert len(ne.task_executions[0].inputs) == 1
assert len(ne.task_executions[0].outputs) == 1
assert ne.task_executions[0].error is None
assert ex.node_executions['a'].inputs['a'] == 500
assert ex.node_executions['a'].outputs['b'] == 501
assert ex.node_executions['b'].inputs['a'] == 100
assert ex.node_executions['b'].outputs['b'] == 101
assert ex.node_executions['c'].inputs['a'] == 1000
assert ex.node_executions['c'].outputs['b'] == 1001```
important imports:
```from flytekit.sdk.types import Types
from flytekit.sdk.workflow import workflow, Input, Output
from flytekit.common import schedules
from flytekit.configuration import TemporaryConfiguration
from flytekit.models import launch_plan as _launch_plan
from flytekit.models.core import execution as _core_execution```
and you’ll want to set the config somewhere beforehand too
|
here is an example script for a pure client approach. note: i copy-pasted different elements so there _might_ be a copy-paste error (and the interface definitions for your tasks/workflows will be different):
```my_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset", "1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216")
input_dict = {
'input_1': Input(Types.Integer, default=10, help='Not required input.'),
'input_2': Input(Types.Integer, help='Required.')
}
nodes = {
'a': my_task(a=input_dict['input_1']),
'b': my_task(a=input_dict['input_2']),
'c': my_task(a=1000)
}
outputs = {
'wf_out': Output(
[
nodes['a'].outputs.b,
nodes['b'].outputs.b,
nodes['c'].outputs.b,
],
sdk_type=[Types.Integer]
)
}
w = workflow(inputs=input_dict, outputs=outputs, nodes=nodes)
w.register('flytedemo', 'development', 'simple_functional', version)))
lp = w.create_launch_plan(
fixed_inputs={'input_2': 100},
schedule=schedules.CronSchedule("0/15 * * * ? *")
lp.register('flytedemo', 'development', 'simple_functional', version)))
ex = lp.execute('flytedemo', 'development', inputs={'input_1': 500})
print(("Execution URN: {}".format(ex.id)))
# wait for execution to complete, then check out the status, inputs, and outputs.
ex.wait_for_completion(timeout=datetime.timedelta(minutes=10))
assert ex.closure.phase == _core_execution.WorkflowExecutionPhase.SUCCEEDED
assert ex.error is None
assert len(ex.inputs) == 2
assert len(ex.outputs) == 1
assert ex.inputs['input_1'] == 500
assert ex.inputs['input_2'] == 100
assert ex.outputs['wf_out'] == [501, 101, 1001]
for k, ne in six.iteritems(ex.node_executions):
if k in {'start-node', 'end-node'}:
continue
ne.sync()
assert ne.closure.phase == _core_execution.NodeExecutionPhase.SUCCEEDED
assert len(ne.inputs) == 1
assert len(ne.outputs) == 1
assert len(ne.executions) == 1
assert len(ne.task_executions) == 1
assert ne.error is None
ne.executions[0].sync()
assert ne.executions[0].error is None
assert len(ne.executions[0].inputs) == 1
assert len(ne.executions[0].outputs) == 1
assert ne.task_executions[0].closure.phase == _core_execution.TaskExecutionPhase.SUCCEEDED
assert len(ne.task_executions[0].inputs) == 1
assert len(ne.task_executions[0].outputs) == 1
assert ne.task_executions[0].error is None
assert ex.node_executions['a'].inputs['a'] == 500
assert ex.node_executions['a'].outputs['b'] == 501
assert ex.node_executions['b'].inputs['a'] == 100
assert ex.node_executions['b'].outputs['b'] == 101
assert ex.node_executions['c'].inputs['a'] == 1000
assert ex.node_executions['c'].outputs['b'] == 1001```
important imports:
```from flytekit.sdk.types import Types
from flytekit.sdk.workflow import workflow, Input, Output
from flytekit.common import schedules
from flytekit.configuration import TemporaryConfiguration
from flytekit.models import launch_plan as _launch_plan
from flytekit.models.core import execution as _core_execution```
and you’ll want to set the config somewhere beforehand too
| hi matt I got it to work after all
one of the issues was that I renamed the task names but didn’t rename them in the code
the other issue is that the docker image has the flytekit pinned to an older version in the flytesnacks example
that’s why I got the error with fetch_latest
thank you for your help!
|
hi matt I got it to work after all
one of the issues was that I renamed the task names but didn’t rename them in the code
the other issue is that the docker image has the flytekit pinned to an older version in the flytesnacks example
that’s why I got the error with fetch_latest
thank you for your help!
| awesome! glad to hear it!!
|
Hongxin Liang and Haytham Abuelfutuh
I’ve merged the boilerplate change
and updated the idl pr <https://github.com/lyft/flyteidl/pull/27>
thanks honnix for adding me as a collab
mind taking a look one last time before we merge?
| Awesome!!! thank you both
|
Awesome!!! thank you both
| I will take care of the rest repos.
Thanks.
|
I will take care of the rest repos.
Thanks.
| good morning
thank you!
i’ll approve everything tomorrow morning.
will then test, and then post an annoucement to the issue and this channel
|
good morning
thank you!
i’ll approve everything tomorrow morning.
will then test, and then post an annoucement to the issue and this channel
| I have updated all the rest repos accordingly and poked you in those PRs. Please take a look. Thanks.
|
Hello everyone… are there any known issues with running a script that uses the “hyperopt” python library under flyte?
| None that we know off
is not playing well with flytekit
|
None that we know off
is not playing well with flytekit
| I have been trying to run a script that uses it under a workflow
|
I have been trying to run a script that uses it under a workflow
| hmmm and
but you are using the python bindings (flytekit) right?
|
hmmm and
but you are using the python bindings (flytekit) right?
| i don’t even have to actually use the library, as soon as I import it in the script i get this
|
i don’t even have to actually use the library, as soon as I import it in the script i get this
| are you pasting the error?
|
are you pasting the error?
| ```ERROR:root:Error from command '['aws', '--endpoint-url', '<http://minio.yolotrainframework.svc.cluster.local:9000>', 's3', 'cp', '<s3://my-s3-bucket/metadata/propeller/flytedemo-d> │
│ b'' │
│ Traceback (most recent call last): │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/data_proxy.py", line 127, in get_data │
│ proxy.download(remote_path, local_path) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 109, in download │
│ return _update_cmd_config_and_execute(cmd) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 36, in _update_cmd_config_and_execute │
│ return _subprocess.check_call(cmd, env=env) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/tools/subprocess.py", line 34, in check_call │
│ "Called process exited with error code: {}. Stderr dump:\n\n{}".format(ret_code, err_str) │
│ Exception: Called process exited with error code: -9. Stderr dump: │
│ b'' │
│ During handling of the above exception, another exception occurred: │
│ Traceback (most recent call last): │
│ File "/app/venv/bin/pyflyte-execute", line 11, in <module> │
│ sys.exit(execute_task_cmd()) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__ │
│ return self.main(*args, **kwargs) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main │
│ rv = self.invoke(ctx) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke │
│ return ctx.invoke(self.callback, **ctx.params) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke │
│ return callback(*args, **kwargs) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 104, in execute_task_cmd │
│ _execute_task(task_module, task_name, inputs, output_prefix, test) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point │
│ return wrapped(*args, **kwargs) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 83, in _execute_task │
│ _data_proxy.Data.get_data(inputs, local_inputs_file)```
this is from inside the container
if I take out the hyperopt import everything works
|
```ERROR:root:Error from command '['aws', '--endpoint-url', '<http://minio.yolotrainframework.svc.cluster.local:9000>', 's3', 'cp', '<s3://my-s3-bucket/metadata/propeller/flytedemo-d> │
│ b'' │
│ Traceback (most recent call last): │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/data_proxy.py", line 127, in get_data │
│ proxy.download(remote_path, local_path) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 109, in download │
│ return _update_cmd_config_and_execute(cmd) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 36, in _update_cmd_config_and_execute │
│ return _subprocess.check_call(cmd, env=env) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/tools/subprocess.py", line 34, in check_call │
│ "Called process exited with error code: {}. Stderr dump:\n\n{}".format(ret_code, err_str) │
│ Exception: Called process exited with error code: -9. Stderr dump: │
│ b'' │
│ During handling of the above exception, another exception occurred: │
│ Traceback (most recent call last): │
│ File "/app/venv/bin/pyflyte-execute", line 11, in <module> │
│ sys.exit(execute_task_cmd()) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__ │
│ return self.main(*args, **kwargs) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main │
│ rv = self.invoke(ctx) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke │
│ return ctx.invoke(self.callback, **ctx.params) │
│ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke │
│ return callback(*args, **kwargs) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 104, in execute_task_cmd │
│ _execute_task(task_module, task_name, inputs, output_prefix, test) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point │
│ return wrapped(*args, **kwargs) │
│ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 83, in _execute_task │
│ _data_proxy.Data.get_data(inputs, local_inputs_file)```
this is from inside the container
if I take out the hyperopt import everything works
| hmmm
Exit code -9
let me check
Giordano can you share your code
if you dont mind?
|
hmmm
Exit code -9
let me check
Giordano can you share your code
if you dont mind?
| it seems like importing hyperopt createss some issues with connecting to s3
sure
gimme a sec
|
it seems like importing hyperopt createss some issues with connecting to s3
sure
gimme a sec
| ya this is actually connecting to minio in your sandbox cluster
|
ya this is actually connecting to minio in your sandbox cluster
| Looks like the subprocess which was trying to copy from s3 got a SIGKILL signal.
The most common reason for that would be OOM. `aws s3 cp` shouldn't take too much memory, but it might be worth trying to bump the memory in your container using something like:
```@python_task(cpu_limit="10000m", memory_limit="10000Mi")```
|
Looks like the subprocess which was trying to copy from s3 got a SIGKILL signal.
The most common reason for that would be OOM. `aws s3 cp` shouldn't take too much memory, but it might be worth trying to bump the memory in your container using something like:
```@python_task(cpu_limit="10000m", memory_limit="10000Mi")```
| Johnny Burns Ketan Umare just tried what you suggested and it's working now, thanks!
It was indeed a resource issue
|
Johnny Burns Ketan Umare just tried what you suggested and it's working now, thanks!
It was indeed a resource issue
| Woot woot :tada:
|
As of today, we have migrated all golang Flyte repos to use go mod. Thanks to Hongxin Liang for driving this change.
Boilerplate changes: <https://github.com/lyft/boilerplate/pull/4>
The boilerplate code will now default to go.mod.
<https://github.com/lyft/flytestdlib/releases/tag/v0.3.0> (was already on go mod as of v0.2.29)
<https://github.com/lyft/flyteidl/releases/tag/v0.17.0>
<https://github.com/lyft/flyteplugins/releases/tag/v0.3.0>
<https://github.com/lyft/flytepropeller/releases/tag/v0.2.0>
<https://github.com/lyft/flyteadmin/releases/tag/v0.2.0>
<https://github.com/lyft/datacatalog/releases/tag/0.2.0>
We’ve introduced a minor workaround to the boilerplate repo (and hence to all other go repos). Please read through the thorough background discussion that Hongxin Liang posted around go modules in the GitHub issue (<https://github.com/lyft/flyte/issues/129>). In order to persist and standardize on a version of the common golang tools, we’ve added a second set of go.mod/sum files. These are listed in the boilerplate PR and basically separate your code from these tools. That is, your code likely uses golangci-lint but that doesn’t mean it should be part of your project’s go.mod files. Honnix’s comment has more information. This dual go.mod approach is also slated to be introduced more formally and rigorously in go 1.14.
This is a good overview of the new commands:
<https://github.com/golang/go/wiki/Modules#daily-workflow>
Note also that in most repos, we’re now pinning the version of client-go to something compatible with the Lyft fork of K8s (see <https://github.com/lyft/datacatalog/pull/21#issuecomment-579411660> for more information).
| Wow awesome!! Way to finish it off—I know it wasn’t easy! :p
|
Wow awesome!! Way to finish it off—I know it wasn’t easy! :p
| Thank you both! this not only aligns Flyte Repos with what's becoming the standards but also improves our tooling in the process (consistent build tools' versions). Big shout out to Hongxin Liang for going above and beyond to deliver that across all Flyte repos.
|
Thank you both! this not only aligns Flyte Repos with what's becoming the standards but also improves our tooling in the process (consistent build tools' versions). Big shout out to Hongxin Liang for going above and beyond to deliver that across all Flyte repos.
| Great job everyone! It was a good learning process for me to get better understanding of go ecosystem, and I enjoyed working on this. Super excited to see everything in place. :thumbsup:
|
I want to understand:
• Flyte's Monitoring
• Logging
• Alerting
• Integeration with Slack or something: Say this job is started by user X, job X is finished etc..
| This is a good place to start learning about notifications:
<https://lyft.github.io/flyte/user/features/notifications.html>
As far as logging/monitoring, there are two "stories" there:
1. As an administrator of the "Flyte" system, you want logging monitoring about the heath of your Flyte cluster.
2. As a user (workflow writer), you want logs showing what your workflow is doing, so you can debug your workflow code.
|
This is a good place to start learning about notifications:
<https://lyft.github.io/flyte/user/features/notifications.html>
As far as logging/monitoring, there are two "stories" there:
1. As an administrator of the "Flyte" system, you want logging monitoring about the heath of your Flyte cluster.
2. As a user (workflow writer), you want logs showing what your workflow is doing, so you can debug your workflow code.
| Email and Slack is good enough
So What about Logging and Monitoring?
|
Email and Slack is good enough
So What about Logging and Monitoring?
| <https://lyft.github.io/flyte/user/features/observability.html>
Flyte system components use Prometheus (they expose metrics through prometheus port).
We found that prometheus can be a hit or a miss for user containers (since it's a pull model, it can miss data points). So for metrics emitted from user containers, you will need to setup <https://github.com/statsd/statsd> in the cluster and configure this :
this (default env vars): <https://github.com/lyft/flyte/blob/master/deployment/sandbox/flyte_generated.yaml#L636>
Add something like this:
`- FLYTE_STATSD_HOST: stats.statsagent`
where stats.statsagent is the DNS for where you deployed statsd within the cluster...
having said that, I think there is an obvious gap in documentation in how to set this up :slightly_smiling_face:
|
<https://lyft.github.io/flyte/user/features/notifications.html#howto-slack> link is broken?
| Hmmmm.... Looks like the docs for that are missing :disappointed:
The notification code should look something like this:
```@workflow(notify=[Email(when=[States.FAILED, States.TIMED_OUT], who=["<mailto:[email protected]|[email protected]>"],
service_instances=["staging", "production"]),
Email(when=[States.SUCCESS], who=["<mailto:[email protected]|[email protected]>", "<mailto:[email protected]|[email protected]>"])])```
|
<!here> Ketan Umare and myself on TWIML, talking about Flyte, giving a bit more history about how it all started, where it's going and key differentiators: <https://twimlai.com/twiml-talk-343-scalable-and-maintainable-workflows-at-lyft-with-flyte-w-haytham-abuelfutuh-and-ketan-umare/>
| Nice! TWiML is my favorite podcast! (went to Twimlcon recently too). Will listen to it this evening.
Great interview!
|
Nice! TWiML is my favorite podcast! (went to Twimlcon recently too). Will listen to it this evening.
Great interview!
| Thank you Richard
|
Thank you Richard
| Awesome! Congrats!
|
Awesome! Congrats!
| Good listen :+1:. Wish you got more time to talk about the caching and cataloguing stuff though as I think that’s a big differentiator from say Argo and Airflow.
|
Good listen :+1:. Wish you got more time to talk about the caching and cataloguing stuff though as I think that’s a big differentiator from say Argo and Airflow.
| Thank you Jonathon, that's certainly true. Stay tuned for more :wink:
|
Thank you Jonathon, that's certainly true. Stay tuned for more :wink:
| Memoization part sounded really exciting !
|
Memoization part sounded really exciting !
| Adhita Selvaraj how’s the tf operator stuff going
|
Adhita Selvaraj how’s the tf operator stuff going
| Slowly :disappointed: I haven't had time to work on it with the Kubeflow 1.0 sprints
After this week, I'll have more time on my hands
|
Hi guys, I've been working the past few days to implement some DL pipelines that I use at work on Flyte and I've been spending a lot of time debugging...right now i'm working locally on my docker-desktop k8s environment and the workflow is as follows:
edit code -> build container -> register workflow -> launch workflow -> check errors -> repeat
Some code I can debug locally but for example the actual task code needs to run inside flyte. Is there a way to "mock" the inputs and outputs or some way to debug it with an editor (I'm using VSCode).
Perhaps there is a way to attach to the task & workflow processes with ptvsd?
| Are you able to run just the raw container that contains the task in question?
|
Are you able to run just the raw container that contains the task in question?
| Yes, of course
|
Yes, of course
| what are the types of these inputs?
the command that is passed to the container is in the task spec. it’ll look something like this.
the braces are filled in by Propeller before execution to be the location of the s3 inputs file and the path that the outputs should be written to.
|
what are the types of these inputs?
the command that is passed to the container is in the task spec. it’ll look something like this.
the braces are filled in by Propeller before execution to be the location of the s3 inputs file and the path that the outputs should be written to.
| Ok thanks I'll try that
I didn't think I could run the command directly
|
Ok thanks I'll try that
I didn't think I could run the command directly
| yup you absolutely can…. this is probably the most black-box way of debugging. you can also run the task outside of the container entirely, just in a virtualenv (assuming this is python).
this isn’t yet sanitized for outside lyft so i’m gonna copy paste some
|
yup you absolutely can…. this is probably the most black-box way of debugging. you can also run the task outside of the container entirely, just in a virtualenv (assuming this is python).
this isn’t yet sanitized for outside lyft so i’m gonna copy paste some
| Thank you for your help I'll look into that
|
Hello Everyone! :hand: I am interested in a small part of `flytepropeller` in `flyte`.
Default, Propeller container is started with the command `flytepropeller --config "/etc/flyte/config*/config.yaml"`
Source: <https://github.com/lyft/flyte/blob/master/kustomize/base/propeller/deployment.yaml#L35-L39> It works great!
In the next step, I just removed the `access-key` and `secret-key` from `config.yam` and changed the command as follows:
```flytepropeller \
--config "/etc/flyte/config*/config.yaml" \
--storage.connection.access-key 'minio' \
--storage.connection.secret-key 'miniostorage'```
In my opinion, these flags should update the configuration for propeller. (then, I’d like use keys from environment vars)
At least, this approach works for `flyteadmin` and `datacatalog`. But, using this approach with `flytepropeller`, I have the error below:
```{"json":{"src":"root.go:221"},"level":"fatal","msg":"Failed to start Controller - [Failed to create Metadata storage: unable to configure the storage for s3. Error: missing Access Key ID]","ts":"2020-02-05T12:19:16Z"}```
Looks like my flags were ignored. Could you advise what is my mistake?
Thanks in advance and have a great day! :slightly_smiling_face:
sorry if I use wrong channel for this question :pray:
| Nice catch. Yeah, my impression is that this is a bug.
My best guess is that it's related to this:
<https://github.com/lyft/flyteadmin/blob/master/cmd/entrypoints/root.go#L80>
Both Flyteadmin and Datacatalog do this, flytepropeller does not.
Haytham Abuelfutuh, can you confirm?
|
Nice catch. Yeah, my impression is that this is a bug.
My best guess is that it's related to this:
<https://github.com/lyft/flyteadmin/blob/master/cmd/entrypoints/root.go#L80>
Both Flyteadmin and Datacatalog do this, flytepropeller does not.
Haytham Abuelfutuh, can you confirm?
| Can you show me the storage section in config?
|
Can you show me the storage section in config?
| Thank you for quick response!
I did only these small changes in comparison with sandbox example <https://github.com/lyft/flyte/blob/master/kustomize/overlays/sandbox/propeller/config.yaml#L40-L49>
just commented two lines with creds:
|
Thank you for quick response!
I did only these small changes in comparison with sandbox example <https://github.com/lyft/flyte/blob/master/kustomize/overlays/sandbox/propeller/config.yaml#L40-L49>
just commented two lines with creds:
| I also reproduced the issue locally (removed the `endpoint` config from the file, added it as a command line arg).
|
I also reproduced the issue locally (removed the `endpoint` config from the file, added it as a command line arg).
| Awesome!. This should work. I agree this is a bug, mind filing an issue on <http://github.com/lyft/flyte|github.com/lyft/flyte> and assign it to me (@enghabu), Ruslan Stanevich?
|
Awesome!. This should work. I agree this is a bug, mind filing an issue on <http://github.com/lyft/flyte|github.com/lyft/flyte> and assign it to me (@enghabu), Ruslan Stanevich?
| oh of course!
thanks for your responsiveness
|
oh of course!
thanks for your responsiveness
| Haytham Abuelfutuh haven't tested this yet, but if you like how it looks, I can give it a whirl.
<https://github.com/lyft/flytepropeller/pull/63/files>
^ Guess it's never that simple. Looks like this results in a whole new problem:
```* '' has invalid keys: qubolelimit, quboletokenpath, redishostkey, redishostpath, resourcemanagertype
1 error(s) decoding:```
|
Haytham Abuelfutuh haven't tested this yet, but if you like how it looks, I can give it a whirl.
<https://github.com/lyft/flytepropeller/pull/63/files>
^ Guess it's never that simple. Looks like this results in a whole new problem:
```* '' has invalid keys: qubolelimit, quboletokenpath, redishostkey, redishostpath, resourcemanagertype
1 error(s) decoding:```
| <https://github.com/lyft/flyte/issues/167>
hope I missed nothing
Johnny, I don’t know exactly, but from my expirience, at least this error:
```1 error(s) decoding:
* '' has invalid keys: qubolelimit```
can be related to using flytepropeller:v0.2* with old plugin configs
<https://github.com/lyft/flyte/commit/b6c38aed5019677e4fc83b4c160fa3daca29cbc0#diff-9587455136ef88535397e7e8006e0dde>
|
<https://github.com/lyft/flyte/issues/167>
hope I missed nothing
Johnny, I don’t know exactly, but from my expirience, at least this error:
```1 error(s) decoding:
* '' has invalid keys: qubolelimit```
can be related to using flytepropeller:v0.2* with old plugin configs
<https://github.com/lyft/flyte/commit/b6c38aed5019677e4fc83b4c160fa3daca29cbc0#diff-9587455136ef88535397e7e8006e0dde>
| Ah, you might be right! Maybe my fix is working. Let me look.
That helps, but doesn't totally fix the issue. Seems some work will need to go into fixing it
|
Ah, you might be right! Maybe my fix is working. Let me look.
That helps, but doesn't totally fix the issue. Seems some work will need to go into fixing it
| oh, as I see there is `secrets` section in pfopeller’s config
<https://github.com/lyft/flytepropeller/blob/b1595306d38404c41eb3e6bf7dbabd8c8978544b/pkg/controller/nodes/task/secretmanager/config.go>
does it mean we are able to use it in the workflow task? Sorry, maybe silly question, I didn’t manage to find it in doc )
A few use cases would be extremely useful
Thanks in advance!
|
oh, as I see there is `secrets` section in pfopeller’s config
<https://github.com/lyft/flytepropeller/blob/b1595306d38404c41eb3e6bf7dbabd8c8978544b/pkg/controller/nodes/task/secretmanager/config.go>
does it mean we are able to use it in the workflow task? Sorry, maybe silly question, I didn’t manage to find it in doc )
A few use cases would be extremely useful
Thanks in advance!
| Great question.
The secret manager in propeller is designed for the secrets of `flytepropeller` (and its plugins). Not so much for individual tasks.
That might be a bit confusing, so I can elaborate a bit.
Flytepropeller launches tasks based on their task "type". In order to launch those tasks, propeller might need access to secrets.
For example, one of our task types can launch hive queries in a remote (3rd party) cluster.
In order for flytepropeller to access that remote cluster (to launch the query), propeller needs a secret "access" token. All tasks of this type need that token.
This plugin uses the secret manager to retrieve the token, and uses that token to make the query.
<https://github.com/lyft/flyteplugins/blob/e5ab7319502a0a69d4825b5abe08764d24133811/go/tasks/plugins/hive/executor.go#L79>
|
Great question.
The secret manager in propeller is designed for the secrets of `flytepropeller` (and its plugins). Not so much for individual tasks.
That might be a bit confusing, so I can elaborate a bit.
Flytepropeller launches tasks based on their task "type". In order to launch those tasks, propeller might need access to secrets.
For example, one of our task types can launch hive queries in a remote (3rd party) cluster.
In order for flytepropeller to access that remote cluster (to launch the query), propeller needs a secret "access" token. All tasks of this type need that token.
This plugin uses the secret manager to retrieve the token, and uses that token to make the query.
<https://github.com/lyft/flyteplugins/blob/e5ab7319502a0a69d4825b5abe08764d24133811/go/tasks/plugins/hive/executor.go#L79>
| Oh, cool
It makes sense!
Thank you!
|