output
stringlengths
1
18.7k
input
stringlengths
1
18.7k
jinx
<#CP2HDHKE1|onboarding>
oh you beat me to it :slightly_smiling_face:
jinx
Hello Sergio! thank you for joining
Hello
glad to hear it! welcome and please don’t hesitate to ask any questions!!
Hello everyone, I just started playing with Flyte and I'm very excited, it looks like a very promising project
Thank you Matt, as a matter of fact I do have a few…what would be the best channel to discuss them?
glad to hear it! welcome and please don’t hesitate to ask any questions!!
I think you can ask here and if it’s better to be in another channel, we can help re-route it for you Though i’m about to head off to bed so someone will have to pick it up tomorrow morning.
Thank you Matt, as a matter of fact I do have a few…what would be the best channel to discuss them?
Hi Giordano sorry for the late reply. So in the local sandbox cluster mode we do not have a scheduler today. When you deploy to a cloud environment we use either AWS cloud watch schedules or on GCP cloud scheduler. This is an area of active interest and we prefer to write a simple one using k8s operator or better option work with our partner Spotify to use Styx scheduler to run the cron schedules So to understand, we just don’t have the cron runner, the runner just triggers Flyte control plane and we do the rest. - let me know if you would love to contribute or have any other questions Let me know if you need help with anything else
ok I got a couple of questions: 1) I followed the examples in the flytesnacks repository and created a simple workflow with a few tasks, now I am trying to schedule it to run every 5 minutes by using this code that I found in the docs (<https://lyft.github.io/flyte/user/features/lanuchplans.html>) which I added after the workflow definition ```my_fixed_rate_launch_plan = myworkflow.create_launch_plan( default_inputs={'string_in': Input(Types.String, default="aaaa")}, schedule=schedules.FixedRate(datetime.timedelta(minutes=5)), )``` when I rebuild the container and re-register the workflow i see that the launchplan gets registered ```Flyte Admin URL 127.0.0.1:30081 Running task, workflow, and launch plan registration for flytedemo, development, ['fk_tasks'] with version 62ffc8b36883d18f4cf424f8d09510a5dd3db46d Registering Task: fk_tasks.workflows.download_dataset Registering Task: fk_tasks.workflows.read_pickle Registering Workflow: fk_tasks.workflows.myworkflow Registering Launch Plan: fk_tasks.workflows.my_fixed_rate_launch_plan Registering Launch Plan: fk_tasks.workflows.myworkflow Registering Task: fk_tasks.workflows.reverse_task Registering Task: fk_tasks.workflows.uppercase_task``` unfortunately in the gui I get a “This workflow has no schedules.” What am I missing?
Hi Ketan Umare thanks for the info, i was looking at flyte to replace airflow running on a barematal k8s cluster so I guess I’ll have to manage without the scheduling part
Hi Giordano sorry for the late reply. So in the local sandbox cluster mode we do not have a scheduler today. When you deploy to a cloud environment we use either AWS cloud watch schedules or on GCP cloud scheduler. This is an area of active interest and we prefer to write a simple one using k8s operator or better option work with our partner Spotify to use Styx scheduler to run the cron schedules So to understand, we just don’t have the cron runner, the runner just triggers Flyte control plane and we do the rest. - let me know if you would love to contribute or have any other questions Let me know if you need help with anything else
Giordano we would love to work with you on a replacement for the cloud scheduler Try out the other features and e See is you like everything else The reason for not packaging a scheduler is reinvention of wheel And building a scheduler for our scale will take sometime 100ks execution per day
Hi Ketan Umare thanks for the info, i was looking at flyte to replace airflow running on a barematal k8s cluster so I guess I’ll have to manage without the scheduling part
There is documentation for this, will share a link. It’s as simple as importing the workflow and there are bunch of API’s on them Also you can use Flyte-cli
2) are there any examples on how to connect to flyteadmin and launch/manage and query workflow results via the flytesdk?
great thanks looking forward to the link.
There is documentation for this, will share a link. It’s as simple as importing the workflow and there are bunch of API’s on them Also you can use Flyte-cli
<https://lyft.github.io/flyte/user/features/flytecli.html|https://lyft.github.io/flyte/user/features/flytecli.html> Flyte cli <https://lyft.github.io/flyte/flytekit/flytekit.common.html#module-flytekit.common.workflow_execution|https://lyft.github.io/flyte/flytekit/flytekit.common.html#module-flytekit.common.workflow_execution> We don’t have better docs on the interactive part yet
great thanks looking forward to the link.
is your code pushed anywhere we can take a look?
ok one more question, I was trying to create a workflow that would call some tasks aready defined in another workflow using an example that I found on the kubecon 19 presentation ```from flytekit.models.launch_plan import LaunchPlanState from flytekit.common import utils, schedules, tasks from flytekit.sdk.tasks import python_task, outputs, inputs from flytekit.sdk.types import Types from flytekit.sdk.workflow import workflow_class, Output, Input from flytekit.common.tasks.task import SdkTask @workflow_class class myworkflow(object): string_in = Input(Types.String, required=True, help="input string") dataset = Input(Types.CSV, default=Types.CSV.create_at_known_location( "<http://172.16.140.171:8000/label_summary.csv>"), help="A CSV File") return_dataset = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset","470cae526836f7a05d41ef81faa07a3b275b9de9")(dataset=dataset) return_pickle = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","470cae526836f7a05d41ef81faa07a3b275b9de9")(dataset=return_dataset.outputs.out) myoutput = Output(return_pickle.outputs.csv_head, sdk_type=Types.String)``` unfortunately when I register I get an error ```flytekit.common.exceptions.user.FlyteAssertion: An entity was not found in modules accessible from the workflow packages configuration. Please ensure that entities in 'fk_workflow.workflow' are moved to a configured packaged, or adjust the configuration.``` Does my code look ok? am I missing something? What I am trying to achieve is to de-couple the tasks from the workflow so that I can dynamically create a workflow from pre-existing tasks. my use case would be some kind of application where a user can choose different pre-built tasks, build a workflow definition through some kind of gui and then run a script that would dynamically create the workflow with those tasks. I understand that the tasks need to be created inside a container, but from what I understand a workflow does not need to be related to a specific container build, is my assumption correct?
No but i can create a quick repo
is your code pushed anywhere we can take a look?
sure, if you don’t mind. always helpful
No but i can create a quick repo
hmm yeh, i’m not seeing where `fk_workflow.workflow` is referenced (the code you copied originally looks great!)
sure, if you don’t mind. always helpful
hi Matt thanks, i just pushed to the repo my example <https://github.com/giordyb/flyte_demo> the workflow inside the fk_tasks folder works just fine (it’s 4 tasks and 1 workflow) in the workflow inside the fk_workflow folder I’m trying fetch the tasks created with the previous workflow but doesn’t work
hmm yeh, i’m not seeing where `fk_workflow.workflow` is referenced (the code you copied originally looks great!)
what happens if you run `flyte-cli -h &lt;your-url&gt; list-task-names -p flytedemo -d development`
hi Matt thanks, i just pushed to the repo my example <https://github.com/giordyb/flyte_demo> the workflow inside the fk_tasks folder works just fine (it’s 4 tasks and 1 workflow) in the workflow inside the fk_workflow folder I’m trying fetch the tasks created with the previous workflow but doesn’t work
Welcome to Flyte CLI! Version: 0.4.4 Task Names Found in flytedemo:development fk_tasks.tasks_and_workflow.download_dataset fk_tasks.tasks_and_workflow.read_pickle fk_tasks.tasks_and_workflow.reverse_task fk_tasks.tasks_and_workflow.uppercase_task btw I also tried fetch_latest without the version number but i get a ```AttributeError: type object 'SdkTask' has no attribute 'fetch_latest'```
what happens if you run `flyte-cli -h &lt;your-url&gt; list-task-names -p flytedemo -d development`
cool and also: `flyte-cli -h &lt;your-url&gt; list-task-versions -p flytedemo -d development -n fk_tasks.tasks_and_workflow.download_dataset`
Welcome to Flyte CLI! Version: 0.4.4 Task Names Found in flytedemo:development fk_tasks.tasks_and_workflow.download_dataset fk_tasks.tasks_and_workflow.read_pickle fk_tasks.tasks_and_workflow.reverse_task fk_tasks.tasks_and_workflow.uppercase_task btw I also tried fetch_latest without the version number but i get a ```AttributeError: type object 'SdkTask' has no attribute 'fetch_latest'```
Welcome to Flyte CLI! Version: 0.4.4 Task Versions Found for flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset Version Urn 1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216 tsk:flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset:1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216
cool and also: `flyte-cli -h &lt;your-url&gt; list-task-versions -p flytedemo -d development -n fk_tasks.tasks_and_workflow.download_dataset`
ok cool, the SHA `470cae526836f7a05d41ef81faa07a3b275b9de9` doesn’t look like it was ever registered if you swapped that with `1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216`, it should work
Welcome to Flyte CLI! Version: 0.4.4 Task Versions Found for flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset Version Urn 1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216 tsk:flytedemo:development:fk_tasks.tasks_and_workflow.download_dataset:1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216
sorry i had rebuilt the deployment in the meantime just tried with that version and it still throws ```flytekit.common.exceptions.user.FlyteAssertion: An entity was not found in modules accessible from the workflow packages configuration. Please ensure that entities in 'fk_workflow.workflow' are moved to a configured packaged, or adjust the configuration.```
ok cool, the SHA `470cae526836f7a05d41ef81faa07a3b275b9de9` doesn’t look like it was ever registered if you swapped that with `1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216`, it should work
```read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") download_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") @workflow_class class myworkflow(object): string_in = Input(Types.String, required=True, help="input string") dataset = Input(Types.CSV, default=Types.CSV.create_at_known_location( "<http://172.16.140.171:8000/label_summary.csv>"), help="A CSV File") return_dataset = download_task(dataset=dataset) return_pickle = read_task(dataset=return_dataset.outputs.out) myoutput = Output(return_pickle.outputs.csv_head, sdk_type=Types.String)``` hmm that might be a bug in the registration tool. …sorry still typing can you try above? i think there is a bug which incorrectly triggers a sanity-check mechanism anyway, this is actually easier in a ‘pure client’ case. The case where you have a service that is dynamically generating workflows, you can use the client to find task versions, weave them into a workflow, and then simply call register in a pure script like: ```t = SdkTask.fetch(...) wf = workflow(nodes={'n1': t()}, inputs=...) lp = wf.create_launch_plan() wf.register(..) lp.register(..)```
sorry i had rebuilt the deployment in the meantime just tried with that version and it still throws ```flytekit.common.exceptions.user.FlyteAssertion: An entity was not found in modules accessible from the workflow packages configuration. Please ensure that entities in 'fk_workflow.workflow' are moved to a configured packaged, or adjust the configuration.```
```Running task, workflow, and launch plan registration for flytedemo, development, ['fk_workflow'] with version 1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216 Traceback (most recent call last): File "/app/venv/bin/pyflyte", line 11, in &lt;module&gt; sys.exit(main()) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 97, in workflows register_all(project, domain, pkgs, test, version) File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 21, in register_all for m, k, o in iterate_registerable_entities_in_order(pkgs): File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 112, in iterate_registerable_entities_in_order for m in iterate_modules(pkgs): File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 16, in iterate_modules yield importlib.import_module(name) File "/usr/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "&lt;frozen importlib._bootstrap&gt;", line 994, in _gcd_import File "&lt;frozen importlib._bootstrap&gt;", line 971, in _find_and_load File "&lt;frozen importlib._bootstrap&gt;", line 955, in _find_and_load_unlocked File "&lt;frozen importlib._bootstrap&gt;", line 665, in _load_unlocked File "&lt;frozen importlib._bootstrap_external&gt;", line 678, in exec_module File "&lt;frozen importlib._bootstrap&gt;", line 219, in _call_with_frames_removed File "/app/fk_workflow/workflow.py", line 9, in &lt;module&gt; read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point return wrapped(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/flytekit/common/tasks/task.py", line 159, in fetch admin_task = _engine_loader.get_engine().fetch_task(task_id=task_id) File "/app/venv/lib/python3.6/site-packages/flytekit/engines/flyte/engine.py", line 108, in fetch_task ).client.get_task(task_id) File "/app/venv/lib/python3.6/site-packages/flytekit/clients/friendly.py", line 162, in get_task id=id.to_flyte_idl() File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 12, in handler return fn(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 136, in get_task return self._stub.GetTask(get_object_request) File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 824, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 726, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: &lt;_InactiveRpcError of RPC that terminated with: status = StatusCode.NOT_FOUND details = "entry not found" debug_error_string = "{"created":"@1579804960.139722800","description":"Error received from peer ipv4:127.0.0.1:30081","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"entry not found","grpc_status":5}" &gt; make: *** [register_workflow] Error 1``` nope still throws an error the pure client solution would be even better i think there is something that I haven’t quite understood yet, maybe you could shed some light I understand that in order to register a task i need to do it from a container (or at least set an env variable that points to the container that will run it) do I still need to reference a container when creating a workflow?
```read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") download_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") @workflow_class class myworkflow(object): string_in = Input(Types.String, required=True, help="input string") dataset = Input(Types.CSV, default=Types.CSV.create_at_known_location( "<http://172.16.140.171:8000/label_summary.csv>"), help="A CSV File") return_dataset = download_task(dataset=dataset) return_pickle = read_task(dataset=return_dataset.outputs.out) myoutput = Output(return_pickle.outputs.csv_head, sdk_type=Types.String)``` hmm that might be a bug in the registration tool. …sorry still typing can you try above? i think there is a bug which incorrectly triggers a sanity-check mechanism anyway, this is actually easier in a ‘pure client’ case. The case where you have a service that is dynamically generating workflows, you can use the client to find task versions, weave them into a workflow, and then simply call register in a pure script like: ```t = SdkTask.fetch(...) wf = workflow(nodes={'n1': t()}, inputs=...) lp = wf.create_launch_plan() wf.register(..) lp.register(..)```
no, definitely not for a workflow…and you don’t actually technically need to be in a container for a task either. Being in a container that is configured in a certain way just makes it easy to auto-fill in information when creating tasks.
```Running task, workflow, and launch plan registration for flytedemo, development, ['fk_workflow'] with version 1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216 Traceback (most recent call last): File "/app/venv/bin/pyflyte", line 11, in &lt;module&gt; sys.exit(main()) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 1137, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 97, in workflows register_all(project, domain, pkgs, test, version) File "/app/venv/lib/python3.6/site-packages/flytekit/clis/sdk_in_container/register.py", line 21, in register_all for m, k, o in iterate_registerable_entities_in_order(pkgs): File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 112, in iterate_registerable_entities_in_order for m in iterate_modules(pkgs): File "/app/venv/lib/python3.6/site-packages/flytekit/tools/module_loader.py", line 16, in iterate_modules yield importlib.import_module(name) File "/usr/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "&lt;frozen importlib._bootstrap&gt;", line 994, in _gcd_import File "&lt;frozen importlib._bootstrap&gt;", line 971, in _find_and_load File "&lt;frozen importlib._bootstrap&gt;", line 955, in _find_and_load_unlocked File "&lt;frozen importlib._bootstrap&gt;", line 665, in _load_unlocked File "&lt;frozen importlib._bootstrap_external&gt;", line 678, in exec_module File "&lt;frozen importlib._bootstrap&gt;", line 219, in _call_with_frames_removed File "/app/fk_workflow/workflow.py", line 9, in &lt;module&gt; read_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.read_pickle","1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point return wrapped(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/flytekit/common/tasks/task.py", line 159, in fetch admin_task = _engine_loader.get_engine().fetch_task(task_id=task_id) File "/app/venv/lib/python3.6/site-packages/flytekit/engines/flyte/engine.py", line 108, in fetch_task ).client.get_task(task_id) File "/app/venv/lib/python3.6/site-packages/flytekit/clients/friendly.py", line 162, in get_task id=id.to_flyte_idl() File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 12, in handler return fn(*args, **kwargs) File "/app/venv/lib/python3.6/site-packages/flytekit/clients/raw.py", line 136, in get_task return self._stub.GetTask(get_object_request) File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 824, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/app/venv/lib/python3.6/site-packages/grpc/_channel.py", line 726, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: &lt;_InactiveRpcError of RPC that terminated with: status = StatusCode.NOT_FOUND details = "entry not found" debug_error_string = "{"created":"@1579804960.139722800","description":"Error received from peer ipv4:127.0.0.1:30081","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"entry not found","grpc_status":5}" &gt; make: *** [register_workflow] Error 1``` nope still throws an error the pure client solution would be even better i think there is something that I haven’t quite understood yet, maybe you could shed some light I understand that in order to register a task i need to do it from a container (or at least set an env variable that points to the container that will run it) do I still need to reference a container when creating a workflow?
right that’s what I figured if you could provide some simple example of a workflow created with the pure client I would appreciate it another thing that’s not very clear from the flytekit documentation is how to connect to the admin service i know that with flyte-cli I can just pass the host parameters but I couldn’t find the equivalent for the flytekit i think it’s looking for some kind of configuration file because I get an error similar to this ```FlyteAssertion: No configuration set for [platform] url. This is a required configuration.```
no, definitely not for a workflow…and you don’t actually technically need to be in a container for a task either. Being in a container that is configured in a certain way just makes it easy to auto-fill in information when creating tasks.
do you get this error when using the pyflyte tool? or just when writing a script yourself? if the latter, take a look at this method: <https://github.com/lyft/flytekit/blob/master/flytekit/configuration/__init__.py#L11>
right that’s what I figured if you could provide some simple example of a workflow created with the pure client I would appreciate it another thing that’s not very clear from the flytekit documentation is how to connect to the admin service i know that with flyte-cli I can just pass the host parameters but I couldn’t find the equivalent for the flytekit i think it’s looking for some kind of configuration file because I get an error similar to this ```FlyteAssertion: No configuration set for [platform] url. This is a required configuration.```
when creating a script myself
do you get this error when using the pyflyte tool? or just when writing a script yourself? if the latter, take a look at this method: <https://github.com/lyft/flytekit/blob/master/flytekit/configuration/__init__.py#L11>
alternatively, you can set env vars that follow the format: `FLYTE_{SECTION}_{KEY}` . so in this case FLYTE_PLATFORM_URL, but i recommend creating a `flytekit.conf` file and i will find an example of a pure client use case
when creating a script myself
is the flytekit.config supposed to be created when running flyte-cli setup-config?
alternatively, you can set env vars that follow the format: `FLYTE_{SECTION}_{KEY}` . so in this case FLYTE_PLATFORM_URL, but i recommend creating a `flytekit.conf` file and i will find an example of a pure client use case
it is the same config file format, so it can be.
is the flytekit.config supposed to be created when running flyte-cli setup-config?
because I tried running ```flyte-cli setup-config -h localhost:30081 -i``` but I got a jsondecode error…should I open an issue about that?
it is the same config file format, so it can be.
yes please. cc: Yee
because I tried running ```flyte-cli setup-config -h localhost:30081 -i``` but I got a jsondecode error…should I open an issue about that?
:+1:
yes please. cc: Yee
there should be a barebones config in the demo repository you can copy
:+1:
ok thanks I’ll take a look
there should be a barebones config in the demo repository you can copy
here is an example script for a pure client approach. note: i copy-pasted different elements so there _might_ be a copy-paste error (and the interface definitions for your tasks/workflows will be different): ```my_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset", "1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") input_dict = { 'input_1': Input(Types.Integer, default=10, help='Not required input.'), 'input_2': Input(Types.Integer, help='Required.') } nodes = { 'a': my_task(a=input_dict['input_1']), 'b': my_task(a=input_dict['input_2']), 'c': my_task(a=1000) } outputs = { 'wf_out': Output( [ nodes['a'].outputs.b, nodes['b'].outputs.b, nodes['c'].outputs.b, ], sdk_type=[Types.Integer] ) } w = workflow(inputs=input_dict, outputs=outputs, nodes=nodes) w.register('flytedemo', 'development', 'simple_functional', version))) lp = w.create_launch_plan( fixed_inputs={'input_2': 100}, schedule=schedules.CronSchedule("0/15 * * * ? *") lp.register('flytedemo', 'development', 'simple_functional', version))) ex = lp.execute('flytedemo', 'development', inputs={'input_1': 500}) print(("Execution URN: {}".format(ex.id))) # wait for execution to complete, then check out the status, inputs, and outputs. ex.wait_for_completion(timeout=datetime.timedelta(minutes=10)) assert ex.closure.phase == _core_execution.WorkflowExecutionPhase.SUCCEEDED assert ex.error is None assert len(ex.inputs) == 2 assert len(ex.outputs) == 1 assert ex.inputs['input_1'] == 500 assert ex.inputs['input_2'] == 100 assert ex.outputs['wf_out'] == [501, 101, 1001] for k, ne in six.iteritems(ex.node_executions): if k in {'start-node', 'end-node'}: continue ne.sync() assert ne.closure.phase == _core_execution.NodeExecutionPhase.SUCCEEDED assert len(ne.inputs) == 1 assert len(ne.outputs) == 1 assert len(ne.executions) == 1 assert len(ne.task_executions) == 1 assert ne.error is None ne.executions[0].sync() assert ne.executions[0].error is None assert len(ne.executions[0].inputs) == 1 assert len(ne.executions[0].outputs) == 1 assert ne.task_executions[0].closure.phase == _core_execution.TaskExecutionPhase.SUCCEEDED assert len(ne.task_executions[0].inputs) == 1 assert len(ne.task_executions[0].outputs) == 1 assert ne.task_executions[0].error is None assert ex.node_executions['a'].inputs['a'] == 500 assert ex.node_executions['a'].outputs['b'] == 501 assert ex.node_executions['b'].inputs['a'] == 100 assert ex.node_executions['b'].outputs['b'] == 101 assert ex.node_executions['c'].inputs['a'] == 1000 assert ex.node_executions['c'].outputs['b'] == 1001``` important imports: ```from flytekit.sdk.types import Types from flytekit.sdk.workflow import workflow, Input, Output from flytekit.common import schedules from flytekit.configuration import TemporaryConfiguration from flytekit.models import launch_plan as _launch_plan from flytekit.models.core import execution as _core_execution``` and you’ll want to set the config somewhere beforehand too
ok thanks I’ll take a look
hi matt I got it to work after all one of the issues was that I renamed the task names but didn’t rename them in the code the other issue is that the docker image has the flytekit pinned to an older version in the flytesnacks example that’s why I got the error with fetch_latest thank you for your help!
here is an example script for a pure client approach. note: i copy-pasted different elements so there _might_ be a copy-paste error (and the interface definitions for your tasks/workflows will be different): ```my_task = SdkTask.fetch("flytedemo","development","fk_tasks.tasks.download_dataset", "1e0d95cc82b85cdd96feab3c6f9b6e2f7baba216") input_dict = { 'input_1': Input(Types.Integer, default=10, help='Not required input.'), 'input_2': Input(Types.Integer, help='Required.') } nodes = { 'a': my_task(a=input_dict['input_1']), 'b': my_task(a=input_dict['input_2']), 'c': my_task(a=1000) } outputs = { 'wf_out': Output( [ nodes['a'].outputs.b, nodes['b'].outputs.b, nodes['c'].outputs.b, ], sdk_type=[Types.Integer] ) } w = workflow(inputs=input_dict, outputs=outputs, nodes=nodes) w.register('flytedemo', 'development', 'simple_functional', version))) lp = w.create_launch_plan( fixed_inputs={'input_2': 100}, schedule=schedules.CronSchedule("0/15 * * * ? *") lp.register('flytedemo', 'development', 'simple_functional', version))) ex = lp.execute('flytedemo', 'development', inputs={'input_1': 500}) print(("Execution URN: {}".format(ex.id))) # wait for execution to complete, then check out the status, inputs, and outputs. ex.wait_for_completion(timeout=datetime.timedelta(minutes=10)) assert ex.closure.phase == _core_execution.WorkflowExecutionPhase.SUCCEEDED assert ex.error is None assert len(ex.inputs) == 2 assert len(ex.outputs) == 1 assert ex.inputs['input_1'] == 500 assert ex.inputs['input_2'] == 100 assert ex.outputs['wf_out'] == [501, 101, 1001] for k, ne in six.iteritems(ex.node_executions): if k in {'start-node', 'end-node'}: continue ne.sync() assert ne.closure.phase == _core_execution.NodeExecutionPhase.SUCCEEDED assert len(ne.inputs) == 1 assert len(ne.outputs) == 1 assert len(ne.executions) == 1 assert len(ne.task_executions) == 1 assert ne.error is None ne.executions[0].sync() assert ne.executions[0].error is None assert len(ne.executions[0].inputs) == 1 assert len(ne.executions[0].outputs) == 1 assert ne.task_executions[0].closure.phase == _core_execution.TaskExecutionPhase.SUCCEEDED assert len(ne.task_executions[0].inputs) == 1 assert len(ne.task_executions[0].outputs) == 1 assert ne.task_executions[0].error is None assert ex.node_executions['a'].inputs['a'] == 500 assert ex.node_executions['a'].outputs['b'] == 501 assert ex.node_executions['b'].inputs['a'] == 100 assert ex.node_executions['b'].outputs['b'] == 101 assert ex.node_executions['c'].inputs['a'] == 1000 assert ex.node_executions['c'].outputs['b'] == 1001``` important imports: ```from flytekit.sdk.types import Types from flytekit.sdk.workflow import workflow, Input, Output from flytekit.common import schedules from flytekit.configuration import TemporaryConfiguration from flytekit.models import launch_plan as _launch_plan from flytekit.models.core import execution as _core_execution``` and you’ll want to set the config somewhere beforehand too
Awesome!!! thank you both
Hongxin Liang and Haytham Abuelfutuh I’ve merged the boilerplate change and updated the idl pr <https://github.com/lyft/flyteidl/pull/27> thanks honnix for adding me as a collab mind taking a look one last time before we merge?
I will take care of the rest repos. Thanks.
Awesome!!! thank you both
good morning thank you! i’ll approve everything tomorrow morning. will then test, and then post an annoucement to the issue and this channel
I will take care of the rest repos. Thanks.
None that we know off is not playing well with flytekit
Hello everyone… are there any known issues with running a script that uses the “hyperopt” python library under flyte?
I have been trying to run a script that uses it under a workflow
None that we know off is not playing well with flytekit
hmmm and but you are using the python bindings (flytekit) right?
I have been trying to run a script that uses it under a workflow
i don’t even have to actually use the library, as soon as I import it in the script i get this
hmmm and but you are using the python bindings (flytekit) right?
are you pasting the error?
i don’t even have to actually use the library, as soon as I import it in the script i get this
```ERROR:root:Error from command '['aws', '--endpoint-url', '<http://minio.yolotrainframework.svc.cluster.local:9000>', 's3', 'cp', '<s3://my-s3-bucket/metadata/propeller/flytedemo-d> │ │ b'' │ │ Traceback (most recent call last): │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/data_proxy.py", line 127, in get_data │ │ proxy.download(remote_path, local_path) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 109, in download │ │ return _update_cmd_config_and_execute(cmd) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 36, in _update_cmd_config_and_execute │ │ return _subprocess.check_call(cmd, env=env) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/tools/subprocess.py", line 34, in check_call │ │ "Called process exited with error code: {}. Stderr dump:\n\n{}".format(ret_code, err_str) │ │ Exception: Called process exited with error code: -9. Stderr dump: │ │ b'' │ │ During handling of the above exception, another exception occurred: │ │ Traceback (most recent call last): │ │ File "/app/venv/bin/pyflyte-execute", line 11, in &lt;module&gt; │ │ sys.exit(execute_task_cmd()) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__ │ │ return self.main(*args, **kwargs) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main │ │ rv = self.invoke(ctx) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke │ │ return ctx.invoke(self.callback, **ctx.params) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke │ │ return callback(*args, **kwargs) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 104, in execute_task_cmd │ │ _execute_task(task_module, task_name, inputs, output_prefix, test) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point │ │ return wrapped(*args, **kwargs) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 83, in _execute_task │ │ _data_proxy.Data.get_data(inputs, local_inputs_file)``` this is from inside the container if I take out the hyperopt import everything works
are you pasting the error?
hmmm Exit code -9 let me check Giordano can you share your code if you dont mind?
```ERROR:root:Error from command '['aws', '--endpoint-url', '<http://minio.yolotrainframework.svc.cluster.local:9000>', 's3', 'cp', '<s3://my-s3-bucket/metadata/propeller/flytedemo-d> │ │ b'' │ │ Traceback (most recent call last): │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/data_proxy.py", line 127, in get_data │ │ proxy.download(remote_path, local_path) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 109, in download │ │ return _update_cmd_config_and_execute(cmd) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/interfaces/data/s3/s3proxy.py", line 36, in _update_cmd_config_and_execute │ │ return _subprocess.check_call(cmd, env=env) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/tools/subprocess.py", line 34, in check_call │ │ "Called process exited with error code: {}. Stderr dump:\n\n{}".format(ret_code, err_str) │ │ Exception: Called process exited with error code: -9. Stderr dump: │ │ b'' │ │ During handling of the above exception, another exception occurred: │ │ Traceback (most recent call last): │ │ File "/app/venv/bin/pyflyte-execute", line 11, in &lt;module&gt; │ │ sys.exit(execute_task_cmd()) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 764, in __call__ │ │ return self.main(*args, **kwargs) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 717, in main │ │ rv = self.invoke(ctx) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 956, in invoke │ │ return ctx.invoke(self.callback, **ctx.params) │ │ File "/app/venv/lib/python3.6/site-packages/click/core.py", line 555, in invoke │ │ return callback(*args, **kwargs) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 104, in execute_task_cmd │ │ _execute_task(task_module, task_name, inputs, output_prefix, test) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/common/exceptions/scopes.py", line 161, in system_entry_point │ │ return wrapped(*args, **kwargs) │ │ File "/app/venv/lib/python3.6/site-packages/flytekit/bin/entrypoint.py", line 83, in _execute_task │ │ _data_proxy.Data.get_data(inputs, local_inputs_file)``` this is from inside the container if I take out the hyperopt import everything works
it seems like importing hyperopt createss some issues with connecting to s3 sure gimme a sec
hmmm Exit code -9 let me check Giordano can you share your code if you dont mind?
ya this is actually connecting to minio in your sandbox cluster
it seems like importing hyperopt createss some issues with connecting to s3 sure gimme a sec
Looks like the subprocess which was trying to copy from s3 got a SIGKILL signal. The most common reason for that would be OOM. `aws s3 cp` shouldn't take too much memory, but it might be worth trying to bump the memory in your container using something like: ```@python_task(cpu_limit="10000m", memory_limit="10000Mi")```
ya this is actually connecting to minio in your sandbox cluster
Johnny Burns Ketan Umare just tried what you suggested and it's working now, thanks! It was indeed a resource issue
Looks like the subprocess which was trying to copy from s3 got a SIGKILL signal. The most common reason for that would be OOM. `aws s3 cp` shouldn't take too much memory, but it might be worth trying to bump the memory in your container using something like: ```@python_task(cpu_limit="10000m", memory_limit="10000Mi")```
Wow awesome!! Way to finish it off—I know it wasn’t easy! :p
As of today, we have migrated all golang Flyte repos to use go mod. Thanks to Hongxin Liang for driving this change. Boilerplate changes: <https://github.com/lyft/boilerplate/pull/4> The boilerplate code will now default to go.mod. <https://github.com/lyft/flytestdlib/releases/tag/v0.3.0> (was already on go mod as of v0.2.29) <https://github.com/lyft/flyteidl/releases/tag/v0.17.0> <https://github.com/lyft/flyteplugins/releases/tag/v0.3.0> <https://github.com/lyft/flytepropeller/releases/tag/v0.2.0> <https://github.com/lyft/flyteadmin/releases/tag/v0.2.0> <https://github.com/lyft/datacatalog/releases/tag/0.2.0> We’ve introduced a minor workaround to the boilerplate repo (and hence to all other go repos).  Please read through the thorough background discussion that Hongxin Liang posted around go modules in the GitHub issue (<https://github.com/lyft/flyte/issues/129>). In order to persist and standardize on a version of the common golang tools, we’ve added a second set of go.mod/sum files. These are listed in the boilerplate PR and basically separate your code from these tools. That is, your code likely uses golangci-lint but that doesn’t mean it should be part of your project’s go.mod files. Honnix’s comment has more information. This dual go.mod approach is also slated to be introduced more formally and rigorously in go 1.14. This is a good overview of the new commands: <https://github.com/golang/go/wiki/Modules#daily-workflow> Note also that in most repos, we’re now pinning the version of client-go to something compatible with the Lyft fork of K8s (see <https://github.com/lyft/datacatalog/pull/21#issuecomment-579411660> for more information).
Thank you both! this not only aligns Flyte Repos with what's becoming the standards but also improves our tooling in the process (consistent build tools' versions). Big shout out to Hongxin Liang for going above and beyond to deliver that across all Flyte repos.
Wow awesome!! Way to finish it off—I know it wasn’t easy! :p
Great job everyone! It was a good learning process for me to get better understanding of go ecosystem, and I enjoyed working on this. Super excited to see everything in place. :thumbsup:
Thank you both! this not only aligns Flyte Repos with what's becoming the standards but also improves our tooling in the process (consistent build tools' versions). Big shout out to Hongxin Liang for going above and beyond to deliver that across all Flyte repos.
This is a good place to start learning about notifications: <https://lyft.github.io/flyte/user/features/notifications.html> As far as logging/monitoring, there are two "stories" there: 1. As an administrator of the "Flyte" system, you want logging monitoring about the heath of your Flyte cluster. 2. As a user (workflow writer), you want logs showing what your workflow is doing, so you can debug your workflow code.
I want to understand: • Flyte's Monitoring • Logging • Alerting • Integeration with Slack or something: Say this job is started by user X, job X is finished etc..
Email and Slack is good enough So What about Logging and Monitoring?
This is a good place to start learning about notifications: <https://lyft.github.io/flyte/user/features/notifications.html> As far as logging/monitoring, there are two "stories" there: 1. As an administrator of the "Flyte" system, you want logging monitoring about the heath of your Flyte cluster. 2. As a user (workflow writer), you want logs showing what your workflow is doing, so you can debug your workflow code.
<https://lyft.github.io/flyte/user/features/observability.html> Flyte system components use Prometheus (they expose metrics through prometheus port). We found that prometheus can be a hit or a miss for user containers (since it's a pull model, it can miss data points). So for metrics emitted from user containers, you will need to setup <https://github.com/statsd/statsd> in the cluster and configure this : this (default env vars): <https://github.com/lyft/flyte/blob/master/deployment/sandbox/flyte_generated.yaml#L636> Add something like this: `- FLYTE_STATSD_HOST: stats.statsagent` where stats.statsagent is the DNS for where you deployed statsd within the cluster... having said that, I think there is an obvious gap in documentation in how to set this up :slightly_smiling_face:
Email and Slack is good enough So What about Logging and Monitoring?
Hmmmm.... Looks like the docs for that are missing :disappointed: The notification code should look something like this: ```@workflow(notify=[Email(when=[States.FAILED, States.TIMED_OUT], who=["<mailto:[email protected]|[email protected]>"], service_instances=["staging", "production"]), Email(when=[States.SUCCESS], who=["<mailto:[email protected]|[email protected]>", "<mailto:[email protected]|[email protected]>"])])```
<https://lyft.github.io/flyte/user/features/notifications.html#howto-slack> link is broken?
Nice! TWiML is my favorite podcast! (went to Twimlcon recently too). Will listen to it this evening. Great interview!
<!here> Ketan Umare and myself on TWIML, talking about Flyte, giving a bit more history about how it all started, where it's going and key differentiators: <https://twimlai.com/twiml-talk-343-scalable-and-maintainable-workflows-at-lyft-with-flyte-w-haytham-abuelfutuh-and-ketan-umare/>
Thank you Richard
Nice! TWiML is my favorite podcast! (went to Twimlcon recently too). Will listen to it this evening. Great interview!
Awesome! Congrats!
Thank you Richard
Good listen :+1:. Wish you got more time to talk about the caching and cataloguing stuff though as I think that’s a big differentiator from say Argo and Airflow.
Awesome! Congrats!
Thank you Jonathon, that's certainly true. Stay tuned for more :wink:
Good listen :+1:. Wish you got more time to talk about the caching and cataloguing stuff though as I think that’s a big differentiator from say Argo and Airflow.
Memoization part sounded really exciting !
Thank you Jonathon, that's certainly true. Stay tuned for more :wink:
Adhita Selvaraj how’s the tf operator stuff going
Memoization part sounded really exciting !
Are you able to run just the raw container that contains the task in question?
Hi guys, I've been working the past few days to implement some DL pipelines that I use at work on Flyte and I've been spending a lot of time debugging...right now i'm working locally on my docker-desktop k8s environment and the workflow is as follows: edit code -&gt; build container -&gt; register workflow -&gt; launch workflow -&gt; check errors -&gt; repeat Some code I can debug locally but for example the actual task code needs to run inside flyte. Is there a way to "mock" the inputs and outputs or some way to debug it with an editor (I'm using VSCode). Perhaps there is a way to attach to the task &amp; workflow processes with ptvsd?
Yes, of course
Are you able to run just the raw container that contains the task in question?
what are the types of these inputs? the command that is passed to the container is in the task spec. it’ll look something like this. the braces are filled in by Propeller before execution to be the location of the s3 inputs file and the path that the outputs should be written to.
Yes, of course
Ok thanks I'll try that I didn't think I could run the command directly
what are the types of these inputs? the command that is passed to the container is in the task spec. it’ll look something like this. the braces are filled in by Propeller before execution to be the location of the s3 inputs file and the path that the outputs should be written to.
yup you absolutely can…. this is probably the most black-box way of debugging. you can also run the task outside of the container entirely, just in a virtualenv (assuming this is python). this isn’t yet sanitized for outside lyft so i’m gonna copy paste some
Ok thanks I'll try that I didn't think I could run the command directly
Nice catch. Yeah, my impression is that this is a bug. My best guess is that it's related to this: <https://github.com/lyft/flyteadmin/blob/master/cmd/entrypoints/root.go#L80> Both Flyteadmin and Datacatalog do this, flytepropeller does not. Haytham Abuelfutuh, can you confirm?
Hello Everyone! :hand: I am interested in a small part of `flytepropeller` in `flyte`. Default, Propeller container is started with the command `flytepropeller --config "/etc/flyte/config*/config.yaml"` Source: <https://github.com/lyft/flyte/blob/master/kustomize/base/propeller/deployment.yaml#L35-L39> It works great! In the next step, I just removed the `access-key` and `secret-key` from `config.yam` and changed the command as follows: ```flytepropeller \ --config "/etc/flyte/config*/config.yaml" \ --storage.connection.access-key 'minio' \ --storage.connection.secret-key 'miniostorage'``` In my opinion, these flags should update the configuration for propeller. (then, I’d like use keys from environment vars) At least, this approach works for `flyteadmin` and `datacatalog`. But, using this approach with `flytepropeller`, I have the error below: ```{"json":{"src":"root.go:221"},"level":"fatal","msg":"Failed to start Controller - [Failed to create Metadata storage: unable to configure the storage for s3. Error: missing Access Key ID]","ts":"2020-02-05T12:19:16Z"}``` Looks like my flags were ignored. Could you advise what is my mistake? Thanks in advance and have a great day! :slightly_smiling_face: sorry if I use wrong channel for this question :pray:
Can you show me the storage section in config?
Nice catch. Yeah, my impression is that this is a bug. My best guess is that it's related to this: <https://github.com/lyft/flyteadmin/blob/master/cmd/entrypoints/root.go#L80> Both Flyteadmin and Datacatalog do this, flytepropeller does not. Haytham Abuelfutuh, can you confirm?
Thank you for quick response! I did only these small changes in comparison with sandbox example <https://github.com/lyft/flyte/blob/master/kustomize/overlays/sandbox/propeller/config.yaml#L40-L49> just commented two lines with creds:
Can you show me the storage section in config?
I also reproduced the issue locally (removed the `endpoint` config from the file, added it as a command line arg).
Thank you for quick response! I did only these small changes in comparison with sandbox example <https://github.com/lyft/flyte/blob/master/kustomize/overlays/sandbox/propeller/config.yaml#L40-L49> just commented two lines with creds:
Awesome!. This should work. I agree this is a bug, mind filing an issue on <http://github.com/lyft/flyte|github.com/lyft/flyte> and assign it to me (@enghabu), Ruslan Stanevich?
I also reproduced the issue locally (removed the `endpoint` config from the file, added it as a command line arg).
oh of course! thanks for your responsiveness
Awesome!. This should work. I agree this is a bug, mind filing an issue on <http://github.com/lyft/flyte|github.com/lyft/flyte> and assign it to me (@enghabu), Ruslan Stanevich?
Haytham Abuelfutuh haven't tested this yet, but if you like how it looks, I can give it a whirl. <https://github.com/lyft/flytepropeller/pull/63/files> ^ Guess it's never that simple. Looks like this results in a whole new problem: ```* '' has invalid keys: qubolelimit, quboletokenpath, redishostkey, redishostpath, resourcemanagertype 1 error(s) decoding:```
oh of course! thanks for your responsiveness
<https://github.com/lyft/flyte/issues/167> hope I missed nothing Johnny, I don’t know exactly, but from my expirience, at least this error: ```1 error(s) decoding: * '' has invalid keys: qubolelimit``` can be related to using flytepropeller:v0.2* with old plugin configs <https://github.com/lyft/flyte/commit/b6c38aed5019677e4fc83b4c160fa3daca29cbc0#diff-9587455136ef88535397e7e8006e0dde>
Haytham Abuelfutuh haven't tested this yet, but if you like how it looks, I can give it a whirl. <https://github.com/lyft/flytepropeller/pull/63/files> ^ Guess it's never that simple. Looks like this results in a whole new problem: ```* '' has invalid keys: qubolelimit, quboletokenpath, redishostkey, redishostpath, resourcemanagertype 1 error(s) decoding:```
Ah, you might be right! Maybe my fix is working. Let me look. That helps, but doesn't totally fix the issue. Seems some work will need to go into fixing it
<https://github.com/lyft/flyte/issues/167> hope I missed nothing Johnny, I don’t know exactly, but from my expirience, at least this error: ```1 error(s) decoding: * '' has invalid keys: qubolelimit``` can be related to using flytepropeller:v0.2* with old plugin configs <https://github.com/lyft/flyte/commit/b6c38aed5019677e4fc83b4c160fa3daca29cbc0#diff-9587455136ef88535397e7e8006e0dde>
oh, as I see there is `secrets` section in pfopeller’s config <https://github.com/lyft/flytepropeller/blob/b1595306d38404c41eb3e6bf7dbabd8c8978544b/pkg/controller/nodes/task/secretmanager/config.go> does it mean we are able to use it in the workflow task? Sorry, maybe silly question, I didn’t manage to find it in doc ) A few use cases would be extremely useful Thanks in advance!
Ah, you might be right! Maybe my fix is working. Let me look. That helps, but doesn't totally fix the issue. Seems some work will need to go into fixing it
Great question. The secret manager in propeller is designed for the secrets of `flytepropeller` (and its plugins). Not so much for individual tasks. That might be a bit confusing, so I can elaborate a bit. Flytepropeller launches tasks based on their task "type". In order to launch those tasks, propeller might need access to secrets. For example, one of our task types can launch hive queries in a remote (3rd party) cluster. In order for flytepropeller to access that remote cluster (to launch the query), propeller needs a secret "access" token. All tasks of this type need that token. This plugin uses the secret manager to retrieve the token, and uses that token to make the query. <https://github.com/lyft/flyteplugins/blob/e5ab7319502a0a69d4825b5abe08764d24133811/go/tasks/plugins/hive/executor.go#L79>
oh, as I see there is `secrets` section in pfopeller’s config <https://github.com/lyft/flytepropeller/blob/b1595306d38404c41eb3e6bf7dbabd8c8978544b/pkg/controller/nodes/task/secretmanager/config.go> does it mean we are able to use it in the workflow task? Sorry, maybe silly question, I didn’t manage to find it in doc ) A few use cases would be extremely useful Thanks in advance!
Oh, cool It makes sense! Thank you!
Great question. The secret manager in propeller is designed for the secrets of `flytepropeller` (and its plugins). Not so much for individual tasks. That might be a bit confusing, so I can elaborate a bit. Flytepropeller launches tasks based on their task "type". In order to launch those tasks, propeller might need access to secrets. For example, one of our task types can launch hive queries in a remote (3rd party) cluster. In order for flytepropeller to access that remote cluster (to launch the query), propeller needs a secret "access" token. All tasks of this type need that token. This plugin uses the secret manager to retrieve the token, and uses that token to make the query. <https://github.com/lyft/flyteplugins/blob/e5ab7319502a0a69d4825b5abe08764d24133811/go/tasks/plugins/hive/executor.go#L79>
If you'd like to have secrets on a per-task basis, we can discuss that too. We definitely do that here at Lyft. You should be able to use "vault" or any other secret manager to get those secrets into your container.
Oh, cool It makes sense! Thank you!
Yeah, currently we use "vault" with serviceaccount token authentication. Thank you for advice! We just looked for the most "native" approach to use it in Flyte
If you'd like to have secrets on a per-task basis, we can discuss that too. We definitely do that here at Lyft. You should be able to use "vault" or any other secret manager to get those secrets into your container.
So Flyte doesn't really prescribe a way to handle secrets (it's up to you), but I here is one way you might do it with vault (we're not using vault, so I could be overlooking some details): Vault has a secret injector: <https://www.vaultproject.io/docs/platform/k8s/injector/index.html> If you annotate your pods correctly, the secrets should get injected into the pod. Flyte allows you to add annotations via launch plans: ```annotations=Annotations({"<http://vault.hashicorp.com/agent-inject-secret-foo|vault.hashicorp.com/agent-inject-secret-foo>": 'bar/baz}),``` I'm sure there are other ways of doing this as well. Just one example.
Yeah, currently we use "vault" with serviceaccount token authentication. Thank you for advice! We just looked for the most "native" approach to use it in Flyte
fetching of an SdkWorkflow object is being worked on in an PR. it’s not quite ready yet. hopefully in a week it’ll be done, along with some other features we’ve been meaning to push out. <https://github.com/lyft/flytekit/pull/75/files>
hello! what would be the best way with the SDK to “fetch” an existing workflow, create a new launch plan with new inputs and execute it? i tried something like this: ```import flytekit.configuration import flytekit.common.workflow flytekit.configuration.set_flyte_config_file("flytekit.conf") wf = flytekit.common.workflow.SdkWorkflow.fetch( project="yolotrain", domain="development", name="train.single.workflow_yolo_single", version="71a60ca9fa75497968bb09fe8c4ba8d3aee042cb", )``` but I’m getting ```FlyteAssertion: An SDK node must have one underlying entity specified at once. Received the following entities: []```
Ok thanks, looking forward to the pr :+1:
fetching of an SdkWorkflow object is being worked on in an PR. it’s not quite ready yet. hopefully in a week it’ll be done, along with some other features we’ve been meaning to push out. <https://github.com/lyft/flytekit/pull/75/files>
i think there’s a way to create the launchplan without the class code present as well but will wait for Matt Smith to answer that (who should be back today)
Ok thanks, looking forward to the pr :+1:
Yee there is a way Giordano you can use flyte-cli that is the simplest way to create a launchplan or am i still jet lagged :slightly_smiling_face:
i think there’s a way to create the launchplan without the class code present as well but will wait for Matt Smith to answer that (who should be back today)
Hey, Giordano quick question: are you trying to execute the workflow here? Or trying to ‘specialize’ the workflow interface by creating defaults, schedules, different service accounts, etc.? for more context: a launch plan can be thought of as a specialized ‘executable’ for a workflow. Then a launch plan can be executed as many times as one wants. For example, if i had a workflow that takes inputs `country` and `time`, I could use the same workflow to create two launch plans. One launch plan could freeze `country='USA'` and be scheduled to run daily with an IAM role given write access to `<s3://USA-data>`. The other could freeze `country='Canada'` and be scheduled to run weekly with an IAM role that only accesses `<s3://Canada-data>`. This way a pipeline (workflow) can be generalized, but then specialized at execution to provide data protections, etc.. An execution -&gt; launch plan, is a many-to-one relationship. So generally, when you are creating a new execution, you don’t need to create a new launch plan. You only need to retrieve an existing one and call execute on it. So if you are trying to launch an execution, the simplest way is: ```lp = SdkLaunchPlan.fetch('project', 'domain', 'name', 'version') ex = lp.execute('project', 'domain' inputs={'a': 1, 'b': 'hello'}, &lt;name='optional idempotency string'&gt;)```
Yee there is a way Giordano you can use flyte-cli that is the simplest way to create a launchplan or am i still jet lagged :slightly_smiling_face:
Hi Matt, right now I’m not looking to customize the workflow, just be able to launch it with different inputs so your suggestion might actually do the trick
Hey, Giordano quick question: are you trying to execute the workflow here? Or trying to ‘specialize’ the workflow interface by creating defaults, schedules, different service accounts, etc.? for more context: a launch plan can be thought of as a specialized ‘executable’ for a workflow. Then a launch plan can be executed as many times as one wants. For example, if i had a workflow that takes inputs `country` and `time`, I could use the same workflow to create two launch plans. One launch plan could freeze `country='USA'` and be scheduled to run daily with an IAM role given write access to `<s3://USA-data>`. The other could freeze `country='Canada'` and be scheduled to run weekly with an IAM role that only accesses `<s3://Canada-data>`. This way a pipeline (workflow) can be generalized, but then specialized at execution to provide data protections, etc.. An execution -&gt; launch plan, is a many-to-one relationship. So generally, when you are creating a new execution, you don’t need to create a new launch plan. You only need to retrieve an existing one and call execute on it. So if you are trying to launch an execution, the simplest way is: ```lp = SdkLaunchPlan.fetch('project', 'domain', 'name', 'version') ex = lp.execute('project', 'domain' inputs={'a': 1, 'b': 'hello'}, &lt;name='optional idempotency string'&gt;)```
cool then it’s an easier answer :slightly_smiling_face: if you use `pyflyte` to register your workflows, there should be a default launch plan created for each workflow with the same name as the workflow
Hi Matt, right now I’m not looking to customize the workflow, just be able to launch it with different inputs so your suggestion might actually do the trick
One thing that is not clear to me is the difference between the active and non-active launchplan for example when I look for “active” launchplans in my environment i don’t see any
cool then it’s an easier answer :slightly_smiling_face: if you use `pyflyte` to register your workflows, there should be a default launch plan created for each workflow with the same name as the workflow
ah yes, so an active launch plan can be thought of a representation of the ‘tip of your deployment’
One thing that is not clear to me is the difference between the active and non-active launchplan for example when I look for “active” launchplans in my environment i don’t see any
kind of like the latest version is that correct?
ah yes, so an active launch plan can be thought of a representation of the ‘tip of your deployment’
yes exactly, except making it a bit easier to rollback if necessary it’s important for two major cases:
kind of like the latest version is that correct?
ok so how do I make a launch plan active then? i couldn’t find it in the docs… i know that there is a “fetch_latest” for tasks…
yes exactly, except making it a bit easier to rollback if necessary it’s important for two major cases:
yeah, that’s a topic of ongoing debate…perhaps the active tag should be applied to all tasks and workflows. or we should go with a more generalized solution where workflows and tasks can be custom tagged and labeled and then have methods like `fetch_tag`. The reason why active is specifically important for launch plans is because of schedules. The admin service needs to know to change the schedule configuration for a launch plan based on which one is active. anywho, to make your launch plan active: `pyflyte -p project -d domain -c flyte.config lp activate-all [--ignore-schedules]`
ok so how do I make a launch plan active then? i couldn’t find it in the docs… i know that there is a “fetch_latest” for tasks…