From ad398e3afa2e0b383e499f9ab336d641aab22385 Mon Sep 17 00:00:00 2001 From: David Wood Date: Thu, 20 Jun 2024 14:05:25 -0400 Subject: [PATCH] Cut release 0.2.0 Signed-off-by: David Wood --- .make.versions | 2 +- data-processing-lib/python/pyproject.toml | 2 +- data-processing-lib/ray/pyproject.toml | 4 ++-- data-processing-lib/spark/pyproject.toml | 4 ++-- .../createRayClusterComponent.yaml | 2 +- .../deleteRayClusterComponent.yaml | 2 +- kfp/kfp_ray_components/executeRayJobComponent.yaml | 2 +- .../executeRayJobComponent_multi_s3.yaml | 2 +- .../executeSubWorkflowComponent.yaml | 2 +- .../kfp_v1_workflow_support/pyproject.toml | 4 ++-- .../kfp_v2_workflow_support/pyproject.toml | 2 +- .../shared_workflow_support/pyproject.toml | 2 +- .../ray/kfp_v1/superworkflow_code_sample_wf.py | 14 +++++++------- .../ray/kfp_v1/superworkflow_dedups_sample_wf.py | 6 +++--- .../code/code_quality/kfp_ray/code_quality_wf.py | 4 ++-- .../kfp_ray/ingest_2_parquet_wf.py | 2 +- transforms/code/malware/kfp_ray/malware_wf.py | 4 ++-- .../proglang_select/kfp_ray/proglang_select_wf.py | 4 ++-- .../lang_id/kfp_ray/lang_id_multiple_wf.py | 4 ++-- transforms/language/lang_id/kfp_ray/lang_id_wf.py | 4 ++-- transforms/language/lang_id/python/pyproject.toml | 4 ++-- transforms/language/lang_id/ray/pyproject.toml | 4 ++-- transforms/universal/doc_id/kfp_ray/doc_id_wf.py | 4 ++-- transforms/universal/ededup/kfp_ray/ededup_wf.py | 4 ++-- transforms/universal/fdedup/kfp_ray/fdedup_wf.py | 4 ++-- transforms/universal/filter/kfp_ray/filter_wf.py | 4 ++-- transforms/universal/filter/python/pyproject.toml | 4 ++-- transforms/universal/filter/ray/pyproject.toml | 6 +++--- .../universal/noop/kfp_ray/noop_multiple_wf.py | 4 ++-- transforms/universal/noop/kfp_ray/noop_wf.py | 4 ++-- transforms/universal/noop/python/pyproject.toml | 4 ++-- transforms/universal/noop/ray/pyproject.toml | 6 +++--- transforms/universal/noop/spark/pyproject.toml | 6 +++--- .../universal/profiler/kfp_ray/profiler_wf.py | 2 +- .../tokenization/kfp_ray/tokenization_wf.py | 4 ++-- .../universal/tokenization/python/pyproject.toml | 4 ++-- .../universal/tokenization/ray/pyproject.toml | 6 +++--- 37 files changed, 73 insertions(+), 73 deletions(-) diff --git a/.make.versions b/.make.versions index e4de65483..79cfd9e35 100644 --- a/.make.versions +++ b/.make.versions @@ -11,7 +11,7 @@ DPK_MAJOR_VERSION=0 DPK_MINOR_VERSION=2 DPK_MICRO_VERSION=0 -DPK_VERSION_SUFFIX=.dev6 +DPK_VERSION_SUFFIX=.dev7 DPK_VERSION=$(DPK_MAJOR_VERSION).$(DPK_MINOR_VERSION).$(DPK_MICRO_VERSION)$(DPK_VERSION_SUFFIX) diff --git a/data-processing-lib/python/pyproject.toml b/data-processing-lib/python/pyproject.toml index ff2525c4a..4c7110dc0 100644 --- a/data-processing-lib/python/pyproject.toml +++ b/data-processing-lib/python/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "data_prep_toolkit" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" keywords = ["data", "data preprocessing", "data preparation", "llm", "generative", "ai", "fine-tuning", "llmapps" ] description = "Data Preparation Toolkit Library" diff --git a/data-processing-lib/ray/pyproject.toml b/data-processing-lib/ray/pyproject.toml index 5aa05535a..516bb2460 100644 --- a/data-processing-lib/ray/pyproject.toml +++ b/data-processing-lib/ray/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "data_prep_toolkit_ray" -version = "0.2.0.dev6" +version = "0.2.0.dev7" keywords = ["data", "data preprocessing", "data preparation", "llm", "generative", "ai", "fine-tuning", "llmapps" ] requires-python = ">=3.10" description = "Data Preparation Toolkit Library for Ray" @@ -11,7 +11,7 @@ authors = [ { name = "Boris Lublinsky", email = "blublinsky@ibm.com" }, ] dependencies = [ - "data-prep-toolkit==0.2.0.dev6", + "data-prep-toolkit==0.2.0.dev7", "ray[default]==2.9.3", # These two are to fix security issues identified by quay.io "fastapi>=0.109.1", diff --git a/data-processing-lib/spark/pyproject.toml b/data-processing-lib/spark/pyproject.toml index 488113f25..6523d0c50 100644 --- a/data-processing-lib/spark/pyproject.toml +++ b/data-processing-lib/spark/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "data_prep_toolkit_spark" -version = "0.2.0.dev6" +version = "0.2.0.dev7" keywords = ["data", "data preprocessing", "data preparation", "llm", "generative", "ai", "fine-tuning", "llmapps" ] requires-python = ">=3.10" description = "Data Preparation Toolkit Library for Spark" @@ -11,7 +11,7 @@ authors = [ { name = "Constantin Adam", email = "cmadam@us.ibm.com" }, ] dependencies = [ - "data-prep-toolkit==0.2.0.dev6", + "data-prep-toolkit==0.2.0.dev7", "pyspark>=3.5.1", "pyyaml>=6.0.1", ] diff --git a/kfp/kfp_ray_components/createRayClusterComponent.yaml b/kfp/kfp_ray_components/createRayClusterComponent.yaml index 7c81abdbf..864129fed 100644 --- a/kfp/kfp_ray_components/createRayClusterComponent.yaml +++ b/kfp/kfp_ray_components/createRayClusterComponent.yaml @@ -11,7 +11,7 @@ inputs: implementation: container: - image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" + image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # command is a list of strings (command-line arguments). # The YAML language has two syntaxes for lists and you can use either of them. # Here we use the "flow syntax" - comma-separated strings inside square brackets. diff --git a/kfp/kfp_ray_components/deleteRayClusterComponent.yaml b/kfp/kfp_ray_components/deleteRayClusterComponent.yaml index a4e9e89aa..e70c5d096 100644 --- a/kfp/kfp_ray_components/deleteRayClusterComponent.yaml +++ b/kfp/kfp_ray_components/deleteRayClusterComponent.yaml @@ -8,7 +8,7 @@ inputs: implementation: container: - image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" + image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # command is a list of strings (command-line arguments). # The YAML language has two syntaxes for lists and you can use either of them. # Here we use the "flow syntax" - comma-separated strings inside square brackets. diff --git a/kfp/kfp_ray_components/executeRayJobComponent.yaml b/kfp/kfp_ray_components/executeRayJobComponent.yaml index 3a928d598..15f878efe 100644 --- a/kfp/kfp_ray_components/executeRayJobComponent.yaml +++ b/kfp/kfp_ray_components/executeRayJobComponent.yaml @@ -12,7 +12,7 @@ inputs: implementation: container: - image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" + image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # command is a list of strings (command-line arguments). # The YAML language has two syntaxes for lists and you can use either of them. # Here we use the "flow syntax" - comma-separated strings inside square brackets. diff --git a/kfp/kfp_ray_components/executeRayJobComponent_multi_s3.yaml b/kfp/kfp_ray_components/executeRayJobComponent_multi_s3.yaml index cd3d3c806..63afe98b9 100644 --- a/kfp/kfp_ray_components/executeRayJobComponent_multi_s3.yaml +++ b/kfp/kfp_ray_components/executeRayJobComponent_multi_s3.yaml @@ -13,7 +13,7 @@ inputs: implementation: container: - image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" + image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # command is a list of strings (command-line arguments). # The YAML language has two syntaxes for lists and you can use either of them. # Here we use the "flow syntax" - comma-separated strings inside square brackets. diff --git a/kfp/kfp_ray_components/executeSubWorkflowComponent.yaml b/kfp/kfp_ray_components/executeSubWorkflowComponent.yaml index f02a5ac8a..2fca30a82 100644 --- a/kfp/kfp_ray_components/executeSubWorkflowComponent.yaml +++ b/kfp/kfp_ray_components/executeSubWorkflowComponent.yaml @@ -27,7 +27,7 @@ outputs: implementation: container: - image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" + image: "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # command is a list of strings (command-line arguments). # The YAML language has two syntaxes for lists, and you can use either of them. # Here we use the "flow syntax" - comma-separated strings inside square brackets. diff --git a/kfp/kfp_support_lib/kfp_v1_workflow_support/pyproject.toml b/kfp/kfp_support_lib/kfp_v1_workflow_support/pyproject.toml index 3e2690923..77c52aa63 100644 --- a/kfp/kfp_support_lib/kfp_v1_workflow_support/pyproject.toml +++ b/kfp/kfp_support_lib/kfp_v1_workflow_support/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "data_prep_toolkit_kfp_v1" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10,<3.12" description = "Data Preparation Kit Library. KFP support" license = {text = "Apache-2.0"} @@ -15,7 +15,7 @@ dependencies = [ "kfp==1.8.22", "ray==2.9.3", "requests", - "data_prep_toolkit==0.2.0.dev6", + "data_prep_toolkit==0.2.0.dev7", ] [build-system] diff --git a/kfp/kfp_support_lib/kfp_v2_workflow_support/pyproject.toml b/kfp/kfp_support_lib/kfp_v2_workflow_support/pyproject.toml index 86ae43e5d..a6de05ff9 100644 --- a/kfp/kfp_support_lib/kfp_v2_workflow_support/pyproject.toml +++ b/kfp/kfp_support_lib/kfp_v2_workflow_support/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "data_prep_toolkit_kfp_v2" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10,<3.12" description = "Data Preparation Kit Library. KFP support" license = {text = "Apache-2.0"} diff --git a/kfp/kfp_support_lib/shared_workflow_support/pyproject.toml b/kfp/kfp_support_lib/shared_workflow_support/pyproject.toml index 2da001c72..3cb188947 100644 --- a/kfp/kfp_support_lib/shared_workflow_support/pyproject.toml +++ b/kfp/kfp_support_lib/shared_workflow_support/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "data_prep_toolkit_kfp_shared" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10,<3.12" description = "Data Preparation Kit Library. KFP support" license = {text = "Apache-2.0"} diff --git a/kfp/superworkflows/ray/kfp_v1/superworkflow_code_sample_wf.py b/kfp/superworkflows/ray/kfp_v1/superworkflow_code_sample_wf.py index f7a46fb7e..79289b942 100644 --- a/kfp/superworkflows/ray/kfp_v1/superworkflow_code_sample_wf.py +++ b/kfp/superworkflows/ray/kfp_v1/superworkflow_code_sample_wf.py @@ -15,13 +15,13 @@ run_fuzzy_dedup_op = comp.load_component_from_file(component_spec_path + "executeSubWorkflowComponent.yaml") run_tokenization_op = comp.load_component_from_file(component_spec_path + "executeSubWorkflowComponent.yaml") -proglang_select_image = "quay.io/dataprep1/data-prep-kit/proglang_select-ray:0.2.0.dev6" -code_quality_image = "quay.io/dataprep1/data-prep-kit/code_quality-ray:0.2.0.dev6" -malware_image = "quay.io/dataprep1/data-prep-kit/malware-ray:0.2.0.dev6" -doc_id_image = "quay.io/dataprep1/data-prep-kit/doc_id-ray:0.2.0.dev6" -ededup_image = "quay.io/dataprep1/data-prep-kit/ededup-ray:0.2.0.dev6" -fdedup_image = "quay.io/dataprep1/data-prep-kit/fdedup-ray:0.2.0.dev6" -tokenizer_image = "quay.io/dataprep1/data-prep-kit/tokenization-ray:0.2.0.dev6" +proglang_select_image = "quay.io/dataprep1/data-prep-kit/proglang_select-ray:0.2.0.dev7" +code_quality_image = "quay.io/dataprep1/data-prep-kit/code_quality-ray:0.2.0.dev7" +malware_image = "quay.io/dataprep1/data-prep-kit/malware-ray:0.2.0.dev7" +doc_id_image = "quay.io/dataprep1/data-prep-kit/doc_id-ray:0.2.0.dev7" +ededup_image = "quay.io/dataprep1/data-prep-kit/ededup-ray:0.2.0.dev7" +fdedup_image = "quay.io/dataprep1/data-prep-kit/fdedup-ray:0.2.0.dev7" +tokenizer_image = "quay.io/dataprep1/data-prep-kit/tokenization-ray:0.2.0.dev7" # Pipeline to invoke execution on remote resource diff --git a/kfp/superworkflows/ray/kfp_v1/superworkflow_dedups_sample_wf.py b/kfp/superworkflows/ray/kfp_v1/superworkflow_dedups_sample_wf.py index fde560b35..690a0fd9b 100644 --- a/kfp/superworkflows/ray/kfp_v1/superworkflow_dedups_sample_wf.py +++ b/kfp/superworkflows/ray/kfp_v1/superworkflow_dedups_sample_wf.py @@ -12,9 +12,9 @@ run_exact_dedup_op = comp.load_component_from_file(component_spec_path + "executeSubWorkflowComponent.yaml") run_fuzzy_dedup_op = comp.load_component_from_file(component_spec_path + "executeSubWorkflowComponent.yaml") -doc_id_image = "quay.io/dataprep1/data-prep-kit/doc_id-ray:0.2.0.dev6" -ededup_image = "quay.io/dataprep1/data-prep-kit/ededup-ray:0.2.0.dev6" -fdedup_image = "quay.io/dataprep1/data-prep-kit/fdedup-ray:0.2.0.dev6" +doc_id_image = "quay.io/dataprep1/data-prep-kit/doc_id-ray:0.2.0.dev7" +ededup_image = "quay.io/dataprep1/data-prep-kit/ededup-ray:0.2.0.dev7" +fdedup_image = "quay.io/dataprep1/data-prep-kit/fdedup-ray:0.2.0.dev7" # Pipeline to invoke execution on remote resource @dsl.pipeline( diff --git a/transforms/code/code_quality/kfp_ray/code_quality_wf.py b/transforms/code/code_quality/kfp_ray/code_quality_wf.py index 0542eb283..8b4a38e68 100644 --- a/transforms/code/code_quality/kfp_ray/code_quality_wf.py +++ b/transforms/code/code_quality/kfp_ray/code_quality_wf.py @@ -21,10 +21,10 @@ EXEC_SCRIPT_NAME: str = "code_quality_transform_ray.py" PREFIX: str = "" -task_image = "quay.io/dataprep1/data-prep-kit/code_quality-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/code_quality-ray:0.2.0.dev7" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/code/ingest_2_parquet/kfp_ray/ingest_2_parquet_wf.py b/transforms/code/ingest_2_parquet/kfp_ray/ingest_2_parquet_wf.py index ad5295411..db1699d13 100644 --- a/transforms/code/ingest_2_parquet/kfp_ray/ingest_2_parquet_wf.py +++ b/transforms/code/ingest_2_parquet/kfp_ray/ingest_2_parquet_wf.py @@ -25,7 +25,7 @@ # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/code/malware/kfp_ray/malware_wf.py b/transforms/code/malware/kfp_ray/malware_wf.py index 4bc26c0e8..8db1a87c3 100644 --- a/transforms/code/malware/kfp_ray/malware_wf.py +++ b/transforms/code/malware/kfp_ray/malware_wf.py @@ -21,10 +21,10 @@ # the name of the job script EXEC_SCRIPT_NAME: str = "malware_transform_ray.py" -task_image = "quay.io/dataprep1/data-prep-kit/malware-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/malware-ray:0.2.0.dev7" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/code/proglang_select/kfp_ray/proglang_select_wf.py b/transforms/code/proglang_select/kfp_ray/proglang_select_wf.py index c8f9c2e98..d0f6bf81a 100644 --- a/transforms/code/proglang_select/kfp_ray/proglang_select_wf.py +++ b/transforms/code/proglang_select/kfp_ray/proglang_select_wf.py @@ -21,10 +21,10 @@ # the name of the job script EXEC_SCRIPT_NAME: str = "proglang_select_transform_ray.py" -task_image = "quay.io/dataprep1/data-prep-kit/proglang_select-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/proglang_select-ray:0.2.0.dev7" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/language/lang_id/kfp_ray/lang_id_multiple_wf.py b/transforms/language/lang_id/kfp_ray/lang_id_multiple_wf.py index 439fb21d2..baae080be 100644 --- a/transforms/language/lang_id/kfp_ray/lang_id_multiple_wf.py +++ b/transforms/language/lang_id/kfp_ray/lang_id_multiple_wf.py @@ -17,13 +17,13 @@ from workflow_support.compile_utils import ONE_HOUR_SEC, ONE_WEEK_SEC, ComponentUtils -task_image = "quay.io/dataprep1/data-prep-kit/lang_id-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/lang_id-ray:0.2.0.dev7" # the name of the job script EXEC_SCRIPT_NAME: str = "lang_id_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/language/lang_id/kfp_ray/lang_id_wf.py b/transforms/language/lang_id/kfp_ray/lang_id_wf.py index b849cdd77..8350e316c 100644 --- a/transforms/language/lang_id/kfp_ray/lang_id_wf.py +++ b/transforms/language/lang_id/kfp_ray/lang_id_wf.py @@ -17,13 +17,13 @@ from workflow_support.compile_utils import ONE_HOUR_SEC, ONE_WEEK_SEC, ComponentUtils -task_image = "quay.io/dataprep1/data-prep-kit/lang_id-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/lang_id-ray:0.2.0.dev7" # the name of the job script EXEC_SCRIPT_NAME: str = "lang_id_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/language/lang_id/python/pyproject.toml b/transforms/language/lang_id/python/pyproject.toml index 126df7fee..1409a25ff 100644 --- a/transforms/language/lang_id/python/pyproject.toml +++ b/transforms/language/lang_id/python/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_lang_id_transform_python" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "Language Identification Python Transform" license = {text = "Apache-2.0"} @@ -9,7 +9,7 @@ authors = [ { name = "Daiki Tsuzuku", email = "dtsuzuku@jp.ibm.com" } ] dependencies = [ - "data-prep-toolkit==0.2.0.dev6", + "data-prep-toolkit==0.2.0.dev7", "fasttext==0.9.2", "langcodes==3.3.0", "huggingface-hub==0.21.4" diff --git a/transforms/language/lang_id/ray/pyproject.toml b/transforms/language/lang_id/ray/pyproject.toml index 80cbb47ff..c861934d5 100644 --- a/transforms/language/lang_id/ray/pyproject.toml +++ b/transforms/language/lang_id/ray/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_lang_id_transform_ray" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "Language Identification Ray Transform" license = {text = "Apache-2.0"} @@ -9,7 +9,7 @@ authors = [ { name = "Daiki Tsuzuku", email = "dtsuzuku@jp.ibm.com" } ] dependencies = [ - "dpk-lang_id-transform-python==0.2.0.dev6", + "dpk-lang_id-transform-python==0.2.0.dev7", "data-prep-toolkit-ray==0.2.0.dev6" ] diff --git a/transforms/universal/doc_id/kfp_ray/doc_id_wf.py b/transforms/universal/doc_id/kfp_ray/doc_id_wf.py index 25c5d779f..da885261f 100644 --- a/transforms/universal/doc_id/kfp_ray/doc_id_wf.py +++ b/transforms/universal/doc_id/kfp_ray/doc_id_wf.py @@ -17,12 +17,12 @@ from workflow_support.compile_utils import ONE_HOUR_SEC, ONE_WEEK_SEC, ComponentUtils -task_image = "quay.io/dataprep1/data-prep-kit/doc_id-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/doc_id-ray:0.2.0.dev7" # the name of the job script EXEC_SCRIPT_NAME: str = "doc_id_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/universal/ededup/kfp_ray/ededup_wf.py b/transforms/universal/ededup/kfp_ray/ededup_wf.py index 6ad1e323e..218c2aaf4 100644 --- a/transforms/universal/ededup/kfp_ray/ededup_wf.py +++ b/transforms/universal/ededup/kfp_ray/ededup_wf.py @@ -18,13 +18,13 @@ from workflow_support.compile_utils import ONE_HOUR_SEC, ONE_WEEK_SEC, ComponentUtils -task_image = "quay.io/dataprep1/data-prep-kit/ededup-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/ededup-ray:0.2.0.dev7" # the name of the job script EXEC_SCRIPT_NAME: str = "ededup_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/universal/fdedup/kfp_ray/fdedup_wf.py b/transforms/universal/fdedup/kfp_ray/fdedup_wf.py index df50a4af3..6d6052d23 100644 --- a/transforms/universal/fdedup/kfp_ray/fdedup_wf.py +++ b/transforms/universal/fdedup/kfp_ray/fdedup_wf.py @@ -18,13 +18,13 @@ from workflow_support.compile_utils import ONE_HOUR_SEC, ONE_WEEK_SEC, ComponentUtils -task_image = "quay.io/dataprep1/data-prep-kit/fdedup-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/fdedup-ray:0.2.0.dev7" # the name of the job script EXEC_SCRIPT_NAME: str = "fdedup_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/universal/filter/kfp_ray/filter_wf.py b/transforms/universal/filter/kfp_ray/filter_wf.py index 9de907c55..53c285a3c 100644 --- a/transforms/universal/filter/kfp_ray/filter_wf.py +++ b/transforms/universal/filter/kfp_ray/filter_wf.py @@ -21,10 +21,10 @@ EXEC_SCRIPT_NAME: str = "filter_transform_ray.py" PREFIX: str = "" -task_image = "quay.io/dataprep1/data-prep-kit/filter-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/filter-ray:0.2.0.dev7" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/universal/filter/python/pyproject.toml b/transforms/universal/filter/python/pyproject.toml index 42ad8ea1c..f0b582df7 100644 --- a/transforms/universal/filter/python/pyproject.toml +++ b/transforms/universal/filter/python/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_filter_transform_python" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "Filter Transform for Python" license = {text = "Apache-2.0"} @@ -9,7 +9,7 @@ authors = [ { name = "Constantin Adam", email = "cmadam@us.ibm.com" }, ] dependencies = [ - "data-prep-toolkit==0.2.0.dev6", + "data-prep-toolkit==0.2.0.dev7", "duckdb==0.10.1", ] diff --git a/transforms/universal/filter/ray/pyproject.toml b/transforms/universal/filter/ray/pyproject.toml index 4773cbcc5..7f5edd343 100644 --- a/transforms/universal/filter/ray/pyproject.toml +++ b/transforms/universal/filter/ray/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_filter_transform_ray" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "Filter Transform for Ray" license = {text = "Apache-2.0"} @@ -9,8 +9,8 @@ authors = [ { name = "Constantin Adam", email = "cmadam@us.ibm.com" }, ] dependencies = [ - "dpk-filter-transform-python==0.2.0.dev6", - "data-prep-toolkit-ray==0.2.0.dev6", + "dpk-filter-transform-python==0.2.0.dev7", + "data-prep-toolkit-ray==0.2.0.dev7", ] [build-system] diff --git a/transforms/universal/noop/kfp_ray/noop_multiple_wf.py b/transforms/universal/noop/kfp_ray/noop_multiple_wf.py index 1aa18fd1a..1b0ede6f6 100644 --- a/transforms/universal/noop/kfp_ray/noop_multiple_wf.py +++ b/transforms/universal/noop/kfp_ray/noop_multiple_wf.py @@ -17,13 +17,13 @@ from workflow_support.compile_utils import ONE_HOUR_SEC, ONE_WEEK_SEC, ComponentUtils -task_image = "quay.io/dataprep1/data-prep-kit/noop-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/noop-ray:0.2.0.dev7" # the name of the job script EXEC_SCRIPT_NAME: str = "noop_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/universal/noop/kfp_ray/noop_wf.py b/transforms/universal/noop/kfp_ray/noop_wf.py index 600e3146c..af6a4b8ef 100644 --- a/transforms/universal/noop/kfp_ray/noop_wf.py +++ b/transforms/universal/noop/kfp_ray/noop_wf.py @@ -17,13 +17,13 @@ from workflow_support.compile_utils import ONE_HOUR_SEC, ONE_WEEK_SEC, ComponentUtils -task_image = "quay.io/dataprep1/data-prep-kit/noop-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/noop-ray:0.2.0.dev7" # the name of the job script EXEC_SCRIPT_NAME: str = "noop_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/universal/noop/python/pyproject.toml b/transforms/universal/noop/python/pyproject.toml index cdf9c2740..62cccd5db 100644 --- a/transforms/universal/noop/python/pyproject.toml +++ b/transforms/universal/noop/python/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_noop_transform_python" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "NOOP Python Transform" license = {text = "Apache-2.0"} @@ -10,7 +10,7 @@ authors = [ { name = "Boris Lublinsky", email = "blublinsky@ibm.com" }, ] dependencies = [ - "data-prep-toolkit==0.2.0.dev6", + "data-prep-toolkit==0.2.0.dev7", ] [build-system] diff --git a/transforms/universal/noop/ray/pyproject.toml b/transforms/universal/noop/ray/pyproject.toml index bc757abe9..8fbd3deb3 100644 --- a/transforms/universal/noop/ray/pyproject.toml +++ b/transforms/universal/noop/ray/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_noop_transform_ray" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "NOOP Ray Transform" license = {text = "Apache-2.0"} @@ -10,8 +10,8 @@ authors = [ { name = "Boris Lublinsky", email = "blublinsky@ibm.com" }, ] dependencies = [ - "dpk-noop-transform-python==0.2.0.dev6", - "data-prep-toolkit-ray==0.2.0.dev6", + "dpk-noop-transform-python==0.2.0.dev7", + "data-prep-toolkit-ray==0.2.0.dev7", ] [build-system] diff --git a/transforms/universal/noop/spark/pyproject.toml b/transforms/universal/noop/spark/pyproject.toml index 5ac47b3da..697d0222a 100644 --- a/transforms/universal/noop/spark/pyproject.toml +++ b/transforms/universal/noop/spark/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_noop_transform_spark" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "NOOP Spark Transform" license = {text = "Apache-2.0"} @@ -10,8 +10,8 @@ authors = [ { name = "Boris Lublinsky", email = "blublinsky@ibm.com" }, ] dependencies = [ - "dpk-noop-transform-python==0.2.0.dev6", - "data-prep-toolkit-spark==0.2.0.dev6", + "dpk-noop-transform-python==0.2.0.dev7", + "data-prep-toolkit-spark==0.2.0.dev7", ] [build-system] diff --git a/transforms/universal/profiler/kfp_ray/profiler_wf.py b/transforms/universal/profiler/kfp_ray/profiler_wf.py index 73fa2c881..e2391458b 100644 --- a/transforms/universal/profiler/kfp_ray/profiler_wf.py +++ b/transforms/universal/profiler/kfp_ray/profiler_wf.py @@ -24,7 +24,7 @@ EXEC_SCRIPT_NAME: str = "profiler_transform_ray.py" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files component_spec_path = "../../../../kfp/kfp_ray_components/" diff --git a/transforms/universal/tokenization/kfp_ray/tokenization_wf.py b/transforms/universal/tokenization/kfp_ray/tokenization_wf.py index f147df186..2cdf51795 100644 --- a/transforms/universal/tokenization/kfp_ray/tokenization_wf.py +++ b/transforms/universal/tokenization/kfp_ray/tokenization_wf.py @@ -20,10 +20,10 @@ # the name of the job script EXEC_SCRIPT_NAME: str = "tokenization_transform_ray.py" -task_image = "quay.io/dataprep1/data-prep-kit/tokenization-ray:0.2.0.dev6" +task_image = "quay.io/dataprep1/data-prep-kit/tokenization-ray:0.2.0.dev7" # components -base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev6" +base_kfp_image = "quay.io/dataprep1/data-prep-kit/kfp-data-processing:0.2.0.dev7" # path to kfp component specifications files # path to kfp component specifications files diff --git a/transforms/universal/tokenization/python/pyproject.toml b/transforms/universal/tokenization/python/pyproject.toml index c5cb72346..429f609da 100644 --- a/transforms/universal/tokenization/python/pyproject.toml +++ b/transforms/universal/tokenization/python/pyproject.toml @@ -1,7 +1,7 @@ [project] name = "dpk_tokenization_transform_python" keywords = ["tokenizer", "data", "data preprocessing", "data preparation", "llm", "generative", "ai", "fine-tuning", "llmapps" ] -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "Tokenization Transform for Python" license = {text = "Apache-2.0"} @@ -10,7 +10,7 @@ authors = [ { name = "Xuan-Hong Dang", email = "xuan-hong.dang@ibm.com"}, ] dependencies = [ - "data-prep-toolkit==0.2.0.dev6", + "data-prep-toolkit==0.2.0.dev7", "transformers==4.38.0", ] diff --git a/transforms/universal/tokenization/ray/pyproject.toml b/transforms/universal/tokenization/ray/pyproject.toml index e09e64932..59f130ea6 100644 --- a/transforms/universal/tokenization/ray/pyproject.toml +++ b/transforms/universal/tokenization/ray/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "dpk_tokenization_transform_ray" -version = "0.2.0.dev6" +version = "0.2.0.dev7" requires-python = ">=3.10" description = "Tokenization Transform for Ray" license = {text = "Apache-2.0"} @@ -9,8 +9,8 @@ authors = [ { name = "Xuan-Hong Dang", email = "xuan-hong.dang@ibm.com"}, ] dependencies = [ - "dpk-tokenization-transform-python==0.2.0.dev6", - "data-prep-toolkit-ray==0.2.0.dev6", + "dpk-tokenization-transform-python==0.2.0.dev7", + "data-prep-toolkit-ray==0.2.0.dev7", ] [build-system]