From f7e4fd1e1e3ed8e10ba37a7248302c25305400ab Mon Sep 17 00:00:00 2001 From: tomer-mamia <125267619+tomerm-iguazio@users.noreply.github.com> Date: Tue, 13 Feb 2024 17:45:49 +0200 Subject: [PATCH] [Demos] Change to local=False in batch infer v2 demo. (#790) --- batch_inference_v2/batch_inference_v2.ipynb | 11 +---------- 1 file changed, 1 insertion(+), 10 deletions(-) diff --git a/batch_inference_v2/batch_inference_v2.ipynb b/batch_inference_v2/batch_inference_v2.ipynb index be7a7722a..bb59221fd 100644 --- a/batch_inference_v2/batch_inference_v2.ipynb +++ b/batch_inference_v2/batch_inference_v2.ipynb @@ -478,10 +478,7 @@ "# Import the `batch_inference_v2` function from the functions hub:\n", "batch_inference_function = mlrun.import_function('hub://batch_inference_v2')\n", "# you can import the function from the current directory as well: \n", - "# batch_inference_function = mlrun.import_function(\"function.yaml\")\n", - "\n", - "# Set the desired artifact path:\n", - "artifact_path = \"./\"" + "# batch_inference_function = mlrun.import_function(\"function.yaml\")\n" ] }, { @@ -1448,23 +1445,18 @@ "# 1. Generate data:\n", "generate_data_run = demo_function.run(\n", " handler=\"generate_data\",\n", - " artifact_path=artifact_path,\n", " returns=[\"training_set : dataset\", \"prediction_set : dataset\"],\n", - " local=True,\n", ")\n", "\n", "# 2. Train a model:\n", "train_run = demo_function.run(\n", " handler=\"train\",\n", - " artifact_path=artifact_path,\n", " inputs={\"training_set\": generate_data_run.outputs[\"training_set\"]},\n", - " local=True,\n", ")\n", "\n", "# 3. Perform batch prediction:\n", "batch_inference_run = batch_inference_function.run(\n", " handler=\"infer\",\n", - " artifact_path=artifact_path,\n", " inputs={\"dataset\": generate_data_run.outputs[\"prediction_set\"]},\n", " params={\n", " \"model_path\": train_run.outputs[\"model\"],\n", @@ -1474,7 +1466,6 @@ " \"model_endpoint_drift_threshold\": 0.2,\n", " \"model_endpoint_possible_drift_threshold\": 0.1,\n", " },\n", - " local=True,\n", ")" ] },