From adf0494fd1aa790410fb6371489f9c800a66e7e0 Mon Sep 17 00:00:00 2001 From: Maximilian <39695405+maxispeicher@users.noreply.github.com> Date: Sat, 11 Nov 2023 21:40:55 +0100 Subject: [PATCH] Fix disabling validation for PySpark DataFrame Schemas (#1407) * Update container.py Signed-off-by: Maximilian Speicher * fix existing test Signed-off-by: Maximilian Speicher --------- Signed-off-by: Maximilian Speicher --- pandera/api/pyspark/container.py | 2 +- tests/pyspark/test_pyspark_config.py | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/pandera/api/pyspark/container.py b/pandera/api/pyspark/container.py index cef8675c9..880c1dd1b 100644 --- a/pandera/api/pyspark/container.py +++ b/pandera/api/pyspark/container.py @@ -327,7 +327,7 @@ def validate( [Row(product='Bread', price=9), Row(product='Butter', price=15)] """ if not CONFIG.validation_enabled: - return + return check_obj error_handler = ErrorHandler(lazy) return self._validate( diff --git a/tests/pyspark/test_pyspark_config.py b/tests/pyspark/test_pyspark_config.py index 6005c0c16..82edb7231 100644 --- a/tests/pyspark/test_pyspark_config.py +++ b/tests/pyspark/test_pyspark_config.py @@ -44,8 +44,8 @@ class TestSchema(DataFrameModel): } assert CONFIG.dict() == expected - assert pandra_schema.validate(input_df) is None - assert TestSchema.validate(input_df) is None + assert pandra_schema.validate(input_df) + assert TestSchema.validate(input_df) # pylint:disable=too-many-locals def test_schema_only(self, spark, sample_spark_schema):