File tree Expand file tree Collapse file tree 9 files changed +11
-25
lines changed
Expand file tree Collapse file tree 9 files changed +11
-25
lines changed Original file line number Diff line number Diff line change 1919from pydeequ .analyzers import AnalysisRunner
2020from pydeequ .checks import Check , CheckLevel
2121
22- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3"
22+ # deequ_maven_coord = "com.amazon.deequ:deequ:1.1.0_spark-2.4-scala-2.11"
23+ deequ_maven_coord = "com.amazon.deequ:deequ:1.1.0_spark-3.0-scala-2.12"
2324f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all"
2425
2526class PyDeequSession :
Original file line number Diff line number Diff line change 11import unittest
22from pyspark .sql import SparkSession , Row , DataFrame
33from pydeequ .analyzers import *
4- from pydeequ import PyDeequSession
4+ from pydeequ import *
55
66class TestAnalyzers (unittest .TestCase ):
77
88 @classmethod
99 def setUpClass (cls ):
10- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3" # TODO: get Maven Coord from Configs
11- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
1210 cls .spark = (SparkSession
1311 .builder
1412 .master ('local[*]' )
Original file line number Diff line number Diff line change 44from pydeequ .anomaly_detection import *
55from pydeequ .repository import *
66from pydeequ .analyzers import *
7+ from pydeequ import *
78
89class TestAnomalies (unittest .TestCase ):
910 @classmethod
1011 def setUpClass (cls ):
11- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3"
12- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
1312 cls .spark = (SparkSession
1413 .builder
1514 .master ('local[*]' )
Original file line number Diff line number Diff line change 22from pyspark .sql import SparkSession , Row , DataFrame
33from pydeequ .verification import *
44from pydeequ .checks import *
5- import py4j
5+ from pydeequ import *
66
77class TestChecks (unittest .TestCase ):
88 @classmethod
99 def setUpClass (cls ):
10- # TODO share spark context between test cases?
11- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3" # TODO get Maven Coord from Configs
12- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
1310 cls .spark = (SparkSession
1411 .builder
1512 .master ('local[*]' )
Original file line number Diff line number Diff line change 66from pydeequ .profiles import ColumnProfilerRunBuilder , ColumnProfilerRunner
77from pydeequ .verification import *
88from pydeequ .checks import *
9- from pydeequ import PyDeequSession
9+ from pydeequ import *
1010from pandas import DataFrame as pandasDF
1111import numpy as np
1212
1313class TestPandasUtils (unittest .TestCase ):
1414
1515 @classmethod
1616 def setUpClass (cls ):
17- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3" # TODO: get Maven Coord from Configs
18- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
1917 cls .spark = (SparkSession
2018 .builder
2119 .master ('local[*]' )
Original file line number Diff line number Diff line change 22from pydeequ .profiles import ColumnProfilerRunBuilder , ColumnProfilerRunner
33from pydeequ .analyzers import KLLParameters
44from pyspark .sql import SparkSession , Row
5-
5+ from pydeequ import *
66
77class TestProfiles (unittest .TestCase ):
88 @classmethod
99 def setUpClass (cls ):
10- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3"
11- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
1210 cls .spark = (SparkSession
1311 .builder
1412 .master ('local[*]' )
Original file line number Diff line number Diff line change 44from pydeequ .repository import *
55from pydeequ .verification import *
66from pydeequ .checks import *
7+ from pydeequ import *
78
89
910class TestRepository (unittest .TestCase ):
1011 @classmethod
1112 def setUpClass (cls ):
12- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3"
13- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
1413 cls .spark = (SparkSession
1514 .builder
1615 .master ('local[*]' )
Original file line number Diff line number Diff line change 11import unittest
22from pydeequ .scala_utils import ScalaFunction1 , ScalaFunction2
33from pyspark .sql import SparkSession
4-
4+ from pydeequ import *
55
66class TestScalaUtils (unittest .TestCase ):
77 @classmethod
88 def setUpClass (cls ):
9- # TODO share spark context between test cases?
10- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3-rc2" # TODO get Maven Coord from Configs
11- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
9+ # deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3-rc2" # TODO This ran rc2?
1210 cls .spark = (SparkSession
1311 .builder
1412 .master ('local[*]' )
Original file line number Diff line number Diff line change 22from pyspark .sql import SparkSession , Row , DataFrame
33from pydeequ .suggestions import *
44import json
5-
5+ from pydeequ import *
66
77class TestSuggestions (unittest .TestCase ):
88 @classmethod
99 def setUpClass (cls ):
10- deequ_maven_coord = "com.amazon.deequ:deequ:1.0.3"
11- f2j_maven_coord = "net.sourceforge.f2j:arpack_combined_all" # This package is excluded because it causes an error in the SparkSession fig
1210 cls .spark = (SparkSession
1311 .builder
1412 .master ('local[*]' )
You can’t perform that action at this time.
0 commit comments