Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@
./soda/spark[hive]
./soda/spark[odbc]
./soda/spark[databricks]
./soda/spark_df
./soda/spark_df[pyspark]
./soda/scientific[simulator]
./soda/sqlserver
./soda/mysql
Expand Down
4 changes: 3 additions & 1 deletion soda/spark_df/setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,12 +8,14 @@

requires = [
f"soda-core-spark=={package_version}",
"pyspark>=3.4.0",
]
extras = {"pyspark": ["pyspark>=3.4.0"]}

# TODO Fix the params
setup(
name=package_name,
version=package_version,
install_requires=requires,
extras_require=extras,
packages=find_namespace_packages(include=["soda*"]),
)
8 changes: 6 additions & 2 deletions soda/spark_df/soda/data_sources/spark_df_connection.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,13 @@
from pyspark.sql.session import SparkSession
from typing import TYPE_CHECKING

from soda.data_sources.spark_df_cursor import SparkDfCursor

if TYPE_CHECKING:
from pyspark.sql.session import SparkSession


class SparkDfConnection:
def __init__(self, spark_session: SparkSession):
def __init__(self, spark_session: "SparkSession"):
self.spark_session = spark_session

def cursor(self) -> SparkDfCursor:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,8 @@

import logging
import re
from typing import TYPE_CHECKING

from pyspark.sql import SparkSession
from soda.data_sources.spark_df_connection import SparkDfConnection
from soda.execution.data_type import DataType

Expand All @@ -13,9 +13,11 @@

logger = logging.getLogger(__name__)

if TYPE_CHECKING:
from pyspark.sql import SparkSession

class SparkDfSqlDialect(SqlDialect):

class SparkDfSqlDialect(SqlDialect):
def __init__(self):
super().__init__()

Expand Down Expand Up @@ -87,7 +89,6 @@ def regex_replace_flags(self) -> str:


class SparkDfContractDataSource(FileClContractDataSource):

def __init__(self, data_source_yaml_file: YamlFile, spark_session: SparkSession):
data_source_yaml_dict: dict = data_source_yaml_file.get_dict()
data_source_yaml_dict[self._KEY_TYPE] = "spark_df"
Expand Down
7 changes: 5 additions & 2 deletions soda/spark_df/soda/data_sources/spark_df_cursor.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,10 @@
from __future__ import annotations

from pyspark.sql import DataFrame, SparkSession
from pyspark.sql.types import Row
from typing import TYPE_CHECKING

if TYPE_CHECKING:
from pyspark.sql import DataFrame, SparkSession
from pyspark.sql.types import Row


class SparkDfCursor:
Expand Down
Loading