Skip to content

Commit

Permalink
Merge pull request #1652 from microsoft/simonz/spark-splitters
Browse files Browse the repository at this point in the history
Remove unused imports and parameters in spark_splitters.py
  • Loading branch information
miguelgfierro authored Feb 28, 2022
2 parents 6c53f92 + 1d7c519 commit de4210f
Showing 1 changed file with 0 additions and 4 deletions.
4 changes: 0 additions & 4 deletions recommenders/datasets/spark_splitters.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,6 @@
DEFAULT_ITEM_COL,
DEFAULT_USER_COL,
DEFAULT_TIMESTAMP_COL,
DEFAULT_RATING_COL,
)
from recommenders.datasets.split_utils import (
process_split_ratio,
Expand Down Expand Up @@ -161,7 +160,6 @@ def spark_chrono_split(
training data set; if it is a list of float numbers, the splitter splits
data into several portions corresponding to the split ratios. If a list is
provided and the ratios are not summed to 1, they will be normalized.
seed (int): Seed.
min_rating (int): minimum number of ratings for user or item.
filter_by (str): either "user" or "item", depending on which of the two is to filter
with min_rating.
Expand Down Expand Up @@ -193,7 +191,6 @@ def spark_stratified_split(
filter_by="user",
col_user=DEFAULT_USER_COL,
col_item=DEFAULT_ITEM_COL,
col_rating=DEFAULT_RATING_COL,
seed=42,
):
"""Spark stratified splitter.
Expand All @@ -216,7 +213,6 @@ def spark_stratified_split(
with min_rating.
col_user (str): column name of user IDs.
col_item (str): column name of item IDs.
col_rating (str): column name of ratings.
Returns:
list: Splits of the input data as pyspark.sql.DataFrame.
Expand Down

0 comments on commit de4210f

Please sign in to comment.