Skip to content

Latest commit

 

History

History
194 lines (153 loc) · 10.1 KB

README.md

File metadata and controls

194 lines (153 loc) · 10.1 KB

Spark-Excel

Spark Excel CI codecov

A Spark data source for reading Microsoft Excel workbooks. Initially started to "scratch and itch" and to learn how to write data sources using the Spark DataSourceV2 APIs. This is based on the Apache POI library which provides the means to read Excel files.

N.B. This project is only intended as a reader and is opinionated about this. There are no plans to implement writer functionality into the project.

Important

This repository makes use of Git-LFS to track the Excel files for unit tests. This allows for larger files to be uploaded as part of the tests without bloating the repository. If cloning the code you will need to ensure you have Git-LFS installed and available.

Download

Check out the releases page for the latest jar files to match the version of Spark you're developing against.

Features

  • Handling Excel 97-2003, 2010, and OOXML files (thanks to Apache POI)
  • Multi-line headers
  • Reading from multiple worksheets given a name pattern
  • Glob pattern support for reading multiple files
  • Outputting the sheet name in the result set (set via an option)
  • Schema inference
  • Cleans column names on read to remove invalid characters which can cause issues in Spark (spaces, periods etc...)
  • Handle merged cells (repeats data to all cells in the merged region)
  • Formula evaluation (for those supported by Apache POI)
  • Schema matching validation in output (set via an option)
  • Works in Scala, PySpark, and Spark SQL

Usage

Standard usage of the library works as follows:

Scala

val df = spark.read
  .format("com.elastacloud.spark.excel")
  .option("cellAddress", "A1")
  .load("/path/to/my_file.xlsx")

A short name has been provided for convenience, as well as convenience method (Scala only currently).

val df = spark.read
  .format("excel")
  .load("/path/to/my_file.xlsx")
val df = spark.read
  .excel("/path/to/my_file.xlsx")

Python

df = (spark.read
    .format("com.elastacloud.spark.excel")
    .option("cellAddress", "A1")
    .load("/path/to/my_file.xlsx")
)

SQL

Accessing Excel files directly (N.B. this style of access requires back-ticks instead of single-quotes)

SELECT
  *
FROM
  excel.`/path/to/my_file.xlsx`

Accessing using options by creating a view

CREATE TEMPORARY VIEW sample_excel
USING excel
OPTIONS (
  path '/path/to/my_file.xlsx',
  schemaMatchColumnName 'is_value'
);


SELECT * FROM sample_excel;

All of these methods accept glob patterns and multiple path values.

// Providing multiple paths
val multiFileDF = spark.read
  .format("com.elastacloud.spark.excel")
  .excel("/path/to/my_file.xlsx", "/path/to/my_second_file.xlsx")

// Using glob patterns
val globFileDF = spark.read
  .format("com.elastacloud.spark.excel")
  .excel("/landing/20*/*.xlsx")

Options

The library supports the following options:

Option Type Default Description
cellAddress String A1 Location of the first cell of the table (including header)
headerRowCount Int 1 Number of rows which make up the header. If no header is available then set this value to 0 (zero)
includeSheetName Boolean False Includes the name of the worksheet the data has come from when set to true. Uses the column _SheetName
workbookPassword String Empty Password required to open Excel workbook
sheetNamePattern String Empty Regular expression to use to match worksheet names
maxRowCount Int 1000 Number of records to read to infer the schema. If set to 0 (zero) then all available rows will be read
nullValue String Sets the value to be treated as null, e.g. "NA"
maxBytesForTempFiles Int 10000000 Sets the number of bytes at which a workbook is (ooxml format) is regarded as too large to hold in memory and the data is put into temp files instead. Whilst the cluster may have large volumes of memory, the node processing the file will be limited.
thresholdBytesForTempFiles Int 10000000 Alias for maxBytesForTempFiles
schemaMatchColumnName Boolean False Defines the column name to write the flag indicating if the current record matches the provided or inferred schema. If the schema is provided then the column name must exist within that provided schema.
evaluateFormulae Boolean True Instructs the parser to evaluate formulas in the source (True), or to extract only the formula itself (False, e.g. A7*2).
val df = spark.read
  .format("com.elastacloud.spark.excel")
  .option("cellAddress", "C3") // The first line of the table starts at cell C3
  .option("headerRowCount", 2) // The first 2 lines of the table make up the header row
  .option("includeSheetName", value = true) // Include the sheet name(s) the data has come from
  .option("workbookPassword", "AP@55w0rd") // Use this password to open the workbook with
  .option("sheetNamePattern", """Sheet[13]""") // Read data from all sheets matching this pattern (e.g. Sheet1 and Sheet3)
  .option("maxRowCount", 10) // Read only the first 10 records to determine the schema of the data
  .option("thresholdBytesForTempFiles", 50000000) // Set size limit before temp files are used
  .option("schemaMatchColumnName", "_isValid") // Write a flag to the '_isValid' column indicating if the records matches the schema
  .option("evaluateFormulae", "true") // Evaluate the formula in the spreadsheet
  .load("/path/to/file.xlsx")

Multi row headers

When reading multiple rows as a header the parser attempts to combine the content of all cells in that column (including merged regions) to create an overall name. For example.

+--------+--------------------+
|        | 2021               |
+--------+--------------------+
| Id     | Target   | Actual  |
+--------+--------------------+
| 1      | 100      | 98      |
| 2      | 200      | 230     |
+--------+----------+---------+

In this case the resulting DataFrame would look as follows.

+--------+--------------+-------------+
| Id     | 2021_Target  | 2021_Actual |
+--------+----------------------------+
| 1      | 100          | 98          |
| 2      | 200          | 230         |
+--------+--------------+-------------+

Sheet name patterns

Often when reading in data from Excel we'll find that there are multiple sheets containing the same shape data, but have been split apart to add context, such as the month, year, department name and so on. To avoid having to add multiple reads of the same file to access difference sheets, the data source provides the sheetNamePattern option, which takes a regular expression. This pattern evaluates for any sheets matching the pattern which are then included in the result set.

If, for example, you wanted to include sheet names for the next 3 years (assuming the year is currently 2021) then you could provide a pattern of 202[234]. This would ignore any sheets for other years, summary sheets etc...

Often when using this option it is useful to include the sheet name so that the original context is not lost, in the same way you might want to include the source file name when reading from multiple files. And so the code might read like.

val df = spark.read
  .format("com.elastacloud.spark.excel")
  .option("sheetNamePattern", "202[234]")
  .option("includeSheetName", "true")
  .load("/path/to/my_file.xlsx")