Skip to content

A standalone web service that pushes data into the CKAN Datastore fast & reliably. It pushes real good!

License

Notifications You must be signed in to change notification settings

dathere/datapusher-plus

Repository files navigation

DataPusher+

DataPusher+ is a fork of Datapusher that combines the speed and robustness of ckanext-xloader with the data type guessing of Datapusher.

Datapusher+ is droping usage of CKAN Service Provider, with Messytables replaced by qsv.

TNRIS/TWDB provided the use cases that informed and supported the development of Datapusher+, specifically, to support a Resource-first upload workflow.

For a more detailed overview, see the CKAN Monthly Live Jan 2023 presentation.

It features:

  • "Bullet-proof", ultra-fast data type inferencing with qsv

    Unlike Messytables which scans only the the first few rows to guess the type of a column, qsv scans the entire table so its data type inferences are guaranteed1.

    Despite this, qsv is still exponentially faster even if it scans the whole file, not only inferring data types, it also calculates summary statistics as well. For example, scanning a 2.7 million row, 124MB CSV file for types and stats took 0.16 seconds2.

    It is very fast as qsv is written in Rust, is multithreaded, and uses all kinds of performance techniques especially designed for data-wrangling.

  • Exponentially faster loading speed

    Similar to xloader, we use PostgreSQL COPY to directly pipe the data into the datastore, short-circuiting the additional processing/transformation/API calls used by Datapusher Plus.

    But unlike xloader, we load everything using the proper data types and not as text, so there's no need to reload the data again after adjusting the Data Dictionary, as you would with xloader.

  • Far more Storage Efficient AND Performant Datastore with easier to compose SQL queries

    As we create the Datastore tables using the most efficient PostgreSQL data type for each column using qsv's guaranteed type inferences - the Datastore is not only more storage efficient, it is also far more more performant for loading AND querying.

    With its "smartint" data type (with qsv inferring the most efficient integer data type for the range of values in the column); comprehensive date format inferencing (supporting 19 date formats, with each format having several variants & with configurable DMY/MDY preference parsing) & auto-formatting dates to RFC3339 format so they are stored as Postgres timestamps; cardinality-aware, configurable auto-indexing; automatic sanitization of column names to valid PostgreSQL column identifiers; auto PostgreSQL vacuuming & analysis of resources after loading; and more - DP+ enables the Datastore to tap into PostgreSQL's full power.

    Configurable auto-aliasing of resources also makes it easier to compose SQL queries, as you can use more intuitive resource aliases instead of cryptic resource IDs.

  • Production-ready Robustness

    In production, the number one source of support issues is Datapusher - primarily, because of data quality issues and Datapusher's inability to correctly infer data types, gracefully handle errors3, and provide the Data Publisher actionable information to correct the data.

    Datapusher+'s design directly addresses all these issues.

  • More informative datastore loading messages

    Datapusher+ messages are designed to be more verbose and actionable, so the data publisher's user experience is far better and makes it possible to have a resource-first upload workflow.

  • Extended preprocessing with qsv

    qsv is leveraged by Datapusher+ to:

    • create "Smarter" Data Dictionaries, with:
      • guaranteed data type inferences
      • optional ability to automatically choose the best integer PostgreSQL data type ("smartint") based on the range of the numeric column (PostgreSQL's int, bigint and numeric types) for optimal storage/indexing efficiency and SQL query performance.
      • sanitized column names (guaranteeing valid PostgreSQL column identifiers) while preserving the original column name as a label, which is used to label columns in DataTables_view.
      • an optional "summary stats" resource as an extension of the Data Dictionary, with comprehensive summary statistics for each column - sum, min/max/range, min/max length, mean, stddev, variance, nullcount, sparsity, quartiles, IQR, lower/upper fences, skewness, median, mode/s, antimode/s & cardinality.
    • convert Excel & OpenOffice/LibreOffice Calc (ODS) files to CSV, with the ability to choose which sheet to use by default (e.g. 0 is the first sheet, -1 is the last sheet, -2 the second to last sheet, etc.)
    • convert various date formats (19 date formats are recognized with each format having several variants; ~80 date format permutations in total) to a standard RFC 3339 format
    • enable random access of a CSV by creating a CSV index - which also enables parallel processing of different parts of a CSV simultaneously (a major reason type inferencing and stats calculation is so fast)
    • instantaneously count the number of rows with a CSV index
    • validate if an uploaded CSV conforms to the RFC-4180 standard
    • normalizes and transcodes CSV/TSV dialects into a standard UTF-8 encoded RFC-4180 CSV format
    • optionally create a preview subset, with the ability to only download the first n preview rows of a file, and not the entire file (e.g. only download first 1,000 rows of 3 gb CSV file - especially good for harvesting/cataloging external sites where you only want to harvest the metadata and a small sample of each file).
    • optionally create a preview subset from the end of a file (e.g. last 1,000 rows, good for time-series/sensor data)
    • auto-index columns based on its cardinality/format (unique indices created for columns with all unique values, auto-index columns whose cardinality is below a given threshold; auto-index date columns)
    • check for duplicates, and optionally deduplicate rows
    • optionally screen for Personally Identifiable Information (PII), with an option to "quarantine" the PII-candidate rows in a separate resource, while still creating the screened resource.
    • optional ability to specify a custom PII screening regex set, instead of the default PII screening regex set.

    Even with all these pre-processing tasks, qsv typically takes less than 5 seconds to finish all its analysis tasks, even for a 100mb CSV file.

    Future versions of Datapusher+ will further leverage qsv's 80+ commands to do additional preprocessing, data-wrangling and validation. The Roadmap is available here. Ideas, suggestions and your feedback are most welcome!

Development Installation

Datapusher+ from version 1.0.0 onwards will be installed as a extension of CKAN, and will be available as a CKAN plugin. This will allow for easier integration with CKAN and other CKAN extensions.

  1. Install the required packages.

    sudo apt install python3-virtualenv python3-dev python3-pip python3-wheel build-essential libxslt1-dev libxml2-dev zlib1g-dev git libffi-dev libpq-dev file
  2. Activate the CKAN virtual environment using atleast python 3.8.

    . /usr/lib/ckan/default/bin/activate

    ℹ️ NOTE: DP+ requires at least python 3.8 as it makes extensive use of new capabilities introduced in 3.7/3.8 to the subprocess module.

  3. Install the extension using following commands:

    pip install -e "git+https://github.com/dathere/[email protected]#egg=datapusher-plus"
  4. Install the dependencies.

    pip install -r requirements.txt
  5. Install qsv.

    Manual Installation

    Download the appropriate precompiled binaries for your platform and copy it to the appropriate directory, e.g. for Linux:

    wget https://github.com/jqnatividad/qsv/releases/download/0.87.1/qsv-0.87.1-x86_64-unknown-linux-gnu.zip
    unzip qsv-0.87.1-x86_64-unknown-linux-gnu.zip
    rm qsv-0.87.1-x86_64-unknown-linux-gnu.zip
    sudo mv qsv* /usr/local/bin

    Alternatively, if you want to install qsv from source, follow the instructions here. Note that when compiling from source, you may want to look into the Performance Tuning section to squeeze even more performance from qsv.

    Also, if you get glibc errors when starting qsv, your Linux distro may not have the required version of the GNU C Library (This will be the case when running Ubuntu 18.04 or older). If so, use the unknown-linux-musl.zip archive as it is statically linked with the MUSL C Library.

    If you already have qsv, update it to the latest release by using the --update option.

    qsvdp --update

    ℹ️ NOTE: qsv is a general purpose CSV data-wrangling toolkit that gets regular updates. To update to the latest version, just run qsv with the --update option and it will check for the latest version and update as required.

    Linux Installation

    If you are running Debian based distribution, you can install qsv using the following command:

    Add the qsv repository to your sources list:

    echo "deb [signed-by=/etc/apt/trusted.gpg.d/qsv-deb.gpg] https://tino097.github.io/qsv-deb-releases ./" > qsv.list

    Import trusted GPG key:

    wget -O - https://tino097.github.io/qsv-deb-releases/qsv-deb.gpg | sudo apt-key add -

    Install qsv:

    sudo apt update
    sudo apt install qsv
  6. Configure the Datapusher+ database.

    Make sure to create the datapusher PostgreSQL user and the datapusher_jobs database (see DataPusher+ Database Setup).

Configuring

CKAN Configuration

Add datapusher_plus to the plugins in your CKAN configuration file (generally located at /etc/ckan/default/ckan.ini):

ckan.plugins = <other plugins> datapusher_plus

ℹ️ NOTE: DP+ recognizes some additional TSV and spreadsheet subformats - xlsm and xlsb for Excel Spreadsheets, and tab for TSV files. To process these subformats, set ckan.datapusher.formats as follows in your CKAN.INI file:

ckanext.datapusher_plus.copy_readbuffer_size = 1048576
ckanext.datapusher_plus.max_content_length = 1256000000000
ckanext.datapusher_plus.ignore_file_hash = true
ckanext.datapusher_plus.chunk_size = 16384
ckanext.datapusher_plus.download_timeout = 300
ckanext.datapusher_plus.ssl_verify = false
ckanext.datapusher_plus.download_proxy =
ckanext.datapusher_plus.types = csv xls xlsx tsv application/csv application/vnd.ms-excel application/vnd.openxmlformats-officedocument.spreadsheetml.sheet
ckanext.datapusher_plus.type_mapping = {"String": "text", "Integer": "numeric","Float": "numeric","DateTime": "timestamp","Date": "timestamp","NULL": "text"}
ckanext.datapusher_plus.pii_screening = false
ckanext.datapusher_plus.pii_quick_screen = false
ckanext.datapusher_plus.pii_found_abort = false
ckanext.datapusher_plus.pii_show_candidates = false
ckanext.datapusher_plus.pii_regex_resource_id_or_alias =
ckanext.datapusher_plus.qsv_bin =  /usr/local/bin/qsvdp
ckanext.datapusher_plus.file_bin = /usr/bin/file
ckanext.datapusher_plus.prefer_dmy = false
ckanext.datapusher_plus.preview_rows = 100
ckanext.datapusher_plus.auto_index_threshold = 3
ckanext.datapusher_plus.auto_unique_index = true
ckanext.datapusher_plus.auto_index_dates = true
ckanext.datapusher_plus.sort_and_dupe_check = true
ckanext.datapusher_plus.dedup = false
ckanext.datapusher_plus.default_excel_sheet = 0
ckanext.datapusher_plus.add_summary_stats_resource = false
ckanext.datapusher_plus.summary_stats_options =
ckanext.datapusher_plus.auto_alias = true
ckanext.datapusher_plus.auto_alias_unique = false
ckanext.datapusher_plus.api_token = 

and add this entry to your CKAN's resource_formats.json file.

["TAB", "Tab Separated Values File", "text/tab-separated-values", []],

DataPusher+ Database Setup

DP+ tables will be created with the command:

ckan -c /etc/ckan/default/ckan.ini datapusher_plus db-init

NOTE: If you are upgrading from a previous version of Datapusher, you will need to run the following command to upgrade the database:

ckan -c /etc/ckan/default/ckan.ini db upgrade -p datapusher_plus

Usage

Any file that has one of the supported formats (defined in ckanext.datapusher_plus.formats) will be attempted to be loaded into the DataStore.

You can also manually trigger resources to be resubmitted. When editing a resource in CKAN (clicking the "Manage" button on a resource page), a new tab named "DataStore" will appear. This will contain a log of the last attempted upload and a button to retry the upload. Once a resource has been "pushed" into the Datastore, a "Data Dictionary" tab will also be available where the data pusblisher can fine-tune the inferred data dictionary.

DataPusher+ UI DataPusher+ UI 2

Command line

Run the following command to submit all resources to datapusher, although it will skip files whose hash of the data file has not changed:

    ckan -c /etc/ckan/default/ckan.ini datapusher_plus resubmit

To Resubmit a specific resource, whether or not the hash of the data file has changed:

    ckan -c /etc/ckan/default/ckan.ini datapusher_plus submit {dataset_id}

License

This material is copyright (c) 2020 Open Knowledge Foundation and other contributors

It is open and licensed under the GNU Affero General Public License (AGPL) v3.0 whose full text may be found at:

http://www.fsf.org/licensing/licenses/agpl-3.0.html

Footnotes

  1. Why use qsv instead of a "proper" python data analysis library like pandas?

  2. It takes 0.16 seconds with an index to run qsv stats against the qsv whirlwind tour sample file on a Ryzen 4800H (8 physical/16 logical cores) with 32 gb memory and a 1 TB SSD. Without an index, it takes 1.3 seconds.

  3. Imagine you have a 1M row CSV, and the last row has an invalid value for a numeric column (e.g. "N/A" instead of a number). After spending hours pushing the data very slowly, legacy datapusher will abort on the last row and the ENTIRE job is invalid. Ok, that's bad, but what makes it worse is that the old table has been deleted already, and Datapusher doesn't tell you what caused the job to fail! YIKES!!!!