You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have started to work on providing a new datasource, targeting semi structured data residing on a distributed file system (HDFS/S3) such as:
csv
json
xml
I have started to work on that by designing a spark API imitating the functionality offered by the already existing datasources.
I would be interested to know if people would appreciate the addition of such a feature.
I also have some ideas regarding the implementation, specifically I am now adding functionality in a new
module under connectors/
Any input is valuable at this point!
The text was updated successfully, but these errors were encountered:
steremma
changed the title
proposed feature: HDFS interface
[feature] HDFS interface
Mar 25, 2017
You'll need an engine that can deserialize, aggregate & filter data as you implement the interface, and Superset isn't the place to do that. Depending on your serde you may want to write a deserializer for Presto or something like Apache Drill.
I have started to work on providing a new datasource, targeting semi structured data residing on a distributed file system (HDFS/S3) such as:
I have started to work on that by designing a spark API imitating the functionality offered by the already existing datasources.
I would be interested to know if people would appreciate the addition of such a feature.
I also have some ideas regarding the implementation, specifically I am now adding functionality in a new
module under connectors/
Any input is valuable at this point!
The text was updated successfully, but these errors were encountered: