-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Regarding the missing site info in the datasets #81
Comments
I don't remember the exact query we made...nor do I know why these are Am 2016-07-13 13:02, schrieb C. Eren Sezener:
|
The hourly DB sites are quite fucked up: >>> h5 = h5py.File('hourly_database.hdf5', 'r'); data = h5['weather_data'][:]
>>> np.unique(data[:,2])
array([ 0.00000000e+00, 1.00000000e+00, 4.00000000e+00,
2.01606212e+11]) So we have sites 1 and 4 and a date (WTF?) |
The DB should be essentially the same. You can check it yourself. |
I am running all the scrapers again such that the outputs will be written to different DBs. Then all scraper authors must review their data since the DB will consist of only their own data. |
If you want scraper authors to review their DB, please provide a clear code snippet, explaining how to access the data or use the query engine and where which data should be stored. |
I have been told by @ge00rg and @clauslang that the data is only from one site in the daily DB. I don't see that this is the case, see the below snippet. But why are there sites like 0.5, 1.5 etc.? Is this expected?
The text was updated successfully, but these errors were encountered: