-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Plans for converting the data #2
Comments
Yes, why not. Let's call it
From my point of view we don't need any folder structure because there will be only three files (see the very first post here: #1).
Yes! Excellent suggestion!
As said above: in terms of files see the very first post here: #1 I am not sure about the table. I don't think @joshmoore has something yet ready to store the table in zarr format?! And ❤️ for helping! |
Ok, in that case I would just add a single root zarr file with three multiscale datasets:
We could just store it as a 2d dataset with column names in the header, but I think there is indeed not a NGF format for tables yet. Anyway, I will start with the volumetric data and let you know once I have something. (I will probably just start with the myosin volume, so @joshmoore can check it out once I have put it on the bucket and after we make sure the format is correct we add the larger files). |
Related to this: #3 If we want to use the MoBIE infrastructure the most straightforward would be if there would be somewhere an |
If we do this there are a few questions about the file layout, because we cannot simply use what I suggested here, because bdv assumes fixed paths inside the dataset ( I see three options:
|
I don't think so.
There is some work now on an initial format:
which briefly looks like this:
Also ok. |
Ok, let's discuss the layout tomorrow in the meeting.
This will produce large jsons in our case :). But we can give it a try; and in the future we can hopefully switch to storing the table as a zarr array. |
For the testing, you could just write one feature value, like |
Personally, if I would like to get something working within one week until i2k, I would do the following:
|
Yup. Definitely aware. I had tried the zarr array solution but ran into saalfeldlab/n5#73 (comment) Also discussed possible integrate with Parquet etc last night on the community call. Open to thoughts. |
@tischi your plan sounds good. I can def. set up 1. :). Will try to do as much as possible there before the meeting tomorrow and then we can finalize the plan before i2k. |
@joshmoore I uploaded one multiscale dataset to our new bucket. Could you please check that you can access it?
If you can access it, can you check if the dataset at Thanks! |
Hi @constantinpape, The .zattrs that's in
|
Thanks for checking! |
I added the data according to what we discussed, see #4 |
@tischi, I wrote a couple of mails with @joshmoore today and as far as I understand the current plan is the following:
We don't ship the data to josh and instead convert and upload it locally.
I have a converter script and I am pretty sure it does the right thing, but I have a couple of other questions:
P.S I made a new issue because #1 got a bit crowded.
The text was updated successfully, but these errors were encountered: