From 293b16ff92655fb7149ffd1599093279ead6f9d1 Mon Sep 17 00:00:00 2001 From: Florian M Date: Mon, 14 Nov 2022 12:11:51 +0100 Subject: [PATCH] Hotfix: increase zarr chunk cache size (#6639) --- .../webknossos/datastore/datareaders/DatasetArray.scala | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/webknossos-datastore/app/com/scalableminds/webknossos/datastore/datareaders/DatasetArray.scala b/webknossos-datastore/app/com/scalableminds/webknossos/datastore/datareaders/DatasetArray.scala index 35ca38cce8b..adfa7c8c62b 100644 --- a/webknossos-datastore/app/com/scalableminds/webknossos/datastore/datareaders/DatasetArray.scala +++ b/webknossos-datastore/app/com/scalableminds/webknossos/datastore/datareaders/DatasetArray.scala @@ -24,11 +24,11 @@ class DatasetArray(relativePath: DatasetPath, protected val chunkReader: ChunkReader = ChunkReader.create(store, header) - // cache currently limited to 100 MB per array + // cache currently limited to 1 GB per array private lazy val chunkContentsCache: Cache[String, MultiArray] = { - val maxSizeBytes = 1000 * 1000 * 100 + val maxSizeBytes = 1000L * 1000 * 1000 val maxEntries = maxSizeBytes / header.bytesPerChunk - AlfuCache(maxEntries) + AlfuCache(maxEntries.toInt) } // @return Byte array in fortran-order with little-endian values