You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jul 6, 2023. It is now read-only.
Implement new index engine which is based on LSM-trie data structure. Given data structure has much less
write amplification than LSM-tree and is ten times faster on benchmarks than LSM-tree implementation.
LSM-trie is capable to handle till 10 TB of data with small memory footprint.
LSM-trie does not support range queries, so it is similar to current hash index implementation.
Non-Goals:
LSM-trie is not supposed to be faster than current implementation of hash index on small and average amount of data but its scalability
on big data should be much better.
Success metrics:
Much lower degree of degradation of index speed on big data.
Motivation:
To minimize random IO overhead during insertion of data into index it is proposed to gather all data changes into memory and write them using single batch operation. All batches will be merged and indexed in a background thread.
Description:
It is proposed to implement index engine which according to measurements of operations throughput
are done on YCSB benchmark using the same database implementation (LevelDB), but different index engines
gives ten times speedup on read and write operations.
First of all, let see results of testing of write throughput on LevelDB ( + LSM trie) vs. the original LevelDB implementation
and vs. RocksDB implementation.
Bellow are results of testing of read performance of the same implementations on two configurations:
In case of presence of 64 GB RAM on server
In the case of presence of 4 GB RAM on a server.
Note that trie based indexes by their nature are hash indexes and do not support range queries.
In our current implementation of ODB, we already use the advantage of trie based indexes over tree-based indexes.
Our hash indexes apparently faster than sb-tree indexes because they ensure
at most one IO operation during read and at most three IO operations during write.
LSM-trie implementation uses the same advantage of trie based indexes over tree-based indexes.
LSM -trie ensures that there are at most 2 IO operations on each level during read of data and if we have enough of RAM (5 GB for 10 TB of data, it is quite a small requirement)
at most one IO operation for all levels except the last one, in the last one, we will have at most two operations (in the worst case).
Let's start with basics.
The most popular implementation of LSM tree which was introduced first in LevelDB is based on concepts of MemTable and SSTable.
SSTable stands for Sorted String Table, by "string" they mean not set of characters but a set of bytes so any string of bytes may be stored in SSTable.
SSTable is implemented as a simple sorted list which consumes fixed amount of space, about 16-32 MB.
MemTable is in memory presentation of SSTalbe.
It worth to note that SSTables are immutable and as a result durable by definition without any WAL support.
How this mix of SSTable and MemTable works.
Any update and insert of data causes an addition of data items in MemTable (deletes are handled by addition specific markers tombstones, they added to indicate that this item is removed). For the sake of durability all operations on MemTable are logged into log segment. Once MemTable is full it is flushed on disk with force sync and log segment is dropped. That is it, very fast and durable.
There are downsides of course:
There are a lot of duplication of data on disk and large space consumption.
Read operations are slow
To speed up read operations from SSTables:
Each SSTable is indexed, and an index is stored in an immutable file along with SSTable. Very often those indexes are cached into memory.
To avoid to read index to find an item on all SSTables bloom filters are used, those bloom filters are placed in memory.
To improve space usage and partially improve the speed of read operations all SSTables are merged into a single sorted list. Merge operation is very fast O(N) on sorted lists and does not require a lot of RAM. This process is called compaction.
To understand how compaction works let split storage by levels on Level 0 we place all SSTables are created as result of a flush of MemTables.
On Level 1 we put all SSTables are created as result of a merge of SSTables from level 0.
Merge or compaction is performed as following:
Several SSTables from Level 0 with overlapped keys are taken.
SSTables from Level 1 which overlaps with SSTables of a level above are taken.
All those tables are merged using merge sort.
Old SSTables are removed, and new SSTables are created and placed on Level 1.
Note that ranges of keys of SSTables on Level 1 are not overlapped. So to read data with a specific key, we may first find SSTable with appropriate key range. Then check bloom filter. If bloom filter indicates that item with given key is present in SSTable, we read the index of SSTable (it may be B-Tree for example, it is relatively fast to create B-tree using sorted list of items) and then read key-value pair from SSTable.
So far so good. But we forget one thing - write amplification. Because an amount of files on Level 1 is growing, we need to read and write more and more files from Level 1 to merge the same amount of SSTables from Level 0. Eventually, write amplification becomes so big that writes block any other IO operations on disk and system is stuck.
To minimize an influence of write amplification, LevelDB uses leveled approach during the merge of SSTables (which is reflected in its name).
The maximum amount of files on Level 1 is limited to 10, on Level 2 the maximum amount of files is limited to 100 and so on.
But algorithm stays the same.
To merge SSTables from Level n to Level n+1 we:
Take one SSTable from Level n (with the exception of Level 0, on this level we take all SSTables with overlapped key ranges)
Read all SSTables which overlaps with SSTable from Level n + 1 and write all merged SSTables in Level n + 1
The key difference between this case and case described above is that in the worst case during compaction only 10 SSTables from Level N + 1 affected.
So far so good. We limited write amplification by an introduction of levels.
But in reality, we only delayed problem (if you google "RocksDB write amplification" for example you may find many references).
Yes during single compaction we merge only 11 SSTables (1 from Level N and ten from Level N + 1), but let's take a look what is write amplification if we want eventually move a single key entry to the lowest level.
Write amplification of compaction of a single level is 10 + 1, if we want to move key down to K level write amplification will be (10 + 1) * K. So for big storages for example with K = 5 write amplification will be 55. Writes again will block all disk IO operations so we will have the same problem as above but later.
So how to solve this problem once and forever.
The solution is quite "simple." If we can involve only upper-level SSTables in compaction write amplification for a single key will be minimal. So how this should work?
We can treat each SSTable on Level 0 as sub-level. So Level 0 consist of sub levels, in sub level key ranges of SSTables are not overlapped (obviously), but between sub-levels key ranges are overlapped.
We may extend this concept to other Levels. In each Level, we have sub levels, and key ranges of SSTables inside of sub-levels are not overlapped but between levels are overlapped.
So to merge SSTables from Level N to Level N + 1:
We read one SSTable from each sub-level of Level N
Merge them together.
Put merged SSTables in sub level of Leve N + 1
But there is one problem, all compactions are targeted to the same sub-level of Level N+1 should not have overlapped key ranges (otherwise we will fall to an algorithm is used to the first approach) which are highly unlikely.
We also can not create sub-level for each single merge because we still will have enormous space usage and slow reads and big memory consumption (because presence of many sub levels implies presence of many data duplication)
So how to solve given the problem?
First of all, let's create a hash code for each key which is going to be added into an index using SHA - 1 function (that will guarantee uniformity of distribution of hash codes ).
Then we will use a prefix of a hash code to determine sub level of level.
Let's look at picture to get what is going on:
We create minuscule trie, which consists of table containers.
All tables which are placed in the same position inside of table containers (for example we have 2 SSTables on position 2 of first and second container) are constitute sub level.
All tables which are contained inside table containers of the same depth constitute a level in LevelDB terminology.
So what we do during compaction:
Take all nodes from single table container (it is the same like we take one SSTable for each sub-level in the previous algorithm).
Merge them into non-overlapped SSTables.
There is an interesting difference in compaction of SSTables in LevelDB and LSM-trie. Items of SSTables are merged and then placed in different containers according to hash codes.
The result of compaction process is described in picture bellow:
As a result, all key ranges (regarding LSM-trie) of SSTables will not be overlapped.
What is interesting despite the fact that LSM-trie minimizes write amplification, it also allows removing the need to have SSTable indexes.
To remove SSTable indexes, SSTable is converted into HTable. HTable is SSTable, but instead of sorting them by items we put keys into buckets by their hash code.
Of course in such case, some buckets will be overflown, and some of them will not have enough elements to be full.
In this case, we do following:
We sort keys which are put into a bucket by their hash code.
Fill bucket to its capacity
In bucket reserve 8 bytes for the following information (source bucket id, maximum value of hash code after which items will be moved to another bucket, bucket destination id)
Move all items which hash code above recorded in step 3 to destination bucket.
Using such approach we solve the issue with overflown buckets and remove the need to have separate indexes for a bucket.
To minimize amount of IO operations are needed to read data from index we are going to cache bucket migration information into memory, but because size of bucket is 64 or 16 K (in new version) memory overhead of such data caching
will be minimal.
How do we read items in case of request?
First of all, we detect tables container by key hash code, then in this tables container, we read Bloom filters to check which HTable's bucket contains the item and read data from a concrete bucket of HTable with single IO operation without any additional indexes. Each bloom filter requires about 16 bit per element or 2 GB for one billion items. 2 GB is not big memory consumption for such amount of data, but some of our users work with big disks and a small amount of RAM. To resolve this issue too, we cluster all bloom filters for the same buckets of HTables in a single page (Bloom filters are created for each bucket, a bucket is a cache page in our implementation).
As a result, if we do not have enough memory to keep bloom filters we read Bloom filters with one disk IO and then check do we need to load specific bucket or not and then load given bucket.
Alternatives:
None
Risks and assumptions:
We need to understand how to integrate LSM-trie mechanics in our transactional system.
Impact matrix
Storage engine
SQL
Protocols
Indexes
Console
Java API
Geospatial
Lucene
Security
Hooks
EE
The text was updated successfully, but these errors were encountered:
Reference:
https://github.com/orientechnologies/orientdb-labs/blob/master/OEP_17.md
Summary:
LSM-trie based index
Goals:
Implement new index engine which is based on LSM-trie data structure. Given data structure has much less
write amplification than LSM-tree and is ten times faster on benchmarks than LSM-tree implementation.
LSM-trie is capable to handle till 10 TB of data with small memory footprint.
LSM-trie does not support range queries, so it is similar to current hash index implementation.
Non-Goals:
LSM-trie is not supposed to be faster than current implementation of hash index on small and average amount of data but its scalability
on big data should be much better.
Success metrics:
Much lower degree of degradation of index speed on big data.
Motivation:
To minimize random IO overhead during insertion of data into index it is proposed to gather all data changes into memory and write them using single batch operation. All batches will be merged and indexed in a background thread.
Description:
It is proposed to implement index engine which according to measurements of operations throughput
are done on YCSB benchmark using the same database implementation (LevelDB), but different index engines
gives ten times speedup on read and write operations.
First of all, let see results of testing of write throughput on LevelDB ( + LSM trie) vs. the original LevelDB implementation
and vs. RocksDB implementation.
Bellow are results of testing of read performance of the same implementations on two configurations:
Note that trie based indexes by their nature are hash indexes and do not support range queries.
In our current implementation of ODB, we already use the advantage of trie based indexes over tree-based indexes.
Our hash indexes apparently faster than sb-tree indexes because they ensure
at most one IO operation during read and at most three IO operations during write.
LSM-trie implementation uses the same advantage of trie based indexes over tree-based indexes.
LSM -trie ensures that there are at most 2 IO operations on each level during read of data and if we have enough of RAM (5 GB for 10 TB of data, it is quite a small requirement)
at most one IO operation for all levels except the last one, in the last one, we will have at most two operations (in the worst case).
Let's start with basics.
The most popular implementation of LSM tree which was introduced first in LevelDB is based on concepts of MemTable and SSTable.
SSTable stands for Sorted String Table, by "string" they mean not set of characters but a set of bytes so any string of bytes may be stored in SSTable.
SSTable is implemented as a simple sorted list which consumes fixed amount of space, about 16-32 MB.
MemTable is in memory presentation of SSTalbe.
It worth to note that SSTables are immutable and as a result durable by definition without any WAL support.
How this mix of SSTable and MemTable works.
Any update and insert of data causes an addition of data items in MemTable (deletes are handled by addition specific markers tombstones, they added to indicate that this item is removed). For the sake of durability all operations on MemTable are logged into log segment. Once MemTable is full it is flushed on disk with force sync and log segment is dropped. That is it, very fast and durable.
There are downsides of course:
To speed up read operations from SSTables:
To improve space usage and partially improve the speed of read operations all SSTables are merged into a single sorted list. Merge operation is very fast O(N) on sorted lists and does not require a lot of RAM. This process is called compaction.
To understand how compaction works let split storage by levels on Level 0 we place all SSTables are created as result of a flush of MemTables.
On Level 1 we put all SSTables are created as result of a merge of SSTables from level 0.
Merge or compaction is performed as following:
Note that ranges of keys of SSTables on Level 1 are not overlapped. So to read data with a specific key, we may first find SSTable with appropriate key range. Then check bloom filter. If bloom filter indicates that item with given key is present in SSTable, we read the index of SSTable (it may be B-Tree for example, it is relatively fast to create B-tree using sorted list of items) and then read key-value pair from SSTable.
So far so good. But we forget one thing - write amplification. Because an amount of files on Level 1 is growing, we need to read and write more and more files from Level 1 to merge the same amount of SSTables from Level 0. Eventually, write amplification becomes so big that writes block any other IO operations on disk and system is stuck.
To minimize an influence of write amplification, LevelDB uses leveled approach during the merge of SSTables (which is reflected in its name).
The maximum amount of files on Level 1 is limited to 10, on Level 2 the maximum amount of files is limited to 100 and so on.
But algorithm stays the same.
To merge SSTables from Level n to Level n+1 we:
The key difference between this case and case described above is that in the worst case during compaction only 10 SSTables from Level N + 1 affected.
So far so good. We limited write amplification by an introduction of levels.
But in reality, we only delayed problem (if you google "RocksDB write amplification" for example you may find many references).
Yes during single compaction we merge only 11 SSTables (1 from Level N and ten from Level N + 1), but let's take a look what is write amplification if we want eventually move a single key entry to the lowest level.
Write amplification of compaction of a single level is 10 + 1, if we want to move key down to K level write amplification will be (10 + 1) * K. So for big storages for example with K = 5 write amplification will be 55. Writes again will block all disk IO operations so we will have the same problem as above but later.
So how to solve this problem once and forever.
The solution is quite "simple." If we can involve only upper-level SSTables in compaction write amplification for a single key will be minimal. So how this should work?
We can treat each SSTable on Level 0 as sub-level. So Level 0 consist of sub levels, in sub level key ranges of SSTables are not overlapped (obviously), but between sub-levels key ranges are overlapped.
We may extend this concept to other Levels. In each Level, we have sub levels, and key ranges of SSTables inside of sub-levels are not overlapped but between levels are overlapped.
So to merge SSTables from Level N to Level N + 1:
But there is one problem, all compactions are targeted to the same sub-level of Level N+1 should not have overlapped key ranges (otherwise we will fall to an algorithm is used to the first approach) which are highly unlikely.
We also can not create sub-level for each single merge because we still will have enormous space usage and slow reads and big memory consumption (because presence of many sub levels implies presence of many data duplication)
So how to solve given the problem?
First of all, let's create a hash code for each key which is going to be added into an index using SHA - 1 function (that will guarantee uniformity of distribution of hash codes ).
Then we will use a prefix of a hash code to determine sub level of level.
Let's look at picture to get what is going on:
We create minuscule trie, which consists of table containers.
All tables which are placed in the same position inside of table containers (for example we have 2 SSTables on position 2 of first and second container) are constitute sub level.
All tables which are contained inside table containers of the same depth constitute a level in LevelDB terminology.
So what we do during compaction:
There is an interesting difference in compaction of SSTables in LevelDB and LSM-trie. Items of SSTables are merged and then placed in different containers according to hash codes.
The result of compaction process is described in picture bellow:
As a result, all key ranges (regarding LSM-trie) of SSTables will not be overlapped.
What is interesting despite the fact that LSM-trie minimizes write amplification, it also allows removing the need to have SSTable indexes.
To remove SSTable indexes, SSTable is converted into HTable. HTable is SSTable, but instead of sorting them by items we put keys into buckets by their hash code.
Of course in such case, some buckets will be overflown, and some of them will not have enough elements to be full.
In this case, we do following:
Using such approach we solve the issue with overflown buckets and remove the need to have separate indexes for a bucket.
To minimize amount of IO operations are needed to read data from index we are going to cache bucket migration information into memory, but because size of bucket is 64 or 16 K (in new version) memory overhead of such data caching
will be minimal.
How do we read items in case of request?
First of all, we detect tables container by key hash code, then in this tables container, we read Bloom filters to check which HTable's bucket contains the item and read data from a concrete bucket of HTable with single IO operation without any additional indexes. Each bloom filter requires about 16 bit per element or 2 GB for one billion items. 2 GB is not big memory consumption for such amount of data, but some of our users work with big disks and a small amount of RAM. To resolve this issue too, we cluster all bloom filters for the same buckets of HTables in a single page (Bloom filters are created for each bucket, a bucket is a cache page in our implementation).
As a result, if we do not have enough memory to keep bloom filters we read Bloom filters with one disk IO and then check do we need to load specific bucket or not and then load given bucket.
Alternatives:
None
Risks and assumptions:
We need to understand how to integrate LSM-trie mechanics in our transactional system.
Impact matrix
The text was updated successfully, but these errors were encountered: