Impact of lossless compression algorithms on in-memory database synchronization

dc.contributor.authorION, ANDREI
dc.contributor.departmentChalmers tekniska högskola / Institutionen för data och informationstekniksv
dc.contributor.departmentChalmers University of Technology / Department of Computer Science and Engineeringen
dc.contributor.examinerTsigas, Philippas
dc.contributor.supervisorDuvignau, Romaric
dc.date.accessioned2022-11-30T14:28:06Z
dc.date.available2022-11-30T14:28:06Z
dc.date.issued2022
dc.date.submitted2020
dc.description.abstractIn-memory databases have gained popularity in the last decades due to the increased demand for high-speed access to data. Redis is one such database that provides a sub-millisecond response time for incoming requests. These speedups are of particular interest in the telecommunication industry, where 5G technology needs to provide multi-Gbps network speeds. Ericsson is a global leader in 5G network equipment that benefits from in-memory database improvements. When a fault occurs in the system, methods to prevent data loss are needed. One such method is the data replication on a secondary node. Synchronization after a fault in the replica node puts pressure on the primary node to withstand the incoming requests from clients. Requests are buffered until the replica is ready. The buffering can demand a lot of main memory space, and if the system runs out of memory, the synchronization restarts. The novelty of the thesis work focuses on minimizing the impact over main memory size when synchronization between primary and replica nodes is taking place. Practically, adding multiple types of compression over received data in the Redis network layer. We gather performance metrics related to memory size reduction, requests per second, CPU utilization, relative time spent on CPU, and Maximum Main Memory used. We show that compression over the random data set, an extreme case, does not provide any memory size reduction, and it has a significant negative impact on performance. Another extreme data set is a single character generated multiple times. Intuitively, this data set is highly compressible and provided unrealistic compression ratios of 71.1. Lastly, we showcase the real data set with a 3.582 compression ratio when using the ZSTD algorithm; furthermore, the above data-set showed a higher maximum transfer rate when using compression. The maximum transfer rate shows how much bandwidth can the system support when synchronization is ongoing. Given these data sets, we showcase the positive impact of adopting compression in a 5G network.
dc.identifier.coursecodeDATX05
dc.identifier.urihttps://odr.chalmers.se/handle/20.500.12380/305854
dc.language.isoeng
dc.setspec.uppsokTechnology
dc.subjectin-memory database
dc.subjectredis
dc.subjectreplication
dc.subjectcompression
dc.subjectperformance
dc.subjectsynchronization
dc.titleImpact of lossless compression algorithms on in-memory database synchronization
dc.type.degreeExamensarbete för masterexamensv
dc.type.degreeMaster's Thesisen
dc.type.uppsokH
local.programmeComputer systems and networks (MPCSN), MSc
Ladda ner
Original bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
CSE 22-93 Ion.pdf
Storlek:
2.15 MB
Format:
Adobe Portable Document Format
Beskrivning:
License bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
license.txt
Storlek:
1.64 KB
Format:
Item-specific license agreed upon to submission
Beskrivning: