Skip to content

A Write-friendly and Cache-optimized Hashing Scheme for Non-volatile Memory Systems (MSST 2017, TPDS 2018)

Notifications You must be signed in to change notification settings

yhuacode/Path-Hashing

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 

Repository files navigation

Path Hashing

Path Hashing: Storage cells in the path hashing are logically organized as an inverted complete binary tree. The last level of the inverted binary tree, i.e., all leaf nodes, is considered as the hash table. All nodes in the remaining levels are considered as the shared positions to deal with hash collisions. Path hashing leverages three techniques, i.e., position sharing, double-path hashing and path shortening, allowing insertion and deletion requests to incur no extra writes to NVMs, while maintaining high performance of hash table in terms of space utilization and request latency.

Cache-optimized Path Hashing: To improve the cache line utilization of memory accesses in path hashing, cache-optimized path hashing divides the binary tree into many subtrees and then packs the cells in each subtree together and stores them in the contiguous memory space. Hence, a single memory access can prefetch multiple cells belonging to the same path, which reduces the number of memory accesses to obtain higher performance.

Purpose

Write-friendly. Path hashing causes no extra writes to non-volatile memories. Insertion and deletion requests in path hashing only need to read the nodes in a path for finding an empty position or the target item, thus causing no extra writes.

Memory-efficient. Path hashing has high space utilization which is as high as 95%, since position sharing and double-path hashing provide multiple standby positions for conflicting items.

Low request latency. Path hashing incurs low latency for insertion, deletion, and query requests, due to only probing the nodes in several levels. The flat-addressed storage structure in path hashing allows all nodes in a read path to be read in parallel from NVMs, which has the time complexity of O(1).

Usage

The code of path hashing can be run in GEM5 with NVMain to evaluate the performance in the context of NVMs, and can also be run in real-world systems with DRAMs to evaluate the performance in DRAMs.

We use the xxHash as an example of the hash function in the implementation code. After cloning the repository, the hash function can be modified to fit your needs.

The Path Hashing Paper

Pengfei Zuo and Yu Hua, "A Write-friendly Hashing Scheme for Non-volatile Memory Systems", in Proceedings of the 33st International Conference on Massive Storage Systems and Technology (MSST), 2017.

Pengfei Zuo and Yu Hua, "A Write-friendly and Cache-optimized Hashing Scheme for Non-volatile Memory Systems", IEEE Transactions on Parallel and Distributed Systems (TPDS), Vol.29, No.5, May 2018, pages: 985-998.

About

A Write-friendly and Cache-optimized Hashing Scheme for Non-volatile Memory Systems (MSST 2017, TPDS 2018)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C 98.2%
  • Makefile 1.8%