hashing algorithms are continually used for large-scale learning and similarity search, with computationally cheap and better algorithms being proposed every year. In this paper we focus on hashing algorithms which in...
详细信息
hashing algorithms are continually used for large-scale learning and similarity search, with computationally cheap and better algorithms being proposed every year. In this paper we focus on hashing algorithms which involve estimating a distance measure d(x(i);x(j)) between two vectors xi;xj. Such hashing algorithms require generation of random variables, and we propose two approaches to reduce the variance of our hashed estimates: control variates and maximum likelihood estimates. We explain how these approaches can be immediately applied to a wide subset of hashing algorithms. Further, we evaluate the impact of these methods on various datasets. We finally run empirical simulations to verify our results.
The goal of Internet-of-Things (IoT) is that every object across the globe be interconnected under the Internet Infrastructure. loT is expanding its application domain to range from environmental monitoring to industr...
详细信息
The goal of Internet-of-Things (IoT) is that every object across the globe be interconnected under the Internet Infrastructure. loT is expanding its application domain to range from environmental monitoring to industrial automation thereby leading to vast research challenges. Vast presence of devices in the Internet has increased the possible challenges faced by devices and data. The devices communicate on a public channel that is more likely to be accessed by unauthorized users and disturb the privacy of genuine users. The existing solutions that ensure data authenticity and user privacy, use MD5 and SHA-family of hashing algorithms under digital signature schemes. These algorithms create a trade-off between the security concern and energy consumption of loT devices. To provide an energy efficient authentication method, we propose a customized BLAKE2b hashing algorithm with modified elliptic curve digital signature scheme (ECDSA). The parameters considered for the evaluation of the proposed methods are signature generation time, signature verification time and hashing time. The experiments are conducted under client server model using Raspberry Pi-3. The proposed method has shown about 0.7-1.91% improvement in the signature generation time and 7.67-9.13 % improvement in signature verification time when compared with BLAKE2b based signature generation/verification. The proposed method is resistant to Man-in-the-Middle attack, Distributed DoS attack (DDoS), pre-image resistance, second pre-image resistance and collision resistance. Based on the performance obtained by the experiments, it can be inferred that the proposed scheme is feasible for resource-constrained IoT devices. (C) 2019 Elsevier B.V. All rights reserved.
The utilization of storage is studied in a two-level memory hierarchy. The first storage level, which is the fast store, is divided into a number of storage areas. When an entry is to be filed in the hierarchy, a hash...
详细信息
The utilization of storage is studied in a two-level memory hierarchy. The first storage level, which is the fast store, is divided into a number of storage areas. When an entry is to be filed in the hierarchy, a hashing algorithm will attempt to place the entry into one of these areas. If this particular area is full, then the entry will be placed into the slower second-level store, even though other areas in the first-level store may have space available. Given that N entries have been filed in the entire hierarchy, an expression is derived for the expected number of entries filed in the first-level store. This expression gives a measure of how effectively the first-level store is being used. By means of examples, storage utilization is then studied as a function of the hashing algorithm, the number of storage areas into which the first-level store is divided and the total size of the first-level store.
We examine a version of the dynamic dictionary problem in which stored items have expiration times and can be removed from the dictionary once they have expired. We show that under several reasonable assumptions about...
详细信息
We examine a version of the dynamic dictionary problem in which stored items have expiration times and can be removed from the dictionary once they have expired. We show that under several reasonable assumptions about the distribution of the items, hashing with lazy deletion uses little more space than methods that use eager deletion. The simple algorithm suggested by this observation was used in a program for analyzing integrated circuit artwork.
Content-based histopathology image retrieval (CBHIR) can assist in the diagnosis of different diseases. The retrieval procedure can be complex and time-consuming if high-dimensional features are required. Thus, hashin...
详细信息
Content-based histopathology image retrieval (CBHIR) can assist in the diagnosis of different diseases. The retrieval procedure can be complex and time-consuming if high-dimensional features are required. Thus, hashing techniques are employed to address these issues by mapping the feature space into binary values of varying lengths. The performance of deep hashing approaches in image retrieval is often superior to that of traditional hashing methods. Among deep hashing approaches, triplet-based models are typically more effective than pairwise ones. Recent studies have demonstrated that incorporating the attention mechanism into a deep hashing approach can improve its effectiveness in retrieving images. This paper presents an innovative triplet deep hashing strategy based on the attention mechanism for retrieving histopathology images, called histopathology attention triplet deep hashing (HATDH). Three deep attention-based hashing models with identical architectures and weights are employed to produce binary values. The proposed attention module can aid the models in extracting features more efficiently. Moreover, we introduce an improved triplet loss function considering pair inputs separately in addition to triplet inputs for increasing efficiency during the training and retrieval steps. Based on experiments conducted on two public histopathology datasets, BreakHis and Kather, HATDH significantly outperforms state-of-the-art hashing algorithms.
Deep hashing has been widely applied to large-scale image retrieval tasks owing to efficient computation and low storage cost by encoding high-dimensional image data into binary codes. Since binary codes do not contai...
详细信息
Deep hashing has been widely applied to large-scale image retrieval tasks owing to efficient computation and low storage cost by encoding high-dimensional image data into binary codes. Since binary codes do not contain as much information as float features, the essence of binary encoding is preserving the main context to guarantee retrieval quality. However, the existing hashing methods have great limitations on suppressing redundant background information and accurately encoding from Euclidean space to Hamming space by a simple sign function. In order to solve these problems, a Cross-Scale Context Extracted hashing Network (CSCE-Net) is proposed in this paper. Firstly, we design a two-branch framework to capture fine-grained local information while maintaining high-level global semantic information. Besides, Attention guided Information Extraction module (AIE) is introduced between two branches, which suppresses areas of low context information cooperated with global sliding windows. Unlike previous methods, our CSCE-Net learns a content-related Dynamic Sign Function (DSF) to replace the original simple sign function. Therefore, the proposed CSCE-Net is context-sensitive and able to perform well on accurate image binary encoding. We further demonstrate that our CSCE-Net is superior to the existing hashing methods, which improves retrieval performance on standard benchmarks.
Ethereum is a popular public blockchain platform and currently the most significant featuring smart contract functionality. Smart contracts are small programs that are executed on the blockchain nodes, which can be us...
详细信息
ISBN:
(纸本)9789897585234
Ethereum is a popular public blockchain platform and currently the most significant featuring smart contract functionality. Smart contracts are small programs that are executed on the blockchain nodes, which can be used to implement complex transaction logic. Several high-level programming languages are available for writing Ethereum smart contracts, the most used being Solidity. The high-level code is further translated into a bytecode executed by a dedicated runtime environment, the Ethereum Virtual Machine (EVM). A few operations are, however, externalized as precompiled contracts, and run by the native implementation of the Ethereum node. These are typically computationally intensive operations such as cryptographic hash functions. Various smart contract patterns require hash computations. In such contexts, the current hash functions supported by Ethereum have a direct impact in both the performance and cost inflicted on the blockchain users. In this paper we investigate the available options for hashing in smart contracts, we discuss the implications regarding some patterns and we evaluate possible improvements. In particular we focus on the recent Blake family of cryptographic hash functions, which show promising performance results, but has yet limited support in the Ethereum platform.
Bloom Filter is a data structure for membership filtering that is deployed in various domains to boost up the lookup performance and to lower the memory consumption. Bloom Filter has gained a massive popularity nowada...
详细信息
Bloom Filter is a data structure for membership filtering that is deployed in various domains to boost up the lookup performance and to lower the memory consumption. Bloom Filter has gained a massive popularity nowadays, and thus, it is deployed in diverse domains, namely, Big Data, Cloud Computing, Networking and Security, Bioinformatics, and IoT. Large scale computing uses a huge memory space, on the contrary, Bloom Filter uses a tiny amount of memory space. However, contemporary solution uses large amounts of memory per input item, and uses very complex arithmetic operations. Therefore, in this article, we propose a novel algorithm, called r-Dimensional Bloom Filter which comprises of Two Dimensional Bloom Filter (2DBF), Three Dimensional Bloom Filter (3DBF), Four Dimensional Bloom Filter (4DBF), and Five Dimensional Bloom Filter (5DBF), and it features a) a very fast filtering system, b) less false positive, c) low extra space consumption, d) free from false negative, e) high adaptability, and f) high scalability. Also, multidimensional Bloom Filter is not found in our literature search. But, many solutions claim the development of multidimensional Bloom Filter. However, these are hierarchical Bloom Filters. To evaluate our proposed data structure, we conduct has carried out an extensive experimentation. 2DBF, 3DBF, 4DBF, and 5DBF outperforms Cuckoo Filter in every aspect. We have experimented using Microsoft trace data and Twitter tweets data.
The last few years have witnessed the rise of the big data era, in which approximate nearest neighbor search is a fundamental problem in many applications, such as large-scale image retrieval. Recently, many research ...
详细信息
The last few years have witnessed the rise of the big data era, in which approximate nearest neighbor search is a fundamental problem in many applications, such as large-scale image retrieval. Recently, many research results demonstrate that hashing can achieve promising performance due to its appealing storage and search efficiency. Since the complex optimization problems for loss functions are difficult to solve, most hashing methods decompose the hash codes learning problem into two steps: projection and quantization. In the quantization step, binary codes are widely used because ranking them by Hamming distance is very efficient. However, the huge information loss produced by the quantization step should be reduced in applications, such as image retrieval where high search accuracy is required. Since many two-step hashing methods produce uneven projected dimensions in the projection step, in this paper, we propose a novel dimension analysis based quantization method (DAQ) on two-step hashing methods for image retrieval. We first perform an importance analysis of the projected dimensions and select a subset of them that are more informative than the others, then we divide the selected projected dimensions into several regions with our quantizer. Every region is quantized with its corresponding codebook. Finally, the similarity between two hash codes is estimated by Manhattan distance between their corresponding codebooks, which is also efficient. We conduct experiments on three public benchmarks containing up to one million descriptors and show that the proposed DAQ method consistently leads to significant accuracy improvements over state-of-the-art quantization methods.
This paper introduces the online Blockchain ART Simulation (BARTS) and a set of experiential online exercises that illustrate hashing concepts for validating blockchain transactions. Our premise is that understanding ...
详细信息
ISBN:
(纸本)9781450383974
This paper introduces the online Blockchain ART Simulation (BARTS) and a set of experiential online exercises that illustrate hashing concepts for validating blockchain transactions. Our premise is that understanding the role of secure hashing algorithms in the mining process is the primary conceptual impediment for students to gain an accurate picture of the validation process for blockchain transactions. The BARTS simulation helps develop the conceptual foundation for understanding hashing concepts, and it also leads to understanding blockchain-related concepts. In addition to assisting in understating blockchain concepts, the BARTS simulation and the accompanying exercises are useful in understanding and applying hashing concepts related to password verification, verifying software signatures, digital content signature verification, and averting man-in-the-middle threats.
暂无评论