In information security, fingerprint verification is one of the most common recent approaches for verifying human identity through a distinctive pattern. The verification process works by comparing a pair of fingerprint templates and identifying the similarity/matching among them. Several research studies have utilized different techniques for the matching process such as fuzzy vault and image filtering approaches. Yet, these approaches are still suffering from the imprecise articulation of the biometrics’ interesting patterns. The emergence of deep learning architectures such as the Convolutional Neural Network (CNN) has been extensively used for image processing and object detection tasks and showed an outstanding performance compared to traditional image filtering techniques. This paper aimed to utilize a specific CNN architecture known as AlexNet for the fingerprint-matching task. Using such an architecture, this study has extracted the significant features of the fingerprint image, generated a key based on such a biometric feature of the image, and stored it in a reference database. Then, using Cosine similarity and Hamming Distance measures, the testing fingerprints have been matched with a reference. Using the FVC2002 database, the proposed method showed a False Acceptance Rate (FAR) of 2.09% and a False Rejection Rate (FRR) of 2.81%. Comparing these results against other studies that utilized traditional approaches such as the Fuzzy Vault has demonstrated the efficacy of CNN in terms of fingerprint matching. It is also emphasizing the usefulness of using Cosine similarity and Hamming Distance in terms of matching.
Image compression is one of the data compression types applied to digital images in order to reduce their high cost for storage and/or transmission. Image compression algorithms may take the benefit of visual sensitivity and statistical properties of image data to deliver superior results in comparison with generic data compression schemes, which are used for other digital data. In the first approach, the input image is divided into blocks, each of which is 16 x 16, 32 x 32, or 64 x 64 pixels. The blocks are converted first into a string; then, encoded by using a lossless and dictionary-based algorithm known as arithmetic coding. The more occurrence of the pixels values is codded in few bits compare with pixel values of less occurre
... Show MoreThe aim of the current research is to identify the personal distance between members of society, as well as, to identify the feelings of satisfaction and positivity from respecting the permissible personal distance. The study also aims to identify the feelings of annoyance and comfort from approaching and going beyond personal distance and not respecting it. To achieve these goals, the researchers reviewed previous literature, theories, and studies that dealt with personal distance. The researchers reached a number of results; first, personal distance is not a law but rather a cultural guiding principle for social and professional morals. Second, there are four distances (intimate distance, the distance between friends, social distance,
... Show MoreSurvival analysis is widely applied in data describing for the life time of item until the occurrence of an event of interest such as death or another event of understudy . The purpose of this paper is to use the dynamic approach in the deep learning neural network method, where in this method a dynamic neural network that suits the nature of discrete survival data and time varying effect. This neural network is based on the Levenberg-Marquardt (L-M) algorithm in training, and the method is called Proposed Dynamic Artificial Neural Network (PDANN). Then a comparison was made with another method that depends entirely on the Bayes methodology is called Maximum A Posterior (MAP) method. This method was carried out using numerical algorithms re
... Show MoreWhen optimizing the performance of neural network-based chatbots, determining the optimizer is one of the most important aspects. Optimizers primarily control the adjustment of model parameters such as weight and bias to minimize a loss function during training. Adaptive optimizers such as ADAM have become a standard choice and are widely used for their invariant parameter updates' magnitudes concerning gradient scale variations, but often pose generalization problems. Alternatively, Stochastic Gradient Descent (SGD) with Momentum and the extension of ADAM, the ADAMW, offers several advantages. This study aims to compare and examine the effects of these optimizers on the chatbot CST dataset. The effectiveness of each optimizer is evaluat
... Show MoreThe recent development in communication technologies between individuals allows for the establishment of more informal collaborative map data projects which are called volunteered geographic information (VGI). These projects, such as OpenStreetMap (OSM) project, seek to create free alternative maps which let users add or input new materials to the data of others. The information of different VGI data sources is often not compliant to any standard and each organization is producing a dataset at various level of richness. In this research the assessment of semantic data quality provided by web sources, e.g. OSM will depend on a comparison with the information from standard sources. This will include the validity of semantic accuracy a
... Show MoreThe recent development in communication technologies between individuals allows for the establishment of more informal collaborative map data projects which are called volunteered geographic information (VGI). These projects, such as OpenStreetMap (OSM) project, seek to create free alternative maps which let users add or input new materials to the data of others. The information of different VGI data sources is often not compliant to any standard and each organization is producing a dataset at various level of richness. In this research the assessment of semantic data quality provided by web sources, e.g. OSM will depend on a comparison with the information from standard sources. This will include the validity of semanti
... Show MoreThe art of preventing the detection of hidden information messages is the way that steganography work. Several algorithms have been proposed for steganographic techniques. A major portion of these algorithms is specified for image steganography because the image has a high level of redundancy. This paper proposed an image steganography technique using a dynamic threshold produced by the discrete cosine coefficient. After dividing the green and blue channel of the cover image into 1*3-pixel blocks, check if any bits of green channel block less or equal to threshold then start to store the secret bits in blue channel block, and to increase the security not all bits in the chosen block used to store the secret bits. Firstly, store in the cente
... Show MoreWireless Multimedia Sensor Networks (WMSNs) are a type of sensor network that contains sensor nodes equipped with cameras, microphones; therefore the WMSNS are able to produce multimedia data such as video and audio streams, still images, and scalar data from the surrounding environment. Most multimedia applications typically produce huge volumes of data, this leads to congestion. To address this challenge, This paper proposes Modify Spike Neural Network control for Traffic Load Parameter with Exponential Weight of Priority Based Rate Control algorithm (MSNTLP with EWBPRC). The Modify Spike Neural Network controller (MSNC) can calculate the appropriate traffi
... Show More