Document source identification in printer forensics involves determining the origin of a printed document based on characteristics such as the printer model, serial number, defects, or unique printing artifacts. This process is crucial in forensic investigations, particularly in cases involving counterfeit documents or unauthorized printing. However, consistent pattern identification across various printer types remains challenging, especially when efforts are made to alter printer-generated artifacts. Machine learning models are often used in these tasks, but selecting discriminative features while minimizing noise is essential. Traditional KNN classifiers require a careful selection of distance metrics to capture relevant printing characteristics effectively. This study proposes leveraging quantum-inspired computing to improve KNN classifiers for printer source identification, offering better accuracy even with noisy or variable printing conditions. The proposed approach uses the Gray Level Co-occurrence Matrix (GLCM) for feature extraction, which is resilient to changes in rotation and scale, making it well-suited for texture analysis. Experimental results show that the quantum-inspired KNN classifier captures subtle printing artifacts, leading to improved classification accuracy despite noise and variability.
Today in the digital realm, where images constitute the massive resource of the social media base but unfortunately suffer from two issues of size and transmission, compression is the ideal solution. Pixel base techniques are one of the modern spatially optimized modeling techniques of deterministic and probabilistic bases that imply mean, index, and residual. This paper introduces adaptive pixel-based coding techniques for the probabilistic part of a lossy scheme by incorporating the MMSA of the C321 base along with the utilization of the deterministic part losslessly. The tested results achieved higher size reduction performance compared to the traditional pixel-based techniques and the standard JPEG by about 40% and 50%,
... Show MoreValues are penetrating or interfering in the lives of individuals and groups and are connected to them in the sense of life itself, because they are closely linked to the motives of behavior, hopes and goals, and can be said that values are everything. The concept of values is one of the concepts that have received the attention of researchers from different disciplines. This has resulted in a kind of confusion and variation in use from one specialization to the other, and uses multiple uses within one specialization. Therefore, there is no uniform definition of values because they relate to individuals. Individuals differ in many things, such as perception,
... Show MoreSingle mode-no core-single mode fiber structure with a section of tuned no-core fiber diameter to sense changes in relative humidity has been experimentally demonstrated. The sensor performance with tuned NCF diameter was investigated to maximize the evanescent fields. Different tuned diameters of of (100, 80, and 60)μm were obtained by chemical etching process based on hydrofluoric acid immersion. The highest wavelength sensitivity was obtained 184.57 pm/RH% in the RH range of 30% –100% when the no-core fiber diameter diameter was 60 μm and the sensor response was in real-time measurements
Glassy polymers like Poly Mathyel Metha Acrylate are usually classified as non-porous materials; they are almost considered as fully transparent. Thin samples of these materials reflect color changing followed by porous formation and consequently cracking when exposed to certain level of ?-irradiation. The more the dose is the higher the effect have been observed. The optical microscope and UV-VIS spectroscopy have clearly approved these consequences especially for doped polymers.
This study looks into the many methods that are used in the risk assessment procedure that is used in the construction industry nowadays. As a result of the slow adoption of novel assessment methods, professionals frequently resort to strategies that have previously been validated as being successful. When it comes to risk assessment, having a precise analytical tool that uses the cost of risk as a measurement and draws on the knowledge of professionals could potentially assist bridge the gap between theory and practice. This step will examine relevant literature, sort articles according to their published year, and identify domains and qualities. Consequently, the most significant findings have been presented in a manne
... Show MoreThe drones have become the focus of researchers’ attention because they enter into many details of life. The Tri-copter was chosen because it combines the advantages of the quadcopter in stability and manoeuvrability quickly. In this paper, the nonlinear Tri-copter model is entirely derived and applied three controllers; Proportional-Integral-Derivative (PID), Fractional Order PID (FOPID), and Nonlinear PID (NLPID). The tuning process for the controllers’ parameters had been tuned by using the Grey Wolf Optimization (GWO) algorithm. Then the results obtained had been compared. Where the improvement rate for the Tri-copter model of the nonlinear controller (NLPID) if compared with
One of the most difficult issues in the history of communication technology is the transmission of secure images. On the internet, photos are used and shared by millions of individuals for both private and business reasons. Utilizing encryption methods to change the original image into an unintelligible or scrambled version is one way to achieve safe image transfer over the network. Cryptographic approaches based on chaotic logistic theory provide several new and promising options for developing secure Image encryption methods. The main aim of this paper is to build a secure system for encrypting gray and color images. The proposed system consists of two stages, the first stage is the encryption process, in which the keys are genera
... Show MoreThe aim of this essay is to use a single-index model in developing and adjusting Fama-MacBeth. Penalized smoothing spline regression technique (SIMPLS) foresaw this adjustment. Two generalized cross-validation techniques, Generalized Cross Validation Grid (GGCV) and Generalized Cross Validation Fast (FGCV), anticipated the regular value of smoothing covered under this technique. Due to the two-steps nature of the Fama-MacBeth model, this estimation generated four estimates: SIMPLS(FGCV) - SIMPLS(FGCV), SIMPLS(FGCV) - SIM PLS(GGCV), SIMPLS(GGCV) - SIMPLS(FGCV), SIM PLS(GGCV) - SIM PLS(GGCV). Three-factor Fama-French model—market risk premium, size factor, value factor, and their implication for excess stock returns and portfolio return
... Show More