Digital image is widely used in computer applications. This paper introduces a proposed method of image zooming based upon inverse slantlet transform and image scaling. Slantlet transform (SLT) is based on the principle of designing different filters for different scales.
First we apply SLT on color image, the idea of transform color image into slant, where large coefficients are mainly the signal and smaller one represent the noise. By suitably modifying these coefficients , using scaling up image by box and Bartlett filters so that the image scales up to 2X2 and then inverse slantlet transform from modifying coefficients using to the reconstructed image .
&nbs
... Show MoreSteganography is a mean of hiding information within a more obvious form of
communication. It exploits the use of host data to hide a piece of information in such a way
that it is imperceptible to human observer. The major goals of effective Steganography are
High Embedding Capacity, Imperceptibility and Robustness. This paper introduces a scheme
for hiding secret images that could be as much as 25% of the host image data. The proposed
algorithm uses orthogonal discrete cosine transform for host image. A scaling factor (a) in
frequency domain controls the quality of the stego images. Experimented results of secret
image recovery after applying JPEG coding to the stego-images are included.
Median filter is adopted to match the noise statistics of the degradation seeking good quality smoothing images. Two methods are suggested in this paper(Pentagonal-Hexagonal mask and Scan Window Mask), the study involved modified median filter for improving noise suppression, the modification is considered toward more reliable results. Modification median filter (Pentagonal-Hexagonal mask) was found gave better results (qualitatively and quantitatively ) than classical median filters and another suggested method (Scan Window Mask), but this will be on the account of the time required. But sometimes when the noise is line type the cross 3x3 filter preferred to another one Pentagonal-Hexagonal with few variation. Scan Window Mask gave bett
... Show MoreIn this paper, an algorithm for reconstruction of a completely lost blocks using Modified
Hybrid Transform. The algorithms examined in this paper do not require a DC estimation
method or interpolation. The reconstruction achieved using matrix manipulation based on
Modified Hybrid transform. Also adopted in this paper smart matrix (Detection Matrix) to detect
the missing blocks for the purpose of rebuilding it. We further asses the performance of the
Modified Hybrid Transform in lost block reconstruction application. Also this paper discusses
the effect of using multiwavelet and 3D Radon in lost block reconstruction.
This research presents a method of using MATLAB in analyzing a nonhomogeneous soil (Gibson-type) by
estimating the displacements and stresses under the strip footing during applied incremental loading
sequences. This paper presents a two-dimensional finite element method. In this method, the soil is divided into a number of triangle elements. A model soil (Gibson-type) with linearly increasing modulus of elasticity with depth is presented. The influences of modulus of elasticity, incremental loading, width of footing, and depth of footing are considered in this paper. The results are compared with authors' conclusions of previous studies.
In this article, we introduced a new concept of mappings called δZA - Quasi contractive mapping and we study the K*- iteration process for approximation of fixed points, and we proved that this iteration process is faster than the existing leading iteration processes like Noor iteration process, CR -iteration process, SP and Karahan Two- step iteration process for 𝛿𝒵𝒜 − quasi contraction mappings. We supported our analytic proof by a numerical example.
This study explores the challenges in Artificial Intelligence (AI) systems in generating image captions, a task that requires effective integration of computer vision and natural language processing techniques. A comparative analysis between traditional approaches such as retrieval- based methods and linguistic templates) and modern approaches based on deep learning such as encoder-decoder models, attention mechanisms, and transformers). Theoretical results show that modern models perform better for the accuracy and the ability to generate more complex descriptions, while traditional methods outperform speed and simplicity. The paper proposes a hybrid framework that combines the advantages of both approaches, where conventional methods prod
... Show More