This research discussed and analyzed the formulation of a strategy to manage tax compliance risks, as an applied research in the General commission for Taxes. The questionnaire was used as a research tool to identify the factors that stimulate or retard the research sample from being compliant. The K-means clustering method was also used to enable the classification of the research sample's views into four behaviors, some of these views pose tax-compliance risks. The research concluded that risk management is a continuous process and that all departments of the General commission for Taxes are responsible for its implementation to enable them to deal with the behavior of the taxpayer towards tax compliance. And it recommended
... Show MoreIn light of the development in computer science and modern technologies, the impersonation crime rate has increased. Consequently, face recognition technology and biometric systems have been employed for security purposes in a variety of applications including human-computer interaction, surveillance systems, etc. Building an advanced sophisticated model to tackle impersonation-related crimes is essential. This study proposes classification Machine Learning (ML) and Deep Learning (DL) models, utilizing Viola-Jones, Linear Discriminant Analysis (LDA), Mutual Information (MI), and Analysis of Variance (ANOVA) techniques. The two proposed facial classification systems are J48 with LDA feature extraction method as input, and a one-dimen
... Show MoreAbstract
The aim of the current research is to identify the level of administrative applications of expert systems in educational leadership departments in light of the systems approach. To achieve the objectives of the research, the descriptive-analytical and survey method was adopted. The results showed that the level of availability of the knowledge base for expert systems in educational leadership departments (as inputs) was low. The level of availability of resources and software for expert systems in educational leadership departments (as transformational processes) came to be low, as well as the level of availability of the user interface for expert systems in educational leadership departments (as outputs
... Show MoreThe current research aims to identify the time-management skills based on the post-test of the experimental group as well as to examine the effect of a training program on developing the skills of managing time among the study sample. To achieve the research objectives, the researcher designed a scale of time management skill included (30) paragraphs. The research reached that the training program is significantly effective in managing and organizing time. There are statistically significant differences in pre-posttest between the experimental and control groups.
In this work a study and calculation of the normal approach between two bodies, spherical and rough flat surface, had been conducted by the aid of image processing technique. Four kinds of metals of different work hardening index had been used as a surface specimens and by capturing images of resolution of 0.006565 mm/pixel a good estimate of the normal approach may be obtained the compression tests had been done in strength of material laboratory in mechanical engineering department, a Monsanto tensometer had been used to conduct the indentation tests.
A light section measuring equipment microscope BK 70x50 was used to calculate the surface parameters of the texture profile like standard deviation of asperity peak heights, centre lin
In this work a study and calculation of the normal approach between two bodies,
spherical and rough flat surface, had been conducted by the aid of image processing
technique. Four kinds of metals of different work hardening index had been used as a
surface specimens and by capturing images of resolution of 0.006565 mm/pixel a good estimate of the normal approach may be obtained the compression tests had been done in strength of material laboratory in mechanical engineering department, a Monsanto tensometer had been used to conduct the indentation tests. A light section measuring equipment microscope BK 70x50 was used to calculate the surface parameters of the texture profile like standard deviation of asperity peak heights
In recent years, the performance of Spatial Data Infrastructures for governments and companies is a task that has gained ample attention. Different categories of geospatial data such as digital maps, coordinates, web maps, aerial and satellite images, etc., are required to realize the geospatial data components of Spatial Data Infrastructures. In general, there are two distinct types of geospatial data sources exist over the Internet: formal and informal data sources. Despite the growth of informal geospatial data sources, the integration between different free sources is not being achieved effectively. The adoption of this task can be considered the main advantage of this research. This article addresses the research question of ho
... Show MoreSpot panchromatic satellite image had been employed to study and know the difference Between ground and satellite data( DN ,its values varies from 0-255) where it is necessary to convert these DN values to absolute radiance values through special equations ,later it converted to spectral reflectance values .In this study a monitoring of the environmental effect resulted from throwing the sewage drainages pollutants (industrial and home) into the Tigris river water in Mosul, was achieved, which have an effect mostly on physical characters specially color and turbidity which lead to the variation in Spectral Reflectance of the river water ,and it could be detected by using many remote sensing techniques. The contaminated areas within th
... Show MoreAbstract
The research Compared two methods for estimating fourparametersof the compound exponential Weibull - Poisson distribution which are the maximum likelihood method and the Downhill Simplex algorithm. Depending on two data cases, the first one assumed the original data (Non-polluting), while the second one assumeddata contamination. Simulation experimentswere conducted for different sample sizes and initial values of parameters and under different levels of contamination. Downhill Simplex algorithm was found to be the best method for in the estimation of the parameters, the probability function and the reliability function of the compound distribution in cases of natural and contaminateddata.
... Show More
In this paper, the theoretical cross section in pre-equilibrium nuclear reaction has been studied for the reaction at energy 22.4 MeV. Ericson’s formula of partial level density PLD and their corrections (William’s correction and spin correction) have been substituted in the theoretical cross section and compared with the experimental data for nucleus. It has been found that the theoretical cross section with one-component PLD from Ericson’s formula when doesn’t agree with the experimental value and when . There is little agreement only at the high value of energy range with the experimental cross section. The theoretical cross section that depends on the one-component William's formula and on-component corrected to spi
... Show More