Graphite Coated Electrodes (GCE) based on molecularly imprinted polymers were fabricated for the selective potentiometric determination of Risperidone (Ris). The molecularly imprinted (MIP) and nonimprinted (NIP) polymers were synthesized by bulk polymerization using (Ris.) as a template, acrylic acid (AA) and acrylamide (AAm) as monomers, ethylene glycol dimethacrylate (EGDMA) as a cross-linker and benzoyl peroxide (BPO) as an initiator. The imprinted membranes and the non-imprinted membranes were prepared using dioctyl phthalate (DOP) and Dibutylphthalate (DBP) as plasticizers in PVC matrix. The membranes were coated on graphite electrodes. The MIP electrodes using
... Show MoreIn this research velocity of moving airplane from its recorded digital sound is introduced. The data of sound file is sliced into several frames using overlapping partitions. Then the array of each frame is transformed from time domain to frequency domain using Fourier Transform (FT). To determine the characteristic frequency of the sound, a moving window mechanics is used, the size of that window is made linearly proportional with the value of the tracked frequency. This proportionality is due to the existing linear relationship between the frequency and its Doppler shift. An algorithm was introduced to select the characteristic frequencies, this algorithm allocates the frequencies which satisfy the Doppler relation, beside that the tra
... Show MoreA comparison of double informative and non- informative priors assumed for the parameter of Rayleigh distribution is considered. Three different sets of double priors are included, for a single unknown parameter of Rayleigh distribution. We have assumed three double priors: the square root inverted gamma (SRIG) - the natural conjugate family of priors distribution, the square root inverted gamma – the non-informative distribution, and the natural conjugate family of priors - the non-informative distribution as double priors .The data is generating form three cases from Rayleigh distribution for different samples sizes (small, medium, and large). And Bayes estimators for the parameter is derived under a squared erro
... Show MoreThis research includes the using of statistical to improve the quality of can plastics which is produced at the state company for Vegetable oils (Almaamon factory ) by using the percentage defective control chart ( p-chart ) of a fixed sample. A sample of size (450) cans daily for (30) days was selected to determine the rejected product . Operations research with a (win QSB ) package for ( p-chart ) was used to determine test quality level required for product specification to justify that the process that is statistically controlled.
The results show high degree of accuracy by using the program and the mathematical operations (primary and secondary ) which used to draw the control limits charts and to reject the statistically uncontr
Abstract
In this research will be treated with a healthy phenomenon has a significant impact on different age groups in the community, but a phenomenon tonsillitis where they will be first Tawfiq model slope self moving averages seasonal ARMA Seasonal through systematic Xbox Cengnzla counter with rheumatoid tonsils in the city of Mosul, and for the period 2004-2009 with prediction of these numbers coming twelve months, has found that the specimen is the best representation of the data model is the phenomenon SARMA (1,1) * (2,1) 12 from the other side and explanatory variables using a maximum temperature and minimum temperature, sol
The study aims to analyze computer textbooks content for preparatory stage according to the logical thinking. The researcher followed the descriptive analytical research approach (content analysis), and adopted an explicit idea during the analysis process. One of the content analysis tools which was designed based on mental processes employed during logical thinking has utilized to figure out the study results. The findings revealed that logical thinking skills formed (52%) in fourth preparatory textbook and (47%) in fifth preparatory textbook.
This paper presents a new algorithm in an important research field which is the semantic word similarity estimation. A new feature-based algorithm is proposed for measuring the word semantic similarity for the Arabic language. It is a highly systematic language where its words exhibit elegant and rigorous logic. The score of sematic similarity between two Arabic words is calculated as a function of their common and total taxonomical features. An Arabic knowledge source is employed for extracting the taxonomical features as a set of all concepts that subsumed the concepts containing the compared words. The previously developed Arabic word benchmark datasets are used for optimizing and evaluating the proposed algorithm. In this paper,
... Show More