Classification of imbalanced data is an important issue. Many algorithms have been developed for classification, such as Back Propagation (BP) neural networks, decision tree, Bayesian networks etc., and have been used repeatedly in many fields. These algorithms speak of the problem of imbalanced data, where there are situations that belong to more classes than others. Imbalanced data result in poor performance and bias to a class without other classes. In this paper, we proposed three techniques based on the Over-Sampling (O.S.) technique for processing imbalanced dataset and redistributing it and converting it into balanced dataset. These techniques are (Improved Synthetic Minority Over-Sampling Technique (Improved SMOTE), Borderline-SMOTE + Imbalanced Ratio(IR), Adaptive Synthetic Sampling (ADASYN) +IR) Algorithm, where the work these techniques are generate the synthetic samples for the minority class to achieve balance between minority and majority classes and then calculate the IR between classes of minority and majority. Experimental results show ImprovedSMOTE algorithm outperform the Borderline-SMOTE + IR and ADASYN + IR algorithms because it achieves a high balance between minority and majority classes.
Background: The study's objective was to estimate the effects of radiation on testosterone-related hormones and blood components in prostate cancer patients. N Materials and Method: This study aims to investigate the effects of radiation on 20 male prostate cancer patients at the Middle Euphrates Oncology Centre. Blood samples were collected before and after radiation treatment, with a total dose of 60- 70 Gy, The blood parameters were analyzed. The hospital laboratory conducted the blood analysis using an analyzer (Diagon D-cell5D) to test blood components before and after radiation. Hormonal examinations included testosterone levels, using the VIDASR 30 for Multiparametric immunoassay system Results: The study assessed the socio-demogra
... Show MoreAdministrative procedures in various organizations produce numerous crucial records and data. These
records and data are also used in other processes like customer relationship management and accounting
operations.It is incredibly challenging to use and extract valuable and meaningful information from these data
and records because they are frequently enormous and continuously growing in size and complexity.Data
mining is the act of sorting through large data sets to find patterns and relationships that might aid in the data
analysis process of resolving business issues. Using data mining techniques, enterprises can forecast future
trends and make better business decisions.The Apriori algorithm has bee
This paper presents a numerical simulation of the flow around elliptic groynes by using CFD software. The flow was simulated in a flume with 4m long, 0.4m wide, and 0.175m high with a constant bed slope. Moreover, the first Groyne placed at 1m from the flow inlet with a constant the Groyne height of 10cm and a 1cm thickness, and the width of Groynes equals 7cm. A submergence ratio of the elliptic Groynes of 75% was assumed, corresponding to a discharge of 0.0057m3/sec. The CFD model showed a good ability to simulate the flow around Groynes with good accuracy. The results of CFD software showed that when using double elliptic Groy
... Show MoreThe present study aimed at identifying the effectiveness of Macaton method in improving some sensory and cognitive skills in autistic children. In order to achieve the aims of the study, the researcher used the experimental method. The present study sample was (10) children whose ages ranged between (7-10) years and were diagnosed medically with autism disorder. The researcher randomly selected the sample and divided it into two groups: the first group consisted of (5) children representing the experimental group, and (5) children representing the control group after extracting the equivalence between the two groups in terms of age, intelligence, economic and social level and the degree of communication. The program was implemented for t
... Show MoreToday, there are large amounts of geospatial data available on the web such as Google Map (GM), OpenStreetMap (OSM), Flickr service, Wikimapia and others. All of these services called open source geospatial data. Geospatial data from different sources often has variable accuracy due to different data collection methods; therefore data accuracy may not meet the user requirement in varying organization. This paper aims to develop a tool to assess the quality of GM data by comparing it with formal data such as spatial data from Mayoralty of Baghdad (MB). This tool developed by Visual Basic language, and validated on two different study areas in Baghdad / Iraq (Al-Karada and Al- Kadhumiyah). The positional accuracy was asses
... Show MoreA crucial area of research in nanotechnology is the formation of environmentally benign nanoparticles. Both unicellular and multicellular play an important role in synthesis nanoparticles through the production of inorganic materials either intracellularly or extracellularly. The agents (pigments, siderophores, cell extracted metabolites and reducing compounds) were used to prepare silver nanparticles with different sizes and shapes. The color variations (dark yellow, slightly dark yellow and golden yellow) arising from changes in the composition, size, and shape of nanoparticles, surrounding medium can be monitored using UV-visible spectrophotometer. These effects are due to the phenomena called surface plasmon resonance. The silver nanopa
... Show MoreSurvival analysis is widely applied to data that described by the length of time until the occurrence of an event under interest such as death or other important events. The purpose of this paper is to use the dynamic methodology which provides a flexible method, especially in the analysis of discrete survival time, to estimate the effect of covariate variables through time in the survival analysis on dialysis patients with kidney failure until death occurs. Where the estimations process is completely based on the Bayes approach by using two estimation methods: the maximum A Posterior (MAP) involved with Iteratively Weighted Kalman Filter Smoothing (IWKFS) and in combination with the Expectation Maximization (EM) algorithm. While the other
... Show MoreReservoir fluids properties are very important in reservoir engineering computations such as material balance calculations, well testing analyses, reserve estimates, and numerical reservoir simulations. Isothermal oil compressibility is required in fluid flow problems, extension of fluid properties from values at the bubble point pressure to higher pressures of interest and in material balance calculations (Ramey, Spivey, and McCain). Isothermal oil compressibility is a measure of the fractional change in volume as pressure is changed at constant temperature (McCain). The most accurate method for determining the Isothermal oil compressibility is a laboratory PVT analysis; however, the evaluation of exploratory wells often require an esti
... Show MoreIn this paper we introduce several estimators for Binwidth of histogram estimators' .We use simulation technique to compare these estimators .In most cases, the results proved that the rule of thumb estimator is better than other estimators.
The OpenStreetMap (OSM) project aims to establish a free geospatial database for the entire world which is editable by international volunteers. The OSM database contains a wide range of different types of geographical data and characteristics, including highways, buildings, and land use regions. The varying scientific backgrounds of the volunteers can affect the quality of the spatial data that is produced and shared on the internet as an OSM dataset. This study aims to compare the completeness and attribute accuracy of the OSM road networks with the data supplied by a digitizing process for areas in the Baghdad and Thi-Qar governorates. The analyses are primarily based on calculating the portion of the commission (extr
... Show More