The influx of data in bioinformatics is primarily in the form of DNA, RNA, and protein sequences. This condition places a significant burden on scientists and computers. Some genomics studies depend on clustering techniques to group similarly expressed genes into one cluster. Clustering is a type of unsupervised learning that can be used to divide unknown cluster data into clusters. The k-means and fuzzy c-means (FCM) algorithms are examples of algorithms that can be used for clustering. Consequently, clustering is a common approach that divides an input space into several homogeneous zones; it can be achieved using a variety of algorithms. This study used three models to cluster a brain tumor dataset. The first model uses FCM, which is used to cluster genes. FCM allows an object to belong to two or more clusters with a membership grade between zero and one and the sum of belonging to all clusters of each gene is equal to one. This paradigm is useful when dealing with microarray data. The total time required to implement the first model is 22.2589 s. The second model combines FCM and particle swarm optimization (PSO) to obtain better results. The hybrid algorithm, i.e., FCM–PSO, uses the DB index as objective function. The experimental results show that the proposed hybrid FCM–PSO method is effective. The total time of implementation of this model is 89.6087 s. The third model combines FCM with a genetic algorithm (GA) to obtain better results. This hybrid algorithm also uses the DB index as objective function. The experimental results show that the proposed hybrid FCM–GA method is effective. Its total time of implementation is 50.8021 s. In addition, this study uses cluster validity indexes to determine the best partitioning for the underlying data. Internal validity indexes include the Jaccard, Davies Bouldin, Dunn, Xie–Beni, and silhouette. Meanwhile, external validity indexes include Minkowski, adjusted Rand, and percentage of correctly categorized pairings. Experiments conducted on brain tumor gene expression data demonstrate that the techniques used in this study outperform traditional models in terms of stability and biological significance.
The consumption of dried bananas has increased because they contain essential nutrients. In order to preserve bananas for a longer period, a drying process is carried out, which makes them a light snack that does not spoil quickly. On the other hand, machine learning algorithms can be used to predict the sweetness of dried bananas. The article aimed to study the effect of different drying times (6, 8, and 10 hours) using an air dryer on some physical and chemical characteristics of bananas, including CIE-L*a*b, water content, carbohydrates, and sweetness. Also predicting the sweetness of dried bananas based on the CIE-L*a*b ratios using machine learn- ing algorithms RF, SVM, LDA, KNN, and CART. The results showed that increasing the drying
... Show MoreIn many oil-recovery systems, relative permeabilities (kr) are essential flow factors that affect fluid dispersion and output from petroleum resources. Traditionally, taking rock samples from the reservoir and performing suitable laboratory studies is required to get these crucial reservoir properties. Despite the fact that kr is a function of fluid saturation, it is now well established that pore shape and distribution, absolute permeability, wettability, interfacial tension (IFT), and saturation history all influence kr values. These rock/fluid characteristics vary greatly from one reservoir region to the next, and it would be impossible to make kr measurements in all of them. The unsteady-state approach was used to calculate the relat
... Show MoreAbstract
The research Compared two methods for estimating fourparametersof the compound exponential Weibull - Poisson distribution which are the maximum likelihood method and the Downhill Simplex algorithm. Depending on two data cases, the first one assumed the original data (Non-polluting), while the second one assumeddata contamination. Simulation experimentswere conducted for different sample sizes and initial values of parameters and under different levels of contamination. Downhill Simplex algorithm was found to be the best method for in the estimation of the parameters, the probability function and the reliability function of the compound distribution in cases of natural and contaminateddata.
... Show More
Journal of Theoretical and Applied Information Technology is a peer-reviewed electronic research papers & review papers journal with aim of promoting and publishing original high quality research dealing with theoretical and scientific aspects in all disciplines of IT (Informaiton Technology
Shear wave velocity is an important feature in the seismic exploration that could be utilized in reservoir development strategy and characterization. Its vital applications in petrophysics, seismic, and geomechanics to predict rock elastic and inelastic properties are essential elements of good stability and fracturing orientation, identification of matrix mineral and gas-bearing formations. However, the shear wave velocity that is usually obtained from core analysis which is an expensive and time-consuming process and dipole sonic imager tool is not commonly available in all wells. In this study, a statistical method is presented to predict shear wave velocity from wireline log data. The model concentrated to predict shear wave velocity fr
... Show MoreImage quality has been estimated and predicted using the signal to noise ratio (SNR). The purpose of this study is to investigate the relationships between body mass index (BMI) and SNR measurements in PET imaging using patient studies with liver cancer. Three groups of 59 patients (24 males and 35 females) were divided according to BMI. After intravenous injection of 0.1 mCi of 18F-FDG per kilogram of body weight, PET emission scans were acquired for (1, 1.5, and 3) min/bed position according to the weight of patient. Because liver is an organ of homogenous metabolism, five region of interest (ROI) were made at the same location, five successive slices of the PET/CT scans to determine the mean uptake (signal) values and its standard deviat
... Show MoreBreast cancer is the commonest cancer affecting women worldwide. Different studies have dealt with the etiological factors of that cancer aiming to find a way for early diagnosis and satisfactory therapy. The present study clarified the relationship between genetic polymorphisms of BRCA1 & BRCA2 genes and some etiological risk factors among breast cancer patients in Iraq. This investigation was carried out on 25 patients (all were females) who were diagnosed as breast cancer patients attended AL-Kadhemya Teaching Hospital in Baghdad and 10 apparently healthy women were used as a control, all women (patients and control) aged above 40 years. The Wizard Promega kit was used for DNA isolation from breast patients and normal individuals. B
... Show MoreAbstract—In this study, we present the experimental results of ultra-wideband (UWB) imaging oriented for detecting small malignant breast tumors at an early stage. The technique is based on radar sensing, whereby tissues are differentiated based on the dielectric contrast between the disease and its surrounding healthy tissues. The image reconstruction algorithm referred to herein as the enhanced version of delay and sum (EDAS) algorithm is used to identify the malignant tissue in a cluttered environment and noisy data. The methods and procedures are tested using MRI-derived breast phantoms, and the results are compared with images obtained from classical DAS variant. Incorporating a new filtering technique and multiplication procedure, t
... Show MoreRadiation therapy plays an important role in improving breast cancer cases, in order to obtain an appropriateestimate of radiation doses number given to the patient after tumor removal; some methods of nonparametric regression werecompared. The Kernel method was used by Nadaraya-Watson estimator to find the estimation regression function forsmoothing data based on the smoothing parameter h according to the Normal scale method (NSM), Least Squared CrossValidation method (LSCV) and Golden Rate Method (GRM). These methods were compared by simulation for samples ofthree sizes, the method (NSM) proved to be the best according to average of Mean Squares Error criterion and the method(LSCV) proved to be the best according to Average of Mean Absolu
... Show More