The support vector machine, also known as SVM, is a type of supervised learning model that can be used for classification or regression depending on the datasets. SVM is used to classify data points by determining the best hyperplane between two or more groups. Working with enormous datasets, on the other hand, might result in a variety of issues, including inefficient accuracy and time-consuming. SVM was updated in this research by applying some non-linear kernel transformations, which are: linear, polynomial, radial basis, and multi-layer kernels. The non-linear SVM classification model was illustrated and summarized in an algorithm using kernel tricks. The proposed method was examined using three simulation datasets with different sample sizes (50, 100, 200). A comparison between non-linear SVM and two standard classification methods was illustrated using various compared features. Our study has shown that the non-linear SVM method gives better results by checking: sensitivity, specificity, accuracy, and time-consuming. © 2024 Author(s).
Abstract:
This research aims to compare Bayesian Method and Full Maximum Likelihood to estimate hierarchical Poisson regression model.
The comparison was done by simulation using different sample sizes (n = 30, 60, 120) and different Frequencies (r = 1000, 5000) for the experiments as was the adoption of the Mean Square Error to compare the preference estimation methods and then choose the best way to appreciate model and concluded that hierarchical Poisson regression model that has been appreciated Full Maximum Likelihood Full Maximum Likelihood with sample size (n = 30) is the best to represent the maternal mortality data after it has been reliance value param
... Show MoreRegression testing being expensive, requires optimization notion. Typically, the optimization of test cases results in selecting a reduced set or subset of test cases or prioritizing the test cases to detect potential faults at an earlier phase. Many former studies revealed the heuristic-dependent mechanism to attain optimality while reducing or prioritizing test cases. Nevertheless, those studies were deprived of systematic procedures to manage tied test cases issue. Moreover, evolutionary algorithms such as the genetic process often help in depleting test cases, together with a concurrent decrease in computational runtime. However, when examining the fault detection capacity along with other parameters, is required, the method falls sh
... Show MoreThis study focuses on the slab-beam interaction in one-way systems. In the context of this study, slab-beam interaction means how beam deflection can affect moment distribution in one-way slabs. This interaction is usually neglected in the traditional approximate analysis that is adopted in engineering practice and design codes. Slab positive moments have been considered as indicators on the accuracy of approximate methods, as they overestimate negative moments while underestimating positive moments.
After proposing of effecting parameters in slab-beam interaction including of panel length and width, beam dimensions, and slab thickness, Buckingham’s theorem has been adopted to transform the dimensional-mo
... Show MoreTwitter is becoming an increasingly popular platform used by financial analysts to monitor and forecast financial markets. In this paper we investigate the impact of the sentiments expressed in Twitter on the subsequent market movement, specifically the bitcoin exchange rate. This study is divided into two phases, the first phase is sentiment analysis, and the second phase is correlation and regression. We analyzed tweets associated with the Bitcoin in order to determine if the user’s sentiment contained within those tweets reflects the exchange rate of the currency. The sentiment of users over a 2-month period is classified as having a positive or negative sentiment of the digital currency using the proposed CNN-LSTM
... Show MoreCowpea is a very important legume in Nigeria that is being utilized to Substitute high-cost animal protein for low-income people. The knowledge of some physical properties of various moisture contents is of utmost importance in the design of its handling and processing equipment and machinery, which is the aim of this work, which studied the physical properties of IT99K-573-1-1 (SAMPEA14) variety of Cowpea within 8.77 to 21.58 % db moisture content. The properties studied include Major, Intermediate, and Minor diameters, Sphericity, Surface area, Specific gravity, Volume, Bulk density, 50-tap density, 100-tap density, 1250-tap density, seed mass, Angle of repose, Geometric mean diameter, and Arithmetic mean diameter. The
... Show MoreAdverse drug reactions (ADR) are important information for verifying the view of the patient on a particular drug. Regular user comments and reviews have been considered during the data collection process to extract ADR mentions, when the user reported a side effect after taking a specific medication. In the literature, most researchers focused on machine learning techniques to detect ADR. These methods train the classification model using annotated medical review data. Yet, there are still many challenging issues that face ADR extraction, especially the accuracy of detection. The main aim of this study is to propose LSA with ANN classifiers for ADR detection. The findings show the effectiveness of utilizing LSA with ANN in extracting AD
... Show MoreAdministrative procedures in various organizations produce numerous crucial records and data. These
records and data are also used in other processes like customer relationship management and accounting
operations.It is incredibly challenging to use and extract valuable and meaningful information from these data
and records because they are frequently enormous and continuously growing in size and complexity.Data
mining is the act of sorting through large data sets to find patterns and relationships that might aid in the data
analysis process of resolving business issues. Using data mining techniques, enterprises can forecast future
trends and make better business decisions.The Apriori algorithm has bee
Problem: Cancer is regarded as one of the world's deadliest diseases. Machine learning and its new branch (deep learning) algorithms can facilitate the way of dealing with cancer, especially in the field of cancer prevention and detection. Traditional ways of analyzing cancer data have their limits, and cancer data is growing quickly. This makes it possible for deep learning to move forward with its powerful abilities to analyze and process cancer data. Aims: In the current study, a deep-learning medical support system for the prediction of lung cancer is presented. Methods: The study uses three different deep learning models (EfficientNetB3, ResNet50 and ResNet101) with the transfer learning concept. The three models are trained using a
... Show MoreIn this paper, an algorithm for binary codebook design has been used in vector quantization technique, which is used to improve the acceptability of the absolute moment block truncation coding (AMBTC) method. Vector quantization (VQ) method is used to compress the bitmap (the output proposed from the first method (AMBTC)). In this paper, the binary codebook can be engender for many images depending on randomly chosen to the code vectors from a set of binary images vectors, and this codebook is then used to compress all bitmaps of these images. The chosen of the bitmap of image in order to compress it by using this codebook based on the criterion of the average bitmap replacement error (ABPRE). This paper is suitable to reduce bit rates
... Show More
Regression testing is a crucial phase in the software development lifecycle that makes sure that new changes/updates in the software system don’t introduce defects or don’t affect adversely the existing functionalities. However, as the software systems grow in complexity, the number of test cases in regression suite can become large which results into more testing time and resource consumption. In addition, the presence of redundant and faulty test cases may affect the efficiency of the regression testing process. Therefore, this paper presents a new Hybrid Framework to Exclude Similar & Faulty Test Cases in Regression Testing (ETCPM) that utilizes automated code analysis techniques and historical test execution data to
... Show More