The support vector machine, also known as SVM, is a type of supervised learning model that can be used for classification or regression depending on the datasets. SVM is used to classify data points by determining the best hyperplane between two or more groups. Working with enormous datasets, on the other hand, might result in a variety of issues, including inefficient accuracy and time-consuming. SVM was updated in this research by applying some non-linear kernel transformations, which are: linear, polynomial, radial basis, and multi-layer kernels. The non-linear SVM classification model was illustrated and summarized in an algorithm using kernel tricks. The proposed method was examined using three simulation datasets with different sample sizes (50, 100, 200). A comparison between non-linear SVM and two standard classification methods was illustrated using various compared features. Our study has shown that the non-linear SVM method gives better results by checking: sensitivity, specificity, accuracy, and time-consuming. © 2024 Author(s).
This paper delves into some significant performance measures (PMs) of a bulk arrival queueing system with constant batch size b, according to arrival rates and service rates being fuzzy parameters. The bulk arrival queuing system deals with observation arrival into the queuing system as a constant group size before allowing individual customers entering to the service. This leads to obtaining a new tool with the aid of generating function methods. The corresponding traditional bulk queueing system model is more convenient under an uncertain environment. The α-cut approach is applied with the conventional Zadeh's extension principle (ZEP) to transform the triangular membership functions (Mem. Fs) fuzzy queues into a family of conventional b
... Show MoreIn this paper we used frequentist and Bayesian approaches for the linear regression model to predict future observations for unemployment rates in Iraq. Parameters are estimated using the ordinary least squares method and for the Bayesian approach using the Markov Chain Monte Carlo (MCMC) method. Calculations are done using the R program. The analysis showed that the linear regression model using the Bayesian approach is better and can be used as an alternative to the frequentist approach. Two criteria, the root mean square error (RMSE) and the median absolute deviation (MAD) were used to compare the performance of the estimates. The results obtained showed that the unemployment rates will continue to increase in the next two decade
... Show MoreAbstract
The Non - Homogeneous Poisson process is considered as one of the statistical subjects which had an importance in other sciences and a large application in different areas as waiting raws and rectifiable systems method , computer and communication systems and the theory of reliability and many other, also it used in modeling the phenomenon that occurred by unfixed way over time (all events that changed by time).
This research deals with some of the basic concepts that are related to the Non - Homogeneous Poisson process , This research carried out two models of the Non - Homogeneous Poisson process which are the power law model , and Musa –okumto , to estimate th
... Show MoreThe Purpose of this research is a comparison between two types of multivariate GARCH models BEKK and DVECH to forecast using financial time series which are the series of daily Iraqi dinar exchange rate with dollar, the global daily of Oil price with dollar and the global daily of gold price with dollar for the period from 01/01/2014 till 01/01/2016.The estimation, testing and forecasting process has been computed through the program RATS. Three time series have been transferred to the three asset returns to get the Stationarity, some tests were conducted including Ljung- Box, Multivariate Q and Multivariate ARCH to Returns Series and Residuals Series for both models with comparison between the estimation and for
... Show MoreThis paper deals with constructing a model of fuzzy linear programming with application on fuels product of Dura- refinery , which consist of seven products that have direct effect ondaily consumption . After Building the model which consist of objective function represents the selling prices ofthe products and fuzzy productions constraints and fuzzy demand constraints addition to production requirements constraints , we used program of ( WIN QSB ) to find the optimal solution
This study examined the tagged (the matter, its ruling, and examples of its applications in Ahmadiyya interpretations), the study of devising formulas of commissioning, especially the matter from the book of interpretation, which is: (Ahmadiyyeh’s interpretations), and the contemplative of the nature of (Ahmadiyya’s interpretations) realizes that the cognitive, linguistic and juristic knowledge product and the verses of judgments And other is the focus of this immortal travel, as it is truly a comprehensive book containing what the students of knowledge, scholars and the general public need in terms of interpreting the legal verses, and it is not possible to reach an understanding of the Book of God Almighty, and to clarify his purpo
... Show More<span lang="EN-US">This paper presents the comparison between optimized unscented Kalman filter (UKF) and optimized extended Kalman filter (EKF) for sensorless direct field orientation control induction motor (DFOCIM) drive. The high performance of UKF and EKF depends on the accurate selection of state and noise covariance matrices. For this goal, multi objective function genetic algorithm is used to find the optimal values of state and noise covariance matrices. The main objectives of genetic algorithm to be minimized are the mean square errors (MSE) between actual and estimation of speed, current, and flux. Simulation results show the optimal state and noise covariance matrices can improve the estimation of speed, current, t
... Show MoreSimulation experiments are a means of solving in many fields, and it is the process of designing a model of the real system in order to follow it and identify its behavior through certain models and formulas written according to a repeating software style with a number of iterations. The aim of this study is to build a model that deals with the behavior suffering from the state of (heteroskedasticity) by studying the models (APGARCH & NAGARCH) using (Gaussian) and (Non-Gaussian) distributions for different sample sizes (500,1000,1500,2000) through the stage of time series analysis (identification , estimation, diagnostic checking and prediction). The data was generated using the estimations of the parameters resulting f
... Show MoreThe Artificial Neural Network methodology is a very important & new subjects that build's the models for Analyzing, Data Evaluation, Forecasting & Controlling without depending on an old model or classic statistic method that describe the behavior of statistic phenomenon, the methodology works by simulating the data to reach a robust optimum model that represent the statistic phenomenon & we can use the model in any time & states, we used the Box-Jenkins (ARMAX) approach for comparing, in this paper depends on the received power to build a robust model for forecasting, analyzing & controlling in the sod power, the received power come from
... Show More