يدرس هذا البحث طرائق اختزال الابعاد التي تعمل على تجاوز مشكلة البعدية عندما تفشل الطرائق التقليدية في ايجاد تقدير جيد للمعلمات، لذلك يتوجب التعامل مع هذه المشكلة بشكل مباشر. ومن اجل ذلك، يجب التخلص من هذه المشكلة لذا تم استعمال اسلوبين لحل مشكلة البيانات ذات الابعاد العالية الاسلوب الاول طريقة الانحدار الشرائحي المعكوس SIR ) ) والتي تعتبر طريقة غير كلاسيكية وكذلك طريقة ( WSIR ) المقترحة والاسلوب الثاني طريقة المركبات الرئيسة ( PCA ) وهي الطريقة العامة المستخدمة في اختزال الابعاد , ان عمل طريقة انحدار الشرائحي المعكوس SIR ) ) و طريقة المركبات الرئيسة (PCA) يقوم على عمل توليفات خطية مختزلة من مجموعة جزئية من المتغيرات التوضيحية الأصلية والتي قد تعاني من مشكلة عدم التجانس ومن مشكلة التعدد الخطي بين معظم المتغيرات التوضيحية , وستقوم هذه التوليفات الجديدة المتمثلة بالمركبات الخطية الناتجة من الطريقتين بإختزال أكثر عدد من المتغيرات التوضيحية للوصول الى بُعد جديد واحد او اكثر يسمى بالبعد الفعّال . وسيتم استعمال معيار جذر متوسط مربعات الخطأ للمقارنة بين الاسلوبين لبيان افضلية الطرائق , وقد تم اجراء دراسة محاكاة للمقارنة بين الطرائق المستعملة وقد بينت نتائج المحاكاة ان طريقة weight standard Sir المقترحة هي الافضل .
In this paper, the error distribution function is estimated for the single index model by the empirical distribution function and the kernel distribution function. Refined minimum average variance estimation (RMAVE) method is used for estimating single index model. We use simulation experiments to compare the two estimation methods for error distribution function with different sample sizes, the results show that the kernel distribution function is better than the empirical distribution function.
In this paper, the fuzzy logic and the trapezoidal fuzzy intuitionistic number were presented, as well as some properties of the trapezoidal fuzzy intuitionistic number and semi- parametric logistic regression model when using the trapezoidal fuzzy intuitionistic number. The output variable represents the dependent variable sometimes cannot be determined in only two cases (response, non-response)or (success, failure) and more than two responses, especially in medical studies; therefore so, use a semi parametric logistic regression model with the output variable (dependent variable) representing a trapezoidal fuzzy intuitionistic number.
the model was estimated on simulati
... Show MoreThe last few years witnessed great and increasing use in the field of medical image analysis. These tools helped the Radiologists and Doctors to consult while making a particular diagnosis. In this study, we used the relationship between statistical measurements, computer vision, and medical images, along with a logistic regression model to extract breast cancer imaging features. These features were used to tell the difference between the shape of a mass (Fibroid vs. Fatty) by looking at the regions of interest (ROI) of the mass. The final fit of the logistic regression model showed that the most important variables that clearly affect breast cancer shape images are Skewness, Kurtosis, Center of mass, and Angle, with an AUCROC of
... Show MoreThe using of the parametric models and the subsequent estimation methods require the presence of many of the primary conditions to be met by those models to represent the population under study adequately, these prompting researchers to search for more flexible models of parametric models and these models were nonparametric models.
In this manuscript were compared to the so-called Nadaraya-Watson estimator in two cases (use of fixed bandwidth and variable) through simulation with different models and samples sizes. Through simulation experiments and the results showed that for the first and second models preferred NW with fixed bandwidth fo
... Show MoreCurrently, one of the topical areas of application of machine learning methods is the prediction of material characteristics. The aim of this work is to develop machine learning models for determining the rheological properties of polymers from experimental stress relaxation curves. The paper presents an overview of the main directions of metaheuristic approaches (local search, evolutionary algorithms) to solving combinatorial optimization problems. Metaheuristic algorithms for solving some important combinatorial optimization problems are described, with special emphasis on the construction of decision trees. A comparative analysis of algorithms for solving the regression problem in CatBoost Regressor has been carried out. The object of
... Show MoreThe aim of this research is to use robust technique by trimming, as the analysis of maximum likelihood (ML) often fails in the case of outliers in the studied phenomenon. Where the (MLE) will lose its advantages because of the bad influence caused by the Outliers. In order to address this problem, new statistical methods have been developed so as not to be affected by the outliers. These methods have robustness or resistance. Therefore, maximum trimmed likelihood: (MTL) is a good alternative to achieve more results. Acceptability and analogies, but weights can be used to increase the efficiency of the resulting capacities and to increase the strength of the estimate using the maximum weighted trimmed likelihood (MWTL). In order to perform t
... Show MoreThe support vector machine, also known as SVM, is a type of supervised learning model that can be used for classification or regression depending on the datasets. SVM is used to classify data points by determining the best hyperplane between two or more groups. Working with enormous datasets, on the other hand, might result in a variety of issues, including inefficient accuracy and time-consuming. SVM was updated in this research by applying some non-linear kernel transformations, which are: linear, polynomial, radial basis, and multi-layer kernels. The non-linear SVM classification model was illustrated and summarized in an algorithm using kernel tricks. The proposed method was examined using three simulation datasets with different sample
... Show MoreSpatial data observed on a group of areal units is common in scientific applications. The usual hierarchical approach for modeling this kind of dataset is to introduce a spatial random effect with an autoregressive prior. However, the usual Markov chain Monte Carlo scheme for this hierarchical framework requires the spatial effects to be sampled from their full conditional posteriors one-by-one resulting in poor mixing. More importantly, it makes the model computationally inefficient for datasets with large number of units. In this article, we propose a Bayesian approach that uses the spectral structure of the adjacency to construct a low-rank expansion for modeling spatial dependence. We propose a pair of computationally efficient estimati
... Show MoreA mixture model is used to model data that come from more than one component. In recent years, it became an effective tool in drawing inferences about the complex data that we might come across in real life. Moreover, it can represent a tremendous confirmatory tool in classification observations based on similarities amongst them. In this paper, several mixture regression-based methods were conducted under the assumption that the data come from a finite number of components. A comparison of these methods has been made according to their results in estimating component parameters. Also, observation membership has been inferred and assessed for these methods. The results showed that the flexible mixture model outperformed the
... Show More