In this paper, the Monte-Carlo simulation method was used to compare the robust circular S estimator with the circular Least squares method in the case of no outlier data and in the case of the presence of an outlier in the data through two trends, the first is contaminant with high inflection points that represents contaminant in the circular independent variable, and the second the contaminant in the vertical variable that represents the circular dependent variable using three comparison criteria, the median standard error (Median SE), the median of the mean squares of error (Median MSE), and the median of the mean cosines of the circular residuals (Median A(k)). It was concluded that the method of least squares is better than the methods of the robust circular S method in the case that the data does not contain outlier values because it was recorded the lowest mean criterion, mean squares error (Median MSE), the least median standard error (Median SE) and the largest value of the criterion of the mean cosines of the circular residuals A(K) for all proposed sample sizes (n=20, 50, 100). In the case of the contaminant in the vertical data, it was found that the circular least squares method is not preferred at all contaminant rates and for all sample sizes, and the higher the percentage of contamination in the vertical data, the greater the preference of the validity of estimation methods, where the mean criterion of median squares of error (Median MSE) and criterion of median standard error (Median SE) decrease and the value of the mean criterion of the mean cosines of the circular residuals A(K) increases for all proposed sample sizes. In the case of the contaminant at high lifting points, the circular least squares method is not preferred by a large percentage at all levels of contaminant and for all sample sizes, and the higher the percentage of the contaminant at the lifting points, the greater the preference of the validity estimation methods, so that the mean criterion of mean squares of error (Median MSE) and criterion of median standard error (Median SE) decrease, and the value of the mean criterion increases for the mean cosines of the circular residuals A(K) and for all sample sizes.
Decision making is vital and important activity in field operations research ,engineering ,administration science and economic science with any industrial or service company or organization because the core of management process as well as improve him performance . The research includes decision making process when the objective function is fraction function and solve models fraction programming by using some fraction programming methods and using goal programming method aid programming ( win QSB )and the results explain the effect use the goal programming method in decision making process when the objective function is
fraction .
In this work, results from an optical technique (laser speckle technique) for measuring surface roughness was done by using statistical properties of speckle pattern from the point of view of computer image texture analysis. Four calibration relationships were used to cover wide range of measurement with the same laser speckle technique. The first one is based on intensity contrast of the speckle, the second is based on analysis of speckle binary image, the third is on size of speckle pattern spot, and the latest one is based on characterization of the energy feature of the gray level co-occurrence matrices for the speckle pattern. By these calibration relationships surface roughness of an object surface can be evaluated within the
... Show MoreIn this paper, an enhanced artificial potential field (EAPF) planner is introduced. This planner is proposed to rapidly find online solutions for the mobile robot path planning problems, when the underlying environment contains obstacles with unknown locations and sizes. The classical artificial potential field represents both the repulsive force due to the detected obstacle and the attractive force due to the target. These forces can be considered as the primary directional indicator for the mobile robot. However, the classical artificial potential field has many drawbacks. So, we suggest two secondary forces which are called the midpoint
... Show MoreThis paper proposes a novel method for generating True Random Numbers (TRNs) using electromechanical switches. The proposed generator is implemented using an FPGA board. The system utilizes the phenomenon of electromechanical switch bounce to produce a randomly fluctuated signal that is used to trigger a counter to generate a binary random number. Compared to other true random number generation methods, the proposed approach features a high degree of randomness using a simple circuit that can be easily built using off-the-shelf components. The proposed system is implemented using a commercial relay circuit connected to an FPGA board that is used to process and record the generated random sequences. Applying statistical testing on the exp
... Show Morein this paper the collocation method will be solve ordinary differential equations of retarted arguments also some examples are presented in order to illustrate this approach
This paper is specifically a detailed review of the Spatial Quantile Autoregressive (SARQR) model that refers to the incorporation of quantile regression models into spatial autoregressive models to facilitate an improved analysis of the characteristics of spatially dependent data. The relevance of SARQR is emphasized in most applications, including but not limited to the fields that might need the study of spatial variation and dependencies. In particular, it looks at literature dated from 1971 and 2024 and shows the extent to which SARQR had already been applied previously in other disciplines such as economics, real estate, environmental science, and epidemiology. Accordingly, evidence indicates SARQR has numerous benefits compar
... Show MoreInterval methods for verified integration of initial value problems (IVPs) for ODEs have been used for more than 40 years. For many classes of IVPs, these methods have the ability to compute guaranteed error bounds for the flow of an ODE, where traditional methods provide only approximations to a solution. Overestimation, however, is a potential drawback of verified methods. For some problems, the computed error bounds become overly pessimistic, or integration even breaks down. The dependency problem and the wrapping effect are particular sources of overestimations in interval computations. Berz (see [1]) and his co-workers have developed Taylor model methods, which extend interval arithmetic with symbolic computations. The latter is an ef
... Show More