Research summarized in applying the model of fuzzy goal programming for aggregate production planning , in General Company for hydraulic industries / plastic factory to get an optimal production plan trying to cope with the impact that fluctuations in demand and employs all available resources using two strategies where they are available inventories strategy and the strategy of change in the level of the workforce, these strategies costs are usually imprecise/fuzzy. The plant administration trying to minimize total production costs, minimize carrying costs and minimize changes in labour levels. depending on the gained data from the plant of the major products for the second half of 2015 nd after the dissolution of the model using the software (GAMS) was getting the best level from the production , inventory and labour levels , where the total production costs was (7041700000) dinar Which is good compared to the cost reduction Cost set by the factory administration, which was (7668442100)dinars a difference (626742100) , carrying costs was (4327200 ) dinar Which is good compared to the cost reduction Cost set by the factory administration, which was ( 5583400)dinars a difference (1256200) and It found that the factory does not need to hire or lay off workers for the duration of the production plan. And resulting deviational value for the three fuzzy goal ( are (0.938, 1, 1) respectively. Through results we found that deviational value of the goals were close and equal for (1) that consider good , that were obtained were catered level of ambition of the decision-maker values.
The influx of data in bioinformatics is primarily in the form of DNA, RNA, and protein sequences. This condition places a significant burden on scientists and computers. Some genomics studies depend on clustering techniques to group similarly expressed genes into one cluster. Clustering is a type of unsupervised learning that can be used to divide unknown cluster data into clusters. The k-means and fuzzy c-means (FCM) algorithms are examples of algorithms that can be used for clustering. Consequently, clustering is a common approach that divides an input space into several homogeneous zones; it can be achieved using a variety of algorithms. This study used three models to cluster a brain tumor dataset. The first model uses FCM, whic
... Show MoreStatistical methods and statistical decisions making were used to arrange and analyze the primary data to get norms which are used with Geographic Information Systems (GIS) and spatial analysis programs to identify the animals production and poultry units in strategic nutrition channels, also the priorities of food insecurity through the local production and import when there is no capacity for production. The poultry production is one of the most important commodities that satisfy human body protein requirements, also the most important criteria to measure the development and prosperity of nations. The poultry fields of Babylon Governorate are located in Abi Ghareg and Al_Kifil centers according to many criteria or factors such as the popu
... Show MoreThe grey system model GM(1,1) is the model of the prediction of the time series and the basis of the grey theory. This research presents the methods for estimating parameters of the grey model GM(1,1) is the accumulative method (ACC), the exponential method (EXP), modified exponential method (Mod EXP) and the Particle Swarm Optimization method (PSO). These methods were compared based on the Mean square error (MSE) and the Mean Absolute percentage error (MAPE) as a basis comparator and the simulation method was adopted for the best of the four methods, The best method was obtained and then applied to real data. This data represents the consumption rate of two types of oils a heavy fuel (HFO) and diesel fuel (D.O) and the use of tests to conf
... Show MoreConstruction project life-cycle processes must be managed in a more effective and predictable way to meet project stakeholder needs , However, there is increasing concern about whether know-how effectively improves understanding of underlying theories of project management processes for construction organizations and their project managers Project planning and scheduling are considered as key and challenging tools in controlling and monitoring project performance , but many worldwide construction projects appear to give insufficient attention to effective management and definition of project planning, including pre planning stages . Indeed, some planning issues indeed, some planning issues have been completely ove
... Show More
The great scientific progress has led to widespread Information as information accumulates in large databases is important in trying to revise and compile this vast amount of data and, where its purpose to extract hidden information or classified data under their relations with each other in order to take advantage of them for technical purposes.
And work with data mining (DM) is appropriate in this area because of the importance of research in the (K-Means) algorithm for clustering data in fact applied with effect can be observed in variables by changing the sample size (n) and the number of clusters (K)
... Show MoreThe objectives of this research are to determine and find out the reality of crops structure of greenhouses in association of Al-Watan in order to stand on the optimal use of economic resources available for the purpose of reaching a crop structure optimization of the farm that achieves maximize profit and gross and net farm incomes , using the method of linear programming to choose the farm optimal plan with the highest net income , as well as identifying production plans farm efficient with (income - deviation) optimal (E-A) of the Association and derived, which takes into account the margin risk wich derived from each plan using the model( MOTAD), as a model of models of linear programming alternative programming m
... Show MoreIn this study, we made a comparison between LASSO & SCAD methods, which are two special methods for dealing with models in partial quantile regression. (Nadaraya & Watson Kernel) was used to estimate the non-parametric part ;in addition, the rule of thumb method was used to estimate the smoothing bandwidth (h). Penalty methods proved to be efficient in estimating the regression coefficients, but the SCAD method according to the mean squared error criterion (MSE) was the best after estimating the missing data using the mean imputation method