Big data analysis is essential for modern applications in areas such as healthcare, assistive technology, intelligent transportation, environment and climate monitoring. Traditional algorithms in data mining and machine learning do not scale well with data size. Mining and learning from big data need time and memory efficient techniques, albeit the cost of possible loss in accuracy. We have developed a data aggregation structure to summarize data with large number of instances and data generated from multiple data sources. Data are aggregated at multiple resolutions and resolution provides a trade-off between efficiency and accuracy. The structure is built once, updated incrementally, and serves as a common data input for multiple mining and learning algorithms. Data mining algorithms are modified to accept the aggregated data as input. Hierarchical data aggregation serves as a paradigm under which novel …
Tight reservoirs have attracted the interest of the oil industry in recent years according to its significant impact on the global oil product. Several challenges are present when producing from these reservoirs due to its low to extra low permeability and very narrow pore throat radius. Development strategy selection for these reservoirs such as horizontal well placement, hydraulic fracture design, well completion, and smart production program, wellbore stability all need accurate characterizations of geomechanical parameters for these reservoirs. Geomechanical properties, including uniaxial compressive strength (UCS), static Young’s modulus (Es), and Poisson’s ratio (υs), were measured experimentally using both static and dynamic met
... Show MoreAbstract
The research Compared two methods for estimating fourparametersof the compound exponential Weibull - Poisson distribution which are the maximum likelihood method and the Downhill Simplex algorithm. Depending on two data cases, the first one assumed the original data (Non-polluting), while the second one assumeddata contamination. Simulation experimentswere conducted for different sample sizes and initial values of parameters and under different levels of contamination. Downhill Simplex algorithm was found to be the best method for in the estimation of the parameters, the probability function and the reliability function of the compound distribution in cases of natural and contaminateddata.
... Show More
Newly series of 6,6’-((2-(Aryl)dihydropyrimidine-1,3(2H,4H)-diyl)bis(methylene))bis(2-methoxy phenol) (3a-i) were synthesized from cyclization of 6,6’-((propane-1,3-diylbis (azanediyl)) bis(methylene)) bis(2-methoxyphenol) with several aryl aldehyde in the presence of acetic acid. The newly compounds characterized from their IR, NMR and EIMs spectra. The antioxidant capacity of these compounds screened by utilizing DPPH and FRAP assays. Compounds 3g and 3i exhibited significant antioxidant capability in both assays. Docking study for these compounds as a potential inhibitors of gyrase enzyme were carried out. Compound 3g exhibited significant inhibition with binding free energies (DG) higher than novobiocin. compounds 2, 3a, 3b, 3
... Show MoreThe communication inspiration formed an essential foundations for contribute the influence individuals and recipients, whether negatively or positively, through the messages that were published and presented in them with multiple themes and viewpoints that covered all parts of the world and all age groups; it is directed to children addressing the various stages of childhood, as it simulates many goals, including what is directed through the digital use of educational data in television production, as it is considered an intellectual and mental bag to deliver ideas and expressive and aesthetic connotations to children, where the songs and cartoons carrying data on education; within adjacent relations and in a mutual direction, both of th
... Show MoreExperimental activity coefficients at infinite dilution are particularly useful for calculating the parameters needed in an expression for the excess Gibbs energy. If reliable values of γ∞1 and γ∞2 are available, either from direct experiment or from a correlation, it is possible to predict the composition of the azeotrope and vapor-liquid equilibrium over the entire range of composition. These can be used to evaluate two adjustable constants in any desired expression for G E. In this study MOSCED model and SPACE model are two different methods were used to calculate γ∞1 and γ∞2
OpenStreetMap (OSM), recognised for its current and readily accessible spatial database, frequently serves regions lacking precise data at the necessary granularity. Global collaboration among OSM contributors presents challenges to data quality and uniformity, exacerbated by the sheer volume of input and indistinct data annotation protocols. This study presents a methodological improvement in the spatial accuracy of OSM datasets centred over Baghdad, Iraq, utilising data derived from OSM services and satellite imagery. An analytical focus was placed on two geometric correction methods: a two-dimensional polynomial affine transformation and a two-dimensional polynomial conformal transformation. The former involves twelve coefficients for ad
... Show MoreThe Arabic grammatical theory is characterized by the characteristics that distinguish it from other languages. It is based on the following equation: In its entirety a homogeneous linguistic system that blends with the social nature of the Arab, his beliefs, and his culture.
This means that this theory was born naturally, after the labor of maintaining an integrated inheritance, starting with its legal text (the Koran), and ends with its features of multiple attributes.
Saber was carrying the founding crucible of that theory, which takes over from his teacher, Hebron, to be built on what it has reached. It is redundant to point to his location and the status of his book.
So came to my research tagged: (c
In recent years, data centre (DC) networks have improved their rapid exchanging abilities. Software-defined networking (SDN) is presented to alternate the impression of conventional networks by segregating the control plane from the SDN data plane. The SDN presented overcomes the limitations of traditional DC networks caused by the rapidly incrementing amounts of apps, websites, data storage needs, etc. Software-defined networking data centres (SDN-DC), based on the open-flow (OF) protocol, are used to achieve superior behaviour for executing traffic load-balancing (LB) jobs. The LB function divides the traffic-flow demands between the end devices to avoid links congestion. In short, SDN is proposed to manage more operative configur
... Show MoreAcceptable Bit Error rate can be maintained by adapting some of the design parameters such as modulation, symbol rate, constellation size, and transmit power according to the channel state.
An estimate of HF propagation effects can be used to design an adaptive data transmission system over HF link. The proposed system combines the well known Automatic Link Establishment (ALE) together with variable rate transmission system. The standard ALE is modified to suite the required goal of selecting the best carrier frequency (channel) for a given transmission. This is based on measuring SINAD (Signal plus Noise plus Distortion to Noise plus Distortion), RSL (Received Signal Level), multipath phase distortion and BER (Bit Error Rate) fo
... Show More