Energy savings are very common in IoT sensor networks because IoT sensor nodes operate with their own limited battery. The data transmission in the IoT sensor nodes is very costly and consume much of the energy while the energy usage for data processing is considerably lower. There are several energy-saving strategies and principles, mainly dedicated to reducing the transmission of data. Therefore, with minimizing data transfers in IoT sensor networks, can conserve a considerable amount of energy. In this research, a Compression-Based Data Reduction (CBDR) technique was suggested which works in the level of IoT sensor nodes. The CBDR includes two stages of compression, a lossy SAX Quantization stage which reduces the dynamic range of the sensor data readings, after which a lossless LZW compression to compress the loss quantization output. Quantizing the sensor node data readings down to the alphabet size of SAX results in lowering, to the advantage of the best compression sizes, which contributes to greater compression from the LZW end of things. Also, another improvement was suggested to the CBDR technique which is to add a Dynamic Transmission (DT-CBDR) to decrease both the total number of data sent to the gateway and the processing required. OMNeT++ simulator along with real sensory data gathered at Intel Lab is used to show the performance of the proposed technique. The simulation experiments illustrate that the proposed CBDR technique provides better performance than the other techniques in the literature.
Compression for color image is now necessary for transmission and storage in the data bases since the color gives a pleasing nature and natural for any object, so three composite techniques based color image compression is implemented to achieve image with high compression, no loss in original image, better performance and good image quality. These techniques are composite stationary wavelet technique (S), composite wavelet technique (W) and composite multi-wavelet technique (M). For the high energy sub-band of the 3rd level of each composite transform in each composite technique, the compression parameters are calculated. The best composite transform among the 27 types is the three levels of multi-wavelet
... Show MoreThe problem of research was to identify after the use of cost technology based on specifications in the validity of determining and measuring the costs of the implementation of contracting, by applying to al-Mansour General Construction Contracting Company as an appropriate alternative to the traditional costing system currently adopted, which is characterized by many shortcomings and weaknesses Which has been reflected in the validity and integrity of the calculations. To solve this problem, the research was based on the premise that: (The application of cost technology based on specifications will result in calculating the cost of the product according to the specification required by the customer, to meet his wishes properly and witho
... Show MoreThe rise of Industry 4.0 and smart manufacturing has highlighted the importance of utilizing intelligent manufacturing techniques, tools, and methods, including predictive maintenance. This feature allows for the early identification of potential issues with machinery, preventing them from reaching critical stages. This paper proposes an intelligent predictive maintenance system for industrial equipment monitoring. The system integrates Industrial IoT, MQTT messaging and machine learning algorithms. Vibration, current and temperature sensors collect real-time data from electrical motors which is analyzed using five ML models to detect anomalies and predict failures, enabling proactive maintenance. The MQTT protocol is used for efficient com
... Show MoreIn regression testing, Test case prioritization (TCP) is a technique to arrange all the available test cases. TCP techniques can improve fault detection performance which is measured by the average percentage of fault detection (APFD). History-based TCP is one of the TCP techniques that consider the history of past data to prioritize test cases. The issue of equal priority allocation to test cases is a common problem for most TCP techniques. However, this problem has not been explored in history-based TCP techniques. To solve this problem in regression testing, most of the researchers resort to random sorting of test cases. This study aims to investigate equal priority in history-based TCP techniques. The first objective is to implement
... Show MorePolyaniline Multi wall Carbon nanotube (PANI/MWCNTs) nanocomposite thin films have been prepared by Plasma jet polymerization at low frequency on glass substrate with preliminary deposited aluminum electrodes to form Al/PANI-MWCNT/Al surface-type capacitive humidity sensors, the gap between the electrodes about 50 μm and the MWCNTs weight concentration varied between 0, 1, 2, 3, 4%. The diameter of the MWCNTs was in the range of 8-15 nm and the length 10-55 μm. The capacitance-humidity relationships of the sensors were investigated at humidity levels from 35 to 90% RH. The electrical properties showed that the capacity increased with increasing relative humidity, and that the sensitivity of the sensor increases with the increase of the
... Show MoreLocalization is an essential demand in wireless sensor networks (WSNs). It relies on several types of measurements. This paper focuses on positioning in 3-D space using time-of-arrival- (TOA-) based distance measurements between the target node and a number of anchor nodes. Central localization is assumed and either RF, acoustic or UWB signals are used for distance measurements. This problem is treated by using iterative gradient descent (GD), and an iterative GD-based algorithm for localization of moving sensors in a WSN has been proposed. To localize a node in 3-D space, at least four anchors are needed. In this work, however, five anchors are used to get better accuracy. In GD localization of a moving sensor, the algo
... Show More
The great scientific progress has led to widespread Information as information accumulates in large databases is important in trying to revise and compile this vast amount of data and, where its purpose to extract hidden information or classified data under their relations with each other in order to take advantage of them for technical purposes.
And work with data mining (DM) is appropriate in this area because of the importance of research in the (K-Means) algorithm for clustering data in fact applied with effect can be observed in variables by changing the sample size (n) and the number of clusters (K)
... Show MoreIn this study, we made a comparison between LASSO & SCAD methods, which are two special methods for dealing with models in partial quantile regression. (Nadaraya & Watson Kernel) was used to estimate the non-parametric part ;in addition, the rule of thumb method was used to estimate the smoothing bandwidth (h). Penalty methods proved to be efficient in estimating the regression coefficients, but the SCAD method according to the mean squared error criterion (MSE) was the best after estimating the missing data using the mean imputation method