Shadow removal is crucial for robot and machine vision as the accuracy of object detection is greatly influenced by the uncertainty and ambiguity of the visual scene. In this paper, we introduce a new algorithm for shadow detection and removal based on different shapes, orientations, and spatial extents of Gaussian equations. Here, the contrast information of the visual scene is utilized for shadow detection and removal through five consecutive processing stages. In the first stage, contrast filtering is performed to obtain the contrast information of the image. The second stage involves a normalization process that suppresses noise and generates a balanced intensity at a specific position compared to the neighboring intensities. In the third stage, the boundary of the target object is extracted, and in the fourth and fifth stages, respectively, the region of interest (ROI) is highlighted and reconstructed. Our model was tested and evaluated using realistic scenarios which include outdoor and indoor scenes. The results reflect the ability of our approach to detect and remove shadows and reconstruct a shadow free image with a small error of approximately 6%.
The present work presents design and implementation of an automated two-axis solar tracking system using local materials with minimum cost, light weight and reliable structure. The tracking system consists of two parts, mechanical units (fixed and moving parts) and control units (four LDR sensors and Arduino UNO microcontroller to control two DC servomotors). The tracking system was fitted and assembled together with a parabolic trough solar concentrator (PTSC) system to move it according to information come from the sensors so as to keep the PTSC always perpendicular to sun rays. The experimental tests have been done on the PTSC system to investigate its thermal performance in two cases, with tracking system (case 1) and without trackin
... Show MoreIn this paper we estimate the coefficients and scale parameter in linear regression model depending on the residuals are of type 1 of extreme value distribution for the largest values . This can be regard as an improvement for the studies with the smallest values . We study two estimation methods ( OLS & MLE ) where we resort to Newton – Raphson (NR) and Fisher Scoring methods to get MLE estimate because the difficulty of using the usual approach with MLE . The relative efficiency criterion is considered beside to the statistical inference procedures for the extreme value regression model of type 1 for largest values . Confidence interval , hypothesis testing for both scale parameter and regression coefficients
... Show More The research aims to (identify the applications of pedagogy in art education), the research community included, art education for the primary stage, so the community consisted of (8) main areas in art education, either the research sample was chosen, two main areas (objectives, and content), and included the research methodology (descriptive and analytical), the researcher built the research tool represented (the validity form of the tool) and presented to a group of experts to indicate its validity as well as to measure its stability, To show the results, the researcher used the percentage, and the researcher recommended - modifying the curriculum every period of time, such as every four years, others
The simulation is the oldest theory in art, since it appeared in the Greek aesthetic thought of the philosopher Plato, as we find in many of the thinkers and philosophers over a wide period of time to reach our world today. Our fascination with art in general and design art in particular is due to the creativity and innovations of the artist through the simulation, as well as the peculiarities in this simulation, which give objects signs and signals that may have an echo that sometimes does not exist in their physical reality.
The real representation of life and design construction, descriptions of the expression of each of them in the form of intellectual construction and the ideas of producti
... Show MoreMany approaches of different complexity already exist to edge detection in
color images. Nevertheless, the question remains of how different are the results
when employing computational costly techniques instead of simple ones. This
paper presents a comparative study on two approaches to color edge detection to
reduce noise in image. The approaches are based on the Sobel operator and the
Laplace operator. Furthermore, an efficient algorithm for implementing the two
operators is presented. The operators have been applied to real images. The results
are presented in this paper. It is shown that the quality of the results increases by
using second derivative operator (Laplace operator). And noise reduced in a good
Lost circulation or losses in drilling fluid is one of the most important problems in the oil and gas industry, and it appeared at the beginning of this industry, which caused many problems during the drilling process, which may lead to closing the well and stopping the drilling process. The drilling muds are relatively expensive, especially the muds that contain oil-based mud or that contain special additives, so it is not economically beneficial to waste and lose these muds. The treatment of drilling fluid losses is also somewhat expensive as a result of the wasted time that it caused, as well as the high cost of materials used in the treatment such as heavy materials, cement, and others. The best way to deal with drilling fluid losses
... Show MoreThis paper present a simple and sensitive method for the determination of DL-Histidine using FIA-Chemiluminometric measurement resulted from oxidation of luminol molecule by hydrogen peroxide in alkaline medium in the presence of DL-Histidine. Using 70?l. sample linear plot with a coefficient of determination 95.79% for (5-60) mmol.L-1 while for a quadratic relation C.O.D = 96.44% for (5-80) mmol.L-1 and found that guadratic plot in more representative. Limit of detection was 31.93 ?g DL-Histidine (S/N = 3), repeatability of measurement was less that 5% (n=6). Positive and negative ion interferances was removed by using minicolume containing ion exchange resin located after injection valve position.
To expedite the learning process, a group of algorithms known as parallel machine learning algorithmscan be executed simultaneously on several computers or processors. As data grows in both size andcomplexity, and as businesses seek efficient ways to mine that data for insights, algorithms like thesewill become increasingly crucial. Data parallelism, model parallelism, and hybrid techniques are justsome of the methods described in this article for speeding up machine learning algorithms. We alsocover the benefits and threats associated with parallel machine learning, such as data splitting,communication, and scalability. We compare how well various methods perform on a variety ofmachine learning tasks and datasets, and we talk abo
... Show MoreConvolutional Neural Networks (CNN) have high performance in the fields of object recognition and classification. The strength of CNNs comes from the fact that they are able to extract information from raw-pixel content and learn features automatically. Feature extraction and classification algorithms can be either hand-crafted or Deep Learning (DL) based. DL detection approaches can be either two stages (region proposal approaches) detector or a single stage (non-region proposal approach) detector. Region proposal-based techniques include R-CNN, Fast RCNN, and Faster RCNN. Non-region proposal-based techniques include Single Shot Detector (SSD) and You Only Look Once (YOLO). We are going to compare the speed and accuracy of Faster RCNN,
... Show More<p><span>A Botnet is one of many attacks that can execute malicious tasks and develop continuously. Therefore, current research introduces a comparison framework, called BotDetectorFW, with classification and complexity improvements for the detection of Botnet attack using CICIDS2017 dataset. It is a free online dataset consist of several attacks with high-dimensions features. The process of feature selection is a significant step to obtain the least features by eliminating irrelated features and consequently reduces the detection time. This process implemented inside BotDetectorFW using two steps; data clustering and five distance measure formulas (cosine, dice, driver & kroeber, overlap, and pearson correlation
... Show More