Data scarcity is a major challenge when training deep learning (DL) models. DL demands a large amount of data to achieve exceptional performance. Unfortunately, many applications have small or inadequate data to train DL frameworks. Usually, manual labeling is needed to provide labeled data, which typically involves human annotators with a vast background of knowledge. This annotation process is costly, time-consuming, and error-prone. Usually, every DL framework is fed by a significant amount of labeled data to automatically learn representations. Ultimately, a larger amount of data would generate a better DL model and its performance is also application dependent. This issue is the main barrier for many applications dismissing the use of DL. Having sufficient data is the first step toward any successful and trustworthy DL application. This paper presents a holistic survey on state-of-the-art techniques to deal with training DL models to overcome three challenges including small, imbalanced datasets, and lack of generalization. This survey starts by listing the learning techniques. Next, the types of DL architectures are introduced. After that, state-of-the-art solutions to address the issue of lack of training data are listed, such as Transfer Learning (TL), Self-Supervised Learning (SSL), Generative Adversarial Networks (GANs), Model Architecture (MA), Physics-Informed Neural Network (PINN), and Deep Synthetic Minority Oversampling Technique (DeepSMOTE). Then, these solutions were followed by some related tips about data acquisition needed prior to training purposes, as well as recommendations for ensuring the trustworthiness of the training dataset. The survey ends with a list of applications that suffer from data scarcity, several alternatives are proposed in order to generate more data in each application including Electromagnetic Imaging (EMI), Civil Structural Health Monitoring, Medical imaging, Meteorology, Wireless Communications, Fluid Mechanics, Microelectromechanical system, and Cybersecurity. To the best of the authors’ knowledge, this is the first review that offers a comprehensive overview on strategies to tackle data scarcity in DL.
The Environmental Data Acquisition Telemetry System is a versatile, flexible and economical means to accumulate data from multiple sensors at remote locations over an extended period of time; the data is normally transferred to the final destination and saved for further analysis.
This paper introduces the design and implementation of a simplified, economical and practical telemetry system to collect and transfer the environmental parameters (humidity, temperature, pressure etc.) from a remote location (Rural Area) to the processing and displaying unit.
To get a flexible and practical system, three data transfer methods (three systems) were proposed (including the design and implementation) for rural area services, the fi
... Show MoreThis investigation integrates experimental and numerical approaches to study a novel solar air heater aimed at achieving an efficient design for a solar collector suitable for drying applications under the meteorological conditions of Iraq. The importance of this investigation stems from the lack of optimal exploitation of solar energy reaching the solar collector, primarily attributable to elevated thermal losses despite numerous designs employed in such solar systems. Consequently, enhancing the thermal performance of solar collectors, particularly those employed in crop drying applications, stands as a crucial focal point for researchers within this domain. Two identical double-pass solar air heaters were designed and constructed for
... Show MoreThis paper presents an analytical study for the magnetohydrodynamic (MHD) flow of a generalized Burgers’ fluid in an annular pipe. Closed from solutions for velocity is obtained by using finite Hankel transform and discrete Laplace transform of the sequential fractional derivatives. Finally, the figures are plotted to show the effects of different parameters on the velocity profile.
Despite the principle of separation of powers brought by the French Revolution, which entrusted the task of drafting legislation and its amendment to the legislative authority and the task of settling disputes and settling them in the judiciary. However, since that date, the French judiciary has played a major role in the development of French civil law (In spite of all the economic and social developments that have taken place in French society throughout these years) since its promulgation until February of 2016, the date of the Legislative Decree No. 131 of the year 2016 A modification is the largest in the history of the French Civil Code (which was the judicial precedents in which a significant impact), was assisted by the French judic
... Show MoreThe research aims to indicate the relationship between lean production tools included seven {constant improvement , and Just in time (JIT), and the production smoothing , and quality at the source, and standardized work, Visual management, and activities 5S } and Mass Customization strategy for the model (Pine & Gilomer, 1997) {collaborative, adaptive, cosmetic, transparent}, as well as providing a conceptual framework and applied for variables search to clarify how they will choose a Mass Customization strategy through the lean production tools, , and recognize the reality of the practices of Iraqi industries in such a field. Moreover, aims to highlight the positive aspects that accrue to companies a
... Show MoreNowadays, internet security is a critical concern; the One of the most difficult study issues in network security is "intrusion detection". Fight against external threats. Intrusion detection is a novel method of securing computers and data networks that are already in use. To boost the efficacy of intrusion detection systems, machine learning and deep learning are widely deployed. While work on intrusion detection systems is already underway, based on data mining and machine learning is effective, it requires to detect intrusions by training static batch classifiers regardless considering the time-varying features of a regular data stream. Real-world problems, on the other hand, rarely fit into models that have such constraints. Furthermor
... Show MoreThe study includes collection of data about cholera disease from six health centers from nine locations with 2500km2 and a population of 750000individual. The average of infection for six centers during the 2000-2003 was recorded. There were 3007 cases of diarrhea diagnosed as cholera caused by Vibrio cholerae. The percentage of male infection was 14. 7% while for female were 13. 2%. The percentage of infection for children (less than one year) was 6.1%, it while for the age (1-5 years) was 6.9%and for the ages more than 5 years was 14.5%.The total percentage of the patients stayed in hospital was 7.7%(4.2%for male and 3.4%for female). The bacteria was isolated and identified from 7cases in the Central Laboratory for Health in Baghdad. In
... Show MoreThe objective of the research is to identify the efficiency of risk management in various names at Baghdad International Airport in the face of various risks (financial - technical - human - natural ..) facing the sample of the search of the General Establishment of Civil Aviation and the Iraqi Airways Company where the researcher identified the hypothesis that summarizes There is a significant significant correlation between risk management, risk management and risk review and assessment. The researcher used the means of research from observation and interviews with the relevant officials in this field, as well as used the questionnaire and distributed a sample of 170 employees in the field of risk management (SMS Department) in Iraqi A
... Show More