Search results for: interpolation techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6744

Search results for: interpolation techniques

5754 Experimental and Analytical Study on the Bending Behavior of Concrete-GFRP Hybrid Beams

Authors: Alaa Koaik, Bruno Jurkiewiez, Sylvain Bel

Abstract:

Recently, the use of GFRP pultruded profiles increased in the domain of civil engineering especially in the construction of sandwiched slabs and footbridges. However, under heavy loads, the risk of using these profiles increases due to their high deformability and instability as a result of their weak stiffness and orthotropic nature. A practical solution proposes the assembly of these profiles with concrete slabs to create a stiffer hybrid element to support higher loads. The connection of these two elements is established either by traditional means of steel studs (bolting in our case) or bonding technique. These two techniques have their advantages and disadvantages regarding the mechanical behavior and in-situ implementation. This paper presents experimental results of interface characterization and bending behavior of two hybrid beams, PB7 and PB8, designed and constructed using both connection techniques. The results obtained are exploited to design and build a hybrid footbridge BPBP1 which is tested within service limits (elastic domain). Analytical methods are also developed to analyze the behavior of these structures in the elastic range and the ultimate phase. Comparisons show acceptable differences mainly due to the sensitivity of the GFRP moduli as well as the non-linearity of concrete elements.

Keywords: analytical model, concrete, flexural behavior, GFRP pultruded profile, hybrid structure, interconnection slip, push-out

Procedia PDF Downloads 224
5753 A Comparative Study of Active Release Technique and Myofascial Release Technique in Treatment of Patients with Upper Trapezius Spasm

Authors: Daxa Mishra, R. Harihara, Ankita

Abstract:

Trapezius muscle pain is the most common musculoskeletal disorder occurring in individuals who work with awkward positions, have repetitive movements and movements with precision demands. Treatment techniques like active release technique (ART) and myofascial release (MFR) can be used to relieve muscle spasm. The aim of the study is to compare the effect of ART and MFR on the upper trapezius muscle spasm. Methodology: A series of 60 patients of both sexes between the age group of 20 and 55 with upper trapezius spasm were divided into two groups by computerized randomization. Subjects in each group received treatment in the form of either ART or MFR for the period of seven days. cervical range of motion (ROM), neck disability index scale (NDI) and visual analog scale (VAS) tools were used to measure the outcome. Results: Paired Sample ‘t’ test was used to compare the Outcome differences within each group, while Independent ‘t’ test was used to compare the differences between the two groups for the same outcome measures. The improvement was found in both the groups at 7th day following intervention, but the group which received ART showed significant improvements as compared to group which received MFR. Conclusion: Although both techniques are effective in alleviation of symptoms and associated disability in upper trapezius muscle spasm, ART gave better results as compared to MRF.

Keywords: goniometer, myofascial release, active release, physiotherapy

Procedia PDF Downloads 239
5752 Evaluation of the Self-Organizing Map and the Adaptive Neuro-Fuzzy Inference System Machine Learning Techniques for the Estimation of Crop Water Stress Index of Wheat under Varying Application of Irrigation Water Levels for Efficient Irrigation Scheduling

Authors: Aschalew C. Workneh, K. S. Hari Prasad, C. S. P. Ojha

Abstract:

The crop water stress index (CWSI) is a cost-effective, non-destructive, and simple technique for tracking the start of crop water stress. This study investigated the feasibility of CWSI derived from canopy temperature to detect the water status of wheat crops. Artificial intelligence (AI) techniques have become increasingly popular in recent years for determining CWSI. In this study, the performance of two AI techniques, adaptive neuro-fuzzy inference system (ANFIS) and self-organizing maps (SOM), are compared while determining the CWSI of paddy crops. Field experiments were conducted for varying irrigation water applications during two seasons in 2022 and 2023 at the irrigation field laboratory at the Civil Engineering Department, Indian Institute of Technology Roorkee, India. The ANFIS and SOM-simulated CWSI values were compared with the experimentally calculated CWSI (EP-CWSI). Multiple regression analysis was used to determine the upper and lower CWSI baselines. The upper CWSI baseline was found to be a function of crop height and wind speed, while the lower CWSI baseline was a function of crop height, air vapor pressure deficit, and wind speed. The performance of ANFIS and SOM were compared based on mean absolute error (MAE), mean bias error (MBE), root mean squared error (RMSE), index of agreement (d), Nash-Sutcliffe efficiency (NSE), and coefficient of correlation (R²). Both models successfully estimated the CWSI of the paddy crop with higher correlation coefficients and lower statistical errors. However, the ANFIS (R²=0.81, NSE=0.73, d=0.94, RMSE=0.04, MAE= 0.00-1.76 and MBE=-2.13-1.32) outperformed the SOM model (R²=0.77, NSE=0.68, d=0.90, RMSE=0.05, MAE= 0.00-2.13 and MBE=-2.29-1.45). Overall, the results suggest that ANFIS is a reliable tool for accurately determining CWSI in wheat crops compared to SOM.

Keywords: adaptive neuro-fuzzy inference system, canopy temperature, crop water stress index, self-organizing map, wheat

Procedia PDF Downloads 44
5751 Forecasting Future Demand for Energy Efficient Vehicles: A Review of Methodological Approaches

Authors: Dimitrios I. Tselentis, Simon P. Washington

Abstract:

Considerable literature has been focused over the last few decades on forecasting the consumer demand of Energy Efficient Vehicles (EEVs). These methodological issues range from how to capture recent purchase decisions in revealed choice studies and how to set up experiments in stated preference (SP) studies, and choice of analysis method for analyzing such data. This paper reviews the plethora of published studies on the field of forecasting demand of EEVs since 1980, and provides a review and annotated bibliography of that literature as it pertains to this particular demand forecasting problem. This detailed review addresses the literature not only to Transportation studies, but specifically to the problem and methodologies around forecasting to the time horizons of planning studies which may represent 10 to 20 year forecasts. The objectives of the paper are to identify where existing gaps in literature exist and to articulate where promising methodologies might guide longer term forecasting. One of the key findings of this review is that there are many common techniques used both in the field of new product demand forecasting and the field of predicting future demand for EEV. Apart from SP and RP methods, some of these new techniques that have emerged in the literature in the last few decades are survey related approaches, product diffusion models, time-series modelling, computational intelligence models and other holistic approaches.

Keywords: demand forecasting, Energy Efficient Vehicles (EEVs), forecasting methodologies review, methodological approaches

Procedia PDF Downloads 483
5750 Extracting Therapeutic Grade Essential Oils from the Lamiaceae Plant Family in the United Arab Emirates (UAE): Highlights on Great Possibilities and Sever Difficulties

Authors: Suzan M. Shahin, Mohammed A. Salem

Abstract:

Essential oils are expensive phytochemicals produced and extracted from specific species belonging to particular families in the plant kingdom. In the United Arab Emirates country (UAE), which is located in the arid region of the world, nine species, from the Lamiaceae family, having the capability to produce therapeutic grade essential oils. These species include; Mentha spicata, Ocimum forskolei, Salvia macrosiphon, Salvia aegyptiaca, Salvia macilenta, Salvia spinosa, Teucrium polium, Teucrium stocksianum, and Zataria multiflora. Although, such potential species are indigenous to the UAE, however, there are almost no studies available to investigate the chemical composition and the quality of the extracted essential oils under the UAE climatological conditions. Therefore, great attention has to be given to such valuable natural resources, through conducting highly supported research projects, tailored to the UAE conditions, and investigating different extraction techniques, including the application of the latest available technologies, such as superficial fluid CO2. This is crucially needed; in order to accomplish the greatest possibilities in the medicinal field, specifically in the discovery of new therapeutic chemotypes, as well as, to achieve the sustainability of this natural resource in the country.

Keywords: essential oils, extraction techniques, Lamiaceae, traditional medicine, United Arab Emirates (UAE)

Procedia PDF Downloads 453
5749 High-Resolution ECG Automated Analysis and Diagnosis

Authors: Ayad Dalloo, Sulaf Dalloo

Abstract:

Electrocardiogram (ECG) recording is prone to complications, on analysis by physicians, due to noise and artifacts, thus creating ambiguity leading to possible error of diagnosis. Such drawbacks may be overcome with the advent of high resolution Methods, such as Discrete Wavelet Analysis and Digital Signal Processing (DSP) techniques. This ECG signal analysis is implemented in three stages: ECG preprocessing, features extraction and classification with the aim of realizing high resolution ECG diagnosis and improved detection of abnormal conditions in the heart. The preprocessing stage involves removing spurious artifacts (noise), due to such factors as muscle contraction, motion, respiration, etc. ECG features are extracted by applying DSP and suggested sloping method techniques. These measured features represent peak amplitude values and intervals of P, Q, R, S, R’, and T waves on ECG, and other features such as ST elevation, QRS width, heart rate, electrical axis, QR and QT intervals. The classification is preformed using these extracted features and the criteria for cardiovascular diseases. The ECG diagnostic system is successfully applied to 12-lead ECG recordings for 12 cases. The system is provided with information to enable it diagnoses 15 different diseases. Physician’s and computer’s diagnoses are compared with 90% agreement, with respect to physician diagnosis, and the time taken for diagnosis is 2 seconds. All of these operations are programmed in Matlab environment.

Keywords: ECG diagnostic system, QRS detection, ECG baseline removal, cardiovascular diseases

Procedia PDF Downloads 290
5748 Machine Learning Techniques in Seismic Risk Assessment of Structures

Authors: Farid Khosravikia, Patricia Clayton

Abstract:

The main objective of this work is to evaluate the advantages and disadvantages of various machine learning techniques in two key steps of seismic hazard and risk assessment of different types of structures. The first step is the development of ground-motion models, which are used for forecasting ground-motion intensity measures (IM) given source characteristics, source-to-site distance, and local site condition for future events. IMs such as peak ground acceleration and velocity (PGA and PGV, respectively) as well as 5% damped elastic pseudospectral accelerations at different periods (PSA), are indicators of the strength of shaking at the ground surface. Typically, linear regression-based models, with pre-defined equations and coefficients, are used in ground motion prediction. However, due to the restrictions of the linear regression methods, such models may not capture more complex nonlinear behaviors that exist in the data. Thus, this study comparatively investigates potential benefits from employing other machine learning techniques as statistical method in ground motion prediction such as Artificial Neural Network, Random Forest, and Support Vector Machine. The results indicate the algorithms satisfy some physically sound characteristics such as magnitude scaling distance dependency without requiring pre-defined equations or coefficients. Moreover, it is shown that, when sufficient data is available, all the alternative algorithms tend to provide more accurate estimates compared to the conventional linear regression-based method, and particularly, Random Forest outperforms the other algorithms. However, the conventional method is a better tool when limited data is available. Second, it is investigated how machine learning techniques could be beneficial for developing probabilistic seismic demand models (PSDMs), which provide the relationship between the structural demand responses (e.g., component deformations, accelerations, internal forces, etc.) and the ground motion IMs. In the risk framework, such models are used to develop fragility curves estimating exceeding probability of damage for pre-defined limit states, and therefore, control the reliability of the predictions in the risk assessment. In this study, machine learning algorithms like artificial neural network, random forest, and support vector machine are adopted and trained on the demand parameters to derive PSDMs for them. It is observed that such models can provide more accurate estimates of prediction in relatively shorter about of time compared to conventional methods. Moreover, they can be used for sensitivity analysis of fragility curves with respect to many modeling parameters without necessarily requiring more intense numerical response-history analysis.

Keywords: artificial neural network, machine learning, random forest, seismic risk analysis, seismic hazard analysis, support vector machine

Procedia PDF Downloads 98
5747 Surgical Applied Anatomy: Alive and Kicking

Authors: Jake Hindmarch, Edward Farley, Norman Eizenberg, Mark Midwinter

Abstract:

There is a need to bring the anatomical knowledge of medical students up to the standards required by surgical specialties. Contention exists amongst anatomists, clinicians, and surgeons about the standard of anatomical knowledge medical students need. The aim of this study was to explore the standards which the Royal Australasian College of Surgeons are applying knowledge of anatomy. Furthermore, to align medical school teaching to what the surgical profession requires from graduates.: The 2018 volume of the ANZ Journal of Surgery was narrowed down to 254 articles by applying the search term “Anatomy”. The main topic was then extracted from each paper. The content of the paper was assessed for ‘novel description’ or ‘application’ of anatomical knowledge’ and classified accordingly. The majority of papers with an anatomical focus was from the general surgery specialty, which focused on surgical techniques, outcomes and management. Vascular surgery had the highest percentage of papers with a novel description and application of anatomy. Cardiothoracic and paediatric surgery had no papers with a novel description of anatomy. Finally, a novel application of anatomy was the main focus of each speciality. Firstly, a high proportion of novel applications and descriptions of anatomy are in general surgery. Secondly, vascular surgery had the largest proportion of novel application and description of anatomy, namely due to the rise of therapeutic imaging and endovascular techniques. Finally, all disciplines demonstrated a trend towards having a higher proportion of novel application of anatomical knowledge

Keywords: anatomical knowledge, anatomy, surgery, novel anatomy

Procedia PDF Downloads 113
5746 Investigation of the Effects of Processing Parameters on Pla Based 3D Printed Tensile Samples

Authors: Saifullah Karimullah

Abstract:

Additive manufacturing techniques are becoming more common with the latest technological advancements. It is composed to bring a revolution in the way products are designed, planned, manufactured, and distributed to end users. Fused deposition modeling (FDM) based 3D printing is one of those promising aspects that have revolutionized the prototyping processes. The purpose of this design and study project is to design a customized laboratory-scale FDM-based 3D printer from locally available sources. The primary goal is to design and fabricate the FDM-based 3D printer. After the fabrication, a tensile test specimen would be designed in Solid Works or [Creo computer-aided design (CAD)] software. A .stl file is generated of the tensile test specimen through slicing software and the G-codes are inserted via a computer for the test specimen to be printed. Different parameters were under studies like printing speed, layer thickness and infill density of the printed object. Some parameters were kept constant such as temperature, extrusion rate, raster orientation etc. Different tensile test specimens were printed for a different sets of parameters of the FDM-based 3d printer. The tensile test specimen were subjected to tensile tests using a universal testing machine (UTM). Design Expert software has been used for analyses, So Different results were obtained from the different tensile test specimens. The best, average and worst specimen were also observed under a compound microscope to investigate the layer bonding in between.

Keywords: additive manufacturing techniques, 3D printing, CAD software, UTM machine

Procedia PDF Downloads 95
5745 Technique for Online Condition Monitoring of Surge Arresters

Authors: Anil S. Khopkar, Kartik S. Pandya

Abstract:

Overvoltage in power systems is a phenomenon that cannot be avoided. However, it can be controlled to a certain extent. Power system equipment is to be protected against overvoltage to avoid system failure. Metal Oxide Surge Arresters (MOSA) are connected to the system for the protection of the power system against overvoltages. The MOSA will behave as an insulator under normal working conditions, where it offers a conductive path under voltage conditions. MOSA consists of zinc oxide elements (ZnO Blocks), which have non-linear V-I characteristics. ZnO blocks are connected in series and fitted in ceramic or polymer housing. This degrades due to the aging effect under continuous operation. Degradation of zinc oxide elements increases the leakage current flowing from the surge arresters. This Increased leakage current results in the increased temperature of the surge arrester, which further decreases the resistance of zinc oxide elements. As a result, leakage current increases, which again increases the temperature of a MOSA. This creates thermal runaway conditions for MOSA. Once it reaches the thermal runaway condition, it cannot return to normal working conditions. This condition is a primary cause of premature failure of surge arresters, as MOSA constitutes a core protective device for electrical power systems against transients. It contributes significantly to the reliable operation of the power system network. Hence, the condition monitoring of surge arresters should be done at periodic intervals. Online and Offline condition monitoring techniques are available for surge arresters. Offline condition monitoring techniques are not very popular as they require removing surge arresters from the system, which requires system shutdown. Hence, online condition monitoring techniques are very popular. This paper presents the evaluation technique for the surge arrester condition based on the leakage current analysis. Maximum amplitude of total leakage current (IT), Maximum amplitude of fundamental resistive leakage current (IR) and maximum amplitude of third harmonic resistive leakage current (I3rd) have been analyzed as indicators for surge arrester condition monitoring.

Keywords: metal oxide surge arrester (MOSA), over voltage, total leakage current, resistive leakage current

Procedia PDF Downloads 58
5744 Methods for Restricting Unwanted Access on the Networks Using Firewall

Authors: Bhagwant Singh, Sikander Singh Cheema

Abstract:

This paper examines firewall mechanisms routinely implemented for network security in depth. A firewall can't protect you against all the hazards of unauthorized networks. Consequently, many kinds of infrastructure are employed to establish a secure network. Firewall strategies have already been the subject of significant analysis. This study's primary purpose is to avoid unnecessary connections by combining the capability of the firewall with the use of additional firewall mechanisms, which include packet filtering and NAT, VPNs, and backdoor solutions. There are insufficient studies on firewall potential and combined approaches, but there aren't many. The research team's goal is to build a safe network by integrating firewall strength and firewall methods. The study's findings indicate that the recommended concept can form a reliable network. This study examines the characteristics of network security and the primary danger, synthesizes existing domestic and foreign firewall technologies, and discusses the theories, benefits, and disadvantages of different firewalls. Through synthesis and comparison of various techniques, as well as an in-depth examination of the primary factors that affect firewall effectiveness, this study investigated firewall technology's current application in computer network security, then introduced a new technique named "tight coupling firewall." Eventually, the article discusses the current state of firewall technology as well as the direction in which it is developing.

Keywords: firewall strategies, firewall potential, packet filtering, NAT, VPN, proxy services, firewall techniques

Procedia PDF Downloads 94
5743 Evaluating the Implementation of Machine Learning Techniques in the South African Built Environment

Authors: Peter Adekunle, Clinton Aigbavboa, Matthew Ikuabe, Opeoluwa Akinradewo

Abstract:

The future of machine learning (ML) in building may seem like a distant idea that will take decades to materialize, but it is actually far closer than previously believed. In reality, the built environment has been progressively increasing interest in machine learning. Although it could appear to be a very technical, impersonal approach, it can really make things more personable. Instead of eliminating humans out of the equation, machine learning allows people do their real work more efficiently. It is therefore vital to evaluate the factors influencing the implementation and challenges of implementing machine learning techniques in the South African built environment. The study's design was one of a survey. In South Africa, construction workers and professionals were given a total of one hundred fifty (150) questionnaires, of which one hundred and twenty-four (124) were returned and deemed eligible for study. Utilizing percentage, mean item scores, standard deviation, and Kruskal-Wallis, the collected data was analyzed. The results demonstrate that the top factors influencing the adoption of machine learning are knowledge level and a lack of understanding of its potential benefits. While lack of collaboration among stakeholders and lack of tools and services are the key hurdles to the deployment of machine learning within the South African built environment. The study came to the conclusion that ML adoption should be promoted in order to increase safety, productivity, and service quality within the built environment.

Keywords: machine learning, implementation, built environment, construction stakeholders

Procedia PDF Downloads 121
5742 Biosensor Technologies in Neurotransmitters Detection

Authors: Joanna Cabaj, Sylwia Baluta, Karol Malecha

Abstract:

Catecholamines are vital neurotransmitters that mediate a variety of central nervous system functions, such as motor control, cognition, emotion, memory processing, and endocrine modulation. Dysfunctions in catecholamine neurotransmission are induced in some neurologic and neuropsychiatric diseases. Changeable neurotransmitters level in biological fluids can be a marker of several neurological disorders. Because of its significance in analytical techniques and diagnostics, sensitive and selective detection of neurotransmitters is increasingly attracting a lot of attention in different areas of bio-analysis or biomedical research. Recently, optical techniques for the detection of catecholamines have attracted interests due to their reasonable cost, convenient control, as well as maneuverability in biological environments. Nevertheless, with the observed need for a sensitive and selective catecholamines sensor, the development of a convenient method for this neurotransmitter is still at its basic level. The manipulation of nanostructured materials in conjunction with biological molecules has led to the development of a new class of hybrid-modified enzymatic sensors in which both enhancement of charge transport and biological activity preservation may be obtained. Immobilization of biomaterials on electrode surfaces is the crucial step in fabricating electrochemical as well as optical biosensors and bioelectronic devices. Continuing systematic investigation in manufacturing of enzyme–conducting sensitive systems, here is presented a convenient fluorescence as well as electrochemical sensing strategy for catecholamines detection.

Keywords: biosensors, catecholamines, fluorescence, enzymes

Procedia PDF Downloads 102
5741 Ultrasonic Evaluation of Periodic Rough Inaccessible Surfaces from Back Side

Authors: Chanh Nghia Nguyen, Yu Kurokawa, Hirotsugu Inoue

Abstract:

The surface roughness is an important parameter for evaluating the quality of material surfaces since it affects functions and performance of industrial components. Although stylus and optical techniques are commonly used for measuring the surface roughness, they are applicable only to accessible surfaces. In practice, surface roughness measurement from the back side is sometimes demanded, for example, in inspection of safety-critical parts such as inner surface of pipes. However, little attention has been paid to the measurement of back surface roughness so far. Since back surface is usually inaccessible by stylus or optical techniques, ultrasonic technique is one of the most effective among others. In this research, an ultrasonic pulse-echo technique is considered for evaluating the pitch and the height of back surface having periodic triangular profile as a very first step. The pitch of the surface profile is measured by applying the diffraction grating theory for oblique incidence; then the height is evaluated by numerical analysis based on the Kirchhoff theory for normal incidence. The validity of the proposed method was verified by both numerical simulation and experiment. It was confirmed that the pitch is accurately measured in most cases. The height was also evaluated with good accuracy when it is smaller than a half of the pitch because of the approximation in the Kirchhoff theory.

Keywords: back side, inaccessible surface, periodic roughness, pulse-echo technique, ultrasonic NDE

Procedia PDF Downloads 271
5740 Signal Processing Techniques for Adaptive Beamforming with Robustness

Authors: Ju-Hong Lee, Ching-Wei Liao

Abstract:

Adaptive beamforming using antenna array of sensors is useful in the process of adaptively detecting and preserving the presence of the desired signal while suppressing the interference and the background noise. For conventional adaptive array beamforming, we require a prior information of either the impinging direction or the waveform of the desired signal to adapt the weights. The adaptive weights of an antenna array beamformer under a steered-beam constraint are calculated by minimizing the output power of the beamformer subject to the constraint that forces the beamformer to make a constant response in the steering direction. Hence, the performance of the beamformer is very sensitive to the accuracy of the steering operation. In the literature, it is well known that the performance of an adaptive beamformer will be deteriorated by any steering angle error encountered in many practical applications, e.g., the wireless communication systems with massive antennas deployed at the base station and user equipment. Hence, developing effective signal processing techniques to deal with the problem due to steering angle error for array beamforming systems has become an important research work. In this paper, we present an effective signal processing technique for constructing an adaptive beamformer against the steering angle error. The proposed array beamformer adaptively estimates the actual direction of the desired signal by using the presumed steering vector and the received array data snapshots. Based on the presumed steering vector and a preset angle range for steering mismatch tolerance, we first create a matrix related to the direction vector of signal sources. Two projection matrices are generated from the matrix. The projection matrix associated with the desired signal information and the received array data are utilized to iteratively estimate the actual direction vector of the desired signal. The estimated direction vector of the desired signal is then used for appropriately finding the quiescent weight vector. The other projection matrix is set to be the signal blocking matrix required for performing adaptive beamforming. Accordingly, the proposed beamformer consists of adaptive quiescent weights and partially adaptive weights. Several computer simulation examples are provided for evaluating and comparing the proposed technique with the existing robust techniques.

Keywords: adaptive beamforming, robustness, signal blocking, steering angle error

Procedia PDF Downloads 118
5739 Energy Efficient Lighting in Educational Buildings through the Example of a High School in Istanbul

Authors: Nihan Gurel Ulusan

Abstract:

It is obvious that electrical energy, which is an inseparable part of modern day’s human and also the most important power source of our age, should be generated on a level that will suffice the nation’s requirements. The electrical energy used for a sustainable architectural design should be reduced as much as possible. Designing the buildings as energy efficient systems which aim at reducing the artificial illumination loads has been a current subject of our times as a result of concepts gaining importance like conscious consumption of energy sources, environment-friendly designs and sustainability. Reducing the consumption of electrical energy regarding the artificial lighting carries great significance, especially in the volumes which are used all day long like the educational buildings. Starting out with such an aim in this paper, the educational buildings are explored in terms of energy efficient lighting. Firstly, illumination techniques, illumination systems, light sources, luminaries, illumination controls and 'efficient energy' usage in lighting are mentioned. In addition, natural and artificial lighting systems used in educational buildings and also the spaces building up these kind buildings are examined in terms of energy efficient lighting. Lastly, the illumination properties of the school sample chosen for this study, Kağıthane Anadolu Lisesi, a typical high school in Istanbul, is observed. Suggestions are made in order to improve the system by evaluating the illumination properties of the classes with the survey carried out with the users.

Keywords: educational buildings, energy efficient, illumination techniques, lighting

Procedia PDF Downloads 275
5738 Resilient Machine Learning in the Nuclear Industry: Crack Detection as a Case Study

Authors: Anita Khadka, Gregory Epiphaniou, Carsten Maple

Abstract:

There is a dramatic surge in the adoption of machine learning (ML) techniques in many areas, including the nuclear industry (such as fault diagnosis and fuel management in nuclear power plants), autonomous systems (including self-driving vehicles), space systems (space debris recovery, for example), medical surgery, network intrusion detection, malware detection, to name a few. With the application of learning methods in such diverse domains, artificial intelligence (AI) has become a part of everyday modern human life. To date, the predominant focus has been on developing underpinning ML algorithms that can improve accuracy, while factors such as resiliency and robustness of algorithms have been largely overlooked. If an adversarial attack is able to compromise the learning method or data, the consequences can be fatal, especially but not exclusively in safety-critical applications. In this paper, we present an in-depth analysis of five adversarial attacks and three defence methods on a crack detection ML model. Our analysis shows that it can be dangerous to adopt machine learning techniques in security-critical areas such as the nuclear industry without rigorous testing since they may be vulnerable to adversarial attacks. While common defence methods can effectively defend against different attacks, none of the three considered can provide protection against all five adversarial attacks analysed.

Keywords: adversarial machine learning, attacks, defences, nuclear industry, crack detection

Procedia PDF Downloads 152
5737 Teaching and Learning Jazz Improvisation Using Bloom's Taxonomy of Learning Domains

Authors: Graham Wood

Abstract:

The 20th Century saw the introduction of many new approaches to music making, including the structured and academic study of jazz improvisation. The rise of many school and tertiary jazz programs was rapid and quickly spread around the globe in a matter of decades. It could be said that the curriculum taught in these new programs was often developed in an ad-hoc manner due to the lack of written literature in this new and rapidly expanding area and the vastly different pedagogical principles when compared to classical music education that was prevalent in school and tertiary programs. There is widespread information regarding the theory and techniques used by jazz improvisers, but methods to practice these concepts in order to achieve the best outcomes for students and teachers is much harder to find. This research project explores the authors’ experiences as a studio jazz piano teacher, ensemble teacher and classroom improvisation lecturer over fifteen years and suggests an alignment with Bloom’s taxonomy of learning domains. This alignment categorizes the different tasks that need to be taught and practiced in order for the teacher and the student to devise a well balanced and effective practice routine and for the teacher to develop an effective teaching program. These techniques have been very useful to the teacher and the student to ensure that a good balance of cognitive, psychomotor and affective skills are taught to the students in a range of learning contexts.

Keywords: bloom, education, jazz, learning, music, teaching

Procedia PDF Downloads 250
5736 Short Answer Grading Using Multi-Context Features

Authors: S. Sharan Sundar, Nithish B. Moudhgalya, Nidhi Bhandari, Vineeth Vijayaraghavan

Abstract:

Automatic Short Answer Grading is one of the prime applications of artificial intelligence in education. Several approaches involving the utilization of selective handcrafted features, graphical matching techniques, concept identification and mapping, complex deep frameworks, sentence embeddings, etc. have been explored over the years. However, keeping in mind the real-world application of the task, these solutions present a slight overhead in terms of computations and resources in achieving high performances. In this work, a simple and effective solution making use of elemental features based on statistical, linguistic properties, and word-based similarity measures in conjunction with tree-based classifiers and regressors is proposed. The results for classification tasks show improvements ranging from 1%-30%, while the regression task shows a stark improvement of 35%. The authors attribute these improvements to the addition of multiple similarity scores to provide ensemble of scoring criteria to the models. The authors also believe the work could reinstate that classical natural language processing techniques and simple machine learning models can be used to achieve high results for short answer grading.

Keywords: artificial intelligence, intelligent systems, natural language processing, text mining

Procedia PDF Downloads 128
5735 Histochemical Localization of Hepatitis B Surface Antigen in Hepatocellular Carcinoma: An Evaluation of Two Staining Techniques in a Tertiary Hospital in Calabar, Nigeria

Authors: Imeobong Joseph Inyang, Aniekan-Augusta Okon Eyo, Abel William Essien

Abstract:

Hepatitis B virus (HBV) is one of the known human carcinogens. The presence of HBsAg in liver tissues indicates active viral replication. More than 85% of Hepatocellular Carcinoma (HCC) cases occur in countries with increased rates of chronic HBV infection. An evaluation study to determine the relationship between positivity for HBsAg and development of HCC and its distribution between age and gender of subjects was done. Shikata Orcein and Haematoxylin and Eosin (H&E) staining techniques were performed on liver sections. A total of 50 liver tissue specimens comprising 38 biopsy and 12 post-mortem specimens were processed. Thirty-five of the 50 specimens were positive for HBsAg with Orcein stain whereas only 16 were positive with H&E stain, and these were also positive with Orcein stain, giving an HBsAg prevalence of 70.0% (35/50). The prevalence of HCC in the study was 56.0% (28/50), of which 21 (75.0%) cases were positive for HBsAg, 18 (64.3%) were males while 10 (35.7%) were females distributed within the age range of 20-70 years. The highest number of HBsAg positive HCC cases, 7/21 (33.3%) occurred in the age group 40-49 years. There was no relationship in the pattern of distribution of HCC between age and gender using the Pearson correlation coefficient (r = 0.0474; P < 0.05). HBV infection predisposed to HCC. Orcein technique was more specific and is therefore recommended for screening of liver tissues where facilities for immunohistochemistry are inaccessible.

Keywords: Hepatitis B. surface antigen, hepatocellular carcinoma, orcein, pathology

Procedia PDF Downloads 304
5734 ADA Tool for Satellite InSAR-Based Ground Displacement Analysis: The Granada Region

Authors: M. Cuevas-González, O. Monserrat, A. Barra, C. Reyes-Carmona, R.M. Mateos, J. P. Galve, R. Sarro, M. Cantalejo, E. Peña, M. Martínez-Corbella, J. A. Luque, J. M. Azañón, A. Millares, M. Béjar, J. A. Navarro, L. Solari

Abstract:

Geohazard prone areas require continuous monitoring to detect risks, understand the phenomena occurring in those regions and prevent disasters. Satellite interferometry (InSAR) has come to be a trustworthy technique for ground movement detection and monitoring in the last few years. InSAR based techniques allow to process large areas providing high number of displacement measurements at low cost. However, the results provided by such techniques are usually not easy to interpret by non-experienced users hampering its use for decision makers. This work presents a set of tools developed in the framework of different projects (Momit, Safety, U-Geohaz, Riskcoast) and an example of their use in the Granada Coastal area (Spain) is shown. The ADA (Active Displacement Areas) tool have been developed with the aim of easing the management, use and interpretation of InSAR based results. It provides a semi-automatic extraction of the most significant ADAs through the application ADAFinder tool. This tool aims to support the exploitation of the European Ground Motion Service (EU-GMS), which will provide consistent, regular and reliable information regarding natural and anthropogenic ground motion phenomena all over Europe.

Keywords: ground displacements, InSAR, natural hazards, satellite imagery

Procedia PDF Downloads 202
5733 Cognitive Approach at the Epicenter of Creative Accounting in Cameroonian Companies: The Relevance of the Psycho-Sociological Approach and the Theory of Cognitive Dissonance

Authors: Romuald Temomo Wamba, Robert Wanda

Abstract:

The issue of creative accounting in the psychological and sociological framework has been a mixed subject for over 60 years. The objective of this article is to ensure the existence of creative accounting in Cameroonian entities on the one hand and to understand the strategies used by audit agents to detect errors, omissions, irregularities, or inadequacies in the financial state; optimization techniques used by account preparers to strategically bypass texts on the other hand. To achieve this, we conducted an exploratory study using a cognitive approach, and the data analysis was performed by the software 'decision explorer'. The results obtained challenge the authors' cognition (manifest latent and deceptive behavior). The tax inspectors stress that the entities in Cameroon do not derogate from the rules of piloting in the financial statements. Likewise, they claim a change in current income and net income through depreciation, provisions, inventories, and the spreading of charges over long periods. This suggests the suspicion or intention of manipulating the financial statements. As for the techniques, the account preparers manage the accruals at the end of the year as the basis of the practice of creative accounting. Likewise, management accounts are more favorable to results management.

Keywords: creative accounting, sociocognitive approach, psychological and sociological approach, cognitive dissonance theory, cognitive mapping

Procedia PDF Downloads 183
5732 Predicting Stack Overflow Accepted Answers Using Features and Models with Varying Degrees of Complexity

Authors: Osayande Pascal Omondiagbe, Sherlock a Licorish

Abstract:

Stack Overflow is a popular community question and answer portal which is used by practitioners to solve technology-related challenges during software development. Previous studies have shown that this forum is becoming a substitute for official software programming languages documentation. While tools have looked to aid developers by presenting interfaces to explore Stack Overflow, developers often face challenges searching through many possible answers to their questions, and this extends the development time. To this end, researchers have provided ways of predicting acceptable Stack Overflow answers by using various modeling techniques. However, less interest is dedicated to examining the performance and quality of typically used modeling methods, and especially in relation to models’ and features’ complexity. Such insights could be of practical significance to the many practitioners that use Stack Overflow. This study examines the performance and quality of various modeling methods that are used for predicting acceptable answers on Stack Overflow, drawn from 2014, 2015 and 2016. Our findings reveal significant differences in models’ performance and quality given the type of features and complexity of models used. Researchers examining classifiers’ performance and quality and features’ complexity may leverage these findings in selecting suitable techniques when developing prediction models.

Keywords: feature selection, modeling and prediction, neural network, random forest, stack overflow

Procedia PDF Downloads 129
5731 Sustainability and Clustering: A Bibliometric Assessment

Authors: Fernanda M. Assef, Maria Teresinha A. Steiner, David Gabriel F. Barros

Abstract:

Review researches are useful in terms of analysis of research problems. Between the types of review documents, we commonly find bibliometric studies. This type of application often helps the global visualization of a research problem and helps academics worldwide to understand the context of a research area better. In this document, a bibliometric view surrounding clustering techniques and sustainability problems is presented. The authors aimed at which issues mostly use clustering techniques, and, even which sustainability issue would be more impactful on today’s moment of research. During the bibliometric analysis, we found ten different groups of research in clustering applications for sustainability issues: Energy; Environmental; Non-urban planning; Sustainable Development; Sustainable Supply Chain; Transport; Urban Planning; Water; Waste Disposal; and, Others. And, by analyzing the citations of each group, we discovered that the Environmental group could be classified as the most impactful research cluster in the area mentioned. Now, after the content analysis of each paper classified in the environmental group, we found that the k-means technique is preferred for solving sustainability problems with clustering methods since it appeared the most amongst the documents. The authors finally conclude that a bibliometric assessment could help indicate a gap of researches on waste disposal – which was the group with the least amount of publications – and the most impactful research on environmental problems.

Keywords: bibliometric assessment, clustering, sustainability, territorial partitioning

Procedia PDF Downloads 98
5730 Backward-Facing Step Measurements at Different Reynolds Numbers Using Acoustic Doppler Velocimetry

Authors: Maria Amelia V. C. Araujo, Billy J. Araujo, Brian Greenwood

Abstract:

The flow over a backward-facing step is characterized by the presence of flow separation, recirculation and reattachment, for a simple geometry. This type of fluid behaviour takes place in many practical engineering applications, hence the reason for being investigated. Historically, fluid flows over a backward-facing step have been examined in many experiments using a variety of measuring techniques such as laser Doppler velocimetry (LDV), hot-wire anemometry, particle image velocimetry or hot-film sensors. However, some of these techniques cannot conveniently be used in separated flows or are too complicated and expensive. In this work, the applicability of the acoustic Doppler velocimetry (ADV) technique is investigated to such type of flows, at various Reynolds numbers corresponding to different flow regimes. The use of this measuring technique in separated flows is very difficult to find in literature. Besides, most of the situations where the Reynolds number effect is evaluated in separated flows are in numerical modelling. The ADV technique has the advantage in providing nearly non-invasive measurements, which is important in resolving turbulence. The ADV Nortek Vectrino+ was used to characterize the flow, in a recirculating laboratory flume, at various Reynolds Numbers (Reh = 3738, 5452, 7908 and 17388) based on the step height (h), in order to capture different flow regimes, and the results compared to those obtained using other measuring techniques. To compare results with other researchers, the step height, expansion ratio and the positions upstream and downstream the step were reproduced. The post-processing of the AVD records was performed using a customized numerical code, which implements several filtering techniques. Subsequently, the Vectrino noise level was evaluated by computing the power spectral density for the stream-wise horizontal velocity component. The normalized mean stream-wise velocity profiles, skin-friction coefficients and reattachment lengths were obtained for each Reh. Turbulent kinetic energy, Reynolds shear stresses and normal Reynolds stresses were determined for Reh = 7908. An uncertainty analysis was carried out, for the measured variables, using the moving block bootstrap technique. Low noise levels were obtained after implementing the post-processing techniques, showing their effectiveness. Besides, the errors obtained in the uncertainty analysis were relatively low, in general. For Reh = 7908, the normalized mean stream-wise velocity and turbulence profiles were compared directly with those acquired by other researchers using the LDV technique and a good agreement was found. The ADV technique proved to be able to characterize the flow properly over a backward-facing step, although additional caution should be taken for measurements very close to the bottom. The ADV measurements showed reliable results regarding: a) the stream-wise velocity profiles; b) the turbulent shear stress; c) the reattachment length; d) the identification of the transition from transitional to turbulent flows. Despite being a relatively inexpensive technique, acoustic Doppler velocimetry can be used with confidence in separated flows and thus very useful for numerical model validation. However, it is very important to perform adequate post-processing of the acquired data, to obtain low noise levels, thus decreasing the uncertainty.

Keywords: ADV, experimental data, multiple Reynolds number, post-processing

Procedia PDF Downloads 136
5729 Design of Cylindrical Crawler Robot Inspired by Amoeba Locomotion

Authors: Jun-ya Nagase

Abstract:

Recently, the need of colonoscopy is increasing because of the rise of colonic disorder including cancer of the colon. However, current colonoscopy depends on doctor's skill strongly. Therefore, a large intestine endoscope that does not depend on the techniques of a doctor with high safety is required. In this research, we aim at development a novel large intestine endoscope that can realize safe insertion without specific techniques. A wheel movement type robot, a snake-like robot and an earthworm-like robot are all described in the relevant literature as endoscope robots that are currently studied. Among them, the tracked crawler robot can travel by traversing uneven ground flexibly with a crawler belt attached firmly to the ground surface. Although conventional crawler robots have high efficiency and/or high ground-covering ability, they require a comparatively large space to move. In this study, a small cylindrical crawler robot inspired by amoeba locomotion, which does not need large space to move and which has high ground-covering ability, is proposed. In addition, we developed a prototype of the large intestine endoscope using the proposed crawler mechanism. Experiments have demonstrated smooth operation and a forward movement of the robot by application of voltage to the motor. This paper reports the structure, drive mechanism, prototype, and experimental evaluation.

Keywords: tracked-crawler, endoscopic robot, narrow path, amoeba locomotion.

Procedia PDF Downloads 379
5728 Incorporation of Copper for Performance Enhancement in Metal-Oxides Resistive Switching Device and Its Potential Electronic Application

Authors: B. Pavan Kumar Reddy, P. Michael Preetam Raj, Souri Banerjee, Souvik Kundu

Abstract:

In this work, the fabrication and characterization of copper-doped zinc oxide (Cu:ZnO) based memristor devices with aluminum (Al) and indium tin oxide (ITO) metal electrodes are reported. The thin films of Cu:ZnO was synthesized using low-cost and low-temperature chemical process. The Cu:ZnO was then deposited onto ITO bottom electrodes using spin-coater technique, whereas the top electrode Al was deposited utilizing physical vapor evaporation technique. Ellipsometer was employed in order to measure the Cu:ZnO thickness and it was found to be 50 nm. Several surface and materials characterization techniques were used to study the thin-film properties of Cu:ZnO. To ascertain the efficacy of Cu:ZnO for memristor applications, electrical characterizations such as current-voltage (I-V), data retention and endurance were obtained, all being the critical parameters for next-generation memory. The I-V characteristic exhibits switching behavior with asymmetrical hysteresis loops. This work imputes the resistance switching to the positional drift of oxygen vacancies associated with respect to the Al/Cu:ZnO junction. Further, a non-linear curve fitting regression techniques were utilized to determine the equivalent circuit for the fabricated Cu:ZnO memristors. Efforts were also devoted in order to establish its potentiality for different electronic applications.

Keywords: copper doped, metal-oxides, oxygen vacancies, resistive switching

Procedia PDF Downloads 159
5727 Evaluation and Assessment of Bioinformatics Methods and Their Applications

Authors: Fatemeh Nokhodchi Bonab

Abstract:

Bioinformatics, in its broad sense, involves application of computer processes to solve biological problems. A wide range of computational tools are needed to effectively and efficiently process large amounts of data being generated as a result of recent technological innovations in biology and medicine. A number of computational tools have been developed or adapted to deal with the experimental riches of complex and multivariate data and transition from data collection to information or knowledge. These bioinformatics tools are being evaluated and applied in various medical areas including early detection, risk assessment, classification, and prognosis of cancer. The goal of these efforts is to develop and identify bioinformatics methods with optimal sensitivity, specificity, and predictive capabilities. The recent flood of data from genome sequences and functional genomics has given rise to new field, bioinformatics, which combines elements of biology and computer science. Bioinformatics is conceptualizing biology in terms of macromolecules (in the sense of physical-chemistry) and then applying "informatics" techniques (derived from disciplines such as applied maths, computer science, and statistics) to understand and organize the information associated with these molecules, on a large-scale. Here we propose a definition for this new field and review some of the research that is being pursued, particularly in relation to transcriptional regulatory systems.

Keywords: methods, applications, transcriptional regulatory systems, techniques

Procedia PDF Downloads 118
5726 Computer-Aided Classification of Liver Lesions Using Contrasting Features Difference

Authors: Hussein Alahmer, Amr Ahmed

Abstract:

Liver cancer is one of the common diseases that cause the death. Early detection is important to diagnose and reduce the incidence of death. Improvements in medical imaging and image processing techniques have significantly enhanced interpretation of medical images. Computer-Aided Diagnosis (CAD) systems based on these techniques play a vital role in the early detection of liver disease and hence reduce liver cancer death rate.  This paper presents an automated CAD system consists of three stages; firstly, automatic liver segmentation and lesion’s detection. Secondly, extracting features. Finally, classifying liver lesions into benign and malignant by using the novel contrasting feature-difference approach. Several types of intensity, texture features are extracted from both; the lesion area and its surrounding normal liver tissue. The difference between the features of both areas is then used as the new lesion descriptors. Machine learning classifiers are then trained on the new descriptors to automatically classify liver lesions into benign or malignant. The experimental results show promising improvements. Moreover, the proposed approach can overcome the problems of varying ranges of intensity and textures between patients, demographics, and imaging devices and settings.

Keywords: CAD system, difference of feature, fuzzy c means, lesion detection, liver segmentation

Procedia PDF Downloads 314
5725 Detecting Earnings Management via Statistical and Neural Networks Techniques

Authors: Mohammad Namazi, Mohammad Sadeghzadeh Maharluie

Abstract:

Predicting earnings management is vital for the capital market participants, financial analysts and managers. The aim of this research is attempting to respond to this query: Is there a significant difference between the regression model and neural networks’ models in predicting earnings management, and which one leads to a superior prediction of it? In approaching this question, a Linear Regression (LR) model was compared with two neural networks including Multi-Layer Perceptron (MLP), and Generalized Regression Neural Network (GRNN). The population of this study includes 94 listed companies in Tehran Stock Exchange (TSE) market from 2003 to 2011. After the results of all models were acquired, ANOVA was exerted to test the hypotheses. In general, the summary of statistical results showed that the precision of GRNN did not exhibit a significant difference in comparison with MLP. In addition, the mean square error of the MLP and GRNN showed a significant difference with the multi variable LR model. These findings support the notion of nonlinear behavior of the earnings management. Therefore, it is more appropriate for capital market participants to analyze earnings management based upon neural networks techniques, and not to adopt linear regression models.

Keywords: earnings management, generalized linear regression, neural networks multi-layer perceptron, Tehran stock exchange

Procedia PDF Downloads 416