Search results for: deep vibro techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8509

Search results for: deep vibro techniques

7429 Employing Operations Research at Universities to Build Management Systems

Authors: Abdallah A. Hlayel

Abstract:

Operations research science (OR) deals with good success in developing and applying scientific methods for problem solving and decision-making. However, by using OR techniques, we can enhance the use of computer decision support systems to achieve optimal management for institutions. OR applies comprehensive analysis including all factors that affect on it and builds mathematical modeling to solve business or organizational problems. In addition, it improves decision-making and uses available resources efficiently. The adoption of OR by universities would definitely contributes to the development and enhancement of the performance of OR techniques. This paper provides an understanding of the structures, approaches and models of OR in problem solving and decision-making.

Keywords: best candidates' method, decision making, decision support system, operations research

Procedia PDF Downloads 445
7428 Risk Management in Construction Projects

Authors: Mustafa Dogru, Ruveyda Komurlu

Abstract:

Companies and professionals in the construction sector face various risks in every project depending on the characteristics, size, complexity, the location of the projects and the techniques used. Some risks’ effects may increase as the project progresses whereas new risks may emerge. Because of the ever-changing nature of the risks, risk management is a cyclical process that needs to be repeated throughout the project. Since the risks threaten the success of the project, risk management is an important part of the entire project management process. The aims of this study are to emphasize the importance of risk management in construction projects, summarize the risk identification process, and introduce a number of methods for preventing risks such as alternative design, checklists, prototyping and test-analysis-correction technique etc. Following the literature review conducted to list the techniques for preventing risks, case studies has been performed to compare and evaluate the success of the techniques in a number of completed projects with the same typology, performed domestic and international. Findings of the study suggest that controlling and minimizing the level of the risks in construction projects, taking optimal precautions for different risks, and mitigating or eliminating the effects of risks are important in order to prevent additional costs for the project. Additionally, focusing on the risks that have highest impact is the most rational way to minimize the effects of the risks on projects.

Keywords: construction projects, construction management, project management, risk management

Procedia PDF Downloads 319
7427 Crafting of Paper Cutting Techniques for Embellishment of Fashion Textiles

Authors: A. Vaidya-Soocheta, K. M. Wong-Hon-Lang

Abstract:

Craft and fashion have always been interlinked. The combination of both often gives stunning results. The present study introduces ‘Paper Cutting Craft Techniques’ like the Japanese –Kirigami, Mexican –PapelPicado, German –Scherenschnitte, Polish –Wycinankito in textiles to develop innovative and novel design structures as embellishments and ornamentation. The project studies various ways of using these paper cutting techniques to obtain interesting features and delicate design patterns on fabrics. While paper has its advantages and related uses, it is fragile rigid and thus not appropriate for clothing. Fabric is sturdy, flexible, dimensionally stable and washable. In the present study, the cut out techniques develop creative design motifs and patterns to give an inventive and unique appeal to the fabrics. The beauty and fascination of lace in garments have always given them a nostalgic charm. Laces with their intricate and delicate complexity in combination with other materials add a feminine touch to a garment and give it a romantic, mysterious appeal. Various textured and decorative effects through fabric manipulation are experimented along with the use of paper cutting craft skills as an innovative substitute for developing lace or “Broderie Anglaise” effects on textiles. A number of assorted fabric types with varied textures were selected for the study. Techniques to avoid fraying and unraveling of the design cut fabrics were introduced. Fabrics were further manipulated by use of interesting prints with embossed effects on cut outs. Fabric layering in combination with assorted techniques such as cutting of folded fabric, printing, appliqué, embroidery, crochet, braiding, weaving added a novel exclusivity to the fabrics. The fabrics developed by these innovative methods were then tailored into garments. The study thus tested the feasibility and practicability of using these fabrics by designing a collection of evening wear garments based on the theme ‘Nostalgia’. The prototypes developed were complemented by designing fashion accessories with the crafted fabrics. Prototypes of accessories add interesting features to the study. The adaptation and application of this novel technique of paper cutting craft on textiles can be an innovative start for a new trend in textile and fashion industry. The study anticipates that this technique will open new avenues in the world of fashion to incorporate its use commercially.

Keywords: collection, fabric cutouts, nostalgia, prototypes

Procedia PDF Downloads 357
7426 The Art and Science of Trauma-Informed Psychotherapy: Guidelines for Inter-Disciplinary Clinicians

Authors: Daphne Alroy-Thiberge

Abstract:

Trauma-impacted individuals present unique treatment challenges that include high reactivity, hyper-and hypo-arousal, poor adherence to therapy, as well as powerful transference and counter-transference experiences in therapy. This work provides an overview of the clinical tenets most often encountered in trauma-impacted individuals. Further, it provides readily applicable clinical techniques to optimize therapeutic rapport and facilitate accelerated positive mental health outcomes. Finally, integrated neuroscience and clinical evidence-based data are discussed to shed new light on crisis states in trauma-impacted individuals. This knowledge is utilized to provide effective and concrete interventions towards rapid and successful de-escalation of the impacted individual. A highly interactive, adult-learning-principles-based modality is utilized to provide an organic learning experience for participants. The information and techniques learned aim to increase clinical effectiveness, reduce staff injuries and burnout, and significantly enhance positive mental health outcomes and self-determination for the trauma-impacted individuals treated.

Keywords: clinical competencies, crisis interventions, psychotherapy techniques, trauma informed care

Procedia PDF Downloads 108
7425 A Survey of Semantic Integration Approaches in Bioinformatics

Authors: Chaimaa Messaoudi, Rachida Fissoune, Hassan Badir

Abstract:

Technological advances of computer science and data analysis are helping to provide continuously huge volumes of biological data, which are available on the web. Such advances involve and require powerful techniques for data integration to extract pertinent knowledge and information for a specific question. Biomedical exploration of these big data often requires the use of complex queries across multiple autonomous, heterogeneous and distributed data sources. Semantic integration is an active area of research in several disciplines, such as databases, information-integration, and ontology. We provide a survey of some approaches and techniques for integrating biological data, we focus on those developed in the ontology community.

Keywords: biological ontology, linked data, semantic data integration, semantic web

Procedia PDF Downloads 449
7424 Virtual 3D Environments for Image-Based Navigation Algorithms

Authors: V. B. Bastos, M. P. Lima, P. R. G. Kurka

Abstract:

This paper applies to the creation of virtual 3D environments for the study and development of mobile robot image based navigation algorithms and techniques, which need to operate robustly and efficiently. The test of these algorithms can be performed in a physical way, from conducting experiments on a prototype, or by numerical simulations. Current simulation platforms for robotic applications do not have flexible and updated models for image rendering, being unable to reproduce complex light effects and materials. Thus, it is necessary to create a test platform that integrates sophisticated simulated applications of real environments for navigation, with data and image processing. This work proposes the development of a high-level platform for building 3D model’s environments and the test of image-based navigation algorithms for mobile robots. Techniques were used for applying texture and lighting effects in order to accurately represent the generation of rendered images regarding the real world version. The application will integrate image processing scripts, trajectory control, dynamic modeling and simulation techniques for physics representation and picture rendering with the open source 3D creation suite - Blender.

Keywords: simulation, visual navigation, mobile robot, data visualization

Procedia PDF Downloads 255
7423 A Review of Research on Pre-training Technology for Natural Language Processing

Authors: Moquan Gong

Abstract:

In recent years, with the rapid development of deep learning, pre-training technology for natural language processing has made great progress. The early field of natural language processing has long used word vector methods such as Word2Vec to encode text. These word vector methods can also be regarded as static pre-training techniques. However, this context-free text representation brings very limited improvement to subsequent natural language processing tasks and cannot solve the problem of word polysemy. ELMo proposes a context-sensitive text representation method that can effectively handle polysemy problems. Since then, pre-training language models such as GPT and BERT have been proposed one after another. Among them, the BERT model has significantly improved its performance on many typical downstream tasks, greatly promoting the technological development in the field of natural language processing, and has since entered the field of natural language processing. The era of dynamic pre-training technology. Since then, a large number of pre-trained language models based on BERT and XLNet have continued to emerge, and pre-training technology has become an indispensable mainstream technology in the field of natural language processing. This article first gives an overview of pre-training technology and its development history, and introduces in detail the classic pre-training technology in the field of natural language processing, including early static pre-training technology and classic dynamic pre-training technology; and then briefly sorts out a series of enlightening technologies. Pre-training technology, including improved models based on BERT and XLNet; on this basis, analyze the problems faced by current pre-training technology research; finally, look forward to the future development trend of pre-training technology.

Keywords: natural language processing, pre-training, language model, word vectors

Procedia PDF Downloads 57
7422 Synthesis and Characterization of Poly (N-(Pyridin-2-Ylmethylidene)Pyridin-2-Amine: Thermal and Conductivity Properties

Authors: Nuray Yılmaz Baran

Abstract:

The conjugated Schiff base polymers which are also called as polyazomethines are promising materials for various applications due to their good thermal resistance semiconductive, liquid crystal, fiber forming, nonlinear optical outstanding photo- and electroluminescence and antimicrobial properties. In recent years, polyazomethines have attracted intense attention of researchers especially due to optoelectronic properties which have made its usage possible in organic light emitting diodes (OLEDs), solar cells (SCs), organic field effect transistors (OFETs), and photorefractive holographic materials (PRHMs). In this study, N-(pyridin-2-ylmethylidene)pyridin-2-amine Schiff base was synthesized from condensation reaction of 2-aminopyridine with 2-pyridine carbaldehyde. Polymerization of Schiff base was achieved by polycondensation reaction using NaOCl oxidant in methanol medium at various time and temperatures. The synthesized Schiff base monomer and polymer (Poly(N-(pyridin-2-ylmethylidene)pyridin-2-amine)) was characterized by UV-vis, FT-IR, 1H-NMR, XRD techniques. Molecular weight distribution and the surface morphology of the polymer was determined by GPC and SEM-EDAX techniques. Thermal behaviour of the monomer and polymer was investigated by TG/DTG, DTA and DSC techniques.

Keywords: polyazomethines, polycondensation reaction, Schiff base polymers, thermal stability

Procedia PDF Downloads 232
7421 Comparison of EMG Normalization Techniques Recommended for Back Muscles Used in Ergonomics Research

Authors: Saif Al-Qaisi, Alif Saba

Abstract:

Normalization of electromyography (EMG) data in ergonomics research is a prerequisite for interpreting the data. Normalizing accounts for variability in the data due to differences in participants’ physical characteristics, electrode placement protocols, time of day, and other nuisance factors. Typically, normalized data is reported as a percentage of the muscle’s isometric maximum voluntary contraction (%MVC). Various MVC techniques have been recommended in the literature for normalizing EMG activity of back muscles. This research tests and compares the recommended MVC techniques in the literature for three back muscles commonly used in ergonomics research, which are the lumbar erector spinae (LES), latissimus dorsi (LD), and thoracic erector spinae (TES). Six healthy males from a university population participated in this research. Five different MVC exercises were compared for each muscle using the Tringo wireless EMG system (Delsys Inc.). Since the LES and TES share similar functions in controlling trunk movements, their MVC exercises were the same, which included trunk extension at -60°, trunk extension at 0°, trunk extension while standing, hip extension, and the arch test. The MVC exercises identified in the literature for the LD were chest-supported shoulder extension, prone shoulder extension, lat-pull down, internal shoulder rotation, and abducted shoulder flexion. The maximum EMG signal was recorded during each MVC trial, and then the averages were computed across participants. A one-way analysis of variance (ANOVA) was utilized to determine the effect of MVC technique on muscle activity. Post-hoc analyses were performed using the Tukey test. The MVC technique effect was statistically significant for each of the muscles (p < 0.05); however, a larger sample of participants was needed to detect significant differences in the Tukey tests. The arch test was associated with the highest EMG average at the LES, and also it resulted in the maximum EMG activity more often than the other techniques (three out of six participants). For the TES, trunk extension at 0° was associated with the largest EMG average, and it resulted in the maximum EMG activity the most often (three out of six participants). For the LD, participants obtained their maximum EMG either from chest-supported shoulder extension (three out of six participants) or prone shoulder extension (three out of six participants). Chest-supported shoulder extension, however, had a larger average than prone shoulder extension (0.263 and 0.240, respectively). Although all the aforementioned techniques were superior in their averages, they did not always result in the maximum EMG activity. If an accurate estimate of the true MVC is desired, more than one technique may have to be performed. This research provides additional MVC techniques for each muscle that may elicit the maximum EMG activity.

Keywords: electromyography, maximum voluntary contraction, normalization, physical ergonomics

Procedia PDF Downloads 193
7420 A Review on Parametric Optimization of Casting Processes Using Optimization Techniques

Authors: Bhrugesh Radadiya, Jaydeep Shah

Abstract:

In Indian foundry industry, there is a need of defect free casting with minimum production cost in short lead time. Casting defect is a very large issue in foundry shop which increases the rejection rate of casting and wastage of materials. The various parameters influences on casting process such as mold machine related parameters, green sand related parameters, cast metal related parameters, mold related parameters and shake out related parameters. The mold related parameters are most influences on casting defects in sand casting process. This paper review the casting produced by foundry with shrinkage and blow holes as a major defects was analyzed and identified that mold related parameters such as mold temperature, pouring temperature and runner size were not properly set in sand casting process. These parameters were optimized using different optimization techniques such as Taguchi method, Response surface methodology, Genetic algorithm and Teaching-learning based optimization algorithm. Finally, concluded that a Teaching-learning based optimization algorithm give better result than other optimization techniques.

Keywords: casting defects, genetic algorithm, parametric optimization, Taguchi method, TLBO algorithm

Procedia PDF Downloads 728
7419 Geophysical and Laboratory Evaluation of Aquifer Position, Aquifer Protective Capacity and Groundwater Quality in Selected Dumpsites in Calabar Municipal Local Government Area, South Eastern Nigeria

Authors: Egor Atan Obeten, Abong Augustine Agwul, Bissong A. Samson

Abstract:

The position of the aquifer, its protective capability, and the quality of the groundwater beneath the dumpsite were all investigated. The techniques employed were laboratory, tritium tagging, electrical resistivity tomography (ERT), and vertical electrical sounding (VES). With a maximum electrode spacing of 500 meters, fifteen VES stations were used, and IPI2win software was used to analyze the data collected. The resistivity map of the dumpsite was determined by deploying six ERT stations for the 2 D survey. To ascertain the degree of soil infiltration beneath the dumpsite, the tritium tagging method was used. Using a conventional laboratory procedure, groundwater samples were taken from neighboring boreholes and examined. The findings showed that there were three to five geoelectric layers, with the aquifer position being inferred to be between 24.2 and 75.1 meters deep in the third, fourth, and fifth levels. Siemens with values in the range of 0.0235 to 0.1908 for the load protection capacity were deemed to be, at most, weakly and badly protected. The obtained porosity values ranged from 44.45 to 89.75. Strong calculated values for transmissivity and porosity indicate a permeable aquifer system with considerable storativity. The area has an infiltration value between 8 and 22 percent, according to the results of the tritium tagging technique, which was used to evaluate the level of infiltration from the dumpsite. Groundwater samples that have been analyzed reveal levels of NO2, DO, Pb2+, magnesium, and cadmium that are higher than what the NSDWQ has approved. Overall analysis of the results from the above-described methodologies shows that the study area's aquifer system is porous and that contaminants will circulate through it quickly if they are contaminated.

Keywords: aquifer, transmissivity, dumpsite, groundwater

Procedia PDF Downloads 47
7418 Modeling of Daily Global Solar Radiation Using Ann Techniques: A Case of Study

Authors: Said Benkaciali, Mourad Haddadi, Abdallah Khellaf, Kacem Gairaa, Mawloud Guermoui

Abstract:

In this study, many experiments were carried out to assess the influence of the input parameters on the performance of multilayer perceptron which is one the configuration of the artificial neural networks. To estimate the daily global solar radiation on the horizontal surface, we have developed some models by using seven combinations of twelve meteorological and geographical input parameters collected from a radiometric station installed at Ghardaïa city (southern of Algeria). For selecting of best combination which provides a good accuracy, six statistical formulas (or statistical indicators) have been evaluated, such as the root mean square errors, mean absolute errors, correlation coefficient, and determination coefficient. We noted that multilayer perceptron techniques have the best performance, except when the sunshine duration parameter is not included in the input variables. The maximum of determination coefficient and correlation coefficient are equal to 98.20 and 99.11%. On the other hand, some empirical models were developed to compare their performances with those of multilayer perceptron neural networks. Results obtained show that the neural networks techniques give the best performance compared to the empirical models.

Keywords: empirical models, multilayer perceptron neural network, solar radiation, statistical formulas

Procedia PDF Downloads 345
7417 A Novel Hybrid Deep Learning Architecture for Predicting Acute Kidney Injury Using Patient Record Data and Ultrasound Kidney Images

Authors: Sophia Shi

Abstract:

Acute kidney injury (AKI) is the sudden onset of kidney damage in which the kidneys cannot filter waste from the blood, requiring emergency hospitalization. AKI patient mortality rate is high in the ICU and is virtually impossible for doctors to predict because it is so unexpected. Currently, there is no hybrid model predicting AKI that takes advantage of two types of data. De-identified patient data from the MIMIC-III database and de-identified kidney images and corresponding patient records from the Beijing Hospital of the Ministry of Health were collected. Using data features including serum creatinine among others, two numeric models using MIMIC and Beijing Hospital data were built, and with the hospital ultrasounds, an image-only model was built. Convolutional neural networks (CNN) were used, VGG and Resnet for numeric data and Resnet for image data, and they were combined into a hybrid model by concatenating feature maps of both types of models to create a new input. This input enters another CNN block and then two fully connected layers, ending in a binary output after running through Softmax and additional code. The hybrid model successfully predicted AKI and the highest AUROC of the model was 0.953, achieving an accuracy of 90% and F1-score of 0.91. This model can be implemented into urgent clinical settings such as the ICU and aid doctors by assessing the risk of AKI shortly after the patient’s admission to the ICU, so that doctors can take preventative measures and diminish mortality risks and severe kidney damage.

Keywords: Acute kidney injury, Convolutional neural network, Hybrid deep learning, Patient record data, ResNet, Ultrasound kidney images, VGG

Procedia PDF Downloads 131
7416 Participatory Testing of Precision Fertilizer Management Technologies in Mid-Hills of Nepal

Authors: Kedar Nath Nepal, Dyutiman Choudhary, Naba Raj Pandit, Yam Gahire

Abstract:

Crop fertilizer recommendations are outdated as these are based on the response trails conducted over half a century ago. Further, these recommendations were based on the response trials conducted over large geographical area ignoring the large spatial variability in indigenous nutrient supplying capacity of soils typical of most smallholder systems. Application of fertilizer following such blanket recommendation in fields with varying native nutrient supply capacity leads to under application in some places and over application in others leading to reduced nutrient-use-efficiency (NUE), loss of profitability, and increased environmental risks associated with loss of unutilized nutrient through emissions or leaching. Opportunities exist to further increase yield and profitability through a significant gain in fertilizer use efficiency with commercialization of affordable and precise application technologies. We conducted participatory trails in Maize (Zea Mays), Cauliflower (Brassica oleracea var. botrytis) and Tomato (Solanum lycopersicum) in Mid Hills of Nepal to evaluate the efficacy of Urea Deep Placement (UDP and Polymer Coated Urea (PCU);. UDP contains 46% of N having individual briquette size 2.7 gm each and PCU contains 44% of N . Both PCU and urea briquette applied at reduced amount (100 kg N/ha) during planting produced similar yields (p>0.05) compared with regular urea (200 Kg N/ha). . These fertilizers also reduced N fertilizer by 35 - 50% over government blanket recommendations. Further, PCU and urea briquette increased farmer’s net income by USD 60 to 80.

Keywords: high efficiency fertilizers, urea deep placement, briquette polymer coated urea, zea mays, brassica, lycopersicum, Nepal

Procedia PDF Downloads 173
7415 Subthalamic Nucleus in Adult Human Cadaveric Brain: A Morphometric Study

Authors: Mangala Kohli, P. A. Athira, Reeha Mahajan

Abstract:

The subthalamic nucleus (STN) is a biconvex nucleus situated in the diencephalon. The knowledge of the morphometry of the subthalamic nucleus is essential for accurate targeting of the nucleus during Deep Brain Stimulation. The present study aims to note the morphometry of the subthalamic nucleus in both the cerebral hemispheres which will prove to be of great value to radiologists and neurosurgeons. A cross‐sectional observational study was conducted in the Departments of Anatomy and Forensic Medicine, Lady Hardinge Medical College & Associated Hospitals, New Delhi on thirty adult cadaveric brain specimens of unclaimed and donated corpses. The specimens were categorized into 3 age groups: 20-35, 35-50 and above 50 years. All samples were collected after following the standard protocol for ethical clearance. The morphometric study of 60 subthalamic nucleus was thus conducted. Transverse section of the brain was made at a plane 4mm ventral to the plane containing mid commissural point. The dimensions of the subthalamic nucleus were measured bilaterally with the aid of digital Vernier caliper and magnifying glass. In the present study, the mean length and width and AC-PC length of the subthalamic nucleus was recorded on the right and left side in Group A, B and C. On comparison of mean of subthalamic nucleus dimensions between the right and left side in Group C, no statistically significant difference was observed. The length and width of subthalamic nucleus measured in the 3 age groups were compared with each other and the p value calculated. There was no statistically significant difference between the dimensions of Group A and B, Group B and C as well as Group A and C. The present study reveals that there is no significant reduction in the size of the nucleus was noted with increasing age. Thus, the values obtained in the present study can be used as a reference for various invasive and non-invasive procedures on subthalamic nucleus.

Keywords: cerebral hemisphere, deep brain stimulation, morphometry, subthalamic nucleus

Procedia PDF Downloads 184
7414 Global Optimization Techniques for Optimal Placement of HF Antennas on a Shipboard

Authors: Mustafa Ural, Can Bayseferogulari

Abstract:

In this work, radio frequency (RF) coupling between two HF antennas on a shipboard platform is minimized by determining an optimal antenna placement. Unlike the other works, the coupling is minimized not only at single frequency but over the whole frequency band of operation. Similarly, GAO and PSO, are used in order to determine optimal antenna placement. Throughout this work, outputs of two optimization techniques are compared with each other in terms of antenna placements and coupling results. At the end of the work, far-field radiation pattern performances of the antennas at their optimal places are analyzed in terms of directivity and coverage in order to see that.

Keywords: electromagnetic compatibility, antenna placement, optimization, genetic algorithm optimization, particle swarm optimization

Procedia PDF Downloads 236
7413 Methodology to Affirm Driver Engagement in Dynamic Driving Task (DDT) for a Level 2 Adas Feature

Authors: Praneeth Puvvula

Abstract:

Autonomy in has become increasingly common in modern automotive cars. There are 5 levels of autonomy as defined by SAE. This paper focuses on a SAE level 2 feature which, by definition, is able to control the vehicle longitudinally and laterally at the same time. The system keeps the vehicle centred with in the lane by detecting the lane boundaries while maintaining the vehicle speed. As with the features from SAE level 1 to level 3, the primary responsibility of dynamic driving task lies with the driver. This will need monitoring techniques to ensure the driver is always engaged even while the feature is active. This paper focuses on the these techniques, which would help the safe usage of the feature and provide appropriate warnings to the driver.

Keywords: autonomous driving, safety, adas, automotive technology

Procedia PDF Downloads 89
7412 Optimization Modeling of the Hybrid Antenna Array for the DoA Estimation

Authors: Somayeh Komeylian

Abstract:

The direction of arrival (DoA) estimation is the crucial aspect of the radar technologies for detecting and dividing several signal sources. In this scenario, the antenna array output modeling involves numerous parameters including noise samples, signal waveform, signal directions, signal number, and signal to noise ratio (SNR), and thereby the methods of the DoA estimation rely heavily on the generalization characteristic for establishing a large number of the training data sets. Hence, we have analogously represented the two different optimization models of the DoA estimation; (1) the implementation of the decision directed acyclic graph (DDAG) for the multiclass least-squares support vector machine (LS-SVM), and (2) the optimization method of the deep neural network (DNN) radial basis function (RBF). We have rigorously verified that the LS-SVM DDAG algorithm is capable of accurately classifying DoAs for the three classes. However, the accuracy and robustness of the DoA estimation are still highly sensitive to technological imperfections of the antenna arrays such as non-ideal array design and manufacture, array implementation, mutual coupling effect, and background radiation and thereby the method may fail in representing high precision for the DoA estimation. Therefore, this work has a further contribution on developing the DNN-RBF model for the DoA estimation for overcoming the limitations of the non-parametric and data-driven methods in terms of array imperfection and generalization. The numerical results of implementing the DNN-RBF model have confirmed the better performance of the DoA estimation compared with the LS-SVM algorithm. Consequently, we have analogously evaluated the performance of utilizing the two aforementioned optimization methods for the DoA estimation using the concept of the mean squared error (MSE).

Keywords: DoA estimation, Adaptive antenna array, Deep Neural Network, LS-SVM optimization model, Radial basis function, and MSE

Procedia PDF Downloads 100
7411 Devulcanization of Waste Rubber Tyre Utilizing Deep Eutectic Solvents and Ultrasonic Energy

Authors: Ricky Saputra, Rashmi Walvekar, Mohammad Khalid, Kaveh Shahbaz, Suganti Ramarad

Abstract:

This particular study of interest aims to study the effect of coupling ultrasonic treatment with eutectic solvents in devulcanization process of waste rubber tyre. Specifically, three different types of Deep Eutectic Solvents (DES) were utilized, namely ChCl:Urea (1:2), ChCl:ZnCl₂ (1:2) and ZnCl₂:urea (2:7) in which their physicochemical properties were analysed and proven to have permissible water content that is less than 3.0 wt%, degradation temperature below 200ᵒC and freezing point below 60ᵒC. The mass ratio of rubber to DES was varied from 1:20-1:40, sonicated for 1 hour at 37 kHz and heated at variable time of 5-30 min at 180ᵒC. Energy dispersive x-rays (EDX) results revealed that the first two DESs give the highest degree of sulphur removal at 74.44 and 76.69% respectively with optimum heating time at 15 minutes whereby if prolonged, reformation of crosslink network would be experienced. Such is supported by the evidence shown by both FTIR and FESEM results where di-sulfide peak reappears at 30 minutes and morphological structures from 15 to 30 minutes change from smooth with high voidage to rigid with low voidage respectively. Furthermore, TGA curve reveals similar phenomena whereby at 15 minutes thermal decomposition temperature is at the lowest due to the decrease of molecular weight as a result of sulphur removal but increases back at 30 minutes. Type of bond change was also analysed whereby it was found that only di-sulphide bond was cleaved and which indicates partial-devulcanization. Overall, the results show that DES has a great potential to be used as devulcanizing solvent.

Keywords: crosslink network, devulcanization, eutectic solvents, reformation, ultrasonic

Procedia PDF Downloads 173
7410 Artificial Intelligence for Traffic Signal Control and Data Collection

Authors: Reggie Chandra

Abstract:

Trafficaccidents and traffic signal optimization are correlated. However, 70-90% of the traffic signals across the USA are not synchronized. The reason behind that is insufficient resources to create and implement timing plans. In this work, we will discuss the use of a breakthrough Artificial Intelligence (AI) technology to optimize traffic flow and collect 24/7/365 accurate traffic data using a vehicle detection system. We will discuss what are recent advances in Artificial Intelligence technology, how does AI work in vehicles, pedestrians, and bike data collection, creating timing plans, and what is the best workflow for that. Apart from that, this paper will showcase how Artificial Intelligence makes signal timing affordable. We will introduce a technology that uses Convolutional Neural Networks (CNN) and deep learning algorithms to detect, collect data, develop timing plans and deploy them in the field. Convolutional Neural Networks are a class of deep learning networks inspired by the biological processes in the visual cortex. A neural net is modeled after the human brain. It consists of millions of densely connected processing nodes. It is a form of machine learning where the neural net learns to recognize vehicles through training - which is called Deep Learning. The well-trained algorithm overcomes most of the issues faced by other detection methods and provides nearly 100% traffic data accuracy. Through this continuous learning-based method, we can constantly update traffic patterns, generate an unlimited number of timing plans and thus improve vehicle flow. Convolutional Neural Networks not only outperform other detection algorithms but also, in cases such as classifying objects into fine-grained categories, outperform humans. Safety is of primary importance to traffic professionals, but they don't have the studies or data to support their decisions. Currently, one-third of transportation agencies do not collect pedestrian and bike data. We will discuss how the use of Artificial Intelligence for data collection can help reduce pedestrian fatalities and enhance the safety of all vulnerable road users. Moreover, it provides traffic engineers with tools that allow them to unleash their potential, instead of dealing with constant complaints, a snapshot of limited handpicked data, dealing with multiple systems requiring additional work for adaptation. The methodologies used and proposed in the research contain a camera model identification method based on deep Convolutional Neural Networks. The proposed application was evaluated on our data sets acquired through a variety of daily real-world road conditions and compared with the performance of the commonly used methods requiring data collection by counting, evaluating, and adapting it, and running it through well-established algorithms, and then deploying it to the field. This work explores themes such as how technologies powered by Artificial Intelligence can benefit your community and how to translate the complex and often overwhelming benefits into a language accessible to elected officials, community leaders, and the public. Exploring such topics empowers citizens with insider knowledge about the potential of better traffic technology to save lives and improve communities. The synergies that Artificial Intelligence brings to traffic signal control and data collection are unsurpassed.

Keywords: artificial intelligence, convolutional neural networks, data collection, signal control, traffic signal

Procedia PDF Downloads 169
7409 Manifestations of Tuberculosis in Otorhinolaryngology Practice: A Retrospective Study Conducted in a Coastal City of South India

Authors: Rithika Sriram, Kiran M. Bhojwani

Abstract:

Introduction : Tuberculosis of the head and neck has proved to be a diagnostic challenge for otorhinolarynologists around the world. These lesions are often misdiagnosed as cancer. So in order to contribute to a better understanding of these lesions, we have conducted our study among patients affected by TB in the head and neck region with the objective of assessing the various manifestations, presentations, diagnostic techniques, risk factors such as smoking and alcohol consumption, coexisting illnesses and treatment modalities. Materials and Methods: This was a retrospective study conducted over a three year period (2012-2014) in 2 hospitals affliated to Kasturba Medical College in Mangalore, South India. A semi structured proforma was used to capture information from the medical records pertaining to the various objectives of the study such as clinical features and history of smoking. Data was analysed using SPSS version 16.0 and results obtained were depicted as percentages. Chi square test was used to find association between the variables and p<0.05 was considered statistically significant. Results: 104 patients were found to have TB of the head and neck and among them,the most common manifestation was found to be Tubercular Lymphadenitis (86.53%), followed by laryngeal TB (4.8%), submandibular gland TB (3.8%), deep neck space abscess(3.8%) and adenotonsillar TB. FNAC was found to be the gold standard for the diagnosis of TB disease of the lymph node.26% of the patients had coexisting HIV infection and 16.3% of the patients had associated pulmonary TB. More than 20% of the patients were smokers. Most patients were treated using ATT. Conclusion: Tuberculosis affecting regions of head and neck is no longer uncommon. Sufficient knowledge and appropriate diagnostic means is required while dealing with these lesions and must be included in the differential diagnosis of pathological lesions of head and neck.

Keywords: FNAC, Mangalore, smoking, tuberculosis

Procedia PDF Downloads 278
7408 Mitigating Ruminal Methanogenesis Through Genomic and Transcriptomic Approaches

Authors: Muhammad Adeel Arshad, Faiz-Ul Hassan, Yanfen Cheng

Abstract:

According to FAO, enteric methane (CH4) production is about 44% of all greenhouse gas emissions from the livestock sector. Ruminants produce CH4 as a result of fermentation of feed in the rumen especially from roughages which yield more CH4 per unit of biomass ingested as compared to concentrates. Efficient ruminal fermentation is not possible without abating CO2 and CH4. Methane abatement strategies are required to curb the predicted rise in emissions associated with greater ruminant production in future to meet ever increasing animal protein requirements. Ecology of ruminal methanogenesis and avenues for its mitigation can be identified through various genomic and transcriptomic techniques. Programs such as Hungate1000 and the Global Rumen Census have been launched to enhance our understanding about global ruminal microbial communities. Through Hungate1000 project, a comprehensive reference set of rumen microbial genome sequences has been developed from cultivated rumen bacteria and methanogenic archaea along with representative rumen anaerobic fungi and ciliate protozoa cultures. But still many species of rumen microbes are underrepresented especially uncultivable microbes. Lack of sequence information specific to the rumen's microbial community has inhibited efforts to use genomic data to identify specific set of species and their target genes involved in methanogenesis. Metagenomic and metatranscriptomic study of entire microbial rumen populations offer new perspectives to understand interaction of methanogens with other rumen microbes and their potential association with total gas and methane production. Deep understanding of methanogenic pathway will help to devise potentially effective strategies to abate methane production while increasing feed efficiency in ruminants.

Keywords: Genome sequences, Hungate1000, methanogens, ruminal fermentation

Procedia PDF Downloads 139
7407 Optimize Data Evaluation Metrics for Fraud Detection Using Machine Learning

Authors: Jennifer Leach, Umashanger Thayasivam

Abstract:

The use of technology has benefited society in more ways than one ever thought possible. Unfortunately, though, as society’s knowledge of technology has advanced, so has its knowledge of ways to use technology to manipulate people. This has led to a simultaneous advancement in the world of fraud. Machine learning techniques can offer a possible solution to help decrease this advancement. This research explores how the use of various machine learning techniques can aid in detecting fraudulent activity across two different types of fraudulent data, and the accuracy, precision, recall, and F1 were recorded for each method. Each machine learning model was also tested across five different training and testing splits in order to discover which testing split and technique would lead to the most optimal results.

Keywords: data science, fraud detection, machine learning, supervised learning

Procedia PDF Downloads 195
7406 Probabilistic Crash Prediction and Prevention of Vehicle Crash

Authors: Lavanya Annadi, Fahimeh Jafari

Abstract:

Transportation brings immense benefits to society, but it also has its costs. Costs include such as the cost of infrastructure, personnel and equipment, but also the loss of life and property in traffic accidents on the road, delays in travel due to traffic congestion and various indirect costs in terms of air transport. More research has been done to identify the various factors that affect road accidents, such as road infrastructure, traffic, sociodemographic characteristics, land use, and the environment. The aim of this research is to predict the probabilistic crash prediction of vehicles using machine learning due to natural and structural reasons by excluding spontaneous reasons like overspeeding etc., in the United States. These factors range from weather factors, like weather conditions, precipitation, visibility, wind speed, wind direction, temperature, pressure, and humidity to human made structures like road structure factors like bump, roundabout, no exit, turning loop, give away, etc. Probabilities are dissected into ten different classes. All the predictions are based on multiclass classification techniques, which are supervised learning. This study considers all crashes that happened in all states collected by the US government. To calculate the probability, multinomial expected value was used and assigned a classification label as the crash probability. We applied three different classification models, including multiclass Logistic Regression, Random Forest and XGBoost. The numerical results show that XGBoost achieved a 75.2% accuracy rate which indicates the part that is being played by natural and structural reasons for the crash. The paper has provided in-deep insights through exploratory data analysis.

Keywords: road safety, crash prediction, exploratory analysis, machine learning

Procedia PDF Downloads 111
7405 Exploratory Study of the Influencing Factors for Hotels' Competitors

Authors: Asma Ameur, Dhafer Malouche

Abstract:

Hotel competitiveness research is an essential phase of the marketing strategy for any hotel. Certainly, knowing the hotels' competitors helps the hotelier to grasp its position in the market and the citizen to make the right choice in picking a hotel. Thus, competitiveness is an important indicator that can be influenced by various factors. In fact, the issue of competitiveness, this ability to cope with competition, remains a difficult and complex concept to define and to exploit. Therefore, the purpose of this article is to make an exploratory study to calculate a competitiveness indicator for hotels. Further on, this paper makes it possible to determine the criteria of direct or indirect effect on the image and the perception of a hotel. The actual research is used to look into the right model for hotel ‘competitiveness. For this reason, we exploit different theoretical contributions in the field of machine learning. Thus, we use some statistical techniques such as the Principal Component Analysis (PCA) to reduce the dimensions, as well as other techniques of statistical modeling. This paper presents a survey covering of the techniques and methods in hotel competitiveness research. Furthermore, this study allows us to deduct the significant variables that influence the determination of hotel’s competitors. Lastly, the discussed experiences in this article found that the hotel competitors are influenced by several factors with different rates.

Keywords: competitiveness, e-reputation, hotels' competitors, online hotel’ review, principal component analysis, statistical modeling

Procedia PDF Downloads 119
7404 A Study on the Effectiveness of Translanguaging in EFL Classrooms: The Case of First-year Japanese University Students

Authors: Malainine Ebnou

Abstract:

This study investigates the effectiveness of using translanguaging techniques in EFL classrooms. The interest in this topic stems from the lack of research on the effectiveness of translanguaging techniques in foreign language learning, both domestically in Japan and globally, as research has focused on translanguaging from a teaching perspective but not much on it from a learning perspective. The main question that the study departs from is whether students’ use of translanguaging techniques can produce better learning outcomes when used at the university level. The sample population of the study is first-year Japanese university students. The study takes an experimental approach where translanguaging is introduced to one group, the experimental group, and withheld from another group, the control group. Both groups will then be assessed and compared to see if the use of translanguaging has had a positive impact on learning. The impact of the research could be in three ways: challenging the prevailing argument that using learners' mother tongue in the classroom is detrimental to the learning process, challenging native speaker-centered approaches in the EFL field, and arguing that translanguaging in EFL classrooms can produce more meaningful learning outcomes. If the effectiveness of translanguaging is confirmed, it will be possible to promote the use of translanguaging in English learning at Japanese universities and contribute to the improvement of students' English, and even lay the foundations for extending the use of translanguaging to people of other ages/nationalities and other languages in the future.

Keywords: translanguaging, EFL, language learning and teaching, applied linguistics

Procedia PDF Downloads 58
7403 Offline Signature Verification in Punjabi Based On SURF Features and Critical Point Matching Using HMM

Authors: Rajpal Kaur, Pooja Choudhary

Abstract:

Biometrics, which refers to identifying an individual based on his or her physiological or behavioral characteristics, has the capabilities to the reliably distinguish between an authorized person and an imposter. The Signature recognition systems can categorized as offline (static) and online (dynamic). This paper presents Surf Feature based recognition of offline signatures system that is trained with low-resolution scanned signature images. The signature of a person is an important biometric attribute of a human being which can be used to authenticate human identity. However the signatures of human can be handled as an image and recognized using computer vision and HMM techniques. With modern computers, there is need to develop fast algorithms for signature recognition. There are multiple techniques are defined to signature recognition with a lot of scope of research. In this paper, (static signature) off-line signature recognition & verification using surf feature with HMM is proposed, where the signature is captured and presented to the user in an image format. Signatures are verified depended on parameters extracted from the signature using various image processing techniques. The Off-line Signature Verification and Recognition is implemented using Mat lab platform. This work has been analyzed or tested and found suitable for its purpose or result. The proposed method performs better than the other recently proposed methods.

Keywords: offline signature verification, offline signature recognition, signatures, SURF features, HMM

Procedia PDF Downloads 384
7402 The Reality of the Digital Inequality and Its Negative Impact on Virtual Learning during the COVID-19 Pandemic: The South African Perspective

Authors: Jacob Medupe

Abstract:

Life as we know it has changed since the global outbreak of Coronavirus Disease 2019 (COVID-19) and business as usual will not continue. The human impact of the COVID-19 crisis is already immeasurable. Moreover, COVID-19 has already negatively impacted economies, livelihoods and disrupted food systems around the world. The disruptive nature of the Corona virus has affected every sphere of life including the culture and teaching and learning. Right now the majority of education research is based around classroom management techniques that are no longer necessary with digital delivery. Instead there is a great need for new data about how to make the best use of the one-on-one attention that is now becoming possible (Diamandis & Kotler, 2014). The COVID-19 pandemic has necessitated an environment where the South African learners are focused to adhere to social distancing in order to minimise the wild spread of the Corona virus. This arrangement forces the student to utilise the online classroom technologies to continue with the lessons. The historical reality is that the country has not made much strides on the closing of the digital divide and this is particularly a common status quo in the deep rural areas. This will prove to be a toll order for most of the learners affected by the Corona Virus to be able to have a seamless access to the online learning facilities. The paper will seek to look deeply into this reality and how the Corona virus has brought us to the reality that South Africa remains a deeply unequal society in every sphere of life. The study will also explore the state of readiness for education system around the online classroom environment.

Keywords: virtual learning, virtual classroom, COVID-19, Corona virus, internet connectivity, blended learning, online learning, distance education, e-learning, self-regulated Learning, pedagogy, digital literacy

Procedia PDF Downloads 127
7401 The Use of Methods and Techniques of Drama Education with Kindergarten Teachers

Authors: Vladimira Hornackova, Jana Kottasova, Zuzana Vanova, Anna Jungrova

Abstract:

Present study deals with drama education in preschool education. The research made in this field brings a qualitative comparative survey with the aim to find out the use of methods and techniques of drama education in preschool education at university or secondary school graduate preschool teachers. The research uses a content analysis and an unstandardized questionnaire for preschool teachers and obtained data are processed with the help of descriptive methods and correlations. The results allow a comparison of aspects applied through drama in preschool education. The research brings impulses for education improvement in kindergartens and inspiration for university study programs of drama education in the professional training of preschool teachers.

Keywords: drama education, preschool education, preschool teacher, research

Procedia PDF Downloads 365
7400 Efficacy of a Wiener Filter Based Technique for Speech Enhancement in Hearing Aids

Authors: Ajish K. Abraham

Abstract:

Hearing aid is the most fundamental technology employed towards rehabilitation of persons with sensory neural hearing impairment. Hearing in noise is still a matter of major concern for many hearing aid users and thus continues to be a challenging issue for the hearing aid designers. Several techniques are being currently used to enhance the speech at the hearing aid output. Most of these techniques, when implemented, result in reduction of intelligibility of the speech signal. Thus the dissatisfaction of the hearing aid user towards comprehending the desired speech amidst noise is prevailing. Multichannel Wiener Filter is widely implemented in binaural hearing aid technology for noise reduction. In this study, Wiener filter based noise reduction approach is experimented for a single microphone based hearing aid set up. This method checks the status of the input speech signal in each frequency band and then selects the relevant noise reduction procedure. Results showed that the Wiener filter based algorithm is capable of enhancing speech even when the input acoustic signal has a very low Signal to Noise Ratio (SNR). Performance of the algorithm was compared with other similar algorithms on the basis of improvement in intelligibility and SNR of the output, at different SNR levels of the input speech. Wiener filter based algorithm provided significant improvement in SNR and intelligibility compared to other techniques.

Keywords: hearing aid output speech, noise reduction, SNR improvement, Wiener filter, speech enhancement

Procedia PDF Downloads 247