Search results for: computational methods
15391 Optimization of Proton Exchange Membrane Fuel Cell Parameters Based on Modified Particle Swarm Algorithms
Authors: M. Dezvarei, S. Morovati
Abstract:
In recent years, increasing usage of electrical energy provides a widespread field for investigating new methods to produce clean electricity with high reliability and cost management. Fuel cells are new clean generations to make electricity and thermal energy together with high performance and no environmental pollution. According to the expansion of fuel cell usage in different industrial networks, the identification and optimization of its parameters is really significant. This paper presents optimization of a proton exchange membrane fuel cell (PEMFC) parameters based on modified particle swarm optimization with real valued mutation (RVM) and clonal algorithms. Mathematical equations of this type of fuel cell are presented as the main model structure in the optimization process. Optimized parameters based on clonal and RVM algorithms are compared with the desired values in the presence and absence of measurement noise. This paper shows that these methods can improve the performance of traditional optimization methods. Simulation results are employed to analyze and compare the performance of these methodologies in order to optimize the proton exchange membrane fuel cell parameters.Keywords: clonal algorithm, proton exchange membrane fuel cell (PEMFC), particle swarm optimization (PSO), real-valued mutation (RVM)
Procedia PDF Downloads 35515390 Practice of Applying MIDI Technology to Train Creative Teaching Skills
Authors: Yang Zhuo
Abstract:
This study explores the integration of MIDI technology as one of the important digital technologies in music teaching, from the perspective of teaching practice, into the process of cultivating students' teaching skills. At the same time, the framework elements of the learning environment for music education students are divided into four aspects: digital technology supported learning space, new knowledge learning, teaching methods, and teaching evaluation. In teaching activities, more attention should be paid to students' subjectivity and interaction between them so as to enhance their emotional experience in teaching practice simulation. In the process of independent exploration and cooperative interaction, problems should be discovered and solved, and basic knowledge of music and teaching methods should be exercised in practice.Keywords: music education, educational technology, MIDI, teacher training
Procedia PDF Downloads 8915389 Recommendations Using Online Water Quality Sensors for Chlorinated Drinking Water Monitoring at Drinking Water Distribution Systems Exposed to Glyphosate
Authors: Angela Maria Fasnacht
Abstract:
Detection of anomalies due to contaminants’ presence, also known as early detection systems in water treatment plants, has become a critical point that deserves an in-depth study for their improvement and adaptation to current requirements. The design of these systems requires a detailed analysis and processing of the data in real-time, so it is necessary to apply various statistical methods appropriate to the data generated, such as Spearman’s Correlation, Factor Analysis, Cross-Correlation, and k-fold Cross-validation. Statistical analysis and methods allow the evaluation of large data sets to model the behavior of variables; in this sense, statistical treatment or analysis could be considered a vital step to be able to develop advanced models focused on machine learning that allows optimized data management in real-time, applied to early detection systems in water treatment processes. These techniques facilitate the development of new technologies used in advanced sensors. In this work, these methods were applied to identify the possible correlations between the measured parameters and the presence of the glyphosate contaminant in the single-pass system. The interaction between the initial concentration of glyphosate and the location of the sensors on the reading of the reported parameters was studied.Keywords: glyphosate, emergent contaminants, machine learning, probes, sensors, predictive
Procedia PDF Downloads 12715388 Comparative Study between Classical P-Q Method and Modern Fuzzy Controller Method to Improve the Power Quality of an Electrical Network
Authors: A. Morsli, A. Tlemçani, N. Ould Cherchali, M. S. Boucherit
Abstract:
This article presents two methods for the compensation of harmonics generated by a nonlinear load. The first is the classic method P-Q. The second is the controller by modern method of artificial intelligence specifically fuzzy logic. Both methods are applied to an Active Power Filter shunt (APFs) based on a three-phase voltage converter at five levels NPC topology. In calculating the harmonic currents of reference, we use the algorithm P-Q and pulse generation, we use the intersective PWM. For flexibility and dynamics, we use fuzzy logic. The results give us clear that the rate of Harmonic Distortion issued by fuzzy logic is better than P-Q.Keywords: fuzzy logic controller, P-Q method, pulse width modulation (PWM), shunt active power filter (sAPF), total harmonic distortion (THD)
Procedia PDF Downloads 55115387 Exploration of Cone Foam Breaker Behavior Using Computational Fluid Dynamic
Authors: G. St-Pierre-Lemieux, E. Askari Mahvelati, D. Groleau, P. Proulx
Abstract:
Mathematical modeling has become an important tool for the study of foam behavior. Computational Fluid Dynamic (CFD) can be used to investigate the behavior of foam around foam breakers to better understand the mechanisms leading to the ‘destruction’ of foam. The focus of this investigation was the simple cone foam breaker, whose performance has been identified in numerous studies. While the optimal pumping angle is known from the literature, the contribution of pressure drop, shearing, and centrifugal forces to the foam syneresis are subject to speculation. This work provides a screening of those factors against changes in the cone angle and foam rheology. The CFD simulation was made with the open source OpenFOAM toolkits on a full three-dimensional model discretized using hexahedral cells. The geometry was generated using a python script then meshed with blockMesh. The OpenFOAM Volume Of Fluid (VOF) method was used (interFOAM) to obtain a detailed description of the interfacial forces, and the model k-omega SST was used to calculate the turbulence fields. The cone configuration allows the use of a rotating wall boundary condition. In each case, a pair of immiscible fluids, foam/air or water/air was used. The foam was modeled as a shear thinning (Herschel-Buckley) fluid. The results were compared to our measurements and to results found in the literature, first by computing the pumping rate of the cone, and second by the liquid break-up at the exit of the cone. A 3D printed version of the cones submerged in foam (shaving cream or soap solution) and water, at speeds varying between 400 RPM and 1500 RPM, was also used to validate the modeling results by calculating the torque exerted on the shaft. While most of the literature is focusing on cone behavior using Newtonian fluids, this works explore its behavior in shear thinning fluid which better reflects foam apparent rheology. Those simulations bring new light on the cone behavior within the foam and allow the computation of shearing, pressure, and velocity of the fluid, enabling to better evaluate the efficiency of the cones as foam breakers. This study contributes to clarify the mechanisms behind foam breaker performances, at least in part, using modern CFD techniques.Keywords: bioreactor, CFD, foam breaker, foam mitigation, OpenFOAM
Procedia PDF Downloads 21015386 Two-Level Graph Causality to Detect and Predict Random Cyber-Attacks
Authors: Van Trieu, Shouhuai Xu, Yusheng Feng
Abstract:
Tracking attack trajectories can be difficult, with limited information about the nature of the attack. Even more difficult as attack information is collected by Intrusion Detection Systems (IDSs) due to the current IDSs having some limitations in identifying malicious and anomalous traffic. Moreover, IDSs only point out the suspicious events but do not show how the events relate to each other or which event possibly cause the other event to happen. Because of this, it is important to investigate new methods capable of performing the tracking of attack trajectories task quickly with less attack information and dependency on IDSs, in order to prioritize actions during incident responses. This paper proposes a two-level graph causality framework for tracking attack trajectories in internet networks by leveraging observable malicious behaviors to detect what is the most probable attack events that can cause another event to occur in the system. Technically, given the time series of malicious events, the framework extracts events with useful features, such as attack time and port number, to apply to the conditional independent tests to detect the relationship between attack events. Using the academic datasets collected by IDSs, experimental results show that the framework can quickly detect the causal pairs that offer meaningful insights into the nature of the internet network, given only reasonable restrictions on network size and structure. Without the framework’s guidance, these insights would not be able to discover by the existing tools, such as IDSs. It would cost expert human analysts a significant time if possible. The computational results from the proposed two-level graph network model reveal the obvious pattern and trends. In fact, more than 85% of causal pairs have the average time difference between the causal and effect events in both computed and observed data within 5 minutes. This result can be used as a preventive measure against future attacks. Although the forecast may be short, from 0.24 seconds to 5 minutes, it is long enough to be used to design a prevention protocol to block those attacks.Keywords: causality, multilevel graph, cyber-attacks, prediction
Procedia PDF Downloads 15915385 Cognitive Dissonance in Robots: A Computational Architecture for Emotional Influence on the Belief System
Authors: Nicolas M. Beleski, Gustavo A. G. Lugo
Abstract:
Robotic agents are taking more and increasingly important roles in society. In order to make these robots and agents more autonomous and efficient, their systems have grown to be considerably complex and convoluted. This growth in complexity has led recent researchers to investigate forms to explain the AI behavior behind these systems in search for more trustworthy interactions. A current problem in explainable AI is the inner workings with the logic inference process and how to conduct a sensibility analysis of the process of valuation and alteration of beliefs. In a social HRI (human-robot interaction) setup, theory of mind is crucial to ease the intentionality gap and to achieve that we should be able to infer over observed human behaviors, such as cases of cognitive dissonance. One specific case inspired in human cognition is the role emotions play on our belief system and the effects caused when observed behavior does not match the expected outcome. In such scenarios emotions can make a person wrongly assume the antecedent P for an observed consequent Q, and as a result, incorrectly assert that P is true. This form of cognitive dissonance where an unproven cause is taken as truth induces changes in the belief base which can directly affect future decisions and actions. If we aim to be inspired by human thoughts in order to apply levels of theory of mind to these artificial agents, we must find the conditions to replicate these observable cognitive mechanisms. To achieve this, a computational architecture is proposed to model the modulation effect emotions have on the belief system and how it affects logic inference process and consequently the decision making of an agent. To validate the model, an experiment based on the prisoner's dilemma is currently under development. The hypothesis to be tested involves two main points: how emotions, modeled as internal argument strength modulators, can alter inference outcomes, and how can explainable outcomes be produced under specific forms of cognitive dissonance.Keywords: cognitive architecture, cognitive dissonance, explainable ai, sensitivity analysis, theory of mind
Procedia PDF Downloads 13515384 A Multistep Broyden’s-Type Method for Solving Systems of Nonlinear Equations
Authors: M. Y. Waziri, M. A. Aliyu
Abstract:
The paper proposes an approach to improve the performance of Broyden’s method for solving systems of nonlinear equations. In this work, we consider the information from two preceding iterates rather than a single preceding iterate to update the Broyden’s matrix that will produce a better approximation of the Jacobian matrix in each iteration. The numerical results verify that the proposed method has clearly enhanced the numerical performance of Broyden’s Method.Keywords: mulit-step Broyden, nonlinear systems of equations, computational efficiency, iterate
Procedia PDF Downloads 64115383 Acceleration Techniques of DEM Simulation for Dynamics of Particle Damping
Authors: Masato Saeki
Abstract:
Presented herein is a novel algorithms for calculating the damping performance of particle dampers. The particle damper is a passive vibration control technique and has many practical applications due to simple design. It consists of granular materials constrained to move between two ends in the cavity of a primary vibrating system. The damping effect results from the exchange of momentum during the impact of granular materials against the wall of the cavity. This damping has the advantage of being independent of the environment. Therefore, particle damping can be applied in extreme temperature environments, where most conventional dampers would fail. It was shown experimentally in many papers that the efficiency of the particle dampers is high in the case of resonant vibration. In order to use the particle dampers effectively, it is necessary to solve the equations of motion for each particle, considering the granularity. The discrete element method (DEM) has been found to be effective for revealing the dynamics of particle damping. In this method, individual particles are assumed as rigid body and interparticle collisions are modeled by mechanical elements as springs and dashpots. However, the computational cost is significant since the equation of motion for each particle must be solved at each time step. In order to improve the computational efficiency of the DEM, the new algorithms are needed. In this study, new algorithms are proposed for implementing the high performance DEM. On the assumption that behaviors of the granular particles in the each divided area of the damper container are the same, the contact force of the primary system with all particles can be considered to be equal to the product of the divided number of the damper area and the contact force of the primary system with granular materials per divided area. This convenience makes it possible to considerably reduce the calculation time. The validity of this calculation method was investigated and the calculated results were compared with the experimental ones. This paper also presents the results of experimental studies of the performance of particle dampers. It is shown that the particle radius affect the noise level. It is also shown that the particle size and the particle material influence the damper performance.Keywords: particle damping, discrete element method (DEM), granular materials, numerical analysis, equivalent noise level
Procedia PDF Downloads 45915382 Literature Review: Application of Artificial Intelligence in EOR
Authors: Masoumeh Mofarrah, Amir NahanMoghadam
Abstract:
Higher oil prices and increasing oil demand are main reasons for great attention to Enhanced Oil Recovery (EOR). Comprehensive researches have been accomplished to develop, appraise and improve EOR methods and their application. Recently Artificial Intelligence (AI) gained popularity in petroleum industry that can help petroleum engineers to solve some fundamental petroleum engineering problems such as reservoir simulation, EOR project risk analysis, well log interpretation and well test model selection. This study presents a historical overview of most popular AI tools including neural networks, genetic algorithms, fuzzy logic and expert systems in petroleum industry and discusses two case studies to represent the application of two mentioned AI methods for selecting an appropriate EOR method based on reservoir characterization in feasible and effective way.Keywords: artificial intelligence, EOR, neural networks, expert systems
Procedia PDF Downloads 41415381 Time Delay Estimation Using Signal Envelopes for Synchronisation of Recordings
Authors: Sergei Aleinik, Mikhail Stolbov
Abstract:
In this work, a method of time delay estimation for dual-channel acoustic signals (speech, music, etc.) recorded under reverberant conditions is investigated. Standard methods based on cross-correlation of the signals show poor results in cases involving strong reverberation, large distances between microphones and asynchronous recordings. Under similar conditions, a method based on cross-correlation of temporal envelopes of the signals delivers a delay estimation of acceptable quality. This method and its properties are described and investigated in detail, including its limits of applicability. The method’s optimal parameter estimation and a comparison with other known methods of time delay estimation are also provided.Keywords: cross-correlation, delay estimation, signal envelope, signal processing
Procedia PDF Downloads 48915380 Methodology for Temporary Analysis of Production and Logistic Systems on the Basis of Distance Data
Authors: M. Mueller, M. Kuehn, M. Voelker
Abstract:
In small and medium-sized enterprises (SMEs), the challenge is to create a well-grounded and reliable basis for process analysis, optimization and planning due to a lack of data. SMEs have limited access to methods with which they can effectively and efficiently analyse processes and identify cause-and-effect relationships in order to generate the necessary database and derive optimization potential from it. The implementation of digitalization within the framework of Industry 4.0 thus becomes a particular necessity for SMEs. For these reasons, the abstract presents an analysis methodology that is subject to the objective of developing an SME-appropriate methodology for efficient, temporarily feasible data collection and evaluation in flexible production and logistics systems as a basis for process analysis and optimization. The overall methodology focuses on retrospective, event-based tracing and analysis of material flow objects. The technological basis consists of Bluetooth low energy (BLE)-based transmitters, so-called beacons, and smart mobile devices (SMD), e.g. smartphones as receivers, between which distance data can be measured and derived motion profiles. The distance is determined using the Received Signal Strength Indicator (RSSI), which is a measure of signal field strength between transmitter and receiver. The focus is the development of a software-based methodology for interpretation of relative movements of transmitters and receivers based on distance data. The main research is on selection and implementation of pattern recognition methods for automatic process recognition as well as methods for the visualization of relative distance data. Due to an existing categorization of the database regarding process types, classification methods (e.g. Support Vector Machine) from the field of supervised learning are used. The necessary data quality requires selection of suitable methods as well as filters for smoothing occurring signal variations of the RSSI, the integration of methods for determination of correction factors depending on possible signal interference sources (columns, pallets) as well as the configuration of the used technology. The parameter settings on which respective algorithms are based have a further significant influence on result quality of the classification methods, correction models and methods for visualizing the position profiles used. The accuracy of classification algorithms can be improved up to 30% by selected parameter variation; this has already been proven in studies. Similar potentials can be observed with parameter variation of methods and filters for signal smoothing. Thus, there is increased interest in obtaining detailed results on the influence of parameter and factor combinations on data quality in this area. The overall methodology is realized with a modular software architecture consisting of independently modules for data acquisition, data preparation and data storage. The demonstrator for initialization and data acquisition is available as mobile Java-based application. The data preparation, including methods for signal smoothing, are Python-based with the possibility to vary parameter settings and to store them in the database (SQLite). The evaluation is divided into two separate software modules with database connection: the achievement of an automated assignment of defined process classes to distance data using selected classification algorithms and the visualization as well as reporting in terms of a graphical user interface (GUI).Keywords: event-based tracing, machine learning, process classification, parameter settings, RSSI, signal smoothing
Procedia PDF Downloads 13815379 Production of High-Content Fructo-Oligosaccharides
Authors: C. Nobre, C. C. Castro, A.-L. Hantson, J. A. Teixeira, L. R. Rodrigues, G. De Weireld
Abstract:
Fructo-oligosaccharides (FOS) are produced from sucrose by Aureobasidium pullulans in yields between 40-60% (w/w). To increase the amount of FOS it is necessary to remove the small, non-prebiotic sugars, present. Two methods for producing high-purity FOS have been developed: the use of microorganisms able to consume small saccharides; and the use of continuous chromatography to separate sugars: simulated moving bed (SMB). It is herein proposed the combination of both methods. The aim of this study is to optimize the composition of the fermentative broth (in terms of salts and sugars) that will be further purified by SMB. A yield of 0.63 gFOS.g Sucrose-1 was obtained with A. pullulans using low amounts of salts in the initial fermentative broth. By removing the small sugars, Saccharomyces cerevisiae and Zymomonas mobilis increased the percentage of FOS from around 56.0% to 83% (w/w) in average, losing only 10% (w/w) of FOS during the recovery process.Keywords: fructo-oligosaccharides, microbial treatment, Saccharomyces cerevisiae, Zymomonas mobilis
Procedia PDF Downloads 31015378 Object Detection in Digital Images under Non-Standardized Conditions Using Illumination and Shadow Filtering
Authors: Waqqas-ur-Rehman Butt, Martin Servin, Marion Pause
Abstract:
In recent years, object detection has gained much attention and very encouraging research area in the field of computer vision. The robust object boundaries detection in an image is demanded in numerous applications of human computer interaction and automated surveillance systems. Many methods and approaches have been developed for automatic object detection in various fields, such as automotive, quality control management and environmental services. Inappropriately, to the best of our knowledge, object detection under illumination with shadow consideration has not been well solved yet. Furthermore, this problem is also one of the major hurdles to keeping an object detection method from the practical applications. This paper presents an approach to automatic object detection in images under non-standardized environmental conditions. A key challenge is how to detect the object, particularly under uneven illumination conditions. Image capturing conditions the algorithms need to consider a variety of possible environmental factors as the colour information, lightening and shadows varies from image to image. Existing methods mostly failed to produce the appropriate result due to variation in colour information, lightening effects, threshold specifications, histogram dependencies and colour ranges. To overcome these limitations we propose an object detection algorithm, with pre-processing methods, to reduce the interference caused by shadow and illumination effects without fixed parameters. We use the Y CrCb colour model without any specific colour ranges and predefined threshold values. The segmented object regions are further classified using morphological operations (Erosion and Dilation) and contours. Proposed approach applied on a large image data set acquired under various environmental conditions for wood stack detection. Experiments show the promising result of the proposed approach in comparison with existing methods.Keywords: image processing, illumination equalization, shadow filtering, object detection
Procedia PDF Downloads 21815377 RANS Simulation of the LNG Ship Squat in Shallow Water
Authors: Mehdi Nakisa, Adi Maimun, Yasser M. Ahmed, Fatemeh Behrouzi
Abstract:
Squat is the reduction in under-keel clearance between a vessel at-rest and underway due to the increased flow of water past the moving body. The forward motion of the ship induces a relative velocity between the ship and the surrounding water that causes a water level depression in which the ship sinks. The problem of ship squat is one among the crucial factors affecting the navigation of ships in restricted waters. This article investigates the LNG ship squat, its effects on flow streamlines around the ship hull and ship behavior and motion using computational fluid dynamics which is applied by Ansys-Fluent.Keywords: ship squat, CFD, confined, mechanic
Procedia PDF Downloads 62315376 Assessing the Effects of Entrepreneurship Education and Moderating Variables on Venture Creation Intention of Undergraduate Students in Ghana
Authors: Daniel K. Gameti
Abstract:
The paper explored the effects of active and passive entrepreneurship education methods on the venture creation intention of undergraduate students in Ghana. The study also examined the moderating effect of gender and negative personal characteristics (risk tolerance, stress tolerance and fear of failure) on students’ venture creation intention. Deductive approach was used in collecting quantitative data from 555 business students from one public university and one private university through self-administered questionnaires. Descriptive statistic was used to determine the dominant method of entrepreneurship education used in Ghana. Further, structural equation model was used to test four hypotheses. The results of the study show that the dominant method of education used in Ghana was lectures and the least method used was field trip. The study further revealed that passive methods of education are less effective compared to active methods which were statistically significant in venture creation intention among students. There was also statistical difference between male and female students’ venture creation intention but stronger among male students and finally, the only personal characteristics that influence students’ intention was stress tolerance because risk tolerance and fear of failure were statistically insignificant.Keywords: entrepreneurship education, Ghana, moderating variables, venture creation intention, undergraduate students
Procedia PDF Downloads 45815375 The Marker Active Compound Identification of Calotropis gigantea Roots Extract as an Anticancer
Authors: Roihatul Mutiah, Sukardiman, Aty Widyawaruyanti
Abstract:
Calotropis gigantiea (L.) R. Br (Apocynaceae) commonly called as “Biduri” or “giant milk weed” is a well-known weed to many cultures for treating various disorders. Several studies reported that C.gigantea roots has anticancer activity. The main aim of this research was to isolate and identify an active marker compound of C.gigantea roots for quality control purpose of its extract in the development as anticancer natural product. The isolation methods was bioactivity guided column chromatography, TLC, and HPLC. Evaluated anticancer activity of there substances using MTT assay methods. Identification structure active compound by UV, 1HNMR, 13CNMR, HMBC, HMQC spectral and other references. The result showed that the marker active compound was identical as Calotropin.Keywords: calotropin, Calotropis gigantea, anticancer, marker active
Procedia PDF Downloads 33715374 Applying Computer Simulation Methods to a Molecular Understanding of Flaviviruses Proteins towards Differential Serological Diagnostics and Therapeutic Intervention
Authors: Sergio Alejandro Cuevas, Catherine Etchebest, Fernando Luis Barroso Da Silva
Abstract:
The flavivirus genus has several organisms responsible for generating various diseases in humans. Special in Brazil, Zika (ZIKV), Dengue (DENV) and Yellow Fever (YFV) viruses have raised great health concerns due to the high number of cases affecting the area during the last years. Diagnostic is still a difficult issue since the clinical symptoms are highly similar. The understanding of their common structural/dynamical and biomolecular interactions features and differences might suggest alternative strategies towards differential serological diagnostics and therapeutic intervention. Due to their immunogenicity, the primary focus of this study was on the ZIKV, DENV and YFV non-structural proteins 1 (NS1) protein. By means of computational studies, we calculated the main physical chemical properties of this protein from different strains that are directly responsible for the biomolecular interactions and, therefore, can be related to the differential infectivity of the strains. We also mapped the electrostatic differences at both the sequence and structural levels for the strains from Uganda to Brazil that could suggest possible molecular mechanisms for the increase of the virulence of ZIKV. It is interesting to note that despite the small changes in the protein sequence due to the high sequence identity among the studied strains, the electrostatic properties are strongly impacted by the pH which also impact on their biomolecular interactions with partners and, consequently, the molecular viral biology. African and Asian strains are distinguishable. Exploring the interfaces used by NS1 to self-associate in different oligomeric states, and to interact with membranes and the antibody, we could map the strategy used by the ZIKV during its evolutionary process. This indicates possible molecular mechanisms that can explain the different immunological response. By the comparison with the known antibody structure available for the West Nile virus, we demonstrated that the antibody would have difficulties to neutralize the NS1 from the Brazilian strain. The present study also opens up perspectives to computationally design high specificity antibodies.Keywords: zika, biomolecular interactions, electrostatic interactions, molecular mechanisms
Procedia PDF Downloads 13515373 Computational Analysis of Adaptable Winglets for Improved Morphing Aircraft Performance
Authors: Erdogan Kaygan, Alvin Gatto
Abstract:
An investigation of adaptable winglets for enhancing morphing aircraft performance is described in this paper. The concepts investigated consist of various winglet configurations fundamentally centered on a baseline swept wing. The impetus for the work was to identify and optimize winglets to enhance the aerodynamic efficiency of a morphing aircraft. All computations were performed with Athena Vortex Lattice modelling with varying degrees of twist and cant angle considered. The results from this work indicate that if adaptable winglets were employed on aircraft’s improvements in aircraft performance could be achieved.Keywords: aircraft, drag, twist, winglet
Procedia PDF Downloads 58915372 Evaluation of Drilling-Induced Delamination of Flax/Epoxy Composites by Non-Destructive Testing Methods
Authors: Hadi Rezghimaleki, Masatoshi Kubouchi, Yoshihiko Arao
Abstract:
The use of natural fiber composites (NFCs) is growing at a fast rate regarding industrial applications and principle researches due to their eco-friendly, renewable nature, and low density/costs. Drilling is one of the most important machining operations that are carried out on natural fiber composites. Delamination is a major concern in the drilling process of NFCs that affects the structural integrity and long-term reliability of the machined components. Flax fiber reinforced epoxy composite laminates were prepared by hot press technique. In this research, we evaluated drilling-induced delamination of flax/epoxy composites by X-ray computed tomography (CT), ultrasonic testing (UT), and optical methods and compared the results.Keywords: natural fiber composites, flax/epoxy, X-ray CT, ultrasonic testing
Procedia PDF Downloads 30215371 3D CFD Modelling of the Airflow and Heat Transfer in Cold Room Filled with Dates
Authors: Zina Ghiloufi, Tahar Khir
Abstract:
A transient three-dimensional computational fluid dynamics (CFD) model is developed to determine the velocity and temperature distribution in different positions cold room during pre-cooling of dates. The turbulence model used is the k-ω Shear Stress Transport (SST) with the standard wall function, the air. The numerical results obtained show that cooling rate is not uniform inside the room; the product at the medium of room has a slower cooling rate. This cooling heterogeneity has a large effect on the energy consumption during cold storage.Keywords: CFD, cold room, cooling rate, dDates, numerical simulation, k-ω (SST)
Procedia PDF Downloads 23815370 Designing Directed Network with Optimal Controllability
Authors: Liang Bai, Yandong Xiao, Haorang Wang, Songyang Lao
Abstract:
The directedness of links is crucial to determine the controllability in complex networks. Even the edge directions can determine the controllability of complex networks. Obviously, for a given network, we wish to design its edge directions that make this network approach the optimal controllability. In this work, we firstly introduce two methods to enhance network by assigning edge directions. However, these two methods could not completely mitigate the negative effects of inaccessibility and dilations. Thus, to approach the optimal network controllability, the edge directions must mitigate the negative effects of inaccessibility and dilations as much as possible. Finally, we propose the edge direction for optimal controllability. The optimal method has been found to be successfully useful on real-world and synthetic networks.Keywords: complex network, dynamics, network control, optimization
Procedia PDF Downloads 19015369 Effect of Different Processing Methods on the Quality Attributes of Pigeon Pea Used in Bread Production
Authors: B. F. Olanipekun, O. J. Oyelade, C. O. Osemobor
Abstract:
Pigeon pea is a very good source of protein and micronutrient, but it is being underutilized in Nigeria because of several constraints. This research considered the effect of different processing methods on the quality attributes of pigeon pea used in bread production towards enhancing its utility. Pigeon pea was obtained at a local market and processed into the flour using three processing methods: soaking, sprouting and roasting and were used to bake bread in different proportions. Chemical composition and sensory attributes of the breads were thereafter determined. The highest values of protein and ash contents were obtained from 20 % substitution of sprouted pigeon pea in wheat flour and may be attributable to complex biochemical changes occurring during hydration, to invariably lead to protein constituent being broken down. Hydrolytic activities of the enzymes from the sprouted sample resulted in improvement in the constituent of total protein probably due to reduction in the carbohydrate content. Sensory qualities analyses showed that bread produced with soaked and roasted pigeon pea flours at 5 and 10% inclusion, respectively were mostly accepted than other blends, and products with sprouted pigeon pea flour were least accepted. The findings of this research suggest that supplementing wheat flour with sprouted pigeon peas have more nutritional potentials. However, with sensory analysis indices, the soaked and roasted pigeon peas up to 10% are majorly accepted, and also can improve the nutritional status. Overall, this will be very beneficial to population dependent on plant protein in order to combat malnutrition problems.Keywords: pigeon pea, processing, protein, malnutrition
Procedia PDF Downloads 25715368 Influence of Selected Finishing Technologies on the Roughness Parameters of Stainless Steel Manufactured by Selective Laser Melting Method
Authors: J. Hajnys, M. Pagac, J. Petru, P. Stefek, J. Mesicek, J. Kratochvil
Abstract:
The new progressive method of 3D metal printing SLM (Selective Laser Melting) is increasingly expanded into the normal operation. As a result, greater demands are placed on the surface quality of the parts produced in this way. The article deals with research of selected finishing methods (tumbling, face milling, sandblasting, shot peening and brushing) and their impact on the final surface roughness. The 20 x 20 x 7 mm produced specimens using SLM additive technology on the Renishaw AM400 were subjected to testing of these finishing methods by adjusting various parameters. Surface parameters of roughness Sa, Sz were chosen as the evaluation criteria and profile parameters Ra, Rz were used as additional measurements. Optical measurement of surface roughness was performed on Alicona Infinite Focus 5. An experiment conducted to optimize the surface roughness revealed, as expected, that the best roughness parameters were achieved through a face milling operation. Tumbling is particularly suitable for 3D printing components, as tumbling media are able to reach even complex shapes and, after changing to polishing bodies, achieve a high surface gloss. Surface quality after tumbling depends on the process time. Other methods with satisfactory results are shot peening and tumbling, which should be the focus of further research.Keywords: additive manufacturing, selective laser melting, SLM, surface roughness, stainless steel
Procedia PDF Downloads 13515367 Generation of Knowlege with Self-Learning Methods for Ophthalmic Data
Authors: Klaus Peter Scherer, Daniel Knöll, Constantin Rieder
Abstract:
Problem and Purpose: Intelligent systems are available and helpful to support the human being decision process, especially when complex surgical eye interventions are necessary and must be performed. Normally, such a decision support system consists of a knowledge-based module, which is responsible for the real assistance power, given by an explanation and logical reasoning processes. The interview based acquisition and generation of the complex knowledge itself is very crucial, because there are different correlations between the complex parameters. So, in this project (semi)automated self-learning methods are researched and developed for an enhancement of the quality of such a decision support system. Methods: For ophthalmic data sets of real patients in a hospital, advanced data mining procedures seem to be very helpful. Especially subgroup analysis methods are developed, extended and used to analyze and find out the correlations and conditional dependencies between the structured patient data. After finding causal dependencies, a ranking must be performed for the generation of rule-based representations. For this, anonymous patient data are transformed into a special machine language format. The imported data are used as input for algorithms of conditioned probability methods to calculate the parameter distributions concerning a special given goal parameter. Results: In the field of knowledge discovery advanced methods and applications could be performed to produce operation and patient related correlations. So, new knowledge was generated by finding causal relations between the operational equipment, the medical instances and patient specific history by a dependency ranking process. After transformation in association rules logically based representations were available for the clinical experts to evaluate the new knowledge. The structured data sets take account of about 80 parameters as special characteristic features per patient. For different extended patient groups (100, 300, 500), as well one target value as well multi-target values were set for the subgroup analysis. So the newly generated hypotheses could be interpreted regarding the dependency or independency of patient number. Conclusions: The aim and the advantage of such a semi-automatically self-learning process are the extensions of the knowledge base by finding new parameter correlations. The discovered knowledge is transformed into association rules and serves as rule-based representation of the knowledge in the knowledge base. Even more, than one goal parameter of interest can be considered by the semi-automated learning process. With ranking procedures, the most strong premises and also conjunctive associated conditions can be found to conclude the interested goal parameter. So the knowledge, hidden in structured tables or lists can be extracted as rule-based representation. This is a real assistance power for the communication with the clinical experts.Keywords: an expert system, knowledge-based support, ophthalmic decision support, self-learning methods
Procedia PDF Downloads 25415366 A Review of Self-Healing Concrete and Various Methods of Its Scientific Implementation
Authors: Davoud Beheshtizadeh, Davood Jafari
Abstract:
Concrete, with its special properties and advantages, has caused it to be widely and increasingly used in construction industry, especially in infrastructures of the country. On the other hand, some defects of concrete and, most importantly, micro-cracks in the concrete after setting have caused the cost of repair and maintenance of infrastructure; therefore, self-healing concretes have been of attention in other countries in the recent years. These concretes have been repaired with general mechanisms such as physical, chemical, biological and combined mechanisms, each of which has different subsets and methods of execution and operation. Also, some of these types of mechanisms are of high importance, which has led to a special production method, and as this subject is new in Iran, this knowledge is almost unknown or at least some part of it has not been considered at all. The present article completely introduces various self-healing mechanisms as a review and tries to present the disadvantages and advantages of each method along with its scope of application.Keywords: micro-cracks, self-healing concrete, microcapsules, concrete, cement, self-sensitive
Procedia PDF Downloads 15515365 AHP and TOPSIS Methods for Supplier Selection Problem in Medical Devices Company
Authors: Sevde D. Karayel, Ediz Atmaca
Abstract:
Supplier selection subject is vital because of development competitiveness and performance of firms which have right, rapid and with low cost procurement. Considering the fact that competition between firms is no longer on their supply chains, hence it is very clear that performance of the firms’ not only depend on their own success but also success of all departments in supply chain. For this purpose, firms want to work with suppliers which are cost effective, flexible in terms of demand and high quality level for customer satisfaction. However, diversification and redundancy of their expectations from suppliers, supplier selection problems need to be solved as a hard problem. In this study, supplier selection problem is discussed for critical piece, which is using almost all production of products in and has troubles with lead time from supplier, in a firm that produces medical devices. Analyzing policy in the current situation of the firm in the supplier selection indicates that supplier selection is made based on the purchasing department experience and other authorized persons’ general judgments. Because selection do not make based on the analytical methods, it is caused disruptions in production, lateness and extra cost. To solve the problem, AHP and TOPSIS which are multi-criteria decision making techniques, which are effective, easy to implement and can analyze many criteria simultaneously, are used to make a selection among alternative suppliers.Keywords: AHP-TOPSIS methods, multi-criteria decision making, supplier selection problem, supply chain management
Procedia PDF Downloads 26515364 A Philosophical Study of Men's Rights Discourses in Light of Feminism
Authors: Michael Barker
Abstract:
Men’s rights activists are largely antifeminism. Evaluation of men’s rights discourses, however, shows that men’s rights’ goals would be better achieved by working with feminism. Discussion of men’s rights discourses, though, is prone to confusion because there is no commonly used men’s rights language. In the presentation ‘male sexism’, ‘matriarchy’ and ‘masculism’ will be unpacked as part of a suggested men’s rights language. Once equipped with a men’s rights vocabulary, sustained philosophical assessment of the extent to which several categories of male disadvantages are wrongful will be offered. Following this, conditions that cause each category of male sexism will be discussed. It shall be argued that male sexism is caused more so by matriarchy than by patriarchy or by feminism. In closing, the success at which various methods address the categories of male sexism will be contrasted. Ultimately, it will be shown that male disadvantages are addressed more successfully by methods that work with, than against, feminism.Keywords: gender studies, feminism, patriarchy, men’s rights, male sexism, matriarchy, masculism
Procedia PDF Downloads 37215363 Combining Shallow and Deep Unsupervised Machine Learning Techniques to Detect Bad Actors in Complex Datasets
Authors: Jun Ming Moey, Zhiyaun Chen, David Nicholson
Abstract:
Bad actors are often hard to detect in data that imprints their behaviour patterns because they are comparatively rare events embedded in non-bad actor data. An unsupervised machine learning framework is applied here to detect bad actors in financial crime datasets that record millions of transactions undertaken by hundreds of actors (<0.01% bad). Specifically, the framework combines ‘shallow’ (PCA, Isolation Forest) and ‘deep’ (Autoencoder) methods to detect outlier patterns. Detection performance analysis for both the individual methods and their combination is reported.Keywords: detection, machine learning, deep learning, unsupervised, outlier analysis, data science, fraud, financial crime
Procedia PDF Downloads 9915362 A Comparison of Methods for Estimating Dichotomous Treatment Effects: A Simulation Study
Authors: Jacqueline Y. Thompson, Sam Watson, Lee Middleton, Karla Hemming
Abstract:
Introduction: The odds ratio (estimated via logistic regression) is a well-established and common approach for estimating covariate-adjusted binary treatment effects when comparing a treatment and control group with dichotomous outcomes. Its popularity is primarily because of its stability and robustness to model misspecification. However, the situation is different for the relative risk and risk difference, which are arguably easier to interpret and better suited to specific designs such as non-inferiority studies. So far, there is no equivalent, widely acceptable approach to estimate an adjusted relative risk and risk difference when conducting clinical trials. This is partly due to the lack of a comprehensive evaluation of available candidate methods. Methods/Approach: A simulation study is designed to evaluate the performance of relevant candidate methods to estimate relative risks to represent conditional and marginal estimation approaches. We consider the log-binomial, generalised linear models (GLM) with iteratively weighted least-squares (IWLS) and model-based standard errors (SE); log-binomial GLM with convex optimisation and model-based SEs; log-binomial GLM with convex optimisation and permutation tests; modified-Poisson GLM IWLS and robust SEs; log-binomial generalised estimation equations (GEE) and robust SEs; marginal standardisation and delta method SEs; and marginal standardisation and permutation test SEs. Independent and identically distributed datasets are simulated from a randomised controlled trial to evaluate these candidate methods. Simulations are replicated 10000 times for each scenario across all possible combinations of sample sizes (200, 1000, and 5000), outcomes (10%, 50%, and 80%), and covariates (ranging from -0.05 to 0.7) representing weak, moderate or strong relationships. Treatment effects (ranging from 0, -0.5, 1; on the log-scale) will consider null (H0) and alternative (H1) hypotheses to evaluate coverage and power in realistic scenarios. Performance measures (bias, mean square error (MSE), relative efficiency, and convergence rates) are evaluated across scenarios covering a range of sample sizes, event rates, covariate prognostic strength, and model misspecifications. Potential Results, Relevance & Impact: There are several methods for estimating unadjusted and adjusted relative risks. However, it is unclear which method(s) is the most efficient, preserves type-I error rate, is robust to model misspecification, or is the most powerful when adjusting for non-prognostic and prognostic covariates. GEE estimations may be biased when the outcome distributions are not from marginal binary data. Also, it seems that marginal standardisation and convex optimisation may perform better than GLM IWLS log-binomial.Keywords: binary outcomes, statistical methods, clinical trials, simulation study
Procedia PDF Downloads 118