Search results for: Markovian Decision Process based Adaptive Scheduling
13637 Ontology-Based Systemizing of the Science Information Devoted to Waste Utilizing by Methanogenesis
Authors: Ye. Shapovalov, V. Shapovalov, O. Stryzhak, A. Salyuk
Abstract:
Over the past decades, amount of scientific information has been growing exponentially. It became more complicated to process and systemize this amount of data. The approach to systematization of scientific information on the production of biogas based on the ontological IT platform “T.O.D.O.S.” has been developed. It has been proposed to select semantic characteristics of each work for their further introduction into the IT platform “T.O.D.O.S.”. An ontological graph with a ranking function for previous scientific research and for a system of selection of microorganisms has been worked out. These systems provide high performance of information management of scientific information.
Keywords: Ontology-based analysis, analysis of scientific data, methanogenesys, microorganism hierarchy, T.O.D.O.S.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 73413636 Numerical Simulation of Natural Gas Dispersion from Low Pressure Pipelines
Authors: Omid Adibi, Nategheh Najafpour, Bijan Farhanieh, Hossein Afshin
Abstract:
Gas release from the pipelines is one of the main factors in the gas industry accidents. Released gas ejects from the pipeline as a free jet and in the growth process, the fuel gets mixed with the ambient air. Accordingly, an accidental spark will release the chemical energy of the mixture with an explosion. Gas explosion damages the equipment and endangers the life of staffs. So due to importance of safety in gas industries, prevision of accident can reduce the number of the casualties. In this paper, natural gas leakages from the low pressure pipelines are studied in two steps: 1) the simulation of mixing process and identification of flammable zones and 2) the simulation of wind effects on the mixing process. The numerical simulations were performed by using the finite volume method and the pressure-based algorithm. Also, for the grid generation the structured method was used. The results show that, in just 6.4 s after accident, released natural gas could penetrate to 40 m in vertical and 20 m in horizontal direction. Moreover, the results show that the wind speed is a key factor in dispersion process. In fact, the wind transports the flammable zones into the downstream. Hence, to improve the safety of the people and human property, it is preferable to construct gas facilities and buildings in the opposite side of prevailing wind direction.
Keywords: Flammable zones, gas pipelines, numerical simulation, wind effects.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 115213635 Method of Parameter Calibration for Error Term in Stochastic User Equilibrium Traffic Assignment Model
Authors: Xiang Zhang, David Rey, S. Travis Waller
Abstract:
Stochastic User Equilibrium (SUE) model is a widely used traffic assignment model in transportation planning, which is regarded more advanced than Deterministic User Equilibrium (DUE) model. However, a problem exists that the performance of the SUE model depends on its error term parameter. The objective of this paper is to propose a systematic method of determining the appropriate error term parameter value for the SUE model. First, the significance of the parameter is explored through a numerical example. Second, the parameter calibration method is developed based on the Logit-based route choice model. The calibration process is realized through multiple nonlinear regression, using sequential quadratic programming combined with least square method. Finally, case analysis is conducted to demonstrate the application of the calibration process and validate the better performance of the SUE model calibrated by the proposed method compared to the SUE models under other parameter values and the DUE model.
Keywords: Parameter calibration, sequential quadratic programming, Stochastic User Equilibrium, traffic assignment, transportation planning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 212913634 Measuring Banks’ Antifragility via Fuzzy Logic
Authors: Danielle Sandler dos Passos, Helder Coelho, Flávia Mori Sarti
Abstract:
Analysing the world banking sector, we realize that traditional risk measurement methodologies no longer reflect the actual scenario with uncertainty and leave out events that can change the dynamics of markets. Considering this, regulators and financial institutions began to search more realistic models. The aim is to include external influences and interdependencies between agents, to describe and measure the operationalization of these complex systems and their risks in a more coherent and credible way. Within this context, X-Events are more frequent than assumed and, with uncertainties and constant changes, the concept of antifragility starts to gain great prominence in comparison to others methodologies of risk management. It is very useful to analyse whether a system succumbs (fragile), resists (robust) or gets benefits (antifragile) from disorder and stress. Thus, this work proposes the creation of the Banking Antifragility Index (BAI), which is based on the calculation of a triangular fuzzy number – to "quantify" qualitative criteria linked to antifragility.
Keywords: Complex adaptive systems, X-events, risk management, antifragility, banking antifragility index, triangular fuzzy number.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 89813633 Emergentist Metaphorical Creativity: Towards a Model of Analysing Metaphorical Creativity in Interactive Talk
Authors: Afef Badri
Abstract:
Metaphorical creativity does not constitute a static property of discourse. It is an interactive dynamic process created online. There has been a lack of research concerning online produced metaphorical creativity. This paper intends to account for metaphorical creativity in online talk-in-interaction as a dynamic process that emerges as discourse unfolds. It brings together insights from the emergentist approach to the study of metaphor in verbal interactions and insights from conceptual blending approach as a model for analysing online metaphorical constructions to propose a model for studying metaphorical creativity in interactive talk. The model is based on three focal points. First, metaphorical creativity is a dynamic emergent and open-to-change process that evolves in real time as interlocutors constantly blend and re-blend previous metaphorical contributions. Second, it is not a product of isolated individual minds but a joint achievement that is co-constructed and co-elaborated by interlocutors. The third and most important point is that the emergent process of metaphorical creativity is tightly shaped by contextual variables surrounding talk-in-interaction. It is grounded in the framework of interpretation of interlocutors. It is constrained by preceding contributions in a way that creates textual cohesion of the verbal exchange and it is also a goal-oriented process predefined by the communicative intention of each participant in a way that reveals the ideological coherence/incoherence of the entire conversation.
Keywords: Communicative intention, conceptual blending, contextual variables, the emergentist approach, ideological coherence, metaphorical creativity, textual cohesion
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 104813632 Exploring the Spatial Characteristics of Mortality Map: A Statistical Area Perspective
Authors: Jung-Hong Hong, Jing-Cen Yang, Cai-Yu Ou
Abstract:
The analysis of geographic inequality heavily relies on the use of location-enabled statistical data and quantitative measures to present the spatial patterns of the selected phenomena and analyze their differences. To protect the privacy of individual instance and link to administrative units, point-based datasets are spatially aggregated to area-based statistical datasets, where only the overall status for the selected levels of spatial units is used for decision making. The partition of the spatial units thus has dominant influence on the outcomes of the analyzed results, well known as the Modifiable Areal Unit Problem (MAUP). A new spatial reference framework, the Taiwan Geographical Statistical Classification (TGSC), was recently introduced in Taiwan based on the spatial partition principles of homogeneous consideration of the number of population and households. Comparing to the outcomes of the traditional township units, TGSC provides additional levels of spatial units with finer granularity for presenting spatial phenomena and enables domain experts to select appropriate dissemination level for publishing statistical data. This paper compares the results of respectively using TGSC and township unit on the mortality data and examines the spatial characteristics of their outcomes. For the mortality data between the period of January 1st, 2008 and December 31st, 2010 of the Taitung County, the all-cause age-standardized death rate (ASDR) ranges from 571 to 1757 per 100,000 persons, whereas the 2nd dissemination area (TGSC) shows greater variation, ranged from 0 to 2222 per 100,000. The finer granularity of spatial units of TGSC clearly provides better outcomes for identifying and evaluating the geographic inequality and can be further analyzed with the statistical measures from other perspectives (e.g., population, area, environment.). The management and analysis of the statistical data referring to the TGSC in this research is strongly supported by the use of Geographic Information System (GIS) technology. An integrated workflow that consists of the tasks of the processing of death certificates, the geocoding of street address, the quality assurance of geocoded results, the automatic calculation of statistic measures, the standardized encoding of measures and the geo-visualization of statistical outcomes is developed. This paper also introduces a set of auxiliary measures from a geographic distribution perspective to further examine the hidden spatial characteristics of mortality data and justify the analyzed results. With the common statistical area framework like TGSC, the preliminary results demonstrate promising potential for developing a web-based statistical service that can effectively access domain statistical data and present the analyzed outcomes in meaningful ways to avoid wrong decision making.
Keywords: Mortality map, spatial patterns, statistical area, variation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 99013631 Detecting Major Misconceptions about Employment in ICT: A Study of the Myths about ICT Work among Females
Authors: Eneli Kindsiko, Kulno Türk
Abstract:
The purpose of the current article is to reveal misconceptions about ICT occupations that keep females away from the field. The study focuses on the three phases in one’s career life cycle: pre-university, university and workplace with the aim of investigating how to attract more females into an ICT-related career. By studying nearly 300 secondary school graduates, 102 university students and 18 female ICT specialists, the study revealed six myths that influence the decision-making process of young girls in pursuing an ICT-related education and career. Furthermore, discriminating conception of ICT as a primarily man’s world is developed before the university period. Stereotypical barriers should be brought out to the public debate, so that a remarkable proportion of possible employees (women) would not stay away from the tech-related fields. Countries could make a remarkable leap in efficiency, when turning their attention to the gender-related issues in the labour market structure.Keywords: ICT, women, education, stereotypes, computers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 163913630 Robust Design of Power System Stabilizers Using Adaptive Genetic Algorithms
Authors: H. Alkhatib, J. Duveau
Abstract:
Genetic algorithms (GAs) have been widely used for global optimization problems. The GA performance depends highly on the choice of the search space for each parameter to be optimized. Often, this choice is a problem-based experience. The search space being a set of potential solutions may contain the global optimum and/or other local optimums. A bad choice of this search space results in poor solutions. In this paper, our approach consists in extending the search space boundaries during the GA optimization, only when it is required. This leads to more diversification of GA population by new solutions that were not available with fixed search space boundaries. So, these dynamic search spaces can improve the GA optimization performances. The proposed approach is applied to power system stabilizer optimization for multimachine power system (16-generator and 68-bus). The obtained results are evaluated and compared with those obtained by ordinary GAs. Eigenvalue analysis and nonlinear system simulation results show the effectiveness of the proposed approach to damp out the electromechanical oscillation and enhance the global system stability.Keywords: Genetic Algorithms, Multiobjective Optimization, Power System Stabilizer, Small Signal Stability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 172413629 A Critics Study of Neural Networks Applied to ion-Exchange Process
Authors: John Kabuba, Antoine Mulaba-Bafubiandi, Kim Battle
Abstract:
This paper presents a critical study about the application of Neural Networks to ion-exchange process. Ionexchange is a complex non-linear process involving many factors influencing the ions uptake mechanisms from the pregnant solution. The following step includes the elution. Published data presents empirical isotherm equations with definite shortcomings resulting in unreliable predictions. Although Neural Network simulation technique encounters a number of disadvantages including its “black box", and a limited ability to explicitly identify possible causal relationships, it has the advantage to implicitly handle complex nonlinear relationships between dependent and independent variables. In the present paper, the Neural Network model based on the back-propagation algorithm Levenberg-Marquardt was developed using a three layer approach with a tangent sigmoid transfer function (tansig) at hidden layer with 11 neurons and linear transfer function (purelin) at out layer. The above mentioned approach has been used to test the effectiveness in simulating ion exchange processes. The modeling results showed that there is an excellent agreement between the experimental data and the predicted values of copper ions removed from aqueous solutions.Keywords: Copper, ion-exchange process, neural networks, simulation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 163213628 On Optimum Stratification
Authors: M. G. M. Khan, V. D. Prasad, D. K. Rao
Abstract:
In this manuscript, we discuss the problem of determining the optimum stratification of a study (or main) variable based on the auxiliary variable that follows a uniform distribution. If the stratification of survey variable is made using the auxiliary variable it may lead to substantial gains in precision of the estimates. This problem is formulated as a Nonlinear Programming Problem (NLPP), which turn out to multistage decision problem and is solved using dynamic programming technique.
Keywords: Auxiliary variable, Dynamic programming technique, Nonlinear programming problem, Optimum stratification, Uniform distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 211313627 Evolution of Quality Function Deployment (QFD) via Fuzzy Concepts and Neural Networks
Authors: M. Haghighi, M. Zowghi, B. Zohouri
Abstract:
Quality Function Deployment (QFD) is an expounded, multi-step planning method for delivering commodity, services, and processes to customers, both external and internal to an organization. It is a way to convert between the diverse customer languages expressing demands (Voice of the Customer), and the organization-s languages expressing results that sate those demands. The policy is to establish one or more matrices that inter-relate producer and consumer reciprocal expectations. Due to its visual presence is called the “House of Quality" (HOQ). In this paper, we assumed HOQ in multi attribute decision making (MADM) pattern and through a proposed MADM method, rank technical specifications. Thereafter compute satisfaction degree of customer requirements and for it, we apply vagueness and uncertainty conditions in decision making by fuzzy set theory. This approach would propound supervised neural network (perceptron) for MADM problem solving.
Keywords: MADM, fuzzy set, QFD, supervised neural network (perceptron).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 174213626 Exploring Tree Growth Variables Influencing Carbon Sequestration in the Face of Climate Change
Authors: F. S. Eguakun, P. O. Adesoye
Abstract:
One of the major problems being faced by human society is that the global temperature is believed to be rising due to human activity that releases carbon IV Oxide (CO2) to the atmosphere. Carbon IV Oxide is the most important greenhouse gas influencing global warming and possible climate change. With climate change becoming alarming, reducing CO2 in our atmosphere has become a primary goal of international efforts. Forest lands are major sink and could absorb large quantities of carbon if the trees are judiciously managed. The study aims at estimating the carbon sequestration capacity of Pinus caribaea (pine) and Tectona grandis (Teak) under the prevailing environmental conditions and exploring tree growth variables that influences the carbon sequestration capacity in Omo Forest Reserve, Ogun State, Nigeria. Improving forest management by manipulating growth characteristics that influences carbon sequestration could be an adaptive strategy of forestry to climate change. Random sampling was used to select Temporary Sample Plots (TSPs) in the study area from where complete enumeration of growth variables was carried out within the plots. The data collected were subjected to descriptive and correlational analyses. The results showed that average carbon stored by Pine and Teak are 994.4±188.3 Kg and 1350.7±180.6 Kg respectively. The difference in carbon stored in the species is significant enough to consider choice of species relevant in climate change adaptation strategy. Tree growth variables influence the capacity of the tree to sequester carbon. Height, diameter, volume, wood density and age are positively correlated to carbon sequestration. These tree growth variables could be manipulated by the forest manager as an adaptive strategy for climate change while plantations of high wood density species could be relevant for management strategy to increase carbon storage.
Keywords: Adaptation, carbon sequestration, climate change, growth variables, wood density.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 231613625 Stability and Kinetic Analysis during Vermicomposting of Sewage Sludge
Authors: Ashish Kumar Nayak, Dhamodharan K., Ajay S. Kalamdhad
Abstract:
The present study is aimed at alteration of sewage sludge into stable compost product using vermicomposting of sewage sludge mixed with cattle manure and saw dust in five different proportions based on C/N ratios (C/N 15 (R1), 20 (R2), 25 (R3) and 30 (R4); and control (R5)) by employing an epigeic earthworm Eisenia fetida. Higher reductions in C/N ratio, CO2 evolution and OUR were observed in R4 demonstrated the compost stability. In addition, R4 proved to be best combination for the growth of the earthworms. In order to observe the optimal degradation, kinetics for degradation of organic matter in vermicomposting were quantitatively evaluated. An approach model was developed by assuming that composting process is carried out in a homogeneous way and the kinetics for decomposition reaction is represented by a Monod-type equation. The results exhibit comparable variations in the kinetic constants Km and K3 under varying parameters during vermicomposting process. Results suggested that higher R2 value in R4, enhanced suitability towards Lineweaver-Burke plot. R4 yields higher degradability coefficient (K) reveals that the occurrence of optimal nutrient balance, which not only enhanced the affinity of enzymes towards substrate but also improved its degradation process. Therefore, it can be proved that R4 provided to be the best feed combination for vermicomposting process as compared to other reactors.
Keywords: Vermicomposting, Eisenia fetida, Sewage sludge, C/N ratio, Stability, Enzyme kinetics concept.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 235013624 Simulation of the Reactive Rotational Molding Using Smoothed Particle Hydrodynamics
Authors: A. Hamidi, S. Khelladi, L. Illoul, A. Tcharkhtchi
Abstract:
Reactive rotational molding (RRM) is a process to manufacture hollow plastic parts with reactive material has several advantages compared to conventional roto molding of thermoplastic powders: process cycle time is shorter; raw material is less expensive because polymerization occurs during processing and high-performance polymers may be used such as thermosets, thermoplastics or blends. However, several phenomena occur during this process which makes the optimization of the process quite complex. In this study, we have used a mixture of isocyanate and polyol as a reactive system. The chemical transformation of this system to polyurethane has been studied by thermal analysis and rheology tests. Thanks to these results of the curing process and rheological measurements, the kinetic and rheokinetik of polyurethane was identified. Smoothed Particle Hydrodynamics, a Lagrangian meshless method, was chosen to simulate reactive fluid flow in 2 and 3D configurations of the polyurethane during the process taking into account the chemical, and chemiorehological results obtained experimentally in this study.Keywords: Reactive rotational molding, free surface flows, simulation, smoothed particle hydrodynamics, surface tension.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 107713623 Energy Efficiency Analysis of Crossover Technologies in Industrial Applications
Authors: W. Schellong
Abstract:
Industry accounts for one-third of global final energy demand. Crossover technologies (e.g. motors, pumps, process heat, and air conditioning) play an important role in improving energy efficiency. These technologies are used in many applications independent of the production branch. Especially electrical power is used by drives, pumps, compressors, and lightning. The paper demonstrates the algorithm of the energy analysis by some selected case studies for typical industrial processes. The energy analysis represents an essential part of energy management systems (EMS). Generally, process control system (PCS) can support EMS. They provide information about the production process, and they organize the maintenance actions. Combining these tools into an integrated process allows the development of an energy critical equipment strategy. Thus, asset and energy management can use the same common data to improve the energy efficiency.
Keywords: Crossover technologies, data management, energy analysis, energy efficiency, process control.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 96113622 Optimization Technique in Scheduling Duck Tours
Authors: Norhazwani M. Y., Khoo, C. F., Hasrul Nisham R.
Abstract:
Tourism industries are rapidly increased for the last few years especially in Malaysia. In order to attract more tourists, Malaysian Governance encourages any effort to increase Malaysian tourism industry. One of the efforts in attracting more tourists in Malacca, Malaysia is a duck tour. Duck tour is an amphibious sightseeing tour that works in two types of engines, hence, it required a huge cost to operate and maintain the vehicle. To other country, it is not so new but in Malaysia, it is just introduced, thus it does not have any systematic routing yet. Therefore, this paper proposed an optimization technique to formulate and schedule this tour to minimize the operating costs by considering it into Travelling Salesman Problem (TSP). The problem is then can be solved by one of the optimization technique especially meta-heuristics approach such as Tabu Search (TS) and Reactive Tabu Search (RTS).Keywords: Optimization, Reactive Tabu Search, Tabu Search, Travelling Salesman Problem
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 170113621 Improvements in Material Handling: A Case Study of Cement Manufacturing Plant
Authors: A. Pancharya
Abstract:
The globalization of the Indian economy has thrown a great challenge to the Indian industries in respect of productivity, quality, cost, delivery etc. Achieving success• the global market has required fundamental shift in the way business is conducted and has dramatically affected virtually every aspect of process industry. The internal manufacturing process and supporting infrastructure should be such that it can compete successfully in global markets with better flexibility and delivery. The paper deals with a case study of a reputed process industry, some changes in the process has been suggested, which leads to reduction in labor cost and production cost.
Keywords: Indian cement industry, material handling, plant layout.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 489513620 Using Gaussian Process in Wind Power Forecasting
Authors: Hacene Benkhoula, Mohamed Badreddine Benabdella, Hamid Bouzeboudja, Abderrahmane Asraoui
Abstract:
The wind is a random variable difficult to master, for this, we developed a mathematical and statistical methods enable to modeling and forecast wind power. Gaussian Processes (GP) is one of the most widely used families of stochastic processes for modeling dependent data observed over time, or space or time and space. GP is an underlying process formed by unrecognized operator’s uses to solve a problem. The purpose of this paper is to present how to forecast wind power by using the GP. The Gaussian process method for forecasting are presented. To validate the presented approach, a simulation under the MATLAB environment has been given.Keywords: Forecasting, Gaussian process, modeling, wind power.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 178913619 Customer Churn Prediction Using Four Machine Learning Algorithms Integrating Feature Selection and Normalization in the Telecom Sector
Authors: Alanoud Moraya Aldalan, Abdulaziz Almaleh
Abstract:
A crucial part of maintaining a customer-oriented business in the telecommunications industry is understanding the reasons and factors that lead to customer churn. Competition between telecom companies has greatly increased in recent years, which has made it more important to understand customers’ needs in this strong market. For those who are looking to turn over their service providers, understanding their needs is especially important. Predictive churn is now a mandatory requirement for retaining customers in the telecommunications industry. Machine learning can be used to accomplish this. Churn Prediction has become a very important topic in terms of machine learning classification in the telecommunications industry. Understanding the factors of customer churn and how they behave is very important to building an effective churn prediction model. This paper aims to predict churn and identify factors of customers’ churn based on their past service usage history. Aiming at this objective, the study makes use of feature selection, normalization, and feature engineering. Then, this study compared the performance of four different machine learning algorithms on the Orange dataset: Logistic Regression, Random Forest, Decision Tree, and Gradient Boosting. Evaluation of the performance was conducted by using the F1 score and ROC-AUC. Comparing the results of this study with existing models has proven to produce better results. The results showed the Gradients Boosting with feature selection technique outperformed in this study by achieving a 99% F1-score and 99% AUC, and all other experiments achieved good results as well.
Keywords: Machine Learning, Gradient Boosting, Logistic Regression, Churn, Random Forest, Decision Tree, ROC, AUC, F1-score.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40813618 Improving TNT Curing Process by Using Infrared Camera
Authors: O. Srihakulung, Y. Soongsumal
Abstract:
Among the chemicals used for ammunition production, TNT (Trinitrotoluene) play a significant role since World War I and II. Various types of military weapon utilize TNT in casting process. However, the TNT casting process for warhead is difficult to control the cooling rate of the liquid TNT. This problem occurs because the casting process lacks the equipment to detect the temperature during the casting procedure This study presents the temperature detected by infrared camera to illustrate the cooling rate and cooling zone of curing, and demonstrates the optimization of TNT condition to reduce the risk of air gap occurred in the warhead which can result in the destruction afterward. Premature initiation of explosive-filled projectiles in response to set-back forces during gunfiring cause by casting defects. Finally the study can help improving the process of the TNT casting. The operators can control the curing of TNT inside the case by rising up the heating rod at the proper time. Consequently this can reduce tremendous time of rework if the air gaps occur and increase strength to lower elastic modulus. Therefore, it can be clearly concluded that the use of Infrared Cameras in this process is another method to improve the casting procedure.
Keywords: Infrared camera, TNT casting, warhead, curing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 226513617 Development of a Complex Meteorological Support System for UAVs
Authors: Z. Bottyán, F. Wantuch, A. Z. Gyöngyösi, Z. Tuba, K. Hadobács, P. Kardos, R. Kurunczi
Abstract:
The sensitivity of UAVs to the atmospheric effects are apparent. All the same the meteorological support for the UAVs missions is often non-adequate or partly missing. In our paper we show a new complex meteorological support system for different types of UAVs pilots, specialists and decision makers, too. The mentioned system has two important parts with different forecasts approach such as the statistical and dynamical ones. The statistical prediction approach is based on a large climatological data base and the special analog method which is able to select similar weather situations from the mentioned data base to apply them during the forecasting procedure. The applied dynamic approach uses the specific WRF model runs twice a day and produces 96 hours, high resolution weather forecast for the UAV users over the Hungary. An easy to use web-based system can give important weather information over the Carpathian basin in Central-Europe. The mentioned products can be reached via internet connection.Keywords: Aviation meteorology, statistical weather prediction, unmanned aerial systems, WRF.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 276313616 The Process of Crisis: Model of Its Development in the Organization
Authors: M. Mikušová
Abstract:
The main aim of this paper is to present a clear and comprehensive picture of the process of a crisis in the organization which will help to better understand its possible developments. For a description of the sequence of individual steps and an indication of their causation and possible variants of the developments, a detailed flow diagram with verbal comment is applied. For simplicity, the process of the crisis is observed in four basic phases called: symptoms of the crisis, diagnosis, action and prevention. The model highlights the complexity of the phenomenon of the crisis and that the various phases of the crisis are interweaving.
Keywords: Crisis, management, model, organization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 113513615 Comparison of Process Slaughtered on Beef Cattle Based on Level of Cortisol and Fourier Transform Infrared Spectroscopy (FTIR)
Authors: Pudji Astuti, C. P. C. Putro, C. M. Airin, L. Sjahfirdi, S. Widiyanto, H. Maheshwari
Abstract:
Stress of slaughter animals starting long before until at the time of process of slaughtering which cause misery and decrease of meat quality. Meanwhile, determination of animal stress using hormonal such as cortisol is expensive and less practical so that portable stress indicator for cows based on Fourier Transform Infrared Spectroscopy (FTIR) must be provided. The aims of this research are to find out the comparison process of slaughter between Rope Casting Local (RCL) and Restraining Box Method (RBM) by measuring of cortisol and wavelength in FTIR methods. Thirty two of male Ongole crossbred cattle were used in this experiment. Blood sampling was taken from jugular vein when they were rested and repeated when slaughtered. All of blood samples were centrifuged at 3000 rpm for 20 minutes to get serum, and then divided into two parts for cortisol assayed using ELISA and for measuring the wavelength using FTIR. The serum then measured at the wavelength between 4000-400 cm-1 using MB3000 FTIR. Band data absorption in wavelength of FTIR is analyzed descriptively by using FTIR Horizon MBTM. For RCL, average of serum cortisol when the animals rested were 11.47 ± 4.88 ng/mL, when the time of slaughter were 23.27 ± 7.84 ng/mL. For RBM, level of cortisol when rested animals were 13.67 ± 3.41 ng/mL and 53.47 ± 20.25 ng/mL during the slaughter. Based on student t-Test, there were significantly different between RBM and RCL methods when beef cattle were slaughtered (P<0.05), but no significantly different when animals were rested (P>0.05). Result of FTIR with the various of wavelength such as methyl group (=CH3 ) 2986cm-1, methylene (=CH2 ) 2827 cm-1, hydroxyl (- OH) 3371 cm-1, carbonyl (ketones) (C=O) 1636 cm-1, carboxyl (COO-1) 1408 cm-1, glucosa 1057 cm-1, urea 1011 cm-1have been obtained. It can be concluded that the RCL slaughtered method is better than the RBM method based on the increase of cortisol as an indicator of stress in beef cattle (P<0.05). FTIR is really possible to be used as stub of stress tool due to differentiate of resting and slaughter condition by recognizing the increase of absorption and the separation of component group at the wavelength.
Keywords: Cows, cortisol, FTIR, RBM, RCL, stress indicator.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 246313614 Development of an Ensemble Classification Model Based on Hybrid Filter-Wrapper Feature Selection for Email Phishing Detection
Authors: R. B. Ibrahim, M. S. Argungu, I. M. Mungadi
Abstract:
It is obvious in this present time, internet has become an indispensable part of human life since its inception. The Internet has provided diverse opportunities to make life so easy for human beings, through the adoption of various channels. Among these channels are email, internet banking, video conferencing, and the like. Email is one of the easiest means of communication hugely accepted among individuals and organizations globally. But over decades the security integrity of this platform has been challenged with malicious activities like Phishing. Email phishing is designed by phishers to fool the recipient into handing over sensitive personal information such as passwords, credit card numbers, account credentials, social security numbers, etc. This activity has caused a lot of financial damage to email users globally which has resulted in bankruptcy, sudden death of victims, and other health-related sicknesses. Although many methods have been proposed to detect email phishing, in this research, the results of multiple machine-learning methods for predicting email phishing have been compared with the use of filter-wrapper feature selection. It is worth noting that all three models performed substantially but one outperformed the other. The dataset used for these models is obtained from Kaggle online data repository, while three classifiers: decision tree, Naïve Bayes, and Logistic regression are ensemble (Bagging) respectively. Results from the study show that the Decision Tree (CART) bagging ensemble recorded the highest accuracy of 98.13% using PEF (Phishing Essential Features). This result further demonstrates the dependability of the proposed model.
Keywords: Ensemble, hybrid, filter-wrapper, phishing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17913613 Design Based Performance Prediction of Component Based Software Products
Authors: K. S. Jasmine, R. Vasantha
Abstract:
Component-Based software engineering provides an opportunity for better quality and increased productivity in software development by using reusable software components [10]. One of the most critical aspects of the quality of a software system is its performance. The systematic application of software performance engineering techniques throughout the development process can help to identify design alternatives that preserve desirable qualities such as extensibility and reusability while meeting performance objectives [1]. In the present scenario, software engineering methodologies strongly focus on the functionality of the system, while applying a “fix- it-later" approach to software performance aspects [3]. As a result, lengthy fine-tunings, expensive extra hard ware, or even redesigns are necessary for the system to meet the performance requirements. In this paper, we propose design based, implementation independent, performance prediction approach to reduce the overhead associated in the later phases while developing a performance guaranteed software product with the help of Unified Modeling Language (UML).Keywords: Software Reuse, Component-based development, Unified Modeling Language, Software performance, Software components, Performance engineering, Software engineering.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 186713612 Mathematical Model and Solution Algorithm for Containership Operation/Maintenance Scheduling
Authors: Hun Go, Ji-Su Kim, Dong-Ho Lee
Abstract:
This study considers the problem of determining operation and maintenance schedules for a containership equipped with components during its sailing according to a pre-determined navigation schedule. The operation schedule, which specifies work time of each component, determines the due-date of each maintenance activity, and the maintenance schedule specifies the actual start time of each maintenance activity. The main constraints are component requirements, workforce availability, working time limitation, and inter-maintenance time. To represent the problem mathematically, a mixed integer programming model is developed. Then, due to the problem complexity, we suggest a heuristic for the objective of minimizing the sum of earliness and tardiness between the due-date and the starting time of each maintenance activity. Computational experiments were done on various test instances and the results are reported.Keywords: Containerships, operation and preventive maintenance schedules, integer programming, heuristic
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 161513611 On-line and Off-line POD Assisted Projective Integral for Non-linear Problems: A Case Study with Burgers-Equation
Authors: Montri Maleewong, Sirod Sirisup
Abstract:
The POD-assisted projective integration method based on the equation-free framework is presented in this paper. The method is essentially based on the slow manifold governing of given system. We have applied two variants which are the “on-line" and “off-line" methods for solving the one-dimensional viscous Bergers- equation. For the on-line method, we have computed the slow manifold by extracting the POD modes and used them on-the-fly along the projective integration process without assuming knowledge of the underlying slow manifold. In contrast, the underlying slow manifold must be computed prior to the projective integration process for the off-line method. The projective step is performed by the forward Euler method. Numerical experiments show that for the case of nonperiodic system, the on-line method is more efficient than the off-line method. Besides, the online approach is more realistic when apply the POD-assisted projective integration method to solve any systems. The critical value of the projective time step which directly limits the efficiency of both methods is also shown.
Keywords: Projective integration, POD method, equation-free.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 135513610 Salient Points Reduction for Content-Based Image Retrieval
Authors: Yao-Hong Tsai
Abstract:
Salient points are frequently used to represent local properties of the image in content-based image retrieval. In this paper, we present a reduction algorithm that extracts the local most salient points such that they not only give a satisfying representation of an image, but also make the image retrieval process efficiently. This algorithm recursively reduces the continuous point set by their corresponding saliency values under a top-down approach. The resulting salient points are evaluated with an image retrieval system using Hausdoff distance. In this experiment, it shows that our method is robust and the extracted salient points provide better retrieval performance comparing with other point detectors.Keywords: Barnard detector, Content-based image retrieval, Points reduction, Salient point.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 146913609 Comparative Study of Bending Angle in Laser Forming Process Using Artificial Neural Network and Fuzzy Logic System
Authors: M. Hassani, Y. Hassani, N. Ajudanioskooei, N. N. Benvid
Abstract:
Laser Forming process as a non-contact thermal forming process is widely used to forming and bending of metallic and non-metallic sheets. In this process, according to laser irradiation along a specific path, sheet is bent. One of the most important output parameters in laser forming is bending angle that depends on process parameters such as physical and mechanical properties of materials, laser power, laser travel speed and the number of scan passes. In this paper, Artificial Neural Network and Fuzzy Logic System were used to predict of bending angle in laser forming process. Inputs to these models were laser travel speed and laser power. The comparison between artificial neural network and fuzzy logic models with experimental results has been shown both of these models have high ability to prediction of bending angles with minimum errors.
Keywords: Artificial neural network, bending angle, fuzzy logic, laser forming.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 96113608 Intelligent Multi-Agent Middleware for Ubiquitous Home Networking Environments
Authors: Minwoo Son, Seung-Hun Lee, Dongkyoo Shin, Dongil Shin
Abstract:
The next stage of the home networking environment is supposed to be ubiquitous, where each piece of material is equipped with an RFID (Radio Frequency Identification) tag. To fully support the ubiquitous environment, home networking middleware should be able to recommend home services based on a user-s interests and efficiently manage information on service usage profiles for the users. Therefore, USN (Ubiquitous Sensor Network) technology, which recognizes and manages a appliance-s state-information (location, capabilities, and so on) by connecting RFID tags is considered. The Intelligent Multi-Agent Middleware (IMAM) architecture was proposed to intelligently manage the mobile RFID-based home networking and to automatically supply information about home services that match a user-s interests. Evaluation results for personalization services for IMAM using Bayesian-Net and Decision Trees are presented.Keywords: Intelligent Agents, Home Network, Mobile RFID, Intelligent Middleware.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1444