Search results for: profitability management mechanism.
1700 Drivers of Land Degradation in Trays Ecosystem as Modulated under a Changing Climate: Case Study of Côte d'Ivoire
Authors: Kadio Valere R. Angaman, Birahim Bouna Niang
Abstract:
Land degradation is a serious problem in developing countries including Cote d’Ivoire, which has its economy focused on agriculture. It occurs in all kinds of ecosystems over the world. However, the drivers of land degradation vary from one region to another, and from one ecosystem to another. Thus, identifying these drivers is an essential prerequisite to develop and implement appropriate policies to reverse the trend of land degradation in the country, especially in the trays ecosystem. Using the binary logistic model with primary data obtained through 780 farmers surveyed, we analyze and identify the drivers of land degradation in the trays ecosystem. The descriptive statistics show that 52% of farmers interviewed have stated facing land degradation in their farmland. This high rate shows the extent of land degradation in this ecosystem. Also, the results obtained from the binary logit regression reveal that land degradation is significantly influenced by a set of variables such as sex, education, slope, erosion, pesticide, agricultural activity, deforestation, and temperature. The drivers identified are mostly local, as a result, the government must implement some policies and strategies that facilitate and incentive the adoption of sustainable land management practices by farmers to reverse the negative trend of land degradation.
Keywords: Drivers, land degradation, trays ecosystem, sustainable land management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4191699 Comparative Economic Analysis of Floating Photovoltaic Systems Using a Synthesis Approach
Authors: Ching-Feng Chen, Shih-Kai Chen
Abstract:
The Floating Photovoltaic (FPV) system highlights economic benefits and energy performance to carbon dioxide (CO2) discharges. Due to land resource scarcity and many negligent water territories, such as reservoirs, dams, and lakes in Japan and Taiwan, both countries are actively developing FPV and responding to the pricing of the emissions trading systems (ETS). This paper performs a case study through a synthesis approach to compare the economic indicators between the FPVs of Taiwan’s Agongdian Reservoir and Japan’s Yamakura Dam. The research results show that the metrics of the system capacity, installation costs, bank interest rates, and ETS and Electricity Bills affect FPV operating gains. In the post-Feed-In-Tariff (FIT) phase, investing in FPV in Japan is more profitable than in Taiwan. The former’s positive net present value (NPV), eminent internal rate of return (IRR) (11.6%), and benefit-cost ratio (BCR) above 1 (2.0) at the discount rate of 10% indicate that investing the FPV in Japan is more favorable than in Taiwan. In addition, the breakeven point is modest (about 61.3%). The presented methodology in the study helps investors evaluate schemes’ pros and cons and determine whether a decision is beneficial while funding PV or FPV projects.
Keywords: Carbon Border Adjustment Mechanism, Floating Photovoltaic, Emissions Trading Systems, Net Present Value, NPV, Internal Rate of Return, IRR, Benefit-Cost Ratio.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1521698 Radial Basis Surrogate Model Integrated to Evolutionary Algorithm for Solving Computation Intensive Black-Box Problems
Authors: Abdulbaset Saad, Adel Younis, Zuomin Dong
Abstract:
For design optimization with high-dimensional expensive problems, an effective and efficient optimization methodology is desired. This work proposes a series of modification to the Differential Evolution (DE) algorithm for solving computation Intensive Black-Box Problems. The proposed methodology is called Radial Basis Meta-Model Algorithm Assisted Differential Evolutionary (RBF-DE), which is a global optimization algorithm based on the meta-modeling techniques. A meta-modeling assisted DE is proposed to solve computationally expensive optimization problems. The Radial Basis Function (RBF) model is used as a surrogate model to approximate the expensive objective function, while DE employs a mechanism to dynamically select the best performing combination of parameters such as differential rate, cross over probability, and population size. The proposed algorithm is tested on benchmark functions and real life practical applications and problems. The test results demonstrate that the proposed algorithm is promising and performs well compared to other optimization algorithms. The proposed algorithm is capable of converging to acceptable and good solutions in terms of accuracy, number of evaluations, and time needed to converge.
Keywords: Differential evolution, engineering design, expensive computations, meta-modeling, radial basis function, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11731697 Adopting Flocks of Birds Approach to Predator for Anomalies Detection on Industrial Control Systems
Abstract:
Industrial Control Systems (ICS) such as Supervisory Control And Data Acquisition (SCADA) can be seen in many different critical infrastructures, from nuclear management to utility, medical equipment, power, waste and engine management on ships and planes. The role SCADA plays in critical infrastructure has resulted in a call to secure them. Many lives depend on it for daily activities and the attack vectors are becoming more sophisticated. Hence, the security of ICS is vital as malfunction of it might result in huge risk. This paper describes how the application of Prey Predator (PP) approach in flocks of birds could enhance the detection of malicious activities on ICS. The PP approach explains how these animals in groups or flocks detect predators by following some simple rules. They are not necessarily very intelligent animals but their approach in solving complex issues such as detection through corporation, coordination and communication worth emulating. This paper will emulate flocking behavior seen in birds in detecting predators. The PP approach will adopt six nearest bird approach in detecting any predator. Their local and global bests are based on the individual detection as well as group detection. The PP algorithm was designed following MapReduce methodology that follows a Split Detection Convergence (SDC) approach.
Keywords: Industrial control systems, prey predator, SCADA, SDC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11741696 An Edge Detection and Filtering Mechanism of Two Dimensional Digital Objects Based on Fuzzy Inference
Authors: Ayman A. Aly, Abdallah A. Alshnnaway
Abstract:
The general idea behind the filter is to average a pixel using other pixel values from its neighborhood, but simultaneously to take care of important image structures such as edges. The main concern of the proposed filter is to distinguish between any variations of the captured digital image due to noise and due to image structure. The edges give the image the appearance depth and sharpness. A loss of edges makes the image appear blurred or unfocused. However, noise smoothing and edge enhancement are traditionally conflicting tasks. Since most noise filtering behaves like a low pass filter, the blurring of edges and loss of detail seems a natural consequence. Techniques to remedy this inherent conflict often encompass generation of new noise due to enhancement. In this work a new fuzzy filter is presented for the noise reduction of images corrupted with additive noise. The filter consists of three stages. (1) Define fuzzy sets in the input space to computes a fuzzy derivative for eight different directions (2) construct a set of IFTHEN rules by to perform fuzzy smoothing according to contributions of neighboring pixel values and (3) define fuzzy sets in the output space to get the filtered and edged image. Experimental results are obtained to show the feasibility of the proposed approach with two dimensional objects.Keywords: Additive noise, edge preserving filtering, fuzzy image filtering, noise reduction, two dimensional mechanical images.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15681695 Traffic Congestion on Highways in Nigeria Causes, Effects and Remedies
Authors: Popoola M. O., Abiola S. O., Adeniji W. A.
Abstract:
This study investigates the causes, effects and remedies of traffic congestion which has become a common sight in most highways in Nigeria; Mowe/Ibafo section of the Lagos-Ibadan expressway was used as the case-study. 300 Structured questionnaires were distributed among the road users comprising drivers (Private and Commercial), passengers, pedestrians, traffic officers, church congregations, community leaders, Mowe/Ibafo residents, and other users of the road.
300 questionnaires were given out; the average of 276 well completed returned questionnaires formed the basis of the study and was analyzed by the Relative Importance Index (R.I.I.). The result from the study showed the causes of traffic congestion as inadequate road capacity, poor road pavement, poor traffic management, poor drainage system poor driving habit, poor parking habit, poor design junctions/round-about, presence of heavy trucks, lack of pedestrian facilities, lack of road furniture, lack of parking facilities and others. Effects of road congestion from the study are waste of time, delay movement, stress, accident, inability to forecast travel of time, fuel consumption, road rage, relocation, night driving, and environmental pollution. To drastically reduce these negative effects; there must be provision for adequate parking space, construction of proper drainage, enlarging the width of the road, rehabilitate all roads needing attention, public enlightenment, traffic education, hack down all illegal buildings/shops built on the right of way (ROW), create a separate/alternative root for trucks and heavy vehicles, provision of pedestrian facilities, In-depth training of transport/traffic personnel, ban all form of road trading/hawking, and reduce the number of bus-stop where necessary. It is hoped that this study will become the foundation of further research in the area of improve road traffic management on our major highway.
Keywords: Highways, Congestion, Traffic, Traffic congestion, traffic management, Nigeria.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 124251694 Deep Reinforcement Learning Approach for Trading Automation in the Stock Market
Authors: Taylan Kabbani, Ekrem Duman
Abstract:
Deep Reinforcement Learning (DRL) algorithms can scale to previously intractable problems. The automation of profit generation in the stock market is possible using DRL, by combining the financial assets price ”prediction” step and the ”allocation” step of the portfolio in one unified process to produce fully autonomous systems capable of interacting with its environment to make optimal decisions through trial and error. This work represents a DRL model to generate profitable trades in the stock market, effectively overcoming the limitations of supervised learning approaches. We formulate the trading problem as a Partially observed Markov Decision Process (POMDP) model, considering the constraints imposed by the stock market, such as liquidity and transaction costs. We then solved the formulated POMDP problem using the Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm and achieved a 2.68 Sharpe ratio on the test dataset. From the point of view of stock market forecasting and the intelligent decision-making mechanism, this paper demonstrates the superiority of DRL in financial markets over other types of machine learning and proves its credibility and advantages of strategic decision-making.
Keywords: Autonomous agent, deep reinforcement learning, MDP, sentiment analysis, stock market, technical indicators, twin delayed deep deterministic policy gradient.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5241693 Knowledge Impact on Measurement: A Conceptual Metric for Evaluating Performance Improvement (PI) at the Kuwait Institute for Scientific Research (KISR)
Authors: AlMatrouk H. S., Juszczak M. D.
Abstract:
Research and development R&D work involves enormous amount of work that has to do with data measurement and collection. This process evolves as new information is fed, new technologies are utilized, and eventually new knowledge is created by the stakeholders i.e., researchers, clients, and end-users. When new knowledge is created, procedures of R&D work should evolve and produce better results within improved research skills and improved methods of data measurements and collection. This measurement improvement should then be benchmarked against a metric that should be developed at the organization. In this paper, we are suggesting a conceptual metric for R&D work performance improvement (PI) at the Kuwait Institute for Scientific Research (KISR). This PI is to be measured against a set of variables in the suggested metric, which are more closely correlated to organizational output, as opposed to organizational norms. The paper also mentions and discusses knowledge creation and management as an addedvalue to R&D work and measurement improvement. The research methodology followed in this work is qualitative in nature, based on a survey that was distributed to researchers and interviews held with senior researchers at KISR. Research and analyses in this paper also include looking at and analyzing KISR-s literature.Keywords: Knowledge Creation, Performance Improvement (PI), Conceptual Metric, Knowledge Management (KM) addedvalue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12521692 Influence of Social Factors and Motives on Commitment of Sport Events Volunteers
Authors: Farideh Sharififar, Zahra Jamalian, Reza Nikbakhsh, Zahra Nobakht Ramezani
Abstract:
In sport, human resources management gives special attention to method of applying volunteers, their maintenance, and participation of volunteers with each other and management approaches for better operation of events celebrants. The recognition of volunteers- characteristics and motives is important to notice, because it makes the basis of their participation and commitment at sport environment. The motivation and commitment of 281 volunteers were assessed using the organizational commitment scale, motivation scale and personal characteristics questionnaire.The descriptive results showed that; 64% of volunteers were women with age average 21/24 years old. They were physical education student, single (71/9%), without occupation (53%) and with average of 5 years sport experience. Their most important motivation was career factor and the most important commitment factor was normative factor. The results of examining the hypothesized showed that; age, sport experience and education are effective in the amount of volunteers- commitment. And the motive factors such as career, material, purposive and protective factors also have the power to predict the amount of sports volunteers- commitment value. Therefore it is recommended to provide possible opportunities for volunteers and carrying out appropriate instructional courses by events executive managers.Keywords: Sport Volunteers, Motivation, Organizational Commitment, Sport Event
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28321691 Genetic Algorithm Optimization of the Economical, Ecological and Self-Consumption Impact of the Energy Production of a Single Building
Authors: Ludovic Favre, Thibaut M. Schafer, Jean-Luc Robyr, Elena-Lavinia Niederhäuser
Abstract:
This paper presents an optimization method based on genetic algorithm for the energy management inside buildings developed in the frame of the project Smart Living Lab (SLL) in Fribourg (Switzerland). This algorithm optimizes the interaction between renewable energy production, storage systems and energy consumers. In comparison with standard algorithms, the innovative aspect of this project is the extension of the smart regulation over three simultaneous criteria: the energy self-consumption, the decrease of greenhouse gas emissions and operating costs. The genetic algorithm approach was chosen due to the large quantity of optimization variables and the non-linearity of the optimization function. The optimization process includes also real time data of the building as well as weather forecast and users habits. This information is used by a physical model of the building energy resources to predict the future energy production and needs, to select the best energetic strategy, to combine production or storage of energy in order to guarantee the demand of electrical and thermal energy. The principle of operation of the algorithm as well as typical output example of the algorithm is presented.Keywords: Building’s energy, control system, energy management, modelling, genetic optimization algorithm, renewable energy, greenhouse gases, energy storage.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7921690 The Alliance for Grassland Renewal: A Model for Teaching Endophyte Technology
Authors: C. A. Roberts, J. G. Andrae, S. R. Smith, M. H. Poore, C. A. Young, D. W. Hancock, G. J. Pent
Abstract:
To the author’s best knowledge, there are no published reports of effective methods for teaching fescue toxicosis and grass endophyte technology in the USA. To address this need, a group of university scientists, industry representatives, government agents, and livestock producers formed an organization called the Alliance for Grassland Renewal. One goal of the Alliance was to develop a teaching method that could be employed across all regions in the USA and all sectors of the agricultural community. The first step in developing this method was identification of experts who were familiar with the science and management of fescue toxicosis. The second step was curriculum development. Experts wrote a curriculum that addressed all aspects of toxicosis and management, including toxicology, animal nutrition, pasture management, economics, and mycology. The curriculum was created for presentation in lectures, laboratories, and in the field. The curriculum was in that it could be delivered across state lines, regardless of peculiar, in-state recommendations. The curriculum was also unique as it was unanimously supported by private companies otherwise in competition with each other. The final step in developing this teaching method was formulating a delivery plan. All experts, including university, industry, government, and production, volunteered to travel from any state in the USA, converge in one location, teach a 1-day workshop, then travel to the next location. The results of this teaching method indicate widespread success. Since 2012, experts across the entire USA have converged to teach Alliance workshops in Kansas, Oklahoma, Missouri, Kentucky, Georgia, South Carolina, North Carolina, and Virginia, with ongoing workshops in Arkansas and Tennessee. Data from post-workshop surveys indicate that instruction has been effective, as at least 50% of the participants stated their intention to adopt the endophyte technology presented in these workshops. The teaching method developed by the Alliance for Grassland Renewal has proved to be effective, and the Alliance continues to expand across the USA.
Keywords: Endophyte, Epichloë coenophiala, ergot alkaloids, fescue toxicosis, tall fescue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7841689 Chose the Right Mutation Rate for Better Evolve Combinational Logic Circuits
Authors: Emanuele Stomeo, Tatiana Kalganova, Cyrille Lambert
Abstract:
Evolvable hardware (EHW) is a developing field that applies evolutionary algorithm (EA) to automatically design circuits, antennas, robot controllers etc. A lot of research has been done in this area and several different EAs have been introduced to tackle numerous problems, as scalability, evolvability etc. However every time a specific EA is chosen for solving a particular task, all its components, such as population size, initialization, selection mechanism, mutation rate, and genetic operators, should be selected in order to achieve the best results. In the last three decade the selection of the right parameters for the EA-s components for solving different “test-problems" has been investigated. In this paper the behaviour of mutation rate for designing logic circuits, which has not been done before, has been deeply analyzed. The mutation rate for an EHW system modifies the number of inputs of each logic gates, the functionality (for example from AND to NOR) and the connectivity between logic gates. The behaviour of the mutation has been analyzed based on the number of generations, genotype redundancy and number of logic gates for the evolved circuits. The experimental results found provide the behaviour of the mutation rate during evolution for the design and optimization of simple logic circuits. The experimental results propose the best mutation rate to be used for designing combinational logic circuits. The research presented is particular important for those who would like to implement a dynamic mutation rate inside the evolutionary algorithm for evolving digital circuits. The researches on the mutation rate during the last 40 years are also summarized.Keywords: Design of logic circuit, evolutionary computation, evolvable hardware, mutation rate.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16931688 Seamless Flow of Voluminous Data in High Speed Network without Congestion Using Feedback Mechanism
Abstract:
Continuously growing needs for Internet applications that transmit massive amount of data have led to the emergence of high speed network. Data transfer must take place without any congestion and hence feedback parameters must be transferred from the receiver end to the sender end so as to restrict the sending rate in order to avoid congestion. Even though TCP tries to avoid congestion by restricting the sending rate and window size, it never announces the sender about the capacity of the data to be sent and also it reduces the window size by half at the time of congestion therefore resulting in the decrease of throughput, low utilization of the bandwidth and maximum delay. In this paper, XCP protocol is used and feedback parameters are calculated based on arrival rate, service rate, traffic rate and queue size and hence the receiver informs the sender about the throughput, capacity of the data to be sent and window size adjustment, resulting in no drastic decrease in window size, better increase in sending rate because of which there is a continuous flow of data without congestion. Therefore as a result of this, there is a maximum increase in throughput, high utilization of the bandwidth and minimum delay. The result of the proposed work is presented as a graph based on throughput, delay and window size. Thus in this paper, XCP protocol is well illustrated and the various parameters are thoroughly analyzed and adequately presented.Keywords: Bandwidth-Delay Product, Congestion Control, Congestion Window, TCP/IP
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14881687 Relevance Feedback within CBIR Systems
Authors: Mawloud Mosbah, Bachir Boucheham
Abstract:
We present here the results for a comparative study of some techniques, available in the literature, related to the relevance feedback mechanism in the case of a short-term learning. Only one method among those considered here is belonging to the data mining field which is the K-nearest neighbors algorithm (KNN) while the rest of the methods is related purely to the information retrieval field and they fall under the purview of the following three major axes: Shifting query, Feature Weighting and the optimization of the parameters of similarity metric. As a contribution, and in addition to the comparative purpose, we propose a new version of the KNN algorithm referred to as an incremental KNN which is distinct from the original version in the sense that besides the influence of the seeds, the rate of the actual target image is influenced also by the images already rated. The results presented here have been obtained after experiments conducted on the Wang database for one iteration and utilizing color moments on the RGB space. This compact descriptor, Color Moments, is adequate for the efficiency purposes needed in the case of interactive systems. The results obtained allow us to claim that the proposed algorithm proves good results; it even outperforms a wide range of techniques available in the literature.
Keywords: CBIR, Category Search, Relevance Feedback (RFB), Query Point Movement, Standard Rocchio’s Formula, Adaptive Shifting Query, Feature Weighting, Optimization of the Parameters of Similarity Metric, Original KNN, Incremental KNN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23421686 Comparative Performance of Artificial Bee Colony Based Algorithms for Wind-Thermal Unit Commitment
Authors: P. K. Singhal, R. Naresh, V. Sharma
Abstract:
This paper presents the three optimization models, namely New Binary Artificial Bee Colony (NBABC) algorithm, NBABC with Local Search (NBABC-LS), and NBABC with Genetic Crossover (NBABC-GC) for solving the Wind-Thermal Unit Commitment (WTUC) problem. The uncertain nature of the wind power is incorporated using the Weibull probability density function, which is used to calculate the overestimation and underestimation costs associated with the wind power fluctuation. The NBABC algorithm utilizes a mechanism based on the dissimilarity measure between binary strings for generating the binary solutions in WTUC problem. In NBABC algorithm, an intelligent scout bee phase is proposed that replaces the abandoned solution with the global best solution. The local search operator exploits the neighboring region of the current solutions, whereas the integration of genetic crossover with the NBABC algorithm increases the diversity in the search space and thus avoids the problem of local trappings encountered with the NBABC algorithm. These models are then used to decide the units on/off status, whereas the lambda iteration method is used to dispatch the hourly load demand among the committed units. The effectiveness of the proposed models is validated on an IEEE 10-unit thermal system combined with a wind farm over the planning period of 24 hours.Keywords: Artificial bee colony algorithm, economic dispatch, unit commitment, wind power.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10761685 Comparative Performance of Artificial Bee Colony Based Algorithms for Wind-Thermal Unit Commitment
Authors: P. K. Singhal, R. Naresh, V. Sharma
Abstract:
This paper presents the three optimization models, namely New Binary Artificial Bee Colony (NBABC) algorithm, NBABC with Local Search (NBABC-LS), and NBABC with Genetic Crossover (NBABC-GC) for solving the Wind-Thermal Unit Commitment (WTUC) problem. The uncertain nature of the wind power is incorporated using the Weibull probability density function, which is used to calculate the overestimation and underestimation costs associated with the wind power fluctuation. The NBABC algorithm utilizes a mechanism based on the dissimilarity measure between binary strings for generating the binary solutions in WTUC problem. In NBABC algorithm, an intelligent scout bee phase is proposed that replaces the abandoned solution with the global best solution. The local search operator exploits the neighboring region of the current solutions, whereas the integration of genetic crossover with the NBABC algorithm increases the diversity in the search space and thus avoids the problem of local trappings encountered with the NBABC algorithm. These models are then used to decide the units on/off status, whereas the lambda iteration method is used to dispatch the hourly load demand among the committed units. The effectiveness of the proposed models is validated on an IEEE 10-unit thermal system combined with a wind farm over the planning period of 24 hours.Keywords: Artificial bee colony algorithm, economic dispatch, unit commitment, wind power.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11821684 Development of Genetic-based Machine Learning for Network Intrusion Detection (GBML-NID)
Authors: Wafa' S.Al-Sharafat, Reyadh Naoum
Abstract:
Society has grown to rely on Internet services, and the number of Internet users increases every day. As more and more users become connected to the network, the window of opportunity for malicious users to do their damage becomes very great and lucrative. The objective of this paper is to incorporate different techniques into classier system to detect and classify intrusion from normal network packet. Among several techniques, Steady State Genetic-based Machine Leaning Algorithm (SSGBML) will be used to detect intrusions. Where Steady State Genetic Algorithm (SSGA), Simple Genetic Algorithm (SGA), Modified Genetic Algorithm and Zeroth Level Classifier system are investigated in this research. SSGA is used as a discovery mechanism instead of SGA. SGA replaces all old rules with new produced rule preventing old good rules from participating in the next rule generation. Zeroth Level Classifier System is used to play the role of detector by matching incoming environment message with classifiers to determine whether the current message is normal or intrusion and receiving feedback from environment. Finally, in order to attain the best results, Modified SSGA will enhance our discovery engine by using Fuzzy Logic to optimize crossover and mutation probability. The experiments and evaluations of the proposed method were performed with the KDD 99 intrusion detection dataset.Keywords: MSSGBML, Network Intrusion Detection, SGA, SSGA.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16721683 Development of EPID-based Real time Dose Verification for Dynamic IMRT
Authors: Todsaporn Fuangrod, Daryl J. O'Connor, Boyd MC McCurdy, Peter B. Greer
Abstract:
An electronic portal image device (EPID) has become a method of patient-specific IMRT dose verification for radiotherapy. Research studies have focused on pre and post-treatment verification, however, there are currently no interventional procedures using EPID dosimetry that measure the dose in real time as a mechanism to ensure that overdoses do not occur and underdoses are detected as soon as is practically possible. As a result, an EPID-based real time dose verification system for dynamic IMRT was developed and was implemented with MATLAB/Simulink. The EPID image acquisition was set to continuous acquisition mode at 1.4 images per second. The system defined the time constraint gap, or execution gap at the image acquisition time, so that every calculation must be completed before the next image capture is completed. In addition, the <=-evaluation method was used for dose comparison, with two types of comparison processes; individual image and cumulative dose comparison monitored. The outputs of the system are the <=-map, the percent of <=<1, and mean-<= versus time, all in real time. Two strategies were used to test the system, including an error detection test and a clinical data test. The system can monitor the actual dose delivery compared with the treatment plan data or previous treatment dose delivery that means a radiation therapist is able to switch off the machine when the error is detected.Keywords: real-time dose verification, EPID dosimetry, simulation, dynamic IMRT
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21881682 Three Dimensional Modeling of Mixture Formation and Combustion in a Direct Injection Heavy-Duty Diesel Engine
Authors: A. R. Binesh, S. Hossainpour
Abstract:
Due to the stringent legislation for emission of diesel engines and also increasing demand on fuel consumption, the importance of detailed 3D simulation of fuel injection, mixing and combustion have been increased in the recent years. In the present work, FIRE code has been used to study the detailed modeling of spray and mixture formation in a Caterpillar heavy-duty diesel engine. The paper provides an overview of the submodels implemented, which account for liquid spray atomization, droplet secondary break-up, droplet collision, impingement, turbulent dispersion and evaporation. The simulation was performed from intake valve closing (IVC) to exhaust valve opening (EVO). The predicted in-cylinder pressure is validated by comparing with existing experimental data. A good agreement between the predicted and experimental values ensures the accuracy of the numerical predictions collected with the present work. Predictions of engine emissions were also performed and a good quantitative agreement between measured and predicted NOx and soot emission data were obtained with the use of the present Zeldowich mechanism and Hiroyasu model. In addition, the results reported in this paper illustrate that the numerical simulation can be one of the most powerful and beneficial tools for the internal combustion engine design, optimization and performance analysis.Keywords: Diesel engine, Combustion, Pollution, CFD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19471681 Seasonal Influence on Environmental Indicators of Beach Waste
Authors: Marcus C. Garcia, Giselle C. Guimarães, Luciana H. Yamane, Renato R. Siman
Abstract:
The environmental indicators and the classification of beach waste are essential tools to diagnose the current situation and to indicate ways to improve the quality of this environment. The purpose of this paper was to perform a quali-quantitative analysis of the beach waste on the Curva da Jurema Beach (Espírito Santo - Brazil). Three transects were used with equidistant positioning over the total length of the beach for the solid waste collection. Solid wastes were later classified according to their use and primary raw material from the low and high summer season. During the low season, average values of 7.10 items.m-1, 18.22 g.m-1 and 0.91 g.m-2 were found for the whole beach, and transect 3 contributed the most waste, with the total sum of items equal to 999 (49%), a total mass of 5.62 kg and a total volume of 21.31 L. During the high summer season, average values of 8.22 items.m-1, 54.40 g.m-1 and 2.72 g.m-2 were found, with transect 2 contributing the most to the total sum with 1,212 items (53%), a total mass of 10.76 kg and a total volume of 51.99 L. Of the total collected, plastic materials represented 51.4% of the total number of items, 35.9% of the total mass and 68% of the total volume. The implementation of reactive and proactive measures is necessary so that the management of the solid wastes on Curva da Jurema Beach is in accordance with principles of sustainability.Keywords: Beach solid waste, environmental indicators, quali-quantitative analysis, waste management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13981680 Embryo Transfer as an Assisted Reproductive Technology in Farm Animals
Authors: Diah Tri Widayati
Abstract:
Various assisted reproductive techniques have been developed and refined to obtain a large number of offspring from genetically superior animals or obtain offspring from infertile (or subfertile) animals. The embryo transfer is one assisted reproductive technique developed well, aimed at increased productivity of selected females, disease control, importation and exportation of livestock, rapid screening of AI sires for genetically recessive characteristics, treatment or circumvention of certain types of infertility. Embryo transfer also is a useful research tool for evaluating fetal and maternal interactions. This technique has been applied to nearly every species of domestic animal and many species of wildlife and exotic animals, including humans and non-human primates. The successful of embryo transfers have been limited to within-animal, homologous replacement of the embryos. There are several examples of interspecific and intergeneric embryo transfers in which embryos implanted but did not develop to term: sheep and goat, mouse and rat. An immunological rejections and placental incompatibility between the embryo and the surrogate mother appear to restrict interspecific embryo transfer/interspecific pregnancy. Recently, preimplantation embryo manipulation procedures have been applied, such as technique of inner cell mass transfer. This technique will possible to overcome the reproductive barrier interspecific embryo transfer/interspecific pregnancy, if there is a protective mechanism which prevents recognition of the foreign fetus by the mother of the other speciesKeywords: Embryo Transfer, Assisted Reproductive Techology, Intraspesific-Interspesific Pregnancy, Inner cell mass.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 47231679 An Analysis of Digital Forensic Laboratory Development among Malaysia’s Law Enforcement Agencies
Authors: Sarah K. Taylor, Miratun M. Saharuddin, Zabri A. Talib
Abstract:
Cybercrime is on the rise, and yet many Law Enforcement Agencies (LEAs) in Malaysia have no Digital Forensics Laboratory (DFL) to assist them in the attrition and analysis of digital evidence. From the estimated number of 30 LEAs in Malaysia, sadly, only eight of them owned a DFL. All of the DFLs are concentrated in the capital of Malaysia and none at the state level. LEAs are still depending on the national DFL (CyberSecurity Malaysia) even for simple and straightforward cases. A survey was conducted among LEAs in Malaysia owning a DFL to understand their history of establishing the DFL, the challenges that they faced and the significance of the DFL to their case investigation. The results showed that the while some LEAs faced no challenge in establishing a DFL, some of them took seven to 10 years to do so. The reason was due to the difficulty in convincing their management because of the high costs involved. The results also revealed that with the establishment of a DFL, LEAs were better able to get faster forensic result and to meet agency’s timeline expectation. It is also found that LEAs were also able to get more meaningful forensic results on cases that require niche expertise, compared to sending off cases to the national DFL. Other than that, cases are getting more complex, and hence, a continuous stream of budget for equipment and training is inevitable. The result derived from the study is hoped to be used by other LEAs in justifying to their management the benefits of establishing an in-house DFL.
Keywords: Digital forensics, digital forensics laboratory, digital evidence, law enforcement agency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14981678 A Combined Approach of a Sequential Life Testing and an Accelerated Life Testing Applied to a Low-Alloy High Strength Steel Component
Authors: D. I. De Souza, D. R. Fonseca, G. P. Azevedo
Abstract:
Sometimes the amount of time available for testing could be considerably less than the expected lifetime of the component. To overcome such a problem, there is the accelerated life-testing alternative aimed at forcing components to fail by testing them at much higher-than-intended application conditions. These models are known as acceleration models. One possible way to translate test results obtained under accelerated conditions to normal using conditions could be through the application of the “Maxwell Distribution Law.” In this paper we will apply a combined approach of a sequential life testing and an accelerated life testing to a low alloy high-strength steel component used in the construction of overpasses in Brazil. The underlying sampling distribution will be three-parameter Inverse Weibull model. To estimate the three parameters of the Inverse Weibull model we will use a maximum likelihood approach for censored failure data. We will be assuming a linear acceleration condition. To evaluate the accuracy (significance) of the parameter values obtained under normal conditions for the underlying Inverse Weibull model we will apply to the expected normal failure times a sequential life testing using a truncation mechanism. An example will illustrate the application of this procedure.
Keywords: Sequential Life Testing, Accelerated Life Testing, Underlying Three-Parameter Weibull Model, Maximum Likelihood Approach, Hypothesis Testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16391677 Investigating the Effective Parameters in Determining the Type of Traffic Congestion Pricing Schemes in Urban Streets
Authors: Saeed Sayyad Hagh Shomar
Abstract:
Traffic congestion pricing – as a strategy in travel demand management in urban areas to reduce traffic congestion, air pollution and noise pollution – has drawn many attentions towards itself. Unlike the satisfying findings in this method, there are still problems in determining the best functional congestion pricing scheme with regard to the situation. The so-called problems in this process will result in further complications and even the scheme failure. That is why having proper knowledge of the significance of congestion pricing schemes and the effective factors in choosing them can lead to the success of this strategy. In this study, first, a variety of traffic congestion pricing schemes and their components are introduced; then, their functional usage is discussed. Next, by analyzing and comparing the barriers, limitations and advantages, the selection criteria of pricing schemes are described. The results, accordingly, show that the selection of the best scheme depends on various parameters. Finally, based on examining the effective parameters, it is concluded that the implementation of area-based schemes (cordon and zonal) has been more successful in non-diversion of traffic. That is considering the topology of the cities and the fact that traffic congestion is often created in the city centers, area-based schemes would be notably functional and appropriate.Keywords: Congestion pricing, demand management, flat toll, variable toll.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6091676 OXADM Asymmetrical Optical Device: Extending the Application to FTTH System
Authors: Mohammad Syuhaimi Ab-Rahman, Mohd. Saiful Dzulkefly Zan, Mohd Taufiq Mohd Yusof
Abstract:
With the drastically growth in optical communication technology, a lossless, low-crosstalk and multifunction optical switch is most desirable for large-scale photonic network. To realize such a switch, we have introduced the new architecture of optical switch that embedded many functions on single device. The asymmetrical architecture of OXADM consists of 3 parts; selective port, add/drop operation, and path routing. Selective port permits only the interest wavelength pass through and acts as a filter. While add and drop function can be implemented in second part of OXADM architecture. The signals can then be re-routed to any output port or/and perform an accumulation function which multiplex all signals onto single path and then exit to any interest output port. This will be done by path routing operation. The unique features offered by OXADM has extended its application to Fiber to-the Home Technology (FTTH), here the OXADM is used as a wavelength management element in Optical Line Terminal (OLT). Each port is assigned specifically with the operating wavelengths and with the dynamic routing management to ensure no traffic combustion occurs in OLT.Keywords: OXADM, asymmetrical architecture, optical switch, OLT, FTTH.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15491675 Exploratory Data Analysis of Passenger Movement on Delhi Urban Bus Route
Authors: Sourabh Jain, Sukhvir Singh Jain, Gaurav V. Jain
Abstract:
Intelligent Transportation System is an integrated application of communication, control and monitoring and display process technologies for developing a user–friendly transportation system for urban areas in developing countries. In fact, the development of a country and the progress of its transportation system are complementary to each other. Urban traffic has been growing vigorously due to population growth as well as escalation of vehicle ownership causing congestion, delays, pollution, accidents, high-energy consumption and low productivity of resources. The development and management of urban transport in developing countries like India however, is at tryout stage with very few accumulations. Under the umbrella of ITS, urban corridor management strategy have proven to be one of the most successful system in accomplishing these objectives. The present study interprets and figures out the performance of the 27.4 km long Urban Bus route having six intersections, five flyovers and 29 bus stops that covers significant area of the city by causality analysis. Performance interpretations incorporate Passenger Boarding and Alighting, Dwell time, Distance between Bus Stops and Total trip time taken by bus on selected urban route.
Keywords: Congestion, Dwell time, delay, passengers boarding alighting, travel time.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10801674 Use of Fuzzy Logic in the Corporate Reputation Assessment: Stock Market Investors’ Perspective
Authors: Tomasz L. Nawrocki, Danuta Szwajca
Abstract:
The growing importance of reputation in building enterprise value and achieving long-term competitive advantage creates the need for its measurement and evaluation for the management purposes (effective reputation and its risk management). The paper presents practical application of self-developed corporate reputation assessment model from the viewpoint of stock market investors. The model has a pioneer character and example analysis performed for selected industry is a form of specific test for this tool. In the proposed solution, three aspects - informational, financial and development, as well as social ones - were considered. It was also assumed that the individual sub-criteria will be based on public sources of information, and as the calculation apparatus, capable of obtaining synthetic final assessment, fuzzy logic will be used. The main reason for developing this model was to fulfill the gap in the scope of synthetic measure of corporate reputation that would provide higher degree of objectivity by relying on "hard" (not from surveys) and publicly available data. It should be also noted that results obtained on the basis of proposed corporate reputation assessment method give possibilities of various internal as well as inter-branch comparisons and analysis of corporate reputation impact.Keywords: Corporate reputation, fuzzy logic, fuzzy model, stock market investors.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13711673 Discrete Element Modeling of the Effect of Particle Shape on Creep Behavior of Rockfills
Authors: Yunjia Wang, Zhihong Zhao, Erxiang Song
Abstract:
Rockfills are widely used in civil engineering, such as dams, railways, and airport foundations in mountain areas. A significant long-term post-construction settlement may affect the serviceability or even the safety of rockfill infrastructures. The creep behavior of rockfills is influenced by a number of factors, such as particle size, strength and shape, water condition and stress level. However, the effect of particle shape on rockfill creep still remains poorly understood, which deserves a careful investigation. Particle-based discrete element method (DEM) was used to simulate the creep behavior of rockfills under different boundary conditions. Both angular and rounded particles were considered in this numerical study, in order to investigate the influence of particle shape. The preliminary results showed that angular particles experience more breakages and larger creep strains under one-dimensional compression than rounded particles. On the contrary, larger creep strains were observed in he rounded specimens in the direct shear test. The mechanism responsible for this difference is that the possibility of the existence of key particle in rounded particles is higher than that in angular particles. The above simulations demonstrate that the influence of particle shape on the creep behavior of rockfills can be simulated by DEM properly. The method of DEM simulation may facilitate our understanding of deformation properties of rockfill materials.
Keywords: Rockfills, creep behavior, particle crushing, discrete element method, boundary conditions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10801672 The Contribution of Diet and Lifestyle Factors in the Prevalence of Irritable Bowel Syndrome
Authors: Alexander Dao, Oscar Wambuguh
Abstract:
Irritable Bowel Syndrome (IBS) is a heterogeneous functional bowel disease that is characterized by chronic visceral abdominal pain and abnormal bowel function and habits. Its multifactorial pathophysiology and mechanisms are still largely a mystery to the contemporary biomedical community, although there are many hypotheses to try to explain IBS’s presumed physiological, psychosocial, genetic, and environmental etiologies. IBS’s symptomatic presentation is varied and divided into four major subtypes: IBS-C, IBS-D, IBS-M, and IBS-U. Given its diverse presentation and unclear mechanisms, diagnosis is done through a combination of positive identification utilizing the “Rome IV Irritable Bowel Syndrome Criteria'' (Rome IV) diagnostic criteria while also excluding other potential conditions with similar symptoms. Treatment of IBS is focused on the management of symptoms using an assortment of pharmaceuticals, lifestyle changes, and dietary changes, with future potential in microbial treatment and psychotherapy as other therapy methods. Its chronic, heterogeneous nature and disruptive gastrointestinal (GI) symptoms are negatively impactful on patients’ daily lives, health systems, and society. However, with a better understanding of the gaps in knowledge and technological advances in IBS’s pathophysiology, management, and treatment options, there is optimism for the millions of people worldwide who are suffering from the debilitating effects of IBS.
Keywords: Irritable bowel syndrome, lifestyle, diet, functional gastrointestinal disorder.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2011671 Solving an Extended Resource Leveling Problem with Multiobjective Evolutionary Algorithms
Authors: Javier Roca, Etienne Pugnaghi, Gaëtan Libert
Abstract:
We introduce an extended resource leveling model that abstracts real life projects that consider specific work ranges for each resource. Contrary to traditional resource leveling problems this model considers scarce resources and multiple objectives: the minimization of the project makespan and the leveling of each resource usage over time. We formulate this model as a multiobjective optimization problem and we propose a multiobjective genetic algorithm-based solver to optimize it. This solver consists in a two-stage process: a main stage where we obtain non-dominated solutions for all the objectives, and a postprocessing stage where we seek to specifically improve the resource leveling of these solutions. We propose an intelligent encoding for the solver that allows including domain specific knowledge in the solving mechanism. The chosen encoding proves to be effective to solve leveling problems with scarce resources and multiple objectives. The outcome of the proposed solvers represent optimized trade-offs (alternatives) that can be later evaluated by a decision maker, this multi-solution approach represents an advantage over the traditional single solution approach. We compare the proposed solver with state-of-art resource leveling methods and we report competitive and performing results.
Keywords: Intelligent problem encoding, multiobjective decision making, evolutionary computing, RCPSP, resource leveling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4194