Search results for: Real time Acquisition.
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7785

Search results for: Real time Acquisition.

5145 Phenols and Manganese Removal from Landfill Leachate and Municipal Wastewater Using the Constructed Wetland

Authors: Amin Mojiri, Lou Ziyang

Abstract:

Constructed Wetland (CW) is a reasonable method to treat wastewater. Current study was carried out to co-treat landfill leachate and domestic wastewater using a CW system. Typha domingensis was transplanted to CW, which encloses two substrate layers of adsorbents named ZELIAC and zeolite. Response surface methodology and central composite design were employed to evaluate experimental data. Contact time (h) and leachate-towastewater mixing ratio (%; v/v) were selected as independent factors. Phenols and manganese removal were selected as dependent responses. At optimum contact time (48.7 h) and leachate-towastewater mixing ratio (20.0%), removal efficiencies of phenols and manganese removal efficiencies were 90.5%, and 89.4%, respectively.

Keywords: Constructed wetland, manganese, phenols, Thypha domingensis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2304
5144 Decision Support System for Hospital Selection in Emergency Medical Services: A Discrete Event Simulation Approach

Authors: D. Tedesco, G. Feletti, P. Trucco

Abstract:

The present study aims to develop a Decision Support System (DSS) to support operational decisions in Emergency Medical Service (EMS) systems regarding the assignment of medical emergency requests to Emergency Departments (ED). This problem is called “hospital selection” and concerns the definition of policies for the selection of the ED to which patients who require further treatment are transported by ambulance. The employed research methodology consists of a first phase of review of the technical-scientific literature concerning DSSs to support the EMS management and, in particular, the hospital selection decision. From the literature analysis, it emerged that current studies mainly focused on the EMS phases related to the ambulance service and consider a process that ends when the ambulance is available after completing a mission. Therefore, all the ED-related issues are excluded and considered as part of a separate process. Indeed, the most studied hospital selection policy turned out to be proximity, thus allowing to minimize the travelling time and to free-up the ambulance in the shortest possible time. The purpose of the present study consists in developing an optimization model for assigning medical emergency requests to the EDs also considering the expected time performance in the subsequent phases of the process, such as the case mix, the expected service throughput times, and the operational capacity of different EDs in hospitals. To this end, a Discrete Event Simulation (DES) model was created to compare different hospital selection policies. The model was implemented with the AnyLogic software and finally validated on a realistic case. The hospital selection policy that returned the best results was the minimization of the Time To Provider (TTP), considered as the time from the beginning of the ambulance journey to the ED at the beginning of the clinical evaluation by the doctor. Finally, two approaches were further compared: a static approach, based on a retrospective estimation of the TTP, and a dynamic approach, focused on a predictive estimation of the TTP which is determined with a constantly updated Winters forecasting model. Findings reveal that considering the minimization of TTP is the best hospital selection policy. It allows to significantly reducing service throughput times in the ED with a negligible increase in travel time. Furthermore, an immediate view of the saturation state of the ED is produced and the case mix present in the ED structures (i.e., the different triage codes) is considered, as different severity codes correspond to different service throughput times. Besides, the use of a predictive approach is certainly more reliable in terms on TTP estimation, than a retrospective approach. These considerations can support decision-makers in introducing different hospital selection policies to enhance EMSs performance.

Keywords: Emergency medical services, hospital selection, discrete event simulation, forecast model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 194
5143 Pollutants Removal from Synthetic Wastewater by the Combined Electrochemical Sequencing Batch Reactor

Authors: Amin Mojiri, Akiyoshi Ohashi, Tomonori Kindaichi

Abstract:

Synthetic domestic wastewater was treated via combining treatment methods, including electrochemical oxidation, adsorption, and sequencing batch reactor (SBR). In the upper part of the reactor, an anode and a cathode (Ti/RuO2-IrO2) were organized in parallel for the electrochemical oxidation procedure. Sodium sulfate (Na2SO4) with a concentration of 2.5 g/L was applied as the electrolyte. The voltage and current were fixed on 7.50 V and 0.40 A, respectively. Then, 15% working value of the reactor was filled by activated sludge, and 85% working value of the reactor was added with synthetic wastewater. Powdered cockleshell, 1.5 g/L, was added in the reactor to do ion-exchange. Response surface methodology was employed for statistical analysis. Reaction time (h) and pH were considered as independent factors. A total of 97.0% biochemical oxygen demand, 99.9% phosphorous and 88.6% cadmium were eliminated at the optimum reaction time (80.0 min) and pH (6.4).

Keywords: Adsorption, electrochemical oxidation, metals, sequencing batch reactor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 771
5142 Automatic Tuning for a Systemic Model of Banking Originated Losses (SYMBOL) Tool on Multicore

Authors: Ronal Muresano, Andrea Pagano

Abstract:

Nowadays, the mathematical/statistical applications are developed with more complexity and accuracy. However, these precisions and complexities have brought as result that applications need more computational power in order to be executed faster. In this sense, the multicore environments are playing an important role to improve and to optimize the execution time of these applications. These environments allow us the inclusion of more parallelism inside the node. However, to take advantage of this parallelism is not an easy task, because we have to deal with some problems such as: cores communications, data locality, memory sizes (cache and RAM), synchronizations, data dependencies on the model, etc. These issues are becoming more important when we wish to improve the application’s performance and scalability. Hence, this paper describes an optimization method developed for Systemic Model of Banking Originated Losses (SYMBOL) tool developed by the European Commission, which is based on analyzing the application's weakness in order to exploit the advantages of the multicore. All these improvements are done in an automatic and transparent manner with the aim of improving the performance metrics of our tool. Finally, experimental evaluations show the effectiveness of our new optimized version, in which we have achieved a considerable improvement on the execution time. The time has been reduced around 96% for the best case tested, between the original serial version and the automatic parallel version.

Keywords: Algorithm optimization, Bank Failures, OpenMP, Parallel Techniques, Statistical tool.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1878
5141 Removal of Chromium from Aqueous Solution using Synthesized Polyaniline in Acetonitrile

Authors: Majid Riahi Samani, Seyed Mehdi Borghei

Abstract:

Absorptive characteristics of polyaniline synthesized in mixture of water and acetonitrile in 50/50 volume ratio was studied. Synthesized polyaniline in powder shape is used as an adsorbent to remove toxic hexavalent chromium from aqueous solutions. Experiments were conducted in batch mode with different variables such as agitation time, solution pH and initial concentration of hexavalent chromium. Removal mechanism is the combination of surface adsorption and reduction. The equilibrium time for removal of Cr(T) and Cr(VI) was about 2 and 10 minutes respectively. The optimum pH for total chromium removal occurred at pH 7 and maximum hexavalent chromium removal took place under acidic condition at pH 3. Investigating the isothermal characteristics showed that the equilibrium adsorption data fitted both Freundlich-s and Langmuir-s isotherms. The maximum adsorption of chromium was calculated 36.1 mg/g for polyaniline

Keywords: Polyaniline, Chromium, acetonitrile, Adsorption

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2204
5140 Redefining Field Experiences: Virtual Environments in Teacher Education

Authors: Laurie Mullen, Jayne Beilke, Nancy Brooks

Abstract:

The explosion of interest in online gaming and virtual worlds is leading many universities to investigate possible educational applications of the new environments. In this paper we explore the possibilities of 3D online worlds for teacher education, particularly the field experience component. Drawing upon two pedagogical examples, we suggest that virtual simulations may, with certain limitations, create safe spaces that allow preservice teachers to adopt alternate identities and interact safely with the “other." In so doing they may become aware of the constructed nature of social categories and gain the essential pedagogical skill of perspective-taking. We suggest that, ultimately, the ability to be the principal creators of themselves in virtual environments can increase their ability to do the same in the real world.

Keywords: field experience, pedagogy, simulation, teacher education

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1567
5139 A Mobile Multihop Relay Dynamic TDD Scheme for Cellular Networks

Authors: Jong-Moon Chung, Hyung-Weon Cho, Ki-Yong Jin, Min-Hee Cho

Abstract:

In this paper, we present an analytical framework for the evaluation of the uplink performance of multihop cellular networks based on dynamic time division duplex (TDD). New wireless broadband protocols, such as WiMAX, WiBro, and 3G-LTE apply TDD, and mobile communication protocols under standardization (e.g., IEEE802.16j) are investigating mobile multihop relay (MMR) as a future technology. In this paper a novel MMR TDD scheme is presented, where the dynamic range of the frame is shared to traffic resources of asymmetric nature and multihop relaying. The mobile communication channel interference model comprises of inner and co-channel interference (CCI). The performance analysis focuses on the uplink due to the fact that the effects of dynamic resource allocation show significant performance degradation only in the uplink compared to time division multiple access (TDMA) schemes due to CCI [1-3], where the downlink results to be the same or better.The analysis was based on the signal to interference power ratio (SIR) outage probability of dynamic TDD (D-TDD) and TDMA systems,which are the most widespread mobile communication multi-user control techniques. This paper presents the uplink SIR outage probability with multihop results and shows that the dynamic TDD scheme applying MMR can provide a performance improvement compared to single hop applications if executed properly.

Keywords: Co-Channel Interference, Dynamic TDD, MobileMultihop Reply, Cellular Network, Time Division Multiple Access.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2325
5138 Corporate Social Responsibility and Values in Innovation Management

Authors: J. Maksimainen, P. Saariluoma, P. Jokivuori

Abstract:

Corporate social responsibility (CSR) viewpoint have challenged the traditional perception to understand corporations position. Production- and managerial-centred views are expanding towards reference group-centred policies. Consequently, the significance of new kind of knowledge has emerged. In addition to management of the organisation, the idea of CSR emphasises the importance to recognise the value-expectations of operational environment. It is know that management is often well-aware of corporate social responsibilities, but it is less clear how well these high level goals are understood in practical product design and development work. In this study, the apprehension above proved to be real to some degree. While management was very aware of CSR it was less familiar to designers. The outcome shows that it is essential to raise ethical values and issues higher in corporate communication, if it is wished that they materialize also in products.

Keywords: Corporate social responsibility, management, engineering, values.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1803
5137 A CBR System to New Product Development: An Application for Hearing Devices Design

Authors: J.L. Castro, K. Benghazi, M.V. Hurtado, M. Navarro, J.M. Zurita

Abstract:

Nowadays, quick technological changes force companies to develop innovative products in an increasingly competitive environment. Therefore, how to enhance the time of new product development is very important. This design problem often lacks the exact formula for getting it, and highly depends upon human designers- past experiences. For these reasons, in this work, a Casebased reasoning (CBR) system to assist in new product development is proposed. When a case is recovered from the case base, the system will take into account not only the attribute-s specific value and how important it is. It will also take into account if the attribute has a positive influence over the product development. Hence the manufacturing time will be improved. This information will be introduced as a new concept called “adaptability". An application to this method for hearing instrument new design illustrates the proposed approach.

Keywords: Case based reasoning, Fuzzy logic, New product development, Retrieval stage, Similarity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1459
5136 Empirical Roughness Progression Models of Heavy Duty Rural Pavements

Authors: Nahla H. Alaswadko, Rayya A. Hassan, Bayar N. Mohammed

Abstract:

Empirical deterministic models have been developed to predict roughness progression of heavy duty spray sealed pavements for a dataset representing rural arterial roads. The dataset provides a good representation of the relevant network and covers a wide range of operating and environmental conditions. A sample with a large size of historical time series data for many pavement sections has been collected and prepared for use in multilevel regression analysis. The modelling parameters include road roughness as performance parameter and traffic loading, time, initial pavement strength, reactivity level of subgrade soil, climate condition, and condition of drainage system as predictor parameters. The purpose of this paper is to report the approaches adopted for models development and validation. The study presents multilevel models that can account for the correlation among time series data of the same section and to capture the effect of unobserved variables. Study results show that the models fit the data very well. The contribution and significance of relevant influencing factors in predicting roughness progression are presented and explained. The paper concludes that the analysis approach used for developing the models confirmed their accuracy and reliability by well-fitting to the validation data.

Keywords: Roughness progression, empirical model, pavement performance, heavy duty pavement.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 783
5135 Towards End-To-End Disease Prediction from Raw Metagenomic Data

Authors: Maxence Queyrel, Edi Prifti, Alexandre Templier, Jean-Daniel Zucker

Abstract:

Analysis of the human microbiome using metagenomic sequencing data has demonstrated high ability in discriminating various human diseases. Raw metagenomic sequencing data require multiple complex and computationally heavy bioinformatics steps prior to data analysis. Such data contain millions of short sequences read from the fragmented DNA sequences and stored as fastq files. Conventional processing pipelines consist in multiple steps including quality control, filtering, alignment of sequences against genomic catalogs (genes, species, taxonomic levels, functional pathways, etc.). These pipelines are complex to use, time consuming and rely on a large number of parameters that often provide variability and impact the estimation of the microbiome elements. Training Deep Neural Networks directly from raw sequencing data is a promising approach to bypass some of the challenges associated with mainstream bioinformatics pipelines. Most of these methods use the concept of word and sentence embeddings that create a meaningful and numerical representation of DNA sequences, while extracting features and reducing the dimensionality of the data. In this paper we present an end-to-end approach that classifies patients into disease groups directly from raw metagenomic reads: metagenome2vec. This approach is composed of four steps (i) generating a vocabulary of k-mers and learning their numerical embeddings; (ii) learning DNA sequence (read) embeddings; (iii) identifying the genome from which the sequence is most likely to come and (iv) training a multiple instance learning classifier which predicts the phenotype based on the vector representation of the raw data. An attention mechanism is applied in the network so that the model can be interpreted, assigning a weight to the influence of the prediction for each genome. Using two public real-life data-sets as well a simulated one, we demonstrated that this original approach reaches high performance, comparable with the state-of-the-art methods applied directly on processed data though mainstream bioinformatics workflows. These results are encouraging for this proof of concept work. We believe that with further dedication, the DNN models have the potential to surpass mainstream bioinformatics workflows in disease classification tasks.

Keywords: Metagenomics, phenotype prediction, deep learning, embeddings, multiple instance learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 878
5134 Kinetic Energy Recovery System Using Spring

Authors: Mayuresh Thombre, Prajyot Borkar, Mangirish Bhobe

Abstract:

New advancement of technology and never satisfying demands of the civilization are putting huge pressure on the natural fuel resources and these resources are at a constant threat to its sustainability. To get the best out of the automobile, the optimum balance between performance and fuel economy is important. In the present state of art, either of the above two aspects are taken into mind while designing and development process which puts the other in the loss as increase in fuel economy leads to decrement in performance and vice-versa. In-depth observation of the vehicle dynamics apparently shows that large amount of energy is lost during braking and likewise large amount of fuel is consumed to reclaim the initial state, this leads to lower fuel efficiency to gain the same performance. Current use of Kinetic Energy Recovery System is only limited to sports vehicles only because of the higher cost of this system. They are also temporary in nature as power can be squeezed only during a small time duration and use of superior parts leads to high cost, which results on concentration on performance only and neglecting the fuel economy. In this paper Kinetic Energy Recovery System for storing the power and then using the same while accelerating has been discussed. The major storing element in this system is a Flat Spiral Spring that will store energy by compression and torsion.

The use of spring ensure the permanent storage of energy until used by the driver unlike present mechanical regeneration system in which the energy stored decreases with time and is eventually lost. A combination of internal gears and spur gears will be used in order to make the energy release uniform which will lead to safe usage. The system can be used to improve the fuel efficiency by assisting in overcoming the vehicle’s inertia after braking or to provide instant acceleration whenever required by the driver. The performance characteristics of the system including response time, mechanical efficiency and overall increase in efficiency are demonstrated. This technology makes the KERS (Kinetic Energy Recovery System) more flexible and economical allowing specific application while at the same time increasing the time frame and ease of usage.

Keywords: Electric control unit, Energy, Mechanical KERS, Planetary Gear system, Power, Smart braking, Spiral Spring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8750
5133 Using Perspective Schemata to Model the ETL Process

Authors: Valeria M. Pequeno, Joao Carlos G. M. Pires

Abstract:

Data Warehouses (DWs) are repositories which contain the unified history of an enterprise for decision support. The data must be Extracted from information sources, Transformed and integrated to be Loaded (ETL) into the DW, using ETL tools. These tools focus on data movement, where the models are only used as a means to this aim. Under a conceptual viewpoint, the authors want to innovate the ETL process in two ways: 1) to make clear compatibility between models in a declarative fashion, using correspondence assertions and 2) to identify the instances of different sources that represent the same entity in the real-world. This paper presents the overview of the proposed framework to model the ETL process, which is based on the use of a reference model and perspective schemata. This approach provides the designer with a better understanding of the semantic associated with the ETL process.

Keywords: conceptual data model, correspondence assertions, data warehouse, data integration, ETL process, object relational database.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1489
5132 Prospective Mathematics Teachers' Views about Using Flash Animations in Mathematics Lessons

Authors: Esra Bukova-Güzel, Berna Cantürk-Günhan

Abstract:

The purpose of the study is to determine secondary prospective mathematics teachers- views related to using flash animations in mathematics lessons and to reveal how the sample presentations towards different mathematical concepts altered their views. This is a case study involving three secondary prospective mathematics teachers from a state university in Turkey. The data gathered from two semi-structural interviews. Findings revealed that these animations help understand mathematics meaningfully, relate mathematics and real world, visualization, and comprehend the importance of mathematics. The analysis of the data indicated that the sample presentations enhanced participants- views about using flash animations in mathematics lessons.

Keywords: Instructional technology, animations, prospective mathematics teachers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2327
5131 Dynamic Balance, Pain and Functional Performance in Cruciate Retaining, Posterior Stabilized and Uni-Compartmental Knee Arthroplasty

Authors: Ahmed R. Z. Baghdadi, Amira A. A. Abdallah

Abstract:

Background: With the perceived pain and poor function experienced following knee arthroplasty, patients usually feel un-satisfied. Yet, a controversy still persists on the appropriate operative technique that doesn’t affect proprioception much.

Purpose: This study compared the effects of Cruciate Retaining (CR) and Posterior Stabilized (PS) total knee arthroplasty (TKA) and uni-compartmental knee arthroplasty (UKA) on dynamic balance, pain and functional performance following rehabilitation.

Methods: Fifteen patients with CRTKA (group I), fifteen with PSTKA (group II), fifteen with UKA (group III) and fifteen indicated for arthroplasty but weren’t operated on yet (group IV) participated in the study. The mean age was 54.53±3.44, 55.13±3.48, 52.8±1.93 and 55.33±2.32 years and BMI 35.7±3.03, 35.7±1.99, 35.6±1.88 and 35.73±1.03 kg/m2 for group I, II, III and IV respectively. The Berg Balance Scale (BBS), WOMAC pain subscale and Timed Up-and-Go (TUG) and Stair-Climbing (SC) tests were used for assessment. Assessments were conducted four and eight weeks pre- and post-operatively with the control group being assessed at the same time intervals. The post-operative rehabilitation involved hospitalization (1st week), home-based (2nd-4th weeks), and outpatient clinic (5th-8th weeks) programs.

Results: The Mixed design MANOVA revealed that group III had significantly higher BBS scores, and lower pain scores and TUG and SC time than groups I and II four and eight weeks post-operatively. In addition, group I had significantly lower pain scores and SC time compared with group II eight weeks post-operatively. Moreover, the BBS scores increased significantly and the pain scores and TUG and SC time decreased significantly eight weeks post-operatively compared with the three other assessments in group I, II and III with the opposite being true four weeks post-operatively.

Interpretation/Conclusion: CRTKA is preferable to PSTKA with UKA being generally superior to TKA, possibly due to the preserved human proprioceptors in the un-excised compartmental articular surface.

Keywords: Dynamic Balance, Functional Performance, Knee Arthroplasty, Pain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2255
5130 An Experimental Study on the Effect of Operating Parameters during the Micro-Electro-Discharge Machining of Ni Based Alloy

Authors: Asma Perveen, M. P. Jahan

Abstract:

Ni alloys have managed to cover wide range of applications such as automotive industries, oil gas industries, and aerospace industries. However, these alloys impose challenges while using conventional machining technologies. On the other hand, Micro-Electro-Discharge machining (micro-EDM) is a non-conventional machining method that uses controlled sparks energy to remove material irrespective of the materials hardness. There has been always a huge interest from the industries for developing optimum methodology and parameters in order to enhance the productivity of micro-EDM in terms of reducing machining time and tool wear for different alloys. Therefore, the aims of this study are to investigate the effects of the micro-EDM process parameters, in order to find their optimal values. The input process parameters include voltage, capacitance, and electrode rotational speed, whereas the output parameters considered are machining time, entrance diameter of hole, overcut, tool wear, and crater size. The surface morphology and element characterization are also investigated with the use of SEM and EDX analysis. The experimental result indicates the reduction of machining time with the increment of discharge energy. Discharge energy also contributes to the enlargement of entrance diameter as well as overcut. In addition, tool wears show reduction with the increase of discharge energy. Moreover, crater size is found to be increased in size along with the increment of discharge energy.

Keywords: Micro EDM, Ni alloy, discharge energy, micro-holes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1318
5129 Performance Comparison of Parallel Sorting Algorithms on the Cluster of Workstations

Authors: Lai Lai Win Kyi, Nay Min Tun

Abstract:

Sorting appears the most attention among all computational tasks over the past years because sorted data is at the heart of many computations. Sorting is of additional importance to parallel computing because of its close relation to the task of routing data among processes, which is an essential part of many parallel algorithms. Many parallel sorting algorithms have been investigated for a variety of parallel computer architectures. In this paper, three parallel sorting algorithms have been implemented and compared in terms of their overall execution time. The algorithms implemented are the odd-even transposition sort, parallel merge sort and parallel rank sort. Cluster of Workstations or Windows Compute Cluster has been used to compare the algorithms implemented. The C# programming language is used to develop the sorting algorithms. The MPI (Message Passing Interface) library has been selected to establish the communication and synchronization between processors. The time complexity for each parallel sorting algorithm will also be mentioned and analyzed.

Keywords: Cluster of Workstations, Parallel sorting algorithms, performance analysis, parallel computing and MPI.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1464
5128 Gluten-Free Cookies Enriched with Blueberry Pomace: Optimization of Baking Process

Authors: Aleksandra Mišan, Bojana Šarić, Nataša Nedeljković, Mladenka Pestorić, Pavle Jovanov, Milica Pojić, Jelena Tomić, Bojana Filipčev, Miroslav Hadnađev, Anamarija Mandić

Abstract:

With the aim of improving nutritional profile and antioxidant capacity of gluten-free cookies, blueberry pomace, by-product of juice production, was processed into a new food ingredient by drying and grinding and used for a gluten-free cookie formulation. Since the quality of a baked product is highly influenced by the baking conditions, the objective of this work was to optimize the baking time and thickness of dough pieces, by applying Response Surface Methodology (RSM) in order to obtain the best technological quality of the cookies. The experiments were carried out according to a Central Composite Design (CCD) by selecting the dough thickness and baking time as independent variables, while hardness, color parameters (L*, a* and b* values), water activity, diameter and short/long ratio were response variables. According to the results of RSM analysis, the baking time of 13.74min and dough thickness of 4.08mm was found to be the optimal for the baking temperature of 170°C. As similar optimal parameters were obtained by previously conducted experiment based on sensory analysis, response surface methodology (RSM) can be considered as a suitable approach to optimize the baking process.

Keywords: Baking process, blueberry pomace, gluten-free cookies, Response Surface Methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2543
5127 Research on the Optimization of the Facility Layout of Efficient Cafeterias for Troops

Authors: Qing Zhang, Jiachen Nie, Yujia Wen, Guanyuan Kou, Peng Yu, Kun Xia, Qin Yang, Li Ding

Abstract:

Background: A facility layout problem (FLP) is an NP-complete (non-deterministic polynomial) problem, for which is hard to obtain an exact optimal solution. FLP has been widely studied in various limited spaces and workflows. For example, cafeterias with many types of equipment for troops cause chaotic processes when dining. Objective: This article tried to optimize the layout of a troops’ cafeteria and to improve the overall efficiency of the dining process. Methods: First, the original cafeteria layout design scheme was analyzed from an ergonomic perspective and two new design schemes were generated. Next, three facility layout models were designed, and further simulation was applied to compare the total time and density of troops between each scheme. Last, an experiment of the dining process with video observation and analysis verified the simulation results. Results: In a simulation, the dining time under the second new layout is shortened by 2.25% and 1.89% (p<0.0001, p=0.0001) compared with the other two layouts, while troops-flow density and interference both greatly reduced in the two new layouts. In the experiment, process completing time and the number of interferences reduced as well, which verified corresponding simulation results. Conclusion: Our two new layout schemes are tested to be optimal by a series of simulation and space experiments. In future research, similar approaches could be applied when taking layout-design algorithm calculation into consideration.

Keywords: Troops’ cafeteria, layout optimization, dining efficiency, AnyLogic simulation, field experiment

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 473
5126 Extraction and Characterization of Oil from Avocado Peels

Authors: Tafere Aga Bullo

Abstract:

The peels of avocados, like other fruit peels, are commonly discarded, not knowing their potential use. This study carried out to extract oils from avocado peels and to characterize the extracted oils with the view to determine their suitability for consumption and other uses. Soxhlet apparatus was used for extraction using n-hexane as a solvent, which is chosen based on the fact that it has a very low value of toxicity and a great extraction rate. The proximate analysis and physicochemical properties of the extracted oil were investigated. The percentage yield of oil extracted from the peel was found to be 40.6%. From this study, the optimum operating conditions for the extraction of oil from avocado peel oil for the particle size of 2.6 mm, solvent type N-hexane and extraction time of 3-5 hr. were considered. A general factorial design was applied to investigate the effect of process variables on oil yield. Maximum oil yield of 40.6% was obtained at an extraction time of 5 hr. The extracted avocado peel oil can be widely used in pharmaceutical and energy production.

Keywords: Avocado fruits, avocado oil, avocado peel oil, characterization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 891
5125 An Optimization Algorithm Based on Dynamic Schema with Dissimilarities and Similarities of Chromosomes

Authors: Radhwan Yousif Sedik Al-Jawadi

Abstract:

Optimization is necessary for finding appropriate solutions to a range of real-life problems. In particular, genetic (or more generally, evolutionary) algorithms have proved very useful in solving many problems for which analytical solutions are not available. In this paper, we present an optimization algorithm called Dynamic Schema with Dissimilarity and Similarity of Chromosomes (DSDSC) which is a variant of the classical genetic algorithm. This approach constructs new chromosomes from a schema and pairs of existing ones by exploring their dissimilarities and similarities. To show the effectiveness of the algorithm, it is tested and compared with the classical GA, on 15 two-dimensional optimization problems taken from literature. We have found that, in most cases, our method is better than the classical genetic algorithm.

Keywords: Genetic algorithm, similarity and dissimilarity, chromosome injection, dynamic schema.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1280
5124 Perceived Constraints on Sport Participation among Young Koreans in Australia

Authors: Jae Won Kang

Abstract:

The purpose of this study was to examine a broader range of sport constraints perceived by young Koreans in Australia who may need to adjust to changing behavioral expectations due to the socio-cultural transitions. Regardless of gender, in terms of quantitative findings, the most important participation constraints within the seven categories were resources, access, interpersonal, affective, religious, socio-cultural, and physical in that order. The most important constraining items were a lack of time, access, information, adaptive skills, and parental and family support in that order. Qualitative research found young Korean’s participation constraints among three categories (time, parental control and interpersonal constraints). It is possible that different ethnic groups would be constrained by different factors; however, this is outside the scope of this study.

Keywords: Constraints, cultural adjustment, Sport, Young Koreans in Australia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2606
5123 Reclaiming Pedestrian Space from Car Dominated Neighborhoods

Authors: Andreas L. Savvides

Abstract:

For a long time as a result of accommodating car traffic, planning ideologies in the past put a low priority on public space, pedestrianism and the role of city space as a meeting place for urban dwellers. In addition, according to authors such as Jan Gehl, market forces and changing architectural perceptions began to shift the focus of planning practice from the integration of public space in various pockets around the contemporary city to individual buildings. Eventually, these buildings have become increasingly more isolated and introverted and have turned their backs to the realm of the public space adjoining them. As a result of this practice, the traditional function of public space as a social forum for city dwellers has in many cases been reduced or even phased out. Author Jane Jacobs published her seminal book “The Death and Life of Great American Cities" more than fifty years ago, but her observations and predictions at the time still ring true today, where she pointed out how the dramatic increase in car traffic and its accommodation by the urban planning ideology that was brought about by the Modern movement has prompted a separation of the uses of the city. At the same time it emphasizes free standing buildings that threaten urban space and city life and result in underutilized and lifeless urban cores. In this discussion context, the aim of this paper is to showcase a reversal of just such a situation in the case of the Dasoupolis neighborhood in Strovolos, Cyprus, where enlightened urban design practice has see the reclamation of pedestrian space in a car dominated area.

Keywords: Urban Design, Public Space, Right to the City, Accessibility, Mobility

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1902
5122 Auto Tuning of PID Controller for MIMO Processes

Authors: M. J. Lengare, R. H. Chile, L. M. Waghmare, Bhavesh Parmar

Abstract:

One of the most basic functions of control engineers is tuning of controllers. There are always several process loops in the plant necessitate of tuning. The auto tuned Proportional Integral Derivative (PID) Controllers are designed for applications where large load changes are expected or the need for extreme accuracy and fast response time exists. The algorithm presented in this paper is used for the tuning PID controller to obtain its parameters with a minimum computing complexity. It requires continuous analysis of variation in few parameters, and let the program to do the plant test and calculate the controller parameters to adjust and optimize the variables for the best performance. The algorithm developed needs less time as compared to a normal step response test for continuous tuning of the PID through gain scheduling.

Keywords: Auto tuning; gain scheduling; MIMO Processes; Optimization; PID controller; Process Control.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2978
5121 Ontology for Semantic Enrichment of Radio Frequency Identification Systems

Authors: Haitham S. Hamza, Mohamed Maher, Shourok Alaa, Aya Khattab, Hadeal Ismail, Kamilia Hosny

Abstract:

Radio Frequency Identification (RFID) has become a key technology in the emerging concept of Internet of Things (IoT). Naturally, business applications would require the deployment of various RFID systems developed by different vendors that use different data formats and structures. This heterogeneity poses a challenge in developing real-life IoT systems with RFID, as integration is becoming very complex and challenging. Semantic integration is a key approach to deal with this challenge. To do so, ontology for RFID systems need to be developed in order to annotated semantically RFID systems, and hence, facilitate their integration. Accordingly, in this paper, we propose ontology for RFID systems. The proposed ontology can be used to semantically enrich RFID systems, and hence, improve their usage and reasoning.

Keywords: IoT, RFID, Semantic, sparql, Ontology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1852
5120 Comparative Dynamic Performance of Load Frequency Control of Nonlinear Interconnected Hydro-Thermal System Using Intelligent Techniques

Authors: Banaja Mohanty, Prakash Kumar Hota

Abstract:

This paper demonstrates dynamic performance evaluation of load frequency control (LFC) with different intelligent techniques. All non-linearities and physical constraints have been considered in simulation studies such as governor dead band (GDB), generation rate constraint (GRC) and boiler dynamics. The conventional integral time absolute error has been considered as objective function. The design problem is formulated as an optimisation problem and particle swarm optimisation (PSO), bacterial foraging optimisation algorithm (BFOA) and differential evolution (DE) are employed to search optimal controller parameters. The superiority of the proposed approach has been shown by comparing the results with published fuzzy logic control (FLC) for the same interconnected power system. The comparison is done using various performance measures like overshoot, undershoot, settling time and standard error criteria of frequency and tie-line power deviation following a step load perturbation (SLP). It is noticed that, the dynamic performance of proposed controller is better than FLC. Further, robustness analysis is carried out by varying the time constants of speed governor, turbine, tie-line power in the range of +40% to -40% to demonstrate the robustness of the proposed DE optimized PID controller.

Keywords: Automatic generation control, governor dead band, generation rate constraint, differential evolution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1048
5119 A Preliminary Study on Effects of Community Structures on Epidemic Spreading and Detection in Complex Networks

Authors: Yi Yu, Gaoxi Xiao

Abstract:

Community structures widely exist in almost all real-life networks. Extensive researches have been carried out on detecting community structures in complex networks. However, many aspects of how community structures may affect the dynamics and properties of complex networks still remain unclear. In this work, we examine the impacts of community structures on the epidemic spreading and detection in complex networks. Extensive simulation results show that community structures may not help decrease the infection size at steady state, yet they could indeed help slow down the infection spreading. Also, networks with strong community structures may expect to have a smaller average infection size when equipped with a number of sparsely deployed monitors.

Keywords: Complex network, epidemic spreading, infection size, infection monitoring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1576
5118 Slovenian Spatial Legislation over Time and Its Issues

Authors: Andreja Benko

Abstract:

Article presents a short overview of the architects’ profession over time with outlined work of the architectural theoreticians. In the continuation is described a former affiliation of Slovenia as well as the spatial planning documents that were in use until the Slovenia joint Yugoslavia (last part in 1919). This legislation from former Austro-Hungarian monarchy was valid almost until 1950 in some parts of Yugoslavia even longer. Upon that will be mentioned some valid Slovenian spatial documents which will be compared with the German legislation. Analyzed will be the number of architect and spatial planners in Slovenia and also their number upon certain region in Slovenia. Based on that will be given also the number from statistical office of Slovenia of the number of buildings between years 2007 and 2012, and described also the collapse of the major construction companies in Slovenia and consequences of that. At the end will be outlined the morality and ethics by spatial interventions and lack of the architectural law in Slovenia as well as the problematic of minimal collaboration between the Ministry of infrastructure and spatial planning with the profession.

Keywords: Architect, history, legislation, Slovenia.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1645
5117 Effect of Heat Treatment on the Portevin-Le Chatelier Effect of Al-2.5%Mg Alloy

Authors: A. Chatterjee, A. Sarkar, N. Gayathri, P. Mukherjee, P. Barat

Abstract:

An experimental study is presented on the effect of microstructural change on the Portevin-Le Chatelier effect behaviour of Al-2.5%Mg alloy. Tensile tests are performed on the as received and heat treated (at 400 ºC for 16 hours) samples for a wide range of strain rates. The serrations observed in the stress-time curve are investigated from statistical analysis point of view. Microstructures of the samples are characterized by optical metallography and X-ray diffraction. It is found that the excess vacancy generated due to heat treatment leads to decrease in the strain rate sensitivity and the increase in the number of stress drop occurrences per unit time during the PLC effect. The microstructural parameters like domain size, dislocation density have no appreciable effect on the PLC effect as far as the statistical behavior of the serrations is considered.

Keywords: Dynamic strain ageing, Heat treatment, Portevin-LeChatelier effect

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2209
5116 Dissipation of Higher Mode using Numerical Integration Algorithm in Dynamic Analysis

Authors: Jin Sup Kim, Woo Young Jung, Minho Kwon

Abstract:

In general dynamic analyses, lower mode response is of interest, however the higher modes of spatially discretized equations generally do not represent the real behavior and not affects to global response much. Some implicit algorithms, therefore, are introduced to filter out the high-frequency modes using intended numerical error. The objective of this study is to introduce the P-method and PC α-method to compare that with dissipation method and Newmark method through the stability analysis and numerical example. PC α-method gives more accuracy than other methods because it based on the α-method inherits the superior properties of the implicit α-method. In finite element analysis, the PC α-method is more useful than other methods because it is the explicit scheme and it achieves the second order accuracy and numerical damping simultaneously.

Keywords: Dynamic, α-Method, P-Method, PC α-Method, Newmark method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3058