Search results for: analytical modelling
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4066

Search results for: analytical modelling

2596 Competitor Analysis to Quantify the Benefits and for Different Use of Transport Infrastructure

Authors: Dimitrios J. Dimitriou, Maria F. Sartzetaki

Abstract:

Different transportation modes have key operational advantages and disadvantages, providing a variety of different transport options to users and passengers. This paper reviews key variables for the competition between air transport and other transport modes. The aim of this paper is to review the competition between air transport and other transport modes, providing results in terms of perceived cost for the users, for destinations high competitiveness for all transport modes. The competitor analysis variables include the cost and time outputs for each transport option, highlighting the level of competitiveness on high demanded Origin-Destination corridors. The case study presents the output of a such analysis for the OD corridor in Greece that connects the Capital city (Athens) with the second largest city (Thessaloniki) and the different transport modes have been considered (air, train, road). Conventional wisdom is to present an easy to handle tool for planners, managers and decision makers towards pricing policy effectiveness and demand attractiveness, appropriate to use for other similar cases.

Keywords: competitor analysis, transport economics, transport generalized cost, quantitative modelling

Procedia PDF Downloads 247
2595 A Modelling Analysis of Monetary Policy Rule

Authors: Wael Bakhit, Salma Bakhit

Abstract:

This paper employs a quarterly time series to determine the timing of structural breaks for interest rates in USA over the last 60 years. The Chow test is used for investigating the non-stationary, where the date of the potential break is assumed to be known. Moreover, an empirical examination of the financial sector was made to check if it is positively related to deviations from an assumed interest rate as given in a standard Taylor rule. The empirical analysis is strengthened by analysing the rule from a historical perspective and a look at the effect of setting the interest rate by the central bank on financial imbalances. The empirical evidence indicates that deviation in monetary policy has a potential causal factor in the build-up of financial imbalances and the subsequent crisis where macro prudential intervention could have beneficial effect. Thus, our findings tend to support the view which states that the probable existence of central banks has been a source of global financial crisis since the past decade.

Keywords: Taylor rule, financial imbalances, central banks, econometrics

Procedia PDF Downloads 389
2594 Modelling of Composite Steel and Concrete Beam with the Lightweight Concrete Slab

Authors: Veronika Přivřelová

Abstract:

Well-designed composite steel and concrete structures highlight the good material properties and lower the deficiencies of steel and concrete, in particular they make use of high tensile strength of steel and high stiffness of concrete. The most common composite steel and concrete structure is a simply supported beam, which concrete slab transferring the slab load to a beam is connected to the steel cross-section. The aim of this paper is to find the most adequate numerical model of a simply supported composite beam with the cross-sectional and material parameters based on the results of a processed parametric study and numerical analysis. The paper also evaluates the suitability of using compact concrete with the lightweight aggregates for composite steel and concrete beams. The most adequate numerical model will be used in the resent future to compare the results of laboratory tests.

Keywords: composite beams, high-performance concrete, high-strength steel, lightweight concrete slab, modeling

Procedia PDF Downloads 408
2593 Edmonton Urban Growth Model as a Support Tool for the City Plan Growth Scenarios Development

Authors: Sinisa J. Vukicevic

Abstract:

Edmonton is currently one of the youngest North American cities and has achieved significant growth over the past 40 years. Strong urban shift requires a new approach to how the city is envisioned, planned, and built. This approach is evidence-based scenario development, and an urban growth model was a key support tool in framing Edmonton development strategies, developing urban policies, and assessing policy implications. The urban growth model has been developed using the Metronamica software platform. The Metronamica land use model evaluated the dynamic of land use change under the influence of key development drivers (population and employment), zoning, land suitability, and land and activity accessibility. The model was designed following the Big City Moves ideas: become greener as we grow, develop a rebuildable city, ignite a community of communities, foster a healing city, and create a city of convergence. The Big City Moves were converted to three development scenarios: ‘Strong Central City’, ‘Node City’, and ‘Corridor City’. Each scenario has a narrative story that expressed scenario’s high level goal, scenario’s approach to residential and commercial activities, to transportation vision, and employment and environmental principles. Land use demand was calculated for each scenario according to specific density targets. Spatial policies were analyzed according to their level of importance within the policy set definition for the specific scenario, but also through the policy measures. The model was calibrated on the way to reproduce known historical land use pattern. For the calibration, we used 2006 and 2011 land use data. The validation is done independently, which means we used the data we did not use for the calibration. The model was validated with 2016 data. In general, the modeling process contain three main phases: ‘from qualitative storyline to quantitative modelling’, ‘model development and model run’, and ‘from quantitative modelling to qualitative storyline’. The model also incorporates five spatial indicators: distance from residential to work, distance from residential to recreation, distance to river valley, urban expansion and habitat fragmentation. The major finding of this research could be looked at from two perspectives: the planning perspective and technology perspective. The planning perspective evaluates the model as a tool for scenario development. Using the model, we explored the land use dynamic that is influenced by a different set of policies. The model enables a direct comparison between the three scenarios. We explored the similarities and differences of scenarios and their quantitative indicators: land use change, population change (and spatial allocation), job allocation, density (population, employment, and dwelling unit), habitat connectivity, proximity to objects of interest, etc. From the technology perspective, the model showed one very important characteristic: the model flexibility. The direction for policy testing changed many times during the consultation process and model flexibility in applying all these changes was highly appreciated. The model satisfied our needs as scenario development and evaluation tool, but also as a communication tool during the consultation process.

Keywords: urban growth model, scenario development, spatial indicators, Metronamica

Procedia PDF Downloads 95
2592 Enhancing Residential Architecture through Generative Design: Balancing Aesthetics, Legal Constraints, and Environmental Considerations

Authors: Milena Nanova, Radul Shishkov, Martin Georgiev, Damyan Damov

Abstract:

This research paper presents an in-depth exploration of the use of generative design in urban residential architecture, with a dual focus on aligning aesthetic values with legal and environmental constraints. The study aims to demonstrate how generative design methodologies can innovate residential building designs that are not only legally compliant and environmentally conscious but also aesthetically compelling. At the core of our research is a specially developed generative design framework tailored for urban residential settings. This framework employs computational algorithms to produce diverse design solutions, meticulously balancing aesthetic appeal with practical considerations. By integrating site-specific features, urban legal restrictions, and environmental factors, our approach generates designs that resonate with the unique character of urban landscapes while adhering to regulatory frameworks. The paper explores how modern digital tools, particularly computational design, and algorithmic modelling, can optimize the early stages of residential building design. By creating a basic parametric model of a residential district, the paper investigates how automated design tools can explore multiple design variants based on predefined parameters (e.g., building cost, dimensions, orientation) and constraints. The paper aims to demonstrate how these tools can rapidly generate and refine architectural solutions that meet the required criteria for quality of life, cost efficiency, and functionality. The study utilizes computational design for database processing and algorithmic modelling within the fields of applied geodesy and architecture. It focuses on optimizing the forms of residential development by adjusting specific parameters and constraints. The results of multiple iterations are analysed, refined, and selected based on their alignment with predefined quality and cost criteria. The findings of this research will contribute to a modern, complex approach to residential area design. The paper demonstrates the potential for integrating BIM models into the design process and their application in virtual 3D Geographic Information Systems (GIS) environments. The study also examines the transformation of BIM models into suitable 3D GIS file formats, such as CityGML, to facilitate the visualization and evaluation of urban planning solutions. In conclusion, our research demonstrates that a generative parametric approach based on real geodesic data and collaborative decision-making could be introduced in the early phases of the design process. This gives the designers powerful tools to explore diverse design possibilities, significantly improving the qualities of the investment during its entire lifecycle.

Keywords: architectural design, residential buildings, urban development, geodesic data, generative design, parametric models, workflow optimization

Procedia PDF Downloads 8
2591 Queueing Modeling of M/G/1 Fault Tolerant System with Threshold Recovery and Imperfect Coverage

Authors: Madhu Jain, Rakesh Kumar Meena

Abstract:

This paper investigates a finite M/G/1 fault tolerant multi-component machining system. The system incorporates the features such as standby support, threshold recovery and imperfect coverage make the study closer to real time systems. The performance prediction of M/G/1 fault tolerant system is carried out using recursive approach by treating remaining service time as a supplementary variable. The numerical results are presented to illustrate the computational tractability of analytical results by taking three different service time distributions viz. exponential, 3-stage Erlang and deterministic. Moreover, the cost function is constructed to determine the optimal choice of system descriptors to upgrading the system.

Keywords: fault tolerant, machine repair, threshold recovery policy, imperfect coverage, supplementary variable technique

Procedia PDF Downloads 292
2590 Predicting Shortage of Hospital Beds during COVID-19 Pandemic in United States

Authors: Saba Ebrahimi, Saeed Ahmadian, Hedie Ashrafi

Abstract:

World-wide spread of coronavirus grows the concern about planning for the excess demand of hospital services in response to COVID-19 pandemic. The surge in the hospital services demand beyond the current capacity leads to shortage of ICU beds and ventilators in some parts of US. In this study, we forecast the required number of hospital beds and possible shortage of beds in US during COVID-19 pandemic to be used in the planning and hospitalization of new cases. In this paper, we used a data on COVID-19 deaths and patients’ hospitalization besides the data on hospital capacities and utilization in US from publicly available sources and national government websites. we used a novel ensemble modelling of deep learning networks, based on stacking different linear and non-linear layers to predict the shortage in hospital beds. The results showed that our proposed approach can predict the excess hospital beds demand very well and this can be helpful in developing strategies and plans to mitigate this gap.

Keywords: COVID-19, deep learning, ensembled models, hospital capacity planning

Procedia PDF Downloads 157
2589 Mathematical Model of Cancer Growth under the Influence of Radiation Therapy

Authors: Beata Jackowska-Zduniak

Abstract:

We formulate and analyze a mathematical model describing dynamics of cancer growth under the influence of radiation therapy. The effect of this type of therapy is considered as an additional equation of discussed model. Numerical simulations show that delay, which is added to ordinary differential equations and represent time needed for transformation from one type of cells to the other one, affects the behavior of the system. The validation and verification of proposed model is based on medical data. Analytical results are illustrated by numerical examples of the model dynamics. The model is able to reconstruct dynamics of treatment of cancer and may be used to determine the most effective treatment regimen based on the study of the behavior of individual treatment protocols.

Keywords: mathematical modeling, numerical simulation, ordinary differential equations, radiation therapy

Procedia PDF Downloads 408
2588 Dynamic Model of Heterogeneous Markets with Imperfect Information for the Optimization of Company's Long-Time Strategy

Authors: Oleg Oborin

Abstract:

This paper is dedicated to the development of the model, which can be used to evaluate the effectiveness of long-term corporate strategies and identify the best strategies. The theoretical model of the relatively homogenous product market (such as iron and steel industry, mobile services or road transport) has been developed. In the model, the market consists of a large number of companies with different internal characteristics and objectives. The companies can perform mergers and acquisitions in order to increase their market share. The model allows the simulation of long-time dynamics of the market (for a period longer than 20 years). Therefore, a large number of simulations on random input data was conducted in the framework of the model. After that, the results of the model were compared with the dynamics of real markets, such as the US steel industry from the beginning of the XX century to the present day, and the market of mobile services in Germany for the period between 1990 and 2015.

Keywords: Economic Modelling, Long-Time Strategy, Mergers and Acquisitions, Simulation

Procedia PDF Downloads 367
2587 A Non-Parametric Based Mapping Algorithm for Use in Audio Fingerprinting

Authors: Analise Borg, Paul Micallef

Abstract:

Over the past few years, the online multimedia collection has grown at a fast pace. Several companies showed interest to study the different ways to organize the amount of audio information without the need of human intervention to generate metadata. In the past few years, many applications have emerged on the market which are capable of identifying a piece of music in a short time. Different audio effects and degradation make it much harder to identify the unknown piece. In this paper, an audio fingerprinting system which makes use of a non-parametric based algorithm is presented. Parametric analysis is also performed using Gaussian Mixture Models (GMMs). The feature extraction methods employed are the Mel Spectrum Coefficients and the MPEG-7 basic descriptors. Bin numbers replaced the extracted feature coefficients during the non-parametric modelling. The results show that non-parametric analysis offer potential results as the ones mentioned in the literature.

Keywords: audio fingerprinting, mapping algorithm, Gaussian Mixture Models, MFCC, MPEG-7

Procedia PDF Downloads 421
2586 Synthesis and Evaluation of Photovoltaic Properties of an Organic Dye for Dye-Sensitized Solar Cells

Authors: M. Hosseinnejad, K. Gharanjig

Abstract:

In the present study, metal free organic dyes were prepared and used as photo-sensitizers in dye-sensitized solar cells. Double rhodanine was utilized as the fundamental electron acceptor group to which electron donor aldehyde with varying substituents was attached to produce new organic dye. This dye was first purified and then characterized by analytical techniques. Spectrophotometric evaluations of the prepared dye in solution and on a nano anatase TiO2 substrate were carried out in order to assess possible changes in the status of the dyes in different environments. The results show that the dye form j-type aggregates on the nano TiO2. Additionally, oxidation potential measurements were also carried out. Finally, dye sensitized solar cell based on synthesized dye was fabricated in order to determine the photovoltaic behavior and conversion efficiency of individual dye.

Keywords: conversion efficiency, dye-sensitized solar cell, photovoltaic behavior, sensitizer

Procedia PDF Downloads 183
2585 On the Use of Analytical Performance Models to Design a High-Performance Active Queue Management Scheme

Authors: Shahram Jamali, Samira Hamed

Abstract:

One of the open issues in Random Early Detection (RED) algorithm is how to set its parameters to reach high performance for the dynamic conditions of the network. Although original RED uses fixed values for its parameters, this paper follows a model-based approach to upgrade performance of the RED algorithm. It models the routers queue behavior by using the Markov model and uses this model to predict future conditions of the queue. This prediction helps the proposed algorithm to make some tunings over RED's parameters and provide efficiency and better performance. Widespread packet level simulations confirm that the proposed algorithm, called Markov-RED, outperforms RED and FARED in terms of queue stability, bottleneck utilization and dropped packets count.

Keywords: active queue management, RED, Markov model, random early detection algorithm

Procedia PDF Downloads 539
2584 Towards Optimising Building Information Modelling and Building Management System in Higher Education Institutions Facility Management: A Review

Authors: Zhuoqun Sun, Francisco Sierra, A. Booth

Abstract:

With BIM rapidly implemented in the design and construction stage of a project, researchers begin to focus on improving the operation and maintenance stage with the aid of BIM. Since the increasing amount of electronic equipment installed in the building, building management system becomes mainstream for controlling a building, especially in higher education institutions that can play an important role in terms of reducing carbon emission and improving energy efficiency. Currently, an approach to integrate BIM and BMS to improve HEIs facility management has not been established yet. Thus, this paper aims to analyse the benefits, issues, and trends of BIM and BMS integration and their application in HEIs. A systematic literature review was carried out on SCOPUS by applying the PRISMA methodology. 73 articles have been chosen based on keywords, abstracts, and the full content of the articles. The benefit and existed issues from the articles are analysed. The review shows the need to develop a tool to improve facility management through BIM BMS integration.

Keywords: BIM, BMS, HEIs, review

Procedia PDF Downloads 162
2583 Effect of Filler Size and Shape on Positive Temperature Coefficient Effect

Authors: Eric Asare, Jamie Evans, Mark Newton, Emiliano Bilotti

Abstract:

Two types of filler shapes (sphere and flakes) and three different sizes are employed to study the size effect on PTC. The composite is prepared using a mini-extruder with high-density polyethylene (HDPE) as the matrix. A computer modelling is used to fit the experimental results. The percolation threshold decreases with decreasing filler size and this was observed for both the spherical particles as well as the flakes. This was caused by the decrease in interparticle distance with decreasing filler size. The 100 µm particles showed a larger PTC intensity compared to the 5 µm particles for the metal coated glass sphere and flake. The small particles have a large surface area and agglomeration and this makes it difficult for the conductive network to e disturbed. Increasing the filler content decreased the PTC intensity and this is due to an increase in the conductive network within the polymer matrix hence more energy is needed to disrupt the network.

Keywords: positive temperature coefficient (PTC) effect, conductive polymer composite (CPC), electrical conductivity

Procedia PDF Downloads 427
2582 Modelling of Factors Affecting Bond Strength of Fibre Reinforced Polymer Externally Bonded to Timber and Concrete

Authors: Abbas Vahedian, Rijun Shrestha, Keith Crews

Abstract:

In recent years, fibre reinforced polymers as applications of strengthening materials have received significant attention by civil engineers and environmentalists because of their excellent characteristics. Currently, these composites have become a mainstream technology for strengthening of infrastructures such as steel, concrete and more recently, timber and masonry structures. However, debonding is identified as the main problem which limit the full utilisation of the FRP material. In this paper, a preliminary analysis of factors affecting bond strength of FRP-to-concrete and timber bonded interface has been conducted. A novel theoretical method through regression analysis has been established to evaluate these factors. Results of proposed model are then assessed with results of pull-out tests and satisfactory comparisons are achieved between measured failure loads (R2 = 0.83, P < 0.0001) and the predicted loads (R2 = 0.78, P < 0.0001).

Keywords: debonding, fibre reinforced polymers (FRP), pull-out test, stepwise regression analysis

Procedia PDF Downloads 248
2581 Application of UAS in Forest Firefighting for Detecting Ignitions and 3D Fuel Volume Estimation

Authors: Artur Krukowski, Emmanouela Vogiatzaki

Abstract:

The article presents results from the AF3 project “Advanced Forest Fire Fighting” focused on Unmanned Aircraft Systems (UAS)-based 3D surveillance and 3D area mapping using high-resolution photogrammetric methods from multispectral imaging, also taking advantage of the 3D scanning techniques from the SCAN4RECO project. We also present a proprietary embedded sensor system used for the detection of fire ignitions in the forest using near-infrared based scanner with weight and form factors allowing it to be easily deployed on standard commercial micro-UAVs, such as DJI Inspire or Mavic. Results from real-life pilot trials in Greece, Spain, and Israel demonstrated added-value in the use of UAS for precise and reliable detection of forest fires, as well as high-resolution 3D aerial modeling for accurate quantification of human resources and equipment required for firefighting.

Keywords: forest wildfires, surveillance, fuel volume estimation, firefighting, ignition detectors, 3D modelling, UAV

Procedia PDF Downloads 142
2580 Non Classical Photonic Nanojets in near Field of Metallic and Negative-Index Scatterers, Purely Electric and Magnetic Nanojets

Authors: Dmytro O. Plutenko, Alexei D. Kiselev, Mikhail V. Vasnetsov

Abstract:

We present the results of our analytical and computational study of Laguerre-Gaussian (LG) beams scattering by spherical homogeneous isotropic particles located on the axis of the beam. We consider different types of scatterers (dielectric, metallic and double negative metamaterials) and different polarizations of the LG beams. A possibility to generate photonic nanojets using metallic and double negative metamaterial Mie scatterers is shown. We have studied the properties of such nonclassical nanojets and discovered new types of the nanojets characterized by zero on-axes magnetic (or electric) field with the electric (or magnetic) field polarized along the z-axis.

Keywords: double negative metamaterial, Laguerre-Gaussian beam, Mie scattering, optical vortices, photonic nanojets

Procedia PDF Downloads 221
2579 A Taxonomy Proposal on Criterion Structure for Evaluating Freight Village Concepts in Early-Stage Design Projects

Authors: Rıza Gürhan Korkut, Metin Çelik, Süleyman Özkaynak

Abstract:

The early-stage design and development projects for the freight village initiatives require a comprehensive analysis of both qualitative and quantitative data. Considering the literature review on structural and operational management requirements, this study proposed an original taxonomy on criterion structure to assess freight village conceptualization. The potential challenges and uncertainties of the developed taxonomy are extended. Besides requirement analysis, this study is also expected to contribute to forthcoming research on benchmarking of freight villages in different regions. The methodology used in this research is a systematic review on several articles as per their modelling approaches, sustainability, entities and decisions made together with the uncertainties and features of their models taken into consideration. The major findings of the study that are the categories for assessing the projects attributes on their environmental, socio-economical, accessibility and location aspects.

Keywords: logistics centers, freight village, operational management, taxonomy

Procedia PDF Downloads 181
2578 [Keynote Talk]: Analysis of One Dimensional Advection Diffusion Model Using Finite Difference Method

Authors: Vijay Kumar Kukreja, Ravneet Kaur

Abstract:

In this paper, one dimensional advection diffusion model is analyzed using finite difference method based on Crank-Nicolson scheme. A practical problem of filter cake washing of chemical engineering is analyzed. The model is converted into dimensionless form. For the grid Ω × ω = [0, 1] × [0, T], the Crank-Nicolson spatial derivative scheme is used in space domain and forward difference scheme is used in time domain. The scheme is found to be unconditionally convergent, stable, first order accurate in time and second order accurate in space domain. For a test problem, numerical results are compared with the analytical ones for different values of parameter.

Keywords: Crank-Nicolson scheme, Lax-Richtmyer theorem, stability, consistency, Peclet number, Greschgorin circle

Procedia PDF Downloads 223
2577 Bayesian Meta-Analysis to Account for Heterogeneity in Studies Relating Life Events to Disease

Authors: Elizabeth Stojanovski

Abstract:

Associations between life events and various forms of cancers have been identified. The purpose of a recent random-effects meta-analysis was to identify studies that examined the association between adverse events associated with changes to financial status including decreased income and breast cancer risk. The same association was studied in four separate studies which displayed traits that were not consistent between studies such as the study design, location and time frame. It was of interest to pool information from various studies to help identify characteristics that differentiated study results. Two random-effects Bayesian meta-analysis models are proposed to combine the reported estimates of the described studies. The proposed models allow major sources of variation to be taken into account, including study level characteristics, between study variance, and within study variance and illustrate the ease with which uncertainty can be incorporated using a hierarchical Bayesian modelling approach.

Keywords: random-effects, meta-analysis, Bayesian, variation

Procedia PDF Downloads 160
2576 Predicting Food Waste and Losses Reduction for Fresh Products in Modified Atmosphere Packaging

Authors: Matar Celine, Gaucel Sebastien, Gontard Nathalie, Guilbert Stephane, Guillard Valerie

Abstract:

To increase the very short shelf life of fresh fruits and vegetable, Modified Atmosphere Packaging (MAP) allows an optimal atmosphere composition to be maintained around the product and thus prevent its decay. This technology relies on the modification of internal packaging atmosphere due to equilibrium between production/consumption of gases by the respiring product and gas permeation through the packaging material. While, to the best of our knowledge, benefit of MAP for fresh fruits and vegetable has been widely demonstrated in the literature, its effect on shelf life increase has never been quantified and formalized in a clear and simple manner leading difficult to anticipate its economic and environmental benefit, notably through the decrease of food losses. Mathematical modelling of mass transfers in the food/packaging system is the basis for a better design and dimensioning of the food packaging system. But up to now, existing models did not permit to estimate food quality nor shelf life gain reached by using MAP. However, shelf life prediction is an indispensable prerequisite for quantifying the effect of MAP on food losses reduction. The objective of this work is to propose an innovative approach to predict shelf life of MAP food product and then to link it to a reduction of food losses and wastes. In this purpose, a ‘Virtual MAP modeling tool’ was developed by coupling a new predictive deterioration model (based on visual surface prediction of deterioration encompassing colour, texture and spoilage development) with models of the literature for respiration and permeation. A major input of this modelling tool is the maximal percentage of deterioration (MAD) which was assessed from dedicated consumers’ studies. Strawberries of the variety Charlotte were selected as the model food for its high perishability, high respiration rate; 50-100 ml CO₂/h/kg produced at 20°C, allowing it to be a good representative of challenging post-harvest storage. A value of 13% was determined as a limit of acceptability for the consumers, permitting to define products’ shelf life. The ‘Virtual MAP modeling tool’ was validated in isothermal conditions (5, 10 and 20°C) and in dynamic temperature conditions mimicking commercial post-harvest storage of strawberries. RMSE values were systematically lower than 3% for respectively, O₂, CO₂ and deterioration profiles as a function of time confirming the goodness of model fitting. For the investigated temperature profile, a shelf life gain of 0.33 days was obtained in MAP compared to the conventional storage situation (no MAP condition). Shelf life gain of more than 1 day could be obtained for optimized post-harvest conditions as numerically investigated. Such shelf life gain permitted to anticipate a significant reduction of food losses at the distribution and consumer steps. This food losses' reduction as a function of shelf life gain has been quantified using a dedicated mathematical equation that has been developed for this purpose.

Keywords: food losses and wastes, modified atmosphere packaging, mathematical modeling, shelf life prediction

Procedia PDF Downloads 183
2575 Eye Diagram for a System of Highly Mode Coupled PMD/PDL Fiber

Authors: Suad M. Abuzariba, Liang Chen, Saeed Hadjifaradji

Abstract:

To evaluate the optical eye diagram due to polarization-mode dispersion (PMD), polarization-dependent loss (PDL), and chromatic dispersion (CD) for a system of highly mode coupled fiber with lumped section at any given optical pulse sequence we present an analytical modle. We found that with considering PDL and the polarization direction correlation between PMD and PDL, a system with highly mode coupled fiber with lumped section can have either higher or lower Q-factor than a highly mode coupled system with same root mean square PDL/PMD values. Also we noticed that a system of two highly mode coupled fibers connected together is not equivalent to a system of highly mode coupled fiber when fluctuation is considered

Keywords: polarization mode dispersion, polarization dependent loss, chromatic dispersion, optical eye diagram

Procedia PDF Downloads 865
2574 Stress Variation of Underground Building Structure during Top-Down Construction

Authors: Soo-yeon Seo, Seol-ki Kim, Su-jin Jung

Abstract:

In the construction of a building, it is necessary to minimize construction period and secure enough work space for stacking of materials during the construction especially in city area. In this manner, various top-down construction methods have been developed and widely used in Korea. This paper investigates the stress variation of underground structure of a building constructed by using SPS (Strut as Permanent System) known as a top-down method in Korea through an analytical approach. Various types of earth pressure distribution related to ground condition were considered in the structural analysis of an example structure at each step of the excavation. From the analysis, the most high member force acting on beams was found when the ground type was medium sandy soil and a stress concentration was found in corner area.

Keywords: construction of building, top-down construction method, earth pressure distribution, member force, stress concentration

Procedia PDF Downloads 306
2573 A Performance Model for Designing Network in Reverse Logistic

Authors: S. Dhib, S. A. Addouche, T. Loukil, A. Elmhamedi

Abstract:

In this paper, a reverse supply chain network is investigated for a decision making. This decision is surrounded by complex flows of returned products, due to the increasing quantity, the type of returned products and the variety of recovery option products (reuse, recycling, and refurbishment). The most important problem in the reverse logistic network (RLN) is to orient returned products to the suitable type of recovery option. However, returned products orientations from collect sources to the recovery disposition have not well considered in performance model. In this study, we propose a performance model for designing a network configuration on reverse logistics. Conceptual and analytical models are developed with taking into account operational, economic and environmental factors on designing network.

Keywords: reverse logistics, network design, performance model, open loop configuration

Procedia PDF Downloads 435
2572 Development and Total Error Concept Validation of Common Analytical Method for Quantification of All Residual Solvents Present in Amino Acids by Gas Chromatography-Head Space

Authors: A. Ramachandra Reddy, V. Murugan, Prema Kumari

Abstract:

Residual solvents in Pharmaceutical samples are monitored using gas chromatography with headspace (GC-HS). Based on current regulatory and compendial requirements, measuring the residual solvents are mandatory for all release testing of active pharmaceutical ingredients (API). Generally, isopropyl alcohol is used as the residual solvent in proline and tryptophan; methanol in cysteine monohydrate hydrochloride, glycine, methionine and serine; ethanol in glycine and lysine monohydrate; acetic acid in methionine. In order to have a single method for determining these residual solvents (isopropyl alcohol, ethanol, methanol and acetic acid) in all these 7 amino acids a sensitive and simple method was developed by using gas chromatography headspace technique with flame ionization detection. During development, no reproducibility, retention time variation and bad peak shape of acetic acid peaks were identified due to the reaction of acetic acid with the stationary phase (cyanopropyl dimethyl polysiloxane phase) of column and dissociation of acetic acid with water (if diluent) while applying temperature gradient. Therefore, dimethyl sulfoxide was used as diluent to avoid these issues. But most the methods published for acetic acid quantification by GC-HS uses derivatisation technique to protect acetic acid. As per compendia, risk-based approach was selected as appropriate to determine the degree and extent of the validation process to assure the fitness of the procedure. Therefore, Total error concept was selected to validate the analytical procedure. An accuracy profile of ±40% was selected for lower level (quantitation limit level) and for other levels ±30% with 95% confidence interval (risk profile 5%). The method was developed using DB-Waxetr column manufactured by Agilent contains 530 µm internal diameter, thickness: 2.0 µm, and length: 30 m. A constant flow of 6.0 mL/min. with constant make up mode of Helium gas was selected as a carrier gas. The present method is simple, rapid, and accurate, which is suitable for rapid analysis of isopropyl alcohol, ethanol, methanol and acetic acid in amino acids. The range of the method for isopropyl alcohol is 50ppm to 200ppm, ethanol is 50ppm to 3000ppm, methanol is 50ppm to 400ppm and acetic acid 100ppm to 400ppm, which covers the specification limits provided in European pharmacopeia. The accuracy profile and risk profile generated as part of validation were found to be satisfactory. Therefore, this method can be used for testing of residual solvents in amino acids drug substances.

Keywords: amino acid, head space, gas chromatography, total error

Procedia PDF Downloads 148
2571 Application Water Quality Modelling In Total Maximum Daily Load (TMDL) Management: A Review

Authors: S. A. Che Osmi, W. M. F. W. Ishak, S. F. Che Osmi

Abstract:

Nowadays the issues of water quality and water pollution have been a major problem across the country. A lot of management attempt to develop their own TMDL database in order to control the river pollution. Over the past decade, the mathematical modeling has been used as the tool for the development of TMDL. This paper presents the application of water quality modeling to develop the total maximum daily load (TMDL) information. To obtain the reliable database of TMDL, the appropriate water quality modeling should choose based on the available data provided. This paper will discuss on the use of several water quality modeling such as QUAL2E, QUAL2K, and EFDC to develop TMDL. The attempts to integrate several modeling are also being discussed in this paper. Based on this paper, the differences in the application of water quality modeling based on their properties such as one, two or three dimensional are showing their ability to develop the modeling of TMDL database.

Keywords: TMDL, water quality modeling, QUAL2E, EFDC

Procedia PDF Downloads 439
2570 Micromechanical Determination of the Mechanical Properties of Carbon Nanotube-Polymer Composites with a Functionally Graded Interphase

Authors: Vahidullah Tac, Ercan Gurses

Abstract:

There have been numerous attempts at modelling carbon nanotube – polymer composites micromechanically in recent years, albeit to limited success. One of the major setbacks of the models used in the scientific community is the lack of regard to the different phases present in a nanocomposite. We employ a multi-phase micromechanical model that allows functionally grading certain phases to determine the mechanical properties of nanocomposites. The model has four distinct phases; the nanotube, the interface between the nanotube and polymer, the interphase, and the bulk matrix. Among the four phases, the interphase is functionally graded such that its moduli gradually decrease from some predetermined values to those of the bulk polymer. We find that the interface plays little role in stiffening/softening of the polymer per se , but instead, it is responsible for load transfer between the polymer and the carbon nanotube. Our results indicate that the carbon nanotube, as well as the interphase, have significant roles in stiffening the composite. The results are then compared to experimental findings and the interphase is tuned accordingly.

Keywords: carbon nanotube, composite, interphase, micromechanical modeling

Procedia PDF Downloads 166
2569 Fluorescence-Based Biosensor for Dopamine Detection Using Quantum Dots

Authors: Sylwia Krawiec, Joanna Cabaj, Karol Malecha

Abstract:

Nowadays, progress in the field of the analytical methods is of great interest for reliable biological research and medical diagnostics. Classical techniques of chemical analysis, despite many advantages, do not permit to obtain immediate results or automatization of measurements. Chemical sensors have displaced the conventional analytical methods - sensors combine precision, sensitivity, fast response and the possibility of continuous-monitoring. Biosensor is a chemical sensor, which except of conventer also possess a biologically active material, which is the basis for the detection of specific chemicals in the sample. Each biosensor device mainly consists of two elements: a sensitive element, where is recognition of receptor-analyte, and a transducer element which receives the signal and converts it into a measurable signal. Through these two elements biosensors can be divided in two categories: due to the recognition element (e.g immunosensor) and due to the transducer (e.g optical sensor). Working of optical sensor is based on measurements of quantitative changes of parameters characterizing light radiation. The most often analyzed parameters include: amplitude (intensity), frequency or polarization. Changes in the optical properties one of the compound which reacts with biological material coated on the sensor is analyzed by a direct method, in an indirect method indicators are used, which changes the optical properties due to the transformation of the testing species. The most commonly used dyes in this method are: small molecules with an aromatic ring, like rhodamine, fluorescent proteins, for example green fluorescent protein (GFP), or nanoparticles such as quantum dots (QDs). Quantum dots have, in comparison with organic dyes, much better photoluminescent properties, better bioavailability and chemical inertness. These are semiconductor nanocrystals size of 2-10 nm. This very limited number of atoms and the ‘nano’-size gives QDs these highly fluorescent properties. Rapid and sensitive detection of dopamine is extremely important in modern medicine. Dopamine is very important neurotransmitter, which mainly occurs in the brain and central nervous system of mammals. Dopamine is responsible for the transmission information of moving through the nervous system and plays an important role in processes of learning or memory. Detection of dopamine is significant for diseases associated with the central nervous system such as Parkinson or schizophrenia. In developed optical biosensor for detection of dopamine, are used graphene quantum dots (GQDs). In such sensor dopamine molecules coats the GQD surface - in result occurs quenching of fluorescence due to Resonance Energy Transfer (FRET). Changes in fluorescence correspond to specific concentrations of the neurotransmitter in tested sample, so it is possible to accurately determine the concentration of dopamine in the sample.

Keywords: biosensor, dopamine, fluorescence, quantum dots

Procedia PDF Downloads 364
2568 Passive Neutralization of Acid Mine Drainage Using Locally Produced Limestone

Authors: Reneiloe Seodigeng, Malwandla Hanabe, Haleden Chiririwa, Hilary Rutto, Tumisang Seodigeng

Abstract:

Neutralisation of acid-mine drainage (AMD) using limestone is cost effective, and good results can be obtained. However, this process has its limitations; it cannot be used for highly acidic water which consists of Fe(III). When Fe(III) reacts with CaCO3, it results in armoring. Armoring slows the reaction, and additional alkalinity can no longer be generated. Limestone is easily accessible, so this problem can be easily dealt with. Experiments were carried out to evaluate the effect of PVC pipe length on ferric and ferrous ions. It was found that the shorter the pipe length the more these dissolved metals precipitate. The effect of the pipe length on the hydrogen ions was also studied, and it was found that these two have an inverse relationship. Experimental data were further compared with the model prediction data to see if they behave in a similar fashion. The model was able to predict the behaviour of 1.5m and 2 m pipes in ferric and ferrous ion precipitation.

Keywords: acid mine drainage, neutralisation, limestone, mathematical modelling

Procedia PDF Downloads 364
2567 Hydrodynamics of Shear Layers at River Confluences by Formation of Secondary Circulation

Authors: Ali Aghazadegan, Ali Shokri, Julia Mullarney

Abstract:

River confluences are areas where there is a lot of mixing, which is often caused by the formation of shear layers and helical motions. The hydrodynamics of secondary circulation at river confluences with low flow discharge ratios and a 90° junction angle are investigated in this study. The analysis is based on Delft 3D modelling, which includes a three-dimensional time-averaged velocity field, turbulence, and water surface levels that have been validated using laboratory data. Confluence structure was characterized by shear layer, secondary circulation, and mixing at the junction and post confluence channel. This study analysis formation of the shear layer by generation of secondary circulations in variation discharge ratios. The values of streamwise, cross-wise, and vertical components are used to estimate the secondary circulation observed within and downstream of the tributary mouth. These variables are estimated for three horizontal planes at Z = [0.14; 0.07; 0.02] and for eight cross-sections at X = [-0.1; 0.00; 0.10; 0.2; 0.30; 0.4; 0.5; 0.6] within a range of 0.05 Y 0.30.

Keywords: river confluence, shear layer, secondary circulation, hydrodynamics

Procedia PDF Downloads 96