Search results for: local interpolation method
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9163

Search results for: local interpolation method

4363 Factors Affecting Access to Education: The Experiences of Parents of Children Who Are Deaf or Hard of Hearing

Authors: Hanh Thi My Nguyen

Abstract:

The purpose of this research is to examine the experiences of parents of children who are deaf or hard of hearing in supporting their children to access education in Vietnam. Parents play a crucial role in supporting their children to gain full access to education. It was widely reported that parents of those children confronted a range of problems to support their children to access education. To author’s best knowledge, there has been a lack of research exploring the experiences of those parents in literature. This research examines factors affecting those parents in supporting their children to access education. To conduct the study, qualitative approach using a phenomenological research design was chosen to explore the central phenomena. Ten parents of children who were diagnosed as deaf or hard of hearing and aged 6-9 years were recruited through the support of the Association of Parents of Children with Hearing Impairment. Participants were interviewed via telephone with a mix of open and closed questions; interviews were audio recorded, transcribed and thematically analysed. The research results show that there are nine main factors that affected the parents in this study in making decisions relating to education for their children including: lack of information resources, perspectives of those parents on communication approaches, the families’ financial capacity, the psychological impact on the participants after their children’ diagnosis, the attitude of family members, attitude of school administrators, lack of local schools and qualified teachers, and current education system for the deaf in Vietnam. Apart from those factors, the lack of knowledge of the participants’ partners about deaf education and the partners’ employment are barriers to educational access and successful communication with their child.

Keywords: Access to education, deaf, hard of hearing, parents experience.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1354
4362 RRNS-Convolutional Concatenated Code for OFDM based Wireless Communication with Direct Analog-to-Residue Converter

Authors: Shahana T. K., Babita R. Jose, K. Poulose Jacob, Sreela Sasi

Abstract:

The modern telecommunication industry demands higher capacity networks with high data rate. Orthogonal frequency division multiplexing (OFDM) is a promising technique for high data rate wireless communications at reasonable complexity in wireless channels. OFDM has been adopted for many types of wireless systems like wireless local area networks such as IEEE 802.11a, and digital audio/video broadcasting (DAB/DVB). The proposed research focuses on a concatenated coding scheme that improve the performance of OFDM based wireless communications. It uses a Redundant Residue Number System (RRNS) code as the outer code and a convolutional code as the inner code. Here, a direct conversion of analog signal to residue domain is done to reduce the conversion complexity using sigma-delta based parallel analog-to-residue converter. The bit error rate (BER) performances of the proposed system under different channel conditions are investigated. These include the effect of additive white Gaussian noise (AWGN), multipath delay spread, peak power clipping and frame start synchronization error. The simulation results show that the proposed RRNS-Convolutional concatenated coding (RCCC) scheme provides significant improvement in the system performance by exploiting the inherent properties of RRNS.

Keywords: Analog-to-residue converter, Concatenated codes, OFDM, Redundant Residue Number System, Sigma-delta modulator, Wireless communication

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1922
4361 Digital Transformation of Lean Production: Systematic Approach for the Determination of Digitally Pervasive Value Chains

Authors: Peter Burggräf, Matthias Dannapfel, Hanno Voet, Patrick-Benjamin Bök, Jérôme Uelpenich, Julian Hoppe

Abstract:

The increasing digitalization of value chains can help companies to handle rising complexity in their processes and thereby reduce the steadily increasing planning and control effort in order to raise performance limits. Due to technological advances, companies face the challenge of smart value chains for the purpose of improvements in productivity, handling the increasing time and cost pressure and the need of individualized production. Therefore, companies need to ensure quick and flexible decisions to create self-optimizing processes and, consequently, to make their production more efficient. Lean production, as the most commonly used paradigm for complexity reduction, reaches its limits when it comes to variant flexible production and constantly changing market and environmental conditions. To lift performance limits, which are inbuilt in current value chains, new methods and tools must be applied. Digitalization provides the potential to derive these new methods and tools. However, companies lack the experience to harmonize different digital technologies. There is no practicable framework, which instructs the transformation of current value chains into digital pervasive value chains. Current research shows that a connection between lean production and digitalization exists. This link is based on factors such as people, technology and organization. In this paper, the introduced method for the determination of digitally pervasive value chains takes the factors people, technology and organization into account and extends existing approaches by a new dimension. It is the first systematic approach for the digital transformation of lean production and consists of four steps: The first step of ‘target definition’ describes the target situation and defines the depth of the analysis with regards to the inspection area and the level of detail. The second step of ‘analysis of the value chain’ verifies the lean-ability of processes and lies in a special focus on the integration capacity of digital technologies in order to raise the limits of lean production. Furthermore, the ‘digital evaluation process’ ensures the usefulness of digital adaptions regarding their practicability and their integrability into the existing production system. Finally, the method defines actions to be performed based on the evaluation process and in accordance with the target situation. As a result, the validation and optimization of the proposed method in a German company from the electronics industry shows that the digital transformation of current value chains based on lean production achieves a raise of their inbuilt performance limits.

Keywords: Digitalization, digital transformation, lean production, Industrie 4.0, value chain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2005
4360 Analysis of the Supramolecular Complex of Kinetin with Glycyrrhizic Acid Using the Chromatography Mass Spectrometry Method

Authors: B. Y. Matmuratov, S. D. Madrakhimova. R. S. Esanov. A. D. Matchanov

Abstract:

Supramolecular complexes of glycyrrhizic acid with kinetin in various molar ratios were obtained, physico-chemical parameters and spectral properties of the resulting complexes were studied (UV, IR, mass spectrometry.

Keywords: Monoammonium salt of glycyrrhizic acid, glycyrrhizic acid, supramolecular complex, isomolar series, IR spectroscopy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 314
4359 An Investigation to Effective Parameters on the Damage of Dual Phase Steels by Acoustic Emission Using Energy Ratio

Authors: A. Fallahi, R. Khamedi

Abstract:

Dual phase steels (DPS)s have a microstructure consisting of a hard second phase called Martensite in the soft Ferrite matrix. In recent years, there has been interest in dual-phase steels, because the application of these materials has made significant usage; particularly in the automotive sector Composite microstructure of (DPS)s exhibit interesting characteristic mechanical properties such as continuous yielding, low yield stress to tensile strength ratios(YS/UTS), and relatively high formability; which offer advantages compared with conventional high strength low alloy steels(HSLAS). The research dealt with the characterization of damage in (DPS)s. In this study by review the mechanisms of failure due to volume fraction of martensite second phase; a new method is introduced to identifying the mechanisms of failure in the various phases of these types of steels. In this method the acoustic emission (AE) technique was used to detect damage progression. These failure mechanisms consist of Ferrite-Martensite interface decohesion and/or martensite phase fracture. For this aim, dual phase steels with different volume fraction of martensite second phase has provided by various heat treatment methods on a low carbon steel (0.1% C), and then AE monitoring is used during tensile test of these DPSs. From AE measurements and an energy ratio curve elaborated from the value of AE energy (it was obtained as the ratio between the strain energy to the acoustic energy), that allows detecting important events, corresponding to the sudden drops. These AE signals events associated with various failure mechanisms are classified for ferrite and (DPS)s with various amount of Vm and different martensite morphology. It is found that AE energy increase with increasing Vm. This increasing of AE energy is because of more contribution of martensite fracture in the failure of samples with higher Vm. Final results show a good relationship between the AE signals and the mechanisms of failure.

Keywords: Dual phase steel (DPS)s, Failure mechanisms, Acoustic Emission, Fracture strain energy to the acoustic energy.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1866
4358 A Comparison on Healing Effects of an Ayurvedic Preparation and Silver Sulfadiazine on Burn Wounds in Albino Rats

Authors: S. S. Pathak, M. A. Borkar, S. S. Patel

Abstract:

To compare Healing Effects of an Ayurvedic Preparation and Silver Sulfadiazine on burn wounds in Albino Rats. Methods: Albino rats– 30 male / female rats weighing between 150-200 g were used in the study. They were individually housed and maintained on normal diet and water ad libitum. Partial thickness burn wounds were inflicted, on overnight-starved animals under pentobarbitone (30mg/kg, i.p.) anaesthesia, by pouring hot molten wax at 80oC into a plastic cylinder of 300 mm2 circular openings placed on the shaven back of the animal. Apart from the drugs under investigation no local/ systemic chemotherapeutic cover will be provided to animals. All the animals were assessed for the percentage of wound contraction, signs of infection, scab formation and histopathological examination. Results: Percentage of wound healing was significantly better in the test ointment group compared to the standard. Signs of infection were observed in more animals in the test ointment group compared to the standard. Scab formation also took place earlier in the test ointment group compared to standard. Epithelial regeneration and healing profile was better in the test ointment compared to the standard. Moreover the test ointment group did not show any raised margins in the wound or blackish discoloration as was observed in silver sulfadiazine group. Conclusion: The burn wound healing effect of the ayurvedic ointment under study is better in comparison to standard therapy of silver sulfadiazine. The problem of infection encountered with the test ointment can be overcome by changing the concentrations and proportions of the ingredients in the test ointment which constitutes the further plan of the study.

Keywords: Ayurvedic test ointment, burn wounds, Silver sulfadiazine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2386
4357 Biaxial Testing of Fabrics - A Comparison of Various Testing Methodologies

Authors: O.B. Ozipek, E. Bozdag, E. Sunbuloglu, A. Abdullahoglu, E. Belen, E. Celikkanat

Abstract:

In textile industry, besides the conventional textile products, technical textile goods, that have been brought external functional properties into, are being developed for technical textile industry. Especially these products produced with weaving technology are widely preferred in areas such as sports, geology, medical, automotive, construction and marine sectors. These textile products are exposed to various stresses and large deformations under typical conditions of use. At this point, sufficient and reliable data could not be obtained with uniaxial tensile tests for determination of the mechanical properties of such products due to mainly biaxial stress state. Therefore, the most preferred method is a biaxial tensile test method and analysis. These tests and analysis is applied to fabrics with different functional features in order to establish the textile material with several characteristics and mechanical properties of the product. Planar biaxial tensile test, cylindrical inflation and bulge tests are generally required to apply for textile products that are used in automotive, sailing and sports areas and construction industry to minimize accidents as long as their service life. Airbags, seat belts and car tires in the automotive sector are also subject to the same biaxial stress states, and can be characterized by same types of experiments. In this study, in accordance with the research literature related to the various biaxial test methods are compared. Results with discussions are elaborated mainly focusing on the design of a biaxial test apparatus to obtain applicable experimental data for developing a finite element model. Sample experimental results on a prototype system are expressed.

Keywords: Biaxial Stress, Bulge Test, Cylindrical Inflation, Fabric Testing, Planar Tension.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4124
4356 New Exact Solutions for the (3+1)-Dimensional Breaking Soliton Equation

Authors: Mohammad Taghi Darvishi, Maliheh Najafi, Mohammad Najafi

Abstract:

In this work, we obtain some analytic solutions for the (3+1)-dimensional breaking soliton after obtaining its Hirota-s bilinear form. Our calculations show that, three-wave method is very easy and straightforward to solve nonlinear partial differential equations.

Keywords: (3+1)-dimensional breaking soliton equation, Hirota'sbilinear form.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1648
4355 Optimization by Ant Colony Hybryde for the Bin-Packing Problem

Authors: Ben Mohamed Ahemed Mohamed, Yassine Adnan

Abstract:

The problem of bin-packing in two dimensions (2BP) consists in placing a given set of rectangular items in a minimum number of rectangular and identical containers, called bins. This article treats the case of objects with a free orientation of 90Ôùª. We propose an approach of resolution combining optimization by colony of ants (ACO) and the heuristic method IMA to resolve this NP-Hard problem.

Keywords: Ant colony algorithm, bin-packing problem, heuristics methods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1814
4354 Blueprinting of a Normalized Supply Chain Processes: Results in Implementing Normalized Software Systems

Authors: Bassam Istanbouli

Abstract:

With the technology evolving every day and with the increase in global competition, industries are always under the pressure to be the best. They need to provide good quality products at competitive prices, when and how the customer wants them.  In order to achieve this level of service, products and their respective supply chain processes need to be flexible and evolvable; otherwise changes will be extremely expensive, slow and with many combinatorial effects. Those combinatorial effects impact the whole organizational structure, from a management, financial, documentation, logistics and specially the information system Enterprise Requirement Planning (ERP) perspective. By applying the normalized system concept/theory to segments of the supply chain, we believe minimal effects, especially at the time of launching an organization global software project. The purpose of this paper is to point out that if an organization wants to develop a software from scratch or implement an existing ERP software for their business needs and if their business processes are normalized and modular then most probably this will yield to a normalized and modular software system that can be easily modified when the business evolves. Another important goal of this paper is to increase the awareness regarding the design of the business processes in a software implementation project. If the blueprints created are normalized then the software developers and configurators will use those modular blueprints to map them into modular software. This paper only prepares the ground for further studies;  the above concept will be supported by going through the steps of developing, configuring and/or implementing a software system for an organization by using two methods: The Software Development Lifecycle method (SDLC) and the Accelerated SAP implementation method (ASAP). Both methods start with the customer requirements, then blue printing of its business processes and finally mapping those processes into a software system.  Since those requirements and processes are the starting point of the implementation process, then normalizing those processes will end up in a normalizing software.

Keywords: Blueprint, ERP, SDLC, Modular.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 368
4353 Biodiesel Production from Waste Chicken Fatbased Sources

Authors: Kambiz Tahvildari A., Narges Davari B., Mohammadreza Allahgholi Ghasri C, MasoomehBehrourzinavid D

Abstract:

Chicken fat was employed as a feedstock for producing of biodiesel by trasesterification reaction with methanol and alkali catalyst (KOH). In this study chicken fat biodiesel with 1.4% free fatty acid, methanol and various amount of potassium hydroxide for 2 hour were studied. The progression of reaction and conversion of triglycerides to methyl ester were checked by IR spectrum method.

Keywords: Alkali catalyst, biodiesel, chicken fat, transesterification reaction

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2813
4352 Applying Wavelet Transform to Ferroresonance Detection and Protection

Authors: Chun-Wei Huang, Jyh-Cherng Gu, Ming-Ta Yang

Abstract:

Non-synchronous breakage or line failure in power systems with light or no loads can lead to core saturation in transformers or potential transformers. This can cause component and capacitance matching resulting in the formation of resonant circuits, which trigger ferroresonance. This study employed a wavelet transform for the detection of ferroresonance. Simulation results demonstrate the efficacy of the proposed method.

Keywords: Ferroresonance, Wavelet Transform, Intelligent Electronic Device, Transformer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1635
4351 Machine Learning Techniques in Bank Credit Analysis

Authors: Fernanda M. Assef, Maria Teresinha A. Steiner

Abstract:

The aim of this paper is to compare and discuss better classifier algorithm options for credit risk assessment by applying different Machine Learning techniques. Using records from a Brazilian financial institution, this study uses a database of 5,432 companies that are clients of the bank, where 2,600 clients are classified as non-defaulters, 1,551 are classified as defaulters and 1,281 are temporarily defaulters, meaning that the clients are overdue on their payments for up 180 days. For each case, a total of 15 attributes was considered for a one-against-all assessment using four different techniques: Artificial Neural Networks Multilayer Perceptron (ANN-MLP), Artificial Neural Networks Radial Basis Functions (ANN-RBF), Logistic Regression (LR) and finally Support Vector Machines (SVM). For each method, different parameters were analyzed in order to obtain different results when the best of each technique was compared. Initially the data were coded in thermometer code (numerical attributes) or dummy coding (for nominal attributes). The methods were then evaluated for each parameter and the best result of each technique was compared in terms of accuracy, false positives, false negatives, true positives and true negatives. This comparison showed that the best method, in terms of accuracy, was ANN-RBF (79.20% for non-defaulter classification, 97.74% for defaulters and 75.37% for the temporarily defaulter classification). However, the best accuracy does not always represent the best technique. For instance, on the classification of temporarily defaulters, this technique, in terms of false positives, was surpassed by SVM, which had the lowest rate (0.07%) of false positive classifications. All these intrinsic details are discussed considering the results found, and an overview of what was presented is shown in the conclusion of this study.

Keywords: Artificial Neural Networks, ANNs, classifier algorithms, credit risk assessment, logistic regression, machine learning, support vector machines.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1236
4350 Optimization of a Bioremediation Strategy for an Urban Stream of Matanza-Riachuelo Basin

Authors: María D. Groppa, Andrea Trentini, Myriam Zawoznik, Roxana Bigi, Carlos Nadra, Patricia L. Marconi

Abstract:

In the present work, a remediation bioprocess based on the use of a local isolate of the microalgae Chlorella vulgaris immobilized in alginate beads is proposed. This process was shown to be effective for the reduction of several chemical and microbial contaminants present in Cildáñez stream, a water course that is part of the Matanza-Riachuelo Basin (Buenos Aires, Argentina). The bioprocess, involving the culture of the microalga in autotrophic conditions in a stirred-tank bioreactor supplied with a marine propeller for 6 days, allowed a significant reduction of Escherichia coli and total coliform numbers (over 95%), as well as of ammoniacal nitrogen (96%), nitrates (86%), nitrites (98%), and total phosphorus (53%) contents. Pb content was also significantly diminished after the bioprocess (95%). Standardized cytotoxicity tests using Allium cepa seeds and Cildáñez water pre- and post-remediation were also performed. Germination rate and mitotic index of onion seeds imbibed in Cildáñez water subjected to the bioprocess was similar to that observed in seeds imbibed in distilled water and significantly superior to that registered when untreated Cildáñez water was used for imbibition. Our results demonstrate the potential of this simple and cost-effective technology to remove urban-water contaminants, offering as an additional advantage the possibility of an easy biomass recovery, which may become a source of alternative energy.

Keywords: Bioreactor, bioremediation, Chlorella vulgaris, Matanza-Riachuelo basin, microalgae.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 806
4349 Lighting Consumption Analysis in Retail Industry: Comparative Study

Authors: Elena C. Tamaş, Grațiela M. Țârlea, Gianni Flamaropol, Dragoș Hera

Abstract:

This article is referring to a comparative study regarding the electrical energy consumption for lighting on diverse types of big sizes commercial buildings built in Romania after 2007, having 3, 4, 5 versus 8, 9, 10 operational years. Some buildings have installed building management systems (BMS) to monitor also the lighting performances starting with the opening days till the present days but some have chosen only local meters to implement. Firstly, for each analyzed building, the total required energy power and the energy power consumption for lighting were calculated depending on the lamps number, the unit power and the average daily running hours. All objects and installations were chosen depending on the destination/location of the lighting (exterior parking or access, interior or covering parking, building interior and building perimeter). Secondly, to all lighting objects and installations, mechanical counters were installed, and to the ones linked to BMS there were installed the digital meters as well for a better monitoring. Some efficient solutions are proposed to improve the power consumption, for example the 1/3 lighting functioning for the covered and exterior parking lighting to those buildings if can be done. This type of lighting share can be performed on each level, especially on the night shifts. Another example is to use the dimmers to reduce the light level, depending on the executed work in the respective area, and a 30% power energy saving can be achieved. Using the right BMS to monitor, the energy consumption depending on the average operational daily hours and changing the non-performant unit lights with the ones having LED technology or economical ones might increase significantly the energy performances and reduce the energy consumption of the buildings.

Keywords: Lighting consumption, commercial buildings, maintenance, energy performances.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 953
4348 Stabilization and Observation of Attitude Control Systems for Micro Satellites

Authors: A. Elakkary, A. Echchatbi, N. Elalami

Abstract:

In this paper, we are interested in attitude control of a satellite, which using wheels of reaction, by state feedback. First, we develop a method allowing us to put the control and its integral in the state-feedback form. Then, by using the theorem of Gronwall- Bellman, we put the sufficient conditions so that the nonlinear system modeling the satellite is stabilisable and observed by state feedback.

Keywords: Satellite, attitude control, state feedback, attitude stabilization.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1371
4347 Investigation of Dimethyl Ether Solubility in Liquid Hexadecane by UNIFAC Method

Authors: F. Raouf, M. Taghizadeh

Abstract:

It is shown that a modified UNIFAC model can be applied to predict solubility of hydrocarbon gases and vapors in hydrocarbon solvents. Very good agreement with experimental data has been achieved. In this work we try to find best way for predicting dimethyl ether solubility in liquid paraffin by using group contribution theory.

Keywords: UNIFAC, Henry's law, Group contribution theory, Solubility.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2285
4346 Modified Scaling-Free CORDIC Based Pipelined Parallel MDC FFT and IFFT Architecture for Radix 2^2 Algorithm

Authors: C. Paramasivam, K. B. Jayanthi

Abstract:

An innovative approach to develop modified scaling free CORDIC based two parallel pipelined Multipath Delay Commutator (MDC) FFT and IFFT architectures for radix 22 FFT algorithm is presented. Multipliers and adders are the most important data paths in FFT and IFFT architectures. Multipliers occupy high area and consume more power. In order to optimize the area and power overhead, modified scaling-free CORDIC based complex multiplier is utilized in the proposed design. In general twiddle factor values are stored in RAM block. In the proposed work, modified scaling-free CORDIC based twiddle factor generator unit is used to generate the twiddle factor and efficient switching units are used. In addition to this, four point FFT operations are performed without complex multiplication which helps to reduce area and power in the last two stages of the pipelined architectures. The design proposed in this paper is based on multipath delay commutator method. The proposed design can be extended to any radix 2n based FFT/IFFT algorithm to improve the throughput. The work is synthesized using Synopsys design Compiler using TSMC 90-nm library. The proposed method proves to be better compared to the reference design in terms of area, throughput and power consumption. The comparative analysis of the proposed design with Xilinx FPGA platform is also discussed in the paper.

Keywords: Coordinate Rotational Digital Computer(CORDIC), Complex multiplier, Fast Fourier transform (FFT), Inverse fast Fourier transform (IFFT), Multipath delay Commutator (MDC), modified scaling free CORDIC, complex multiplier, pipelining, parallel processing, radix-2^2.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1787
4345 LINUX Cluster Possibilities in 3-D PHOTO Quality Imaging and Animation

Authors: Arjun Jain, Himanshu Agrawal, Nalini Vasudevan

Abstract:

In this paper we present the PC cluster built at R.V. College of Engineering (with great help from the Department of Computer Science and Electrical Engineering). The structure of the cluster is described and the performance is evaluated by rendering of complex 3D Persistence of Vision (POV) images by the Ray-Tracing algorithm. Here, we propose an unexampled method to render such images, distributedly on a low cost scalable.

Keywords: PC cluster, parallel computations, ray tracing, persistence of vision, rendering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1531
4344 Evaluation of Seismic Damage for Gisha Bridge in Tehran by HAZUS Methodology

Authors: Langroudi B., Salehi E., Keshani S., Baghersad M.

Abstract:

Transportation is of great importance in the current life of human beings. The transportation system plays many roles, from economical development to after-catastrophe aids such as rescue operation in the first hours and days after an earthquake. In after earthquakes response phase, transportation system acts as a basis for ground operations including rescue and relief operation, food providing for victims and etc. It is obvious that partial or complete obstruction of this system results in the stop of these operations. Bridges are one of the most important elements of transportation network. Failure of a bridge, in the most optimistic case, cuts the relation between two regions and in more developed countries, cuts the relation of numerous regions. In this paper, to evaluate the vulnerability and estimate the damage level of Tehran bridges, HAZUS method, developed by Federal Emergency Management Agency (FEMA) with the aid of National Institute of Building Science (NIBS), is used for the first time in Iran. In this method, to evaluate the collapse probability, fragility curves are used. Iran is located on seismic belt and thus, it is vulnerable to earthquakes. Thus, the study of the probability of bridge collapses, as an important part of transportation system, during earthquakes is of great importance. The purpose of this study is to provide fragility curves for Gisha Bridge, one of the longest steel bridges in Tehran, as an important lifeline element. Besides, the damage probability for this bridge during a specific earthquake, introduced as scenario earthquakes, is calculated. The fragility curves show that for the considered scenario, the probability of occurrence of complete collapse for the bridge is 8.6%.

Keywords: Bridge, Damage evaluation, Fragility curve, Lifelines, Seismic vulnerability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2117
4343 An Iterative Updating Method for Damped Gyroscopic Systems

Authors: Yongxin Yuan

Abstract:

The problem of updating damped gyroscopic systems using measured modal data can be mathematically formulated as following two problems. Problem I: Given Ma ∈ Rn×n, Λ = diag{λ1, ··· , λp} ∈ Cp×p, X = [x1, ··· , xp] ∈ Cn×p, where p<n and both Λ and X are closed under complex conjugation in the sense that λ2j = λ¯2j−1 ∈ C, x2j = ¯x2j−1 ∈ Cn for j = 1, ··· , l, and λk ∈ R, xk ∈ Rn for k = 2l + 1, ··· , p, find real-valued symmetric matrices D,K and a real-valued skew-symmetric matrix G (that is, GT = −G) such that MaXΛ2 + (D + G)XΛ + KX = 0. Problem II: Given real-valued symmetric matrices Da, Ka ∈ Rn×n and a real-valued skew-symmetric matrix Ga, find (D, ˆ G, ˆ Kˆ ) ∈ SE such that Dˆ −Da2+Gˆ−Ga2+Kˆ −Ka2 = min(D,G,K)∈SE (D− Da2 + G − Ga2 + K − Ka2), where SE is the solution set of Problem I and · is the Frobenius norm. This paper presents an iterative algorithm to solve Problem I and Problem II. By using the proposed iterative method, a solution of Problem I can be obtained within finite iteration steps in the absence of roundoff errors, and the minimum Frobenius norm solution of Problem I can be obtained by choosing a special kind of initial matrices. Moreover, the optimal approximation solution (D, ˆ G, ˆ Kˆ ) of Problem II can be obtained by finding the minimum Frobenius norm solution of a changed Problem I. A numerical example shows that the introduced iterative algorithm is quite efficient.

Keywords: Model updating, iterative algorithm, gyroscopic system, partially prescribed spectral data, optimal approximation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1424
4342 Comparison of Different Hydrograph Routing Techniques in XPSTORM Modelling Software: A Case Study

Authors: Fatema Akram, Mohammad Golam Rasul, Mohammad Masud Kamal Khan, Md. Sharif Imam Ibne Amir

Abstract:

A variety of routing techniques are available to develop surface runoff hydrographs from rainfall. The selection of runoff routing method is very vital as it is directly related to the type of watershed and the required degree of accuracy. There are different modelling softwares available to explore the rainfall-runoff process in urban areas. XPSTORM, a link-node based, integrated stormwater modelling software, has been used in this study for developing surface runoff hydrograph for a Golf course area located in Rockhampton in Central Queensland in Australia. Four commonly used methods, namely SWMM runoff, Kinematic wave, Laurenson, and Time-Area are employed to generate runoff hydrograph for design storm of this study area. In runoff mode of XPSTORM, the rainfall, infiltration, evaporation and depression storage for subcatchments were simulated and the runoff from the subcatchment to collection node was calculated. The simulation results are presented, discussed and compared. The total surface runoff generated by SWMM runoff, Kinematic wave and Time-Area methods are found to be reasonably close, which indicates any of these methods can be used for developing runoff hydrograph of the study area. Laurenson method produces a comparatively less amount of surface runoff, however, it creates highest peak of surface runoff among all which may be suitable for hilly region. Although the Laurenson hydrograph technique is widely acceptable surface runoff routing technique in Queensland (Australia), extensive investigation is recommended with detailed topographic and hydrologic data in order to assess its suitability for use in the case study area.

Keywords: ARI, design storm, IFD, rainfall temporal pattern, routing techniques, surface runoff, XPSTORM.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5023
4341 Long-Term Durability of Roller-Compacted Concrete Pavement

Authors: Jun Hee Lee, Young Kyu Kim, Seong Jae Hong, Chamroeun Chhorn, Seung Woo Lee

Abstract:

Roller-compacted concrete pavement (RCCP), an environmental friendly pavement of which load carry capacity benefitted from both hydration and aggregate interlock from roller compacting, demonstrated a superb structural performance for a relatively small amount of water and cement content. Even though an excellent structural performance can be secured, it is required to investigate roller-compacted concrete (RCC) under environmental loading and its long-term durability under critical conditions. In order to secure long-term durability, an appropriate internal air-void structure is required for this concrete. In this study, a method for improving the long-term durability of RCCP is suggested by analyzing the internal air-void structure and corresponding durability of RCC. The method of improving the long-term durability involves measurements of air content, air voids, and air-spacing factors in RCC that experiences changes in terms of type of air-entraining agent and its usage amount. This test is conducted according to the testing criteria in ASTM C 457, 672, and KS F 2456. It was found that the freezing-thawing and scaling resistances of RCC without any chemical admixture was quite low. Interestingly, an improvement of freezing-thawing and scaling resistances was observed for RCC with appropriate the air entraining (AE) agent content; Relative dynamic elastic modulus was found to be more than 80% for those mixtures. In RCC with AE agent mixtures, large amount of air was distributed within a range of 2% to 3%, and an air void spacing factor ranging between 200 and 300 μm (close to 250 μm, recommended by PCA) was secured. The long-term durability of RCC has a direct relationship with air-void spacing factor, and thus it can only be secured by ensuring the air void spacing factor through the inclusion of the AE in the mixture.

Keywords: RCCP, durability, air spacing factor, surface scaling resistance test, freezing and thawing resistance test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1139
4340 Preparation of ATO Conductive Particles with Narrow Size Distribution

Authors: Yueying Wu, Fengzhu Lv, Yihe Zhang, Zixian Xu

Abstract:

Antimosy-doped tin oxide (ATO) particles were prepared via chemical coprecipitation and reverse emulsion. The size and size distribution of ATO particles were obviously decreased via reverse microemulsion method. At the relatively high yield the ATO particles were nearly spherical in shape, meanwhile the crystalline structure and excellent conductivity were reserved, which could satisfy the requirement as composite fillers, such as dielectric filler of polyimide film.

Keywords: ATO particle, Conductivity, Distribution, Reverse emulsion

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1861
4339 Lexical Based Method for Opinion Detection on Tripadvisor Collection

Authors: Faiza Belbachir, Thibault Schienhinski

Abstract:

The massive development of online social networks allows users to post and share their opinions on various topics. With this huge volume of opinion, it is interesting to extract and interpret these information for different domains, e.g., product and service benchmarking, politic, system of recommendation. This is why opinion detection is one of the most important research tasks. It consists on differentiating between opinion data and factual data. The difficulty of this task is to determine an approach which returns opinionated document. Generally, there are two approaches used for opinion detection i.e. Lexical based approaches and Machine Learning based approaches. In Lexical based approaches, a dictionary of sentimental words is used, words are associated with weights. The opinion score of document is derived by the occurrence of words from this dictionary. In Machine learning approaches, usually a classifier is trained using a set of annotated document containing sentiment, and features such as n-grams of words, part-of-speech tags, and logical forms. Majority of these works are based on documents text to determine opinion score but dont take into account if these texts are really correct. Thus, it is interesting to exploit other information to improve opinion detection. In our work, we will develop a new way to consider the opinion score. We introduce the notion of trust score. We determine opinionated documents but also if these opinions are really trustable information in relation with topics. For that we use lexical SentiWordNet to calculate opinion and trust scores, we compute different features about users like (numbers of their comments, numbers of their useful comments, Average useful review). After that, we combine opinion score and trust score to obtain a final score. We applied our method to detect trust opinions in TRIPADVISOR collection. Our experimental results report that the combination between opinion score and trust score improves opinion detection.

Keywords: Tripadvisor, Opinion detection, SentiWordNet, trust score.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 729
4338 Capacity Building for Hazmat Transport Emergency Preparedness: 'Hotspot Impact Zone' Mapping from Flammable and Toxic Releases

Authors: U K Chakrabarti, Jigisha Parikh

Abstract:

Hazardous Material transportation by road is coupled with inherent risk of accidents causing loss of lives, grievous injuries, property losses and environmental damages. The most common type of hazmat road accident happens to be the releases (78%) of hazardous substances, followed by fires (28%), explosions (14%) and vapour/ gas clouds (6 %.). The paper is discussing initially the probable 'Impact Zones' likely to be caused by one flammable (LPG) and one toxic (ethylene oxide) chemicals being transported through a sizable segment of a State Highway connecting three notified Industrial zones in Surat district in Western India housing 26 MAH industrial units. Three 'hotspots' were identified along the highway segment depending on the particular chemical traffic and the population distribution within 500 meters on either sides. The thermal radiation and explosion overpressure have been calculated for LPG / Ethylene Oxide BLEVE scenarios along with toxic release scenario for ethylene oxide. Besides, the dispersion calculations for ethylene oxide toxic release have been made for each 'hotspot' location and the impact zones have been mapped for the LOC concentrations. Subsequently, the maximum Initial Isolation and the protective zones were calculated based on ERPG-3 and ERPG-2 values of ethylene oxide respectively which are estimated taking the worst case scenario under worst weather conditions. The data analysis will be helpful to the local administration in capacity building with respect to rescue / evacuation and medical preparedness and quantitative inputs to augment the District Offsite Emergency Plan document.

Keywords: Hotspot, Ethylene Oxide, LPG, MAH (MajorAccident Hazard).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1787
4337 Combination of Different Classifiers for Cardiac Arrhythmia Recognition

Authors: M. R. Homaeinezhad, E. Tavakkoli, M. Habibi, S. A. Atyabi, A. Ghaffari

Abstract:

This paper describes a new supervised fusion (hybrid) electrocardiogram (ECG) classification solution consisting of a new QRS complex geometrical feature extraction as well as a new version of the learning vector quantization (LVQ) classification algorithm aimed for overcoming the stability-plasticity dilemma. Toward this objective, after detection and delineation of the major events of ECG signal via an appropriate algorithm, each QRS region and also its corresponding discrete wavelet transform (DWT) are supposed as virtual images and each of them is divided into eight polar sectors. Then, the curve length of each excerpted segment is calculated and is used as the element of the feature space. To increase the robustness of the proposed classification algorithm versus noise, artifacts and arrhythmic outliers, a fusion structure consisting of five different classifiers namely as Support Vector Machine (SVM), Modified Learning Vector Quantization (MLVQ) and three Multi Layer Perceptron-Back Propagation (MLP–BP) neural networks with different topologies were designed and implemented. The new proposed algorithm was applied to all 48 MIT–BIH Arrhythmia Database records (within–record analysis) and the discrimination power of the classifier in isolation of different beat types of each record was assessed and as the result, the average accuracy value Acc=98.51% was obtained. Also, the proposed method was applied to 6 number of arrhythmias (Normal, LBBB, RBBB, PVC, APB, PB) belonging to 20 different records of the aforementioned database (between– record analysis) and the average value of Acc=95.6% was achieved. To evaluate performance quality of the new proposed hybrid learning machine, the obtained results were compared with similar peer– reviewed studies in this area.

Keywords: Feature Extraction, Curve Length Method, SupportVector Machine, Learning Vector Quantization, Multi Layer Perceptron, Fusion (Hybrid) Classification, Arrhythmia Classification, Supervised Learning Machine.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2202
4336 Optical Flow Technique for Supersonic Jet Measurements

Authors: H. D. Lim, Jie Wu, T. H. New, Shengxian Shi

Abstract:

This paper outlines the development of an experimental technique in quantifying supersonic jet flows, in an attempt to avoid seeding particle problems frequently associated with particle-image velocimetry (PIV) techniques at high Mach numbers. Based on optical flow algorithms, the idea behind the technique involves using high speed cameras to capture Schlieren images of the supersonic jet shear layers, before they are subjected to an adapted optical flow algorithm based on the Horn-Schnuck method to determine the associated flow fields. The proposed method is capable of offering full-field unsteady flow information with potentially higher accuracy and resolution than existing point-measurements or PIV techniques. Preliminary study via numerical simulations of a circular de Laval jet nozzle successfully reveals flow and shock structures typically associated with supersonic jet flows, which serve as useful data for subsequent validation of the optical flow based experimental results. For experimental technique, a Z-type Schlieren setup is proposed with supersonic jet operated in cold mode, stagnation pressure of 4 bar and exit Mach of 1.5. High-speed singleframe or double-frame cameras are used to capture successive Schlieren images. As implementation of optical flow technique to supersonic flows remains rare, the current focus revolves around methodology validation through synthetic images. The results of validation test offers valuable insight into how the optical flow algorithm can be further improved to improve robustness and accuracy. Despite these challenges however, this supersonic flow measurement technique may potentially offer a simpler way to identify and quantify the fine spatial structures within the shock shear layer.

Keywords: Schlieren, optical flow, supersonic jets, shock shear layer.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1879
4335 From Primer Generation to Chromosome Identification: A Primer Generation Genotyping Method for Bacterial Identification and Typing

Authors: Wisam H. Benamer, Ehab A. Elfallah, Mohamed A. Elshaari, Farag A. Elshaari

Abstract:

A challenge for laboratories is to provide bacterial identification and antibiotic sensitivity results within a short time. Hence, advancement in the required technology is desirable to improve timing, accuracy and quality. Even with the current advances in methods used for both phenotypic and genotypic identification of bacteria the need is there to develop method(s) that enhance the outcome of bacteriology laboratories in accuracy and time. The hypothesis introduced here is based on the assumption that the chromosome of any bacteria contains unique sequences that can be used for its identification and typing. The outcome of a pilot study designed to test this hypothesis is reported in this manuscript. Methods: The complete chromosome sequences of several bacterial species were downloaded to use as search targets for unique sequences. Visual basic and SQL server (2014) were used to generate a complete set of 18-base long primers, a process started with reverse translation of randomly chosen 6 amino acids to limit the number of the generated primers. In addition, the software used to scan the downloaded chromosomes using the generated primers for similarities was designed, and the resulting hits were classified according to the number of similar chromosomal sequences, i.e., unique or otherwise. Results: All primers that had identical/similar sequences in the selected genome sequence(s) were classified according to the number of hits in the chromosomes search. Those that were identical to a single site on a single bacterial chromosome were referred to as unique. On the other hand, most generated primers sequences were identical to multiple sites on a single or multiple chromosomes. Following scanning, the generated primers were classified based on ability to differentiate between medically important bacterial and the initial results looks promising. Conclusion: A simple strategy that started by generating primers was introduced; the primers were used to screen bacterial genomes for match. Primer(s) that were uniquely identical to specific DNA sequence on a specific bacterial chromosome were selected. The identified unique sequence can be used in different molecular diagnostic techniques, possibly to identify bacteria. In addition, a single primer that can identify multiple sites in a single chromosome can be exploited for region or genome identification. Although genomes sequences draft of isolates of organism DNA enable high throughput primer design using alignment strategy, and this enhances diagnostic performance in comparison to traditional molecular assays. In this method the generated primers can be used to identify an organism before the draft sequence is completed. In addition, the generated primers can be used to build a bank for easy access of the primers that can be used to identify bacteria.

Keywords: Bacteria chromosome, bacterial identification, sequence, primer generation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1024
4334 Experimental Investigation of Heat Transfer and Flow of Nano Fluids in Horizontal Circular Tube

Authors: Abdulhassan Abd. K, Sattar Al-Jabair, Khalid Sultan

Abstract:

We have measured the pressure drop and convective heat transfer coefficient of water – based AL(25nm),AL2O3(30nm) and CuO(50nm) Nanofluids flowing through a uniform heated circular tube in the fully developed laminar flow regime. The experimental results show that the data for Nanofluids friction factor show a good agreement with analytical prediction from the Darcy's equation for single-phase flow. After reducing the experimental results to the form of Reynolds, Rayleigh and Nusselt numbers. The results show the local Nusselt number and temperature have distribution with the non-dimensional axial distance from the tube entry. Study decided that thenNanofluid as Newtonian fluids through the design of the linear relationship between shear stress and the rate of stress has been the study of three chains of the Nanofluid with different concentrations and where the AL, AL2O3 and CuO – water ranging from (0.25 - 2.5 vol %). In addition to measuring the four properties of the Nanofluid in practice so as to ensure the validity of equations of properties developed by the researchers in this area and these properties is viscosity, specific heat, and density and found that the difference does not exceed 3.5% for the experimental equations between them and the practical. The study also demonstrated that the amount of the increase in heat transfer coefficient for three types of Nano fluid is AL, AL2O3, and CuO – Water and these ratios are respectively (45%, 32%, 25%) with insulation and without insulation (36%, 23%, 19%), and the statement of any of the cases the best increase in heat transfer has been proven that using insulation is better than not using it. I have been using three types of Nano particles and one metallic Nanoparticle and two oxide Nanoparticle and a statement, whichever gives the best increase in heat transfer.

Keywords: Newtonian, NUR factor, Brownian motion

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1843