Search results for: Borne-Gentina criterion
75 A TFETI Domain Decompositon Solver for Von Mises Elastoplasticity Model with Combination of Linear Isotropic-Kinematic Hardening
Authors: Martin Cermak, Stanislav Sysala
Abstract:
In this paper we present the efficient parallel implementation of elastoplastic problems based on the TFETI (Total Finite Element Tearing and Interconnecting) domain decomposition method. This approach allow us to use parallel solution and compute this nonlinear problem on the supercomputers and decrease the solution time and compute problems with millions of DOFs. In our approach we consider an associated elastoplastic model with the von Mises plastic criterion and the combination of linear isotropic-kinematic hardening law. This model is discretized by the implicit Euler method in time and by the finite element method in space. We consider the system of nonlinear equations with a strongly semismooth and strongly monotone operator. The semismooth Newton method is applied to solve this nonlinear system. Corresponding linearized problems arising in the Newton iterations are solved in parallel by the above mentioned TFETI. The implementation of this problem is realized in our in-house MatSol packages developed in MatLab.
Keywords: Isotropic-kinematic hardening, TFETI, domain decomposition, parallel solution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 175974 Assertion-Driven Test Repair Based on Priority Criteria
Authors: Ruilian Zhao, Shukai Zhang, Yan Wang, Weiwei Wang
Abstract:
Repairing broken test cases is an expensive and challenging task in evolving software systems. Although an automated repair technique with intent-preservation has been proposed, it does not take into account the association between test repairs and assertions, leading a large number of irrelevant candidates and decreasing the repair capability. This paper proposes a assertion-driven test repair approach. Furthermore, a intent-oriented priority criterion is raised to guide the repair candidate generation, making the repairs closer to the intent of the test. In more detail, repair targets are determined through post-dominance relations between assertions and the methods that directly cause compilation errors. Then, test repairs are generated from the target in a bottom-up way, guided by the the intent-oriented priority criteria. Finally, the generated repair candidates are prioritized to match the original test intent. The approach is implemented and evaluated on the benchmark of 4 open-source programs and 91 broken test cases. The result shows that the approach can fix 89% (81/91) broken test cases, which are more effective than the existing intent-preserved test repair approach, and our intent-oriented priority criteria work well.
Keywords: Test repair, test intent, software test, test case evolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15573 On the Computation of a Common n-finger Robotic Grasp for a Set of Objects
Authors: Avishai Sintov, Roland Menassa, Amir Shapiro
Abstract:
Industrial robotic arms utilize multiple end-effectors, each for a specific part and for a specific task. We propose a novel algorithm which will define a single end-effector’s configuration able to grasp a given set of objects with different geometries. The algorithm will have great benefit in production lines allowing a single robot to grasp various parts. Hence, reducing the number of endeffectors needed. Moreover, the algorithm will reduce end-effector design and manufacturing time and final product cost. The algorithm searches for a common grasp over the set of objects. The search algorithm maps all possible grasps for each object which satisfy a quality criterion and takes into account possible external wrenches (forces and torques) applied to the object. The mapped grasps are- represented by high-dimensional feature vectors which describes the shape of the gripper. We generate a database of all possible grasps for each object in the feature space. Then we use a search and classification algorithm for intersecting all possible grasps over all parts and finding a single common grasp suitable for all objects. We present simulations of planar and spatial objects to validate the feasibility of the approach.
Keywords: Common Grasping, Search Algorithm, Robotic End-Effector.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 167572 Optimal Retrofit Design of Reinforced Concrete Frame with Infill Wall Using Fiber Reinforced Plastic Materials
Authors: Sang Wook Park, Se Woon Choi, Yousok Kim, Byung Kwan Oh, Hyo Seon Park
Abstract:
Various retrofit techniques for reinforced concrete frame with infill wall have been steadily developed. Among those techniques, strengthening methodology based on diagonal FRP strips (FRP bracings) has numerous advantages such as feasibility of implementing without interrupting the building under operation, reduction of cost and time, and easy application. Considering the safety of structure and retrofit cost, the most appropriate retrofit solution is needed. Thus, the objective of this study is to suggest pareto-optimal solution for existing building using FRP bracings. To find pareto-optimal solution analysis, NSGA-II is applied. Moreover, the seismic performance of retrofit building is evaluated. The example building is 5-storey, 3-bay RC frames with infill wall. Nonlinear static pushover analyses are performed with FEMA 356. The criterion of performance evaluation is inter-story drift ratio at the performance level IO, LS, CP. Optimal retrofit solutions is obtained for 32 individuals and 200 generations. Through the proposed optimal solutions, we confirm the improvement of seismic performance of the example building.
Keywords: Retrofit, FRP bracings, reinforced concrete frame with infill wall, seismic performance evaluation, NSGA-II.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 204171 An Empirical Model to Calculate the Threads Stripping of a Bolt Installed in a Tapped Part
Authors: Manuel Martínez Martínez, Daniel Zavala Ríos
Abstract:
To determine the length of engagement threads of a bolt installed in a tapped part in order to avoid the threads stripping remains a very current problem in the design of the thread assemblies. It does not exist a calculation method formalized for the cases where the bolt is screwed directly in a ductile material. In this article, we study the behavior of the threads stripping of a loaded assembly by using a modelling by finite elements and a rupture criterion by damage. This modelling enables us to study the different parameters likely to influence the behavior of this bolted connection. We study in particular, the influence of couple of materials constituting the connection, of the bolt-s diameter and the geometrical characteristics of the tapped part, like the external diameter and the length of engagement threads. We established an experiments design to know the most significant parameters. That enables us to propose a simple expression making possible to calculate the resistance of the threads whatever the metallic materials of the bolt and the tapped part. We carried out stripping tests in order to validate our model. The estimated results are very close to those obtained by the tests.
Keywords: Bolt, damage, plasticity, stripping, thread assemblies.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 527070 Auto Rickshaw Impacts with Pedestrians: A Computational Analysis of Post-Collision Kinematics and Injury Mechanics
Authors: A. J. Al-Graitti, G. A. Khalid, P. Berthelson, A. Mason-Jones, R. Prabhu, M. D. Jones
Abstract:
Motor vehicle related pedestrian road traffic collisions are a major road safety challenge, since they are a leading cause of death and serious injury worldwide, contributing to a third of the global disease burden. The auto rickshaw, which is a common form of urban transport in many developing countries, plays a major transport role, both as a vehicle for hire and for private use. The most common auto rickshaws are quite unlike ‘typical’ four-wheel motor vehicle, being typically characterised by three wheels, a non-tilting sheet-metal body or open frame construction, a canvas roof and side curtains, a small drivers’ cabin, handlebar controls and a passenger space at the rear. Given the propensity, in developing countries, for auto rickshaws to be used in mixed cityscapes, where pedestrians and vehicles share the roadway, the potential for auto rickshaw impacts with pedestrians is relatively high. Whilst auto rickshaws are used in some Western countries, their limited number and spatial separation from pedestrian walkways, as a result of city planning, has not resulted in significant accident statistics. Thus, auto rickshaws have not been subject to the vehicle impact related pedestrian crash kinematic analyses and/or injury mechanics assessment, typically associated with motor vehicle development in Western Europe, North America and Japan. This study presents a parametric analysis of auto rickshaw related pedestrian impacts by computational simulation, using a Finite Element model of an auto rickshaw and an LS-DYNA 50th percentile male Hybrid III Anthropometric Test Device (dummy). Parametric variables include auto rickshaw impact velocity, auto rickshaw impact region (front, centre or offset) and relative pedestrian impact position (front, side and rear). The output data of each impact simulation was correlated against reported injury metrics, Head Injury Criterion (front, side and rear), Neck injury Criterion (front, side and rear), Abbreviated Injury Scale and reported risk level and adds greater understanding to the issue of auto rickshaw related pedestrian injury risk. The parametric analyses suggest that pedestrians are subject to a relatively high risk of injury during impacts with an auto rickshaw at velocities of 20 km/h or greater, which during some of the impact simulations may even risk fatalities. The present study provides valuable evidence for informing a series of recommendations and guidelines for making the auto rickshaw safer during collisions with pedestrians. Whilst it is acknowledged that the present research findings are based in the field of safety engineering and may over represent injury risk, compared to “Real World” accidents, many of the simulated interactions produced injury response values significantly greater than current threshold curves and thus, justify their inclusion in the study. To reduce the injury risk level and increase the safety of the auto rickshaw, there should be a reduction in the velocity of the auto rickshaw and, or, consideration of engineering solutions, such as retro fitting injury mitigation technologies to those auto rickshaw contact regions which are the subject of the greatest risk of producing pedestrian injury.
Keywords: Auto Rickshaw, finite element analysis, injury risk level, LS-DYNA, pedestrian impact.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 131969 Adaptive Weighted Averaging Filter Using the Appropriate Number of Consecutive Frames
Authors: Mahmoud Saeidi, Ali Nazemipour
Abstract:
In this paper, we propose a novel adaptive spatiotemporal filter that utilizes image sequences in order to remove noise. The consecutive frames include: current, previous and next noisy frames. The filter proposed in this paper is based upon the weighted averaging pixels intensity and noise variance in image sequences. It utilizes the Appropriate Number of Consecutive Frames (ANCF) based on the noisy pixels intensity among the frames. The number of consecutive frames is adaptively calculated for each region in image and its value may change from one region to another region depending on the pixels intensity within the region. The weights are determined by a well-defined mathematical criterion, which is adaptive to the feature of spatiotemporal pixels of the consecutive frames. It is experimentally shown that the proposed filter can preserve image structures and edges under motion while suppressing noise, and thus can be effectively used in image sequences filtering. In addition, the AWA filter using ANCF is particularly well suited for filtering sequences that contain segments with abruptly changing scene content due to, for example, rapid zooming and changes in the view of the camera.Keywords: Appropriate Number of Consecutive Frames, Adaptive Weighted Averaging, Motion Estimation, Noise Variance, Motion Compensation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 181868 Geometry Calibration Factors of Modified Arcan Fracture Test for Welded Joint
Authors: S. R. Hosseini, N. Choupani, A. R. M. Gharabaghi
Abstract:
In this study the mixed mode fracture mechanics parameters were investigated for high tensile steel butt welded joint based on modified Arcan test and finite element analysis was used to evaluate the effect of crack length on fracture criterion. The nondimensional stress intensity factors, strain energy release rates and Jintegral energy on crack tip were obtained for various in-plane loading combinations on Arcan specimen starting from pure mode-I to pure mode-II loading conditions. The specimen and apparatus were modeled by finite element method and analyzed under various loading angles (between 0 to 90 degrees with 15 degree interval) to simulate the pure mode-I, II and mixed mode fracture. Since the analytical results are independent from elasticity modules for isotropic materials, therefore the results in elastic fields can be used for Arcan specimens. The main objective of this study was to evaluate the geometric calibration factors for modified Arcan test specimen in order to obtain fracture toughness under mixed mode loading conditions.Keywords: Arcan specimen, Geometric calibration factors, Mixed Mode, Fracture mechanics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 196667 Influence of Vegetable Oil-Based Controlled Cutting Fluid Impinging Supply System on Micro Hardness in Machining of Ti-6Al-4V
Authors: Salah Gariani, Islam Shyha, Fawad Inam, Dehong Huo
Abstract:
A controlled cutting fluid impinging supply system (CUT-LIST) was developed to deliver an accurate amount of cutting fluid into the machining zone via well-positioned coherent nozzles based on a calculation of the heat generated. The performance of the CUT-LIST was evaluated against a conventional flood cutting fluid supply system during step shoulder milling of Ti-6Al-4V using vegetable oil-based cutting fluid. In this paper, the micro-hardness of the machined surface was used as the main criterion to compare the two systems. CUT-LIST provided significant reductions in cutting fluid consumption (up to 42%). Both systems caused increased micro-hardness value at 100 µm from the machined surface, whereas a slight reduction in micro-hardness of 4.5% was measured when using CUL-LIST. It was noted that the first 50 µm is the soft sub-surface promoted by thermal softening, whereas down to 100 µm is the hard sub-surface caused by the cyclic internal work hardening and then gradually decreased until it reached the base material nominal hardness. It can be concluded that the CUT-LIST has always given lower micro-hardness values near the machined surfaces in all conditions investigated.
Keywords: Impinging supply system, micro-hardness, shoulder milling, Ti-6Al-4V, vegetable oil-based cutting fluid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 101966 The Relationship between Conceptual Organizational Culture and the Level of Tolerance in Employees
Authors: M. Sadoughi, R. Ehsani
Abstract:
The aim of the present study is examining the relationship between conceptual organizational culture and the level of tolerance in employees of Islamic Azad University of Shahre Ghods. This research is a correlational and analytic-descriptive one. The samples included 144 individuals. A 24-item standard questionnaire of organizational culture by Cameron and Queen was used in this study. This questionnaire has six criteria and each criterion includes four items that each item indicates one cultural dimension. Reliability coefficient of this questionnaire was normed using Cronbach's alpha of 0.91. Also, the 25-item questionnaire of tolerance by Conor and Davidson was used. This questionnaire is in a five-degree Likert scale form. It has seven criteria and is designed to measure the power of coping with pressure and threat. It has the needed content reliability and its reliability coefficient is normed using Cronbach's alpha of 0.87. Data were analyzed using Pearson correlation coefficient and multivariable regression. The results showed among various dimensions of organizational culture, there is a positive significant relationship between three dimensions (family, adhocracy, bureaucracy) and tolerance, there is a negative significant relationship between dimension of market and tolerance and components of organizational culture have the power of prediction and explaining the tolerance. In this explanation, the component of family is the most effective and the best predictor of tolerance.
Keywords: Adhocracy, bureaucracy, organizational culture, tolerance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 103165 Chlorophyll Fluorescence as Criterion for the Diagnosis Salt Stress in Wheat (Triticum aestivum) Plants
Authors: M. Abdeshahian, M. Nabipour, M. Meskarbashee
Abstract:
To investigate effect of salt stress on Chlorophyll fluorescence four cultivars (fong,star,chamran and kharchia) of wheat (Triticum aestivum) plants subjected to salinity levels ( control,8,12 and 16 dsm-1 ) from one week after emergence to the end of stem elongation under greenhouse condition . results showed that quantum yield of photosystem II from light adopted leaves (ΦPSII), Photochemical quenching (qP) ,quantum yield of dark adopted leaves (fv/fm) and non photochemical quenching (NPq) were affected by salt stress . Salinity levels affected photosynthetic rate. Star and fong cultivars showed minimum and maximum levels of photosynthetic rate in respectively. Minimum photosynthetic rate differences between levels of salinity were shown in Kharchia. Shoot dry matter of all cultivars decreased by increasing salinity levels. Results showed that non photochemical quenching by salinity levels attribute to the decreases in shoot dry matter.Keywords: salt stress, wheat, chlorophyll fluorescence, photosynthesis , shoot dry matter .
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 247964 The Use of Chlorophyll Meter Readings for the Selection of Maize Inbred Lines under Drought Stress
Authors: F. Gekas, C. Pankou, I. Mylonas, E. Ninou, E. Sinapidou, A. Lithourgidis, F. Papathanasiou, J. –K. Petrevska, F. Papadopoulou, P. Zouliamis, G. Tsaprounis, I. Tokatlidis, C. Dordas
Abstract:
The present study aimed to investigate whether chlorophyll meter readings (SPAD) can be used as criterion of singleplant selection in maize breeding. Experimentation was performed at the ultra-low density of 0.74 plants/m2 in order the potential yield per plant to be fully expressed. R-31 honeycomb experiments were conducted in three different areas in Greece (Thessaloniki, Giannitsa and Florina) using 30 inbred lines at well-watered and water-stressed conditions during the 2012 growing season. The chlorophyll meter readings had higher rates at dry conditions, except location of Giannitsa where differences were not significant. Genotypes of highest chlorophyll meter readings were consistent across areas, emphasizing on the character’s stability. A positive correlation between the chlorophyll meter readings and grain yield was strengthening over time and culminated at the physiological maturity stage. There was a clear sign that the chlorophyll meter readings has the potential to be used for the selection of stress-adaptive genotypes and may permit modern maize to be grown at wider range of environments addressing the climate change scenarios.
Keywords: Drought-prone environments, honeycomb breeding, SPAD, Zea mays.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 283963 A Framework to Assess the Maturity of Customer Involvement in the Service Design of Product-Service Systems
Authors: Taghreed Abu-Salim
Abstract:
This paper develops and investigates a framework for the assessment of customer involvement in the service design process of result oriented product-service systems in order to improve the service offering in a business-to-business (B2B) context. The framework comprises five main criteria and fifteen sub-criteria that contribute to customer involvement in a hierarchy using a maturity grid to highlight the strengths and weaknesses for each criterion. To develop the customer involvement framework, an extensive literature review related to service design, result oriented product-service system (PSS) and customer involvement in service design was carried out. Key factors that significantly influence customer involvement from industry and literature were identified to develop the framework. A major contribution of the developed framework includes a hierarchy of appropriate criteria for assessing customer involvement in the service design process within results oriented PSS; the definition of four maturity levels which are suitable to describe the whole spectrum of customer involvement in the service design process; and finally, The paper concludes by enabling service providers to: take proactive decisions; screen and evaluate new services; improve perceived service quality; and provide barriers against imitation.Keywords: Customer involvement, maturity grid, new service development, result oriented product-service system, service design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 215962 Probe Selection for Pathway-Specific Microarray Probe Design Minimizing Melting Temperature Variance
Authors: Fabian Horn, Reinhard Guthke
Abstract:
In molecular biology, microarray technology is widely and successfully utilized to efficiently measure gene activity. If working with less studied organisms, methods to design custom-made microarray probes are available. One design criterion is to select probes with minimal melting temperature variances thus ensuring similar hybridization properties. If the microarray application focuses on the investigation of metabolic pathways, it is not necessary to cover the whole genome. It is more efficient to cover each metabolic pathway with a limited number of genes. Firstly, an approach is presented which minimizes the overall melting temperature variance of selected probes for all genes of interest. Secondly, the approach is extended to include the additional constraints of covering all pathways with a limited number of genes while minimizing the overall variance. The new optimization problem is solved by a bottom-up programming approach which reduces the complexity to make it computationally feasible. The new method is exemplary applied for the selection of microarray probes in order to cover all fungal secondary metabolite gene clusters for Aspergillus terreus.
Keywords: bottom-up approach, gene clusters, melting temperature, metabolic pathway, microarray probe design, probe selection
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 155961 A Formative Assessment Tool for Effective Feedback
Authors: Rami Rashkovits, Ilana Lavy
Abstract:
In this study we present our developed formative assessment tool for students' assignments. The tool enables lecturers to define assignments for the course and assign each problem in each assignment a list of criteria and weights by which the students' work is evaluated. During assessment, the lecturers feed the scores for each criterion with justifications. When the scores of the current assignment are completely fed in, the tool automatically generates reports for both students and lecturers. The students receive a report by email including detailed description of their assessed work, their relative score and their progress across the criteria along the course timeline. This information is presented via charts generated automatically by the tool based on the scores fed in. The lecturers receive a report that includes summative (e.g., averages, standard deviations) and detailed (e.g., histogram) data of the current assignment. This information enables the lecturers to follow the class achievements and adjust the learning process accordingly. The tool was examined on two pilot groups of college students that study a course in (1) Object-Oriented Programming (2) Plane Geometry. Results reveal that most of the students were satisfied with the assessment process and the reports produced by the tool. The lecturers who used the tool were also satisfied with the reports and their contribution to the learning process.
Keywords: Computer-based formative assessment tool, science education.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 189060 MONPAR - A Page Replacement Algorithm for a Spatiotemporal Database
Authors: U. Kalay, O. Kalıpsız
Abstract:
For a spatiotemporal database management system, I/O cost of queries and other operations is an important performance criterion. In order to optimize this cost, an intense research on designing robust index structures has been done in the past decade. With these major considerations, there are still other design issues that deserve addressing due to their direct impact on the I/O cost. Having said this, an efficient buffer management strategy plays a key role on reducing redundant disk access. In this paper, we proposed an efficient buffer strategy for a spatiotemporal database index structure, specifically indexing objects moving over a network of roads. The proposed strategy, namely MONPAR, is based on the data type (i.e. spatiotemporal data) and the structure of the index structure. For the purpose of an experimental evaluation, we set up a simulation environment that counts the number of disk accesses while executing a number of spatiotemporal range-queries over the index. We reiterated simulations with query sets with different distributions, such as uniform query distribution and skewed query distribution. Based on the comparison of our strategy with wellknown page-replacement techniques, like LRU-based and Prioritybased buffers, we conclude that MONPAR behaves better than its competitors for small and medium size buffers under all used query-distributions.Keywords: Buffer Management, Spatiotemporal databases.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 147559 Fuzzy Uncertainty Theory for Stealth Fighter Aircraft Selection in Entropic Fuzzy TOPSIS Decision Analysis Process
Authors: C. Ardil
Abstract:
The purpose of this paper is to present fuzzy TOPSIS in an entropic fuzzy environment. Due to the ambiguous concepts often represented in decision data, exact values are insufficient to model real-life situations. In this paper, the rating of each alternative is defined in fuzzy linguistic terms, which can be expressed with triangular fuzzy numbers. The weight of each criterion is then derived from the decision matrix using the entropy weighting method. Next, a vertex method is proposed to calculate the distance between two triangular fuzzy numbers. According to the TOPSIS concept, a closeness coefficient is defined to determine the ranking order of all alternatives by simultaneously calculating the distances to both the fuzzy positive-ideal solution (FPIS) and the fuzzy negative-ideal solution (FNIS). Finally, an illustrative example of selecting stealth fighter aircraft is shown at the end of this article to highlight the procedure of the proposed method. Correlation analysis and validation analysis using TOPSIS, WSM, and WPM methods were performed to compare the ranking order of the alternatives.
Keywords: stealth fighter aircraft selection, fuzzy uncertainty theory (FUT), fuzzy entropic decision (FED), fuzzy linguistic variables, triangular fuzzy numbers, multiple criteria decision making analysis, MCDMA, TOPSIS, WSM, WPM
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 60158 Flood Predicting in Karkheh River Basin Using Stochastic ARIMA Model
Authors: Karim Hamidi Machekposhti, Hossein Sedghi, Abdolrasoul Telvari, Hossein Babazadeh
Abstract:
Floods have huge environmental and economic impact. Therefore, flood prediction is given a lot of attention due to its importance. This study analysed the annual maximum streamflow (discharge) (AMS or AMD) of Karkheh River in Karkheh River Basin for flood predicting using ARIMA model. For this purpose, we use the Box-Jenkins approach, which contains four-stage method model identification, parameter estimation, diagnostic checking and forecasting (predicting). The main tool used in ARIMA modelling was the SAS and SPSS software. Model identification was done by visual inspection on the ACF and PACF. SAS software computed the model parameters using the ML, CLS and ULS methods. The diagnostic checking tests, AIC criterion, RACF graph and RPACF graphs, were used for selected model verification. In this study, the best ARIMA models for Annual Maximum Discharge (AMD) time series was (4,1,1) with their AIC value of 88.87. The RACF and RPACF showed residuals’ independence. To forecast AMD for 10 future years, this model showed the ability of the model to predict floods of the river under study in the Karkheh River Basin. Model accuracy was checked by comparing the predicted and observation series by using coefficient of determination (R2).
Keywords: Time series modelling, stochastic processes, ARIMA model, Karkheh River.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 104457 Modified Plastic-Damage Model for Fiber Reinforced Polymer-Confined Repaired Concrete Columns
Authors: I. A Tijani, Y. F Wu, C.W. Lim
Abstract:
Concrete Damaged Plasticity Model (CDPM) is capable of modeling the stress-strain behavior of confined concrete. Nevertheless, the accuracy of the model largely depends on its parameters. To date, most research works mainly focus on the identification and modification of the parameters for fiber reinforced polymer (FRP) confined concrete prior to damage. And, it has been established that the FRP-strengthened concrete behaves differently to FRP-repaired concrete. This paper presents a modified plastic damage model within the context of the CDPM in ABAQUS for modelling of a uniformly FRP-confined repaired concrete under monotonic loading. The proposed model includes infliction damage, elastic stiffness, yield criterion and strain hardening rule. The distinct feature of damaged concrete is elastic stiffness reduction; this is included in the model. Meanwhile, the test results were obtained from a physical testing of repaired concrete. The dilation model is expressed as a function of the lateral stiffness of the FRP-jacket. The finite element predictions are shown to be in close agreement with the obtained test results of the repaired concrete. It was observed from the study that with necessary modifications, finite element method is capable of modeling FRP-repaired concrete structures.
Keywords: Concrete, FRP, damage, repairing, plasticity, and finite element method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 92556 Optimisation of a Dragonfly-Inspired Flapping Wing-Actuation System
Authors: Jia-Ming Kok, Javaan Chahl
Abstract:
An optimisation method using both global and local optimisation is implemented to determine the flapping profile which will produce the most lift for an experimental wing-actuation system. The optimisation method is tested using a numerical quasi-steady analysis. Results of an optimised flapping profile show a 20% increase in lift generated as compared to flapping profiles obtained by high speed cinematography of a Sympetrum frequens dragonfly. Initial optimisation procedures showed 3166 objective function evaluations. The global optimisation parameters - initial sample size and stage one sample size, were altered to reduce the number of function evaluations. Altering the stage one sample size had no significant effect. It was found that reducing the initial sample size to 400 would allow a reduction in computational effort to approximately 1500 function evaluations without compromising the global solvers ability to locate potential minima. To further reduce the optimisation effort required, we increase the local solver’s convergence tolerance criterion. An increase in the tolerance from 0.02N to 0.05N decreased the number of function evaluations by another 20%. However, this potentially reduces the maximum obtainable lift by up to 0.025N.
Keywords: Flapping wing, Optimisation, Quasi-steady model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 240255 Holistic Approach to Teaching Mathematics in Secondary School as a Means of Improving Students’ Comprehension of Study Material
Authors: Natalia Podkhodova, Olga Sheremeteva, Mariia Soldaeva
Abstract:
Creating favourable conditions for students’ comprehension of mathematical content is one of the primary problems in teaching mathematics in secondary school. The fact of comprehension includes the ability to build a working situational model and thus becomes an important means of solving mathematical problems. This paper describes a holistic approach to teaching mathematics designed to address the primary challenges of such teaching; specifically, the challenge of students’ comprehension. Essentially, this approach consists of (1) establishing links between the attributes of the notion: the sense, the meaning, and the term; (2) taking into account the components of student’s subjective experience—value-based emotions, contextual, procedural and communicative—during the educational process; (3) linking together different ways to present mathematical information; (4) identifying and leveraging the relationships between real, perceptual and conceptual (scientific) mathematical spaces by applying real-life situational modelling. The article describes approaches to the practical use of these foundational concepts. Identifying how proposed methods and techniques influence understanding of material used in teaching mathematics was the primary goal. The study included an experiment in which 256 secondary school students took part: 142 in the study group and 114 in the control group. All students in these groups had similar levels of achievement in math and studied math under the same curriculum. In the course of the experiment, comprehension of two topics — “Derivative” and “Trigonometric functions”—was evaluated. Control group participants were taught using traditional methods. Students in the study group were taught using the holistic method: under teacher’s guidance, they carried out assignments designed to establish linkages between notion’s characteristics, to convert information from one mode of presentation to another, as well as assignments that required the ability to operate with all modes of presentation. Identification, accounting for and transformation of subjective experience were associated with methods of stimulating the emotional value component of the studied mathematical content (discussions of lesson titles, assignments aimed to create study dominants, performing theme-related physical exercise ...) The use of techniques that forms inter-subject notions based on linkages between, perceptual real and mathematical conceptual spaces proved to be of special interest to the students. Results of the experiment were analysed by presenting students in each of the groups with a final test in each of the studied topics. The test included assignments that required building real situational models. Statistical analysis was used to aggregate test results. Pierson criterion x2 was used to reveal statistics significance of results (pass-fail the modelling test). Significant difference of results was revealed (p < 0.001), which allowed to conclude that students in the study group showed better comprehension of mathematical information than those in the control group. The total number of completed assignments of each student was analysed as well, with average results calculated for each group. Statistical significance of result differences against the quantitative criterion (number of completed assignments) was determined using Student’s t-test, which showed that students in the study group completed significantly more assignments than those in the control group (p = 0.0001). Authors thus come to the conclusion that suggested increase in the level of comprehension of study material took place as a result of applying implemented methods and techniques.
Keywords: Comprehension of mathematical content, holistic approach to teaching mathematics in secondary school, subjective experience, technology of the formation of inter-subject notions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 60654 A Mathematical Representation for Mechanical Model Assessment: Numerical Model Qualification Method
Authors: Keny Ordaz-Hernandez, Xavier Fischer, Fouad Bennis
Abstract:
This article illustrates a model selection management approach for virtual prototypes in interactive simulations. In those numerical simulations, the virtual prototype and its environment are modelled as a multiagent system, where every entity (prototype,human, etc.) is modelled as an agent. In particular, virtual prototyp ingagents that provide mathematical models of mechanical behaviour inform of computational methods are considered. This work argues that selection of an appropriate model in a changing environment,supported by models? characteristics, can be managed by the deter-mination a priori of specific exploitation and performance measures of virtual prototype models. As different models exist to represent a single phenomenon, it is not always possible to select the best one under all possible circumstances of the environment. Instead the most appropriate shall be selecting according to the use case. The proposed approach consists in identifying relevant metrics or indicators for each group of models (e.g. entity models, global model), formulate their qualification, analyse the performance, and apply the qualification criteria. Then, a model can be selected based on the performance prediction obtained from its qualification. The authors hope that this approach will not only help to inform engineers and researchers about another approach for selecting virtual prototype models, but also assist virtual prototype engineers in the systematic or automatic model selection.
Keywords: Virtual prototype models, domain, qualification criterion, model qualification, model assessment, environmental modelling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 203753 Effect of Progressive Type-I Right Censoring on Bayesian Statistical Inference of Simple Step–Stress Acceleration Life Testing Plan under Weibull Life Distribution
Authors: Saleem Z. Ramadan
Abstract:
This paper discusses the effects of using progressive Type-I right censoring on the design of the Simple Step Accelerated Life testing using Bayesian approach for Weibull life products under the assumption of cumulative exposure model. The optimization criterion used in this paper is to minimize the expected pre-posterior variance of the Pth percentile time of failures. The model variables are the stress changing time and the stress value for the first step. A comparison between the conventional and the progressive Type-I right censoring is provided. The results have shown that the progressive Type-I right censoring reduces the cost of testing on the expense of the test precision when the sample size is small. Moreover, the results have shown that using strong priors or large sample size reduces the sensitivity of the test precision to the censoring proportion. Hence, the progressive Type-I right censoring is recommended in these cases as progressive Type-I right censoring reduces the cost of the test and doesn't affect the precision of the test a lot. Moreover, the results have shown that using direct or indirect priors affects the precision of the test.
Keywords: Reliability, Accelerated life testing, Cumulative exposure model, Bayesian estimation, Progressive Type-I censoring, Weibull distribution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 215952 Through Biometric Card in Romania: Person Identification by Face, Fingerprint and Voice Recognition
Authors: Hariton N. Costin, Iulian Ciocoiu, Tudor Barbu, Cristian Rotariu
Abstract:
In this paper three different approaches for person verification and identification, i.e. by means of fingerprints, face and voice recognition, are studied. Face recognition uses parts-based representation methods and a manifold learning approach. The assessment criterion is recognition accuracy. The techniques under investigation are: a) Local Non-negative Matrix Factorization (LNMF); b) Independent Components Analysis (ICA); c) NMF with sparse constraints (NMFsc); d) Locality Preserving Projections (Laplacianfaces). Fingerprint detection was approached by classical minutiae (small graphical patterns) matching through image segmentation by using a structural approach and a neural network as decision block. As to voice / speaker recognition, melodic cepstral and delta delta mel cepstral analysis were used as main methods, in order to construct a supervised speaker-dependent voice recognition system. The final decision (e.g. “accept-reject" for a verification task) is taken by using a majority voting technique applied to the three biometrics. The preliminary results, obtained for medium databases of fingerprints, faces and voice recordings, indicate the feasibility of our study and an overall recognition precision (about 92%) permitting the utilization of our system for a future complex biometric card.Keywords: Biometry, image processing, pattern recognition, speech analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 194351 A Study of Two Disease Models: With and Without Incubation Period
Authors: H. C. Chinwenyi, H. D. Ibrahim, J. O. Adekunle
Abstract:
The incubation period is defined as the time from infection with a microorganism to development of symptoms. In this research, two disease models: one with incubation period and another without incubation period were studied. The study involves the use of a mathematical model with a single incubation period. The test for the existence and stability of the disease free and the endemic equilibrium states for both models were carried out. The fourth order Runge-Kutta method was used to solve both models numerically. Finally, a computer program in MATLAB was developed to run the numerical experiments. From the results, we are able to show that the endemic equilibrium state of the model with incubation period is locally asymptotically stable whereas the endemic equilibrium state of the model without incubation period is unstable under certain conditions on the given model parameters. It was also established that the disease free equilibrium states of the model with and without incubation period are locally asymptotically stable. Furthermore, results from numerical experiments using empirical data obtained from Nigeria Centre for Disease Control (NCDC) showed that the overall population of the infected people for the model with incubation period is higher than that without incubation period. We also established from the results obtained that as the transmission rate from susceptible to infected population increases, the peak values of the infected population for the model with incubation period decrease and are always less than those for the model without incubation period.
Keywords: Asymptotic stability, incubation period, Routh-Hurwitz criterion, Runge Kutta method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 68450 Featured based Segmentation of Color Textured Images using GLCM and Markov Random Field Model
Authors: Dipti Patra, Mridula J
Abstract:
In this paper, we propose a new image segmentation approach for colour textured images. The proposed method for image segmentation consists of two stages. In the first stage, textural features using gray level co-occurrence matrix(GLCM) are computed for regions of interest (ROI) considered for each class. ROI acts as ground truth for the classes. Ohta model (I1, I2, I3) is the colour model used for segmentation. Statistical mean feature at certain inter pixel distance (IPD) of I2 component was considered to be the optimized textural feature for further segmentation. In the second stage, the feature matrix obtained is assumed to be the degraded version of the image labels and modeled as Markov Random Field (MRF) model to model the unknown image labels. The labels are estimated through maximum a posteriori (MAP) estimation criterion using ICM algorithm. The performance of the proposed approach is compared with that of the existing schemes, JSEG and another scheme which uses GLCM and MRF in RGB colour space. The proposed method is found to be outperforming the existing ones in terms of segmentation accuracy with acceptable rate of convergence. The results are validated with synthetic and real textured images.
Keywords: Texture Image Segmentation, Gray Level Cooccurrence Matrix, Markov Random Field Model, Ohta colour space, ICM algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 217249 Comparison of Finite Difference Schemes for Water Flow in Unsaturated Soils
Authors: H. Taheri Shahraiyni, B. Ataie Ashtiani
Abstract:
Flow movement in unsaturated soil can be expressed by a partial differential equation, named Richards equation. The objective of this study is the finding of an appropriate implicit numerical solution for head based Richards equation. Some of the well known finite difference schemes (fully implicit, Crank Nicolson and Runge-Kutta) have been utilized in this study. In addition, the effects of different approximations of moisture capacity function, convergence criteria and time stepping methods were evaluated. Two different infiltration problems were solved to investigate the performance of different schemes. These problems include of vertical water flow in a wet and very dry soils. The numerical solutions of two problems were compared using four evaluation criteria and the results of comparisons showed that fully implicit scheme is better than the other schemes. In addition, utilizing of standard chord slope method for approximation of moisture capacity function, automatic time stepping method and difference between two successive iterations as convergence criterion in the fully implicit scheme can lead to better and more reliable results for simulation of fluid movement in different unsaturated soils.Keywords: Finite Difference methods, Richards equation, fullyimplicit, Crank-Nicolson, Runge-Kutta.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 237548 Seamless Handover in Urban 5G-UAV Systems Using Entropy Weighted Method
Authors: Anirudh Sunil Warrier, Saba Al-Rubaye, Dimitrios Panagiotakopoulos, Gokhan Inalhan, Antonios Tsourdos
Abstract:
The demand for increased data transfer rate and network traffic capacity has given rise to the concept of heterogeneous networks. Heterogeneous networks are wireless networks, consisting of devices using different underlying radio access technologies (RAT). For Unmanned Aerial Vehicles (UAVs) this enhanced data rate and network capacity are even more critical especially in their applications of medicine, delivery missions and military. In an urban heterogeneous network environment, the UAVs must be able switch seamlessly from one base station (BS) to another for maintaining a reliable link. Therefore, seamless handover in such urban environments has become a major challenge. In this paper, a scheme to achieve seamless handover is developed, an algorithm based on Received Signal Strength (RSS) criterion for network selection is used and Entropy Weighted Method (EWM) is implemented for decision making. Seamless handover using EWM decision-making is demonstrated successfully for a UAV moving across fifth generation (5G) and long-term evolution (LTE) networks via a simulation level analysis. Thus, a solution for UAV-5G communication, specifically the mobility challenge in heterogeneous networks is solved and this work could act as step forward in making UAV-5G architecture integration a possibility.
Keywords: Air to ground, A2G, fifth generation, 5G, handover, mobility, unmanned aerial vehicle, UAV, urban environments.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 43247 Flutter Analysis of Slender Beams with Variable Cross Sections Based on Integral Equation Formulation
Authors: Z. El Felsoufi, L. Azrar
Abstract:
This paper studies a mathematical model based on the integral equations for dynamic analyzes numerical investigations of a non-uniform or multi-material composite beam. The beam is subjected to a sub-tangential follower force and elastic foundation. The boundary conditions are represented by generalized parameterized fixations by the linear and rotary springs. A mathematical formula based on Euler-Bernoulli beam theory is presented for beams with variable cross-sections. The non-uniform section introduces non-uniformity in the rigidity and inertia of beams and consequently, more complicated equilibrium who governs the equation. Using the boundary element method and radial basis functions, the equation of motion is reduced to an algebro-differential system related to internal and boundary unknowns. A generalized formula for the deflection, the slope, the moment and the shear force are presented. The free vibration of non-uniform loaded beams is formulated in a compact matrix form and all needed matrices are explicitly given. The dynamic stability analysis of slender beam is illustrated numerically based on the coalescence criterion. A realistic case related to an industrial chimney is investigated.
Keywords: Chimney, BEM and integral equation formulation, non uniform cross section, vibration and Flutter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 162046 An Axiomatic Model for Development of the Allocated Architecture in Systems Engineering Process
Authors: A. Sharahi, R. Tehrani, A. Mollajan
Abstract:
The final step to complete the “Analytical Systems Engineering Process” is the “Allocated Architecture” in which all Functional Requirements (FRs) of an engineering system must be allocated into their corresponding Physical Components (PCs). At this step, any design for developing the system’s allocated architecture in which no clear pattern of assigning the exclusive “responsibility” of each PC for fulfilling the allocated FR(s) can be found is considered a poor design that may cause difficulties in determining the specific PC(s) which has (have) failed to satisfy a given FR successfully. The present study utilizes the Axiomatic Design method principles to mathematically address this problem and establishes an “Axiomatic Model” as a solution for reaching good alternatives for developing the allocated architecture. This study proposes a “loss Function”, as a quantitative criterion to monetarily compare non-ideal designs for developing the allocated architecture and choose the one which imposes relatively lower cost to the system’s stakeholders. For the case-study, we use the existing design of U. S. electricity marketing subsystem, based on data provided by the U.S. Energy Information Administration (EIA). The result for 2012 shows the symptoms of a poor design and ineffectiveness due to coupling among the FRs of this subsystem.
Keywords: Allocated Architecture, Analytical Systems Engineering Process, Functional Requirements (FRs), Physical Components (PCs), Responsibility of a Physical Component, System’s Stakeholders.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1972