Search results for: analytical method validation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20979

Search results for: analytical method validation

20139 Vibration Behavior of Nanoparticle Delivery in a Single-Walled Carbon Nanotube Using Nonlocal Timoshenko Beam Theory

Authors: Haw-Long Lee, Win-Jin Chang, Yu-Ching Yang

Abstract:

In the paper, the coupled equation of motion for the dynamic displacement of a fullerene moving in a (10,10) single-walled carbon nanotube (SWCNT) is derived using nonlocal Timoshenko beam theory, including the effects of rotary inertia and shear deformation. The effects of confined stiffness between the fullerene and nanotube, foundation stiffness, and nonlocal parameter on the dynamic behavior are analyzed using the Runge-Kutta Method. The numerical solution is in agreement with the analytical result for the special case. The numerical results show that increasing the confined stiffness and foundation stiffness decrease the dynamic displacement of SWCNT. However, the dynamic displacement increases with increasing the nonlocal parameter. In addition, result using the Euler beam theory and the Timoshenko beam theory are compared. It can be found that ignoring the effects of rotary inertia and shear deformation leads to an underestimation of the displacement.

Keywords: single-walled carbon nanotube, nanoparticle delivery, Nonlocal Timoshenko beam theory, Runge-Kutta Method, Van der Waals force

Procedia PDF Downloads 369
20138 Awareness and Utilization of E-Learning Technologies in Teaching and Learning of Human Kinetics and Health Education Courses in Nigeria Universities

Authors: Ibrahim Laro ABUBAKAR

Abstract:

The study examined the Availability and Utilization of E-Learning Technologies in Teaching of Human Kinetics and Health Education courses in Nigerian Universities, specifically, Universities in Kwara State. Two purposes were formulated to guide the study from which two research questions and two hypotheses were raised. The descriptive research design was used in the research. Three Hundred respondents (100 Lecturers and 200 Students) made up the population for the study. There was no sampling, as the population of the study was not much. A structured questionnaire tagged ‘Availability and Utilization of E-Learning Technologies in Teaching and Learning Questionnaire’ (AUETTLQ) was used for data collection. The questionnaire was subjected to face and content validation, and it was equally pilot tested. The validation yielded a reliability coefficient of 0.78. The data collected from the study were statistically analyzed using frequencies and percentage count for personal data of the respondents, mean and standard deviation to answer the research questions. The null hypotheses were tested at 0.05 level of significance using the independent t-test. One among other findings of this study showed that lecturers and Student are aware of synchronous e-learning technologies in teaching and learning of Human Kinetics and Health Education but often utilize the synchronous e-learning technologies. It was recommended among others that lecturers and Students should be sensitized through seminars and workshops on the need to maximally utilize available e-learning technologies in teaching and learning of Human Kinetics and Health Education courses in Universities.

Keywords: awareness, utilization, E-Learning, technologies, human kinetics synchronous

Procedia PDF Downloads 108
20137 Influence of Instrumental Playing on Attachment Type of Musicians and Music Students Using Adult Attachment Scale-R

Authors: Sofia Serra-Dawa

Abstract:

Adult relationships accrue on a variety of past social experiences, intentions, and emotions that might predispose and influence the approach to and construction of subsequent relationships. The Adult Attachment Theory (AAT) proposes four types of adult attachment, where attachment is built over two dimensions of anxiety and avoidance: secure, anxious-preoccupied, dismissive-avoidant, and fearful-avoidant. The AAT has been studied in multiple settings such as personal and therapeutic relationships, educational settings, sexual orientation, health, and religion. In music scholarship, the AAT has been used to frame class learning of student singers and study the relational behavior between voice teachers and students. Building on this study, the present inquiry studies how attachment types might characterize learning relationships of music students (in the Western Conservatory tradition), and whether particular instrumental experiences might correlate to given attachment styles. Given certain behavioral cohesive features of established traditions of instrumental playing and performance modes, it is hypothesized that student musicians will display specific characteristics correlated to instrumental traditions, demonstrating clear tendency of attachment style, which in turn has implications on subsequent professional interactions. This study is informed by the methodological framework of Adult Attachment Scale-R (Collins and Read, 1990), which was particularly chosen given its non-invasive questions and classificatory validation. It is further hypothesized that the analytical comparison of musicians’ profiles has the potential to serve as the baseline for other comparative behavioral observation studies [this component is expected to be verified and completed well before the conference meeting]. This research may have implications for practitioners concerned with matching and improving musical teaching and learning relationships and in (professional and amateur) long-term musical settings.

Keywords: adult attachment, music education, musicians attachment profile, musicians relationships

Procedia PDF Downloads 148
20136 Spirometric Reference Values in 236,606 Healthy, Non-Smoking Chinese Aged 4–90 Years

Authors: Jiashu Shen

Abstract:

Objectives: Spirometry is a basic reference for health evaluation which is widely used in clinical. Previous reference of spirometry is not applicable because of drastic changes of social and natural circumstance in China. A new reference values for the spirometry of the Chinese population is extremely needed. Method: Spirometric reference value was established using the statistical modeling method Generalized Additive Models for Location, Scale and Shape for forced expiratory volume in 1 s (FEV1), forced vital capacity (FVC), FEV1/FVC, and maximal mid-expiratory flow (MMEF). Results: Data from 236,606 healthy non-smokers aged 4–90 years was collected from the MJ Health Check database. Spirometry equations for FEV1, FVC, MMEF, and FEV1/FVC were established, including the predicted values and lower limits of normal (LLNs) by sex. The predictive equations that were developed for the spirometric results elaborated the relationship between spirometry and age, and they eliminated the effects of height as a variable. Most previous predictive equations for Chinese spirometry were significantly overestimated (to be exact, with mean differences of 22.21% in FEV1 and 31.39% in FVC for males, along with differences of 26.93% in FEV1 and 35.76% in FVC for females) or underestimated (with mean differences of -5.81% in MMEF and -14.56% in FEV1/FVC for males, along with a difference of -14.54% in FEV1/FVC for females) the results of lung function measurements as found in this study. Through cross-validation, our equations were established as having good fit, and the means of the measured value and the estimated value were compared, with good results. Conclusions: Our study updates the spirometric reference equations for Chinese people of all ages and provides comprehensive values for both physical examination and clinical diagnosis.

Keywords: Chinese, GAMLSS model, reference values, spirometry

Procedia PDF Downloads 123
20135 Repair and Strengthening of Plain and FRC Shear Deficient Beams Using Externally Bonded CFRP Sheets

Authors: H. S. S. Abou El-Mal, H. E. M. Sallam

Abstract:

This paper presents experimental and analytical study on the behavior of repaired and strengthened shear critical RC beams using externally bonded CFRP bi-directional fabrics. The use of CFRP sheets to repair or strengthen RC beams has been repetitively studied and proven feasible. However, the use of combined repair techniques and applying that method to both plain and FRC beams can maximize the shear capacity of RC shear deficient beams. A total of twelve slender beams were tested under four-point bending. The test parameters included CFRP layout, number of layers and fiber direction, injecting cracks before applying repairing sheets, enhancing the flexural capacity to differentiate between shear repair and strengthening techniques, and concrete matrix types. The findings revealed that applying CFRP sheets increased the overall shear capacity, the amount and orientation of wrapping is of prime importance in both repairing and strengthening, CFRP wrapping could change the failure mode from shear to flexural shear, the use of crack injection combined to CFRP wrapping further improved the shear capacity while, applying the previous method to FRC beams enhanced both shear capacity and failure ductility. Acceptable agreement was found between predicted shear capacities using the Canadian code and the experimental results of the current study.

Keywords: CFRP, FRC, repair, shear strengthening

Procedia PDF Downloads 339
20134 Determination of Verapamil Hydrochloride in Tablets and Injection Solutions With the Verapamil-Selective Electrode and Possibilities of Application in Pharmaceutical Analysis

Authors: Faisal A. Salih

Abstract:

Verapamil hydrochloride (Ver) is a drug used in medicine for arrythmia, angina and hypertension as a calcium channel blocker. For the quantitative determination of Ver in dosage forms, the HPLC method is most often used. A convenient alternative to the chromatographic method is potentiometry using a Verselective electrode, which does not require expensive equipment, can be used without separation from the matrix components, which significantly reduces the analysis time, and does not use toxic organic solvents, being a "green", "environmentally friendly" technique. It has been established in this study that the rational choice of the membrane plasticizer and the preconditioning and measurement algorithms, which prevent nonexchangeable extraction of Ver into the membrane phase, makes it possible to achieve excellent analytical characteristics of Ver-selective electrodes based on commercially available components. In particular, an electrode with the following membrane composition: PVC (32.8 wt %), ortho-nitrophenyloctyl ether (66.6 wt %), and tetrakis-4-chlorophenylborate (0.6 wt % or 0.01 M) have the lower detection limit 4 × 10−8 M and potential reproducibility 0.15–0.22 mV. Both direct potentiometry (DP) and potentiometric titration (PT) methods can be used for the determination of Ver in tablets and injection solutions. Masses of Ver per average tablet weight determined by the methods of DP and PT for the same set of 10 tablets were (80.4±0.2 and80.7±0.2) mg, respectively. The masses of Ver in solutions for injection, determined by DP for two ampoules from one set, were (5.00±0.015 and 5.004±0.006) mg. In all cases, good reproducibility and excellent correspondence with the declared quantities were observed.

Keywords: verapamil, potentiometry, ion-selective electrode, pharmaceutical analysis

Procedia PDF Downloads 82
20133 An Accurate Computer-Aided Diagnosis: CAD System for Diagnosis of Aortic Enlargement by Using Convolutional Neural Networks

Authors: Mahdi Bazarganigilani

Abstract:

Aortic enlargement, also known as an aortic aneurysm, can occur when the walls of the aorta become weak. This disease can become deadly if overlooked and undiagnosed. In this paper, a computer-aided diagnosis (CAD) system was introduced to accurately diagnose aortic enlargement from chest x-ray images. An enhanced convolutional neural network (CNN) was employed and then trained by transfer learning by using three different main areas from the original images. The areas included the left lung, heart, and right lung. The accuracy of the system was then evaluated on 1001 samples by using 4-fold cross-validation. A promising accuracy of 90% was achieved in terms of the F-measure indicator. The results showed using different areas from the original image in the training phase of CNN could increase the accuracy of predictions. This encouraged the author to evaluate this method on a larger dataset and even on different CAD systems for further enhancement of this methodology.

Keywords: computer-aided diagnosis systems, aortic enlargement, chest X-ray, image processing, convolutional neural networks

Procedia PDF Downloads 150
20132 Predicting the Turbulence Intensity, Excess Energy Available and Potential Power Generated by Building Mounted Wind Turbines over Four Major UK City

Authors: Emejeamara Francis

Abstract:

The future of potentials wind energy applications within suburban/urban areas are currently faced with various problems. These include insufficient assessment of urban wind resource, and the effectiveness of commercial gust control solutions as well as unavailability of effective and cheaper valuable tools for scoping the potentials of urban wind applications within built-up environments. In order to achieve effective assessment of the potentials of urban wind installations, an estimation of the total energy that would be available to them were effective control systems to be used, and evaluating the potential power to be generated by the wind system is required. This paper presents a methodology of predicting the power generated by a wind system operating within an urban wind resource. This method was developed by using high temporal resolution wind measurements from eight potential sites within the urban and suburban environment as inputs to a vertical axis wind turbine multiple stream tube model. A relationship between the unsteady performance coefficient obtained from the stream tube model results and turbulence intensity was demonstrated. Hence, an analytical methodology for estimating the unsteady power coefficient at a potential turbine site is proposed. This is combined with analytical models that were developed to predict the wind speed and the excess energy (EEC) available in estimating the potential power generated by wind systems at different heights within a built environment. Estimates of turbulence intensities, wind speed, EEC and turbine performance based on the current methodology allow a more complete assessment of available wind resource and potential urban wind projects. This methodology is applied to four major UK cities namely Leeds, Manchester, London and Edinburgh and the potential to map the turbine performance at different heights within a typical urban city is demonstrated.

Keywords: small-scale wind, turbine power, urban wind energy, turbulence intensity, excess energy content

Procedia PDF Downloads 266
20131 Entropy Generation Minimization in a Porous Pipe Heat Exchanger under Magnetohydrodynamics Using Cattaneo-Christov Heat Flux

Authors: Saima Ijaz, Muhammad Mushtaq, Sufian Munawar

Abstract:

This article is devoted to studying the second law analysis of the Cattaneo-Christov heat flux for non-Newtonian fluid on a moving porous pipe intensification of the magnetic field and heat source/sink. The non-Newtonian fluid is considered to have Maxwell-fluid characteristics. The Cattaneo-Christov model takes into account the specific relaxation time for heat transfer. The main causes that are responsible for creating entropy generation are viscous dissipation, heat transfer, and joule heating. An analytical method, the Homotopy Analysis Method (HAM), is utilized to solve the non-linear governing equations of the underlying model. Mathematical results are shown with graphs and tables. In this work, all those parameters are sorted out which are responsible for an increase or decrease in entropy generation. Namely, the porosity, magnetic field effects, and heat source/sink rate are in the former category, and Cattaneo-Christov relaxation time is in the latter one. These results are new contributions in the case of internal flow in the pipe and would be helpful for reducing the entropy generation strategies.

Keywords: Cattaneo-Christov heat flux, entropy generation analysis, heat source / sink, joule heating, non-newtonian fluid, porous pipe

Procedia PDF Downloads 22
20130 Optimization and Coordination of Organic Product Supply Chains under Competition: An Analytical Modeling Perspective

Authors: Mohammadreza Nematollahi, Bahareh Mosadegh Sedghy, Alireza Tajbakhsh

Abstract:

The last two decades have witnessed substantial attention to organic and sustainable agricultural supply chains. Motivated by real-world practices, this paper aims to address two main challenges observed in organic product supply chains: decentralized decision-making process between farmers and their retailers, and competition between organic products and their conventional counterparts. To this aim, an agricultural supply chain consisting of two farmers, a conventional farmer and an organic farmer who offers an organic version of the same product, is considered. Both farmers distribute their products through a single retailer, where there exists competition between the organic and the conventional product. The retailer, as the market leader, sets the wholesale price, and afterward, the farmers set their production quantity decisions. This paper first models the demand functions of the conventional and organic products by incorporating the effect of asymmetric brand equity, which captures the fact that consumers usually pay a premium for organic due to positive perceptions regarding their health and environmental benefits. Then, profit functions with consideration of some characteristics of organic farming, including crop yield gap and organic cost factor, are modeled. Our research also considers both economies and diseconomies of scale in farming production as well as the effects of organic subsidy paid by the government to support organic farming. This paper explores the investigated supply chain in three scenarios: decentralized, centralized, and coordinated decision-making structures. In the decentralized scenario, the conventional and organic farmers and the retailer maximize their own profits individually. In this case, the interaction between the farmers is modeled under the Bertrand competition, while analyzing the interaction between the retailer and farmers under the Stackelberg game structure. In the centralized model, the optimal production strategies are obtained from the entire supply chain perspective. Analytical models are developed to derive closed-form optimal solutions. Moreover, analytical sensitivity analyses are conducted to explore the effects of main parameters like the crop yield gap, organic cost factor, organic subsidy, and percent price premium of the organic product on the farmers’ and retailer’s optimal strategies. Afterward, a coordination scenario is proposed to convince the three supply chain members to shift from the decentralized to centralized decision-making structure. The results indicate that the proposed coordination scenario provides a win-win-win situation for all three members compared to the decentralized model. Moreover, our paper demonstrates that the coordinated model respectively increases and decreases the production and price of organic produce, which in turn motivates the consumption of organic products in the market. Moreover, the proposed coordination model helps the organic farmer better handle the challenges of organic farming, including the additional cost and crop yield gap. Last but not least, our results highlight the active role of the organic subsidy paid by the government as a means of promoting sustainable organic product supply chains. Our paper shows that although the amount of organic subsidy plays a significant role in the production and sales price of organic products, the allocation method of subsidy between the organic farmer and retailer is not of that importance.

Keywords: analytical game-theoretic model, product competition, supply chain coordination, sustainable organic supply chain

Procedia PDF Downloads 102
20129 Non-Linear Assessment of Chromatographic Lipophilicity of Selected Steroid Derivatives

Authors: Milica Karadžić, Lidija Jevrić, Sanja Podunavac-Kuzmanović, Strahinja Kovačević, Anamarija Mandić, Aleksandar Oklješa, Andrea Nikolić, Marija Sakač, Katarina Penov Gaši

Abstract:

Using chemometric approach, the relationships between the chromatographic lipophilicity and in silico molecular descriptors for twenty-nine selected steroid derivatives were studied. The chromatographic lipophilicity was predicted using artificial neural networks (ANNs) method. The most important in silico molecular descriptors were selected applying stepwise selection (SS) paired with partial least squares (PLS) method. Molecular descriptors with satisfactory variable importance in projection (VIP) values were selected for ANN modeling. The usefulness of generated models was confirmed by detailed statistical validation. High agreement between experimental and predicted values indicated that obtained models have good quality and high predictive ability. Global sensitivity analysis (GSA) confirmed the importance of each molecular descriptor used as an input variable. High-quality networks indicate a strong non-linear relationship between chromatographic lipophilicity and used in silico molecular descriptors. Applying selected molecular descriptors and generated ANNs the good prediction of chromatographic lipophilicity of the studied steroid derivatives can be obtained. This article is based upon work from COST Actions (CM1306 and CA15222), supported by COST (European Cooperation and Science and Technology).

Keywords: artificial neural networks, chemometrics, global sensitivity analysis, liquid chromatography, steroids

Procedia PDF Downloads 334
20128 Crude Oil Electrostatic Mathematical Modelling on an Existing Industrial Plant

Authors: Fatemeh Yazdanmehr, Iulian Nistor

Abstract:

The scope of the current study is the prediction of water separation in a two-stage industrial crude oil desalting plant. This research study was focused on developing a desalting operation in an existing production unit of one Iranian heavy oil field with 75 MBPD capacity. Because of some operational issues, such as oil dehydration at high temperatures, the optimization of the desalter operational parameters was essential. The mathematical desalting is modeled based on the population balance method. The existing operational data is used for tuning and validation of the accuracy of the modeling. The inlet oil temperature to desalter used was decreased from 110°C to 80°C, and the desalted electrical field was increased from 0.75 kv to 2.5 kv. The proposed condition for the desalter also meets the water oil specification. Based on these conditions of desalter, the oil recovery is increased by 574 BBL/D, and the gas flaring decrease by 2.8 MMSCF/D. Depending on the oil price, the additional production of oil can increase the annual income by about $15 MM and reduces greenhouse gas production caused by gas flaring.

Keywords: desalter, demulsification, modelling, water-oil separation, crude oil emulsion

Procedia PDF Downloads 65
20127 Free Vibration Analysis of FG Nanocomposite Sandwich Beams Using Various Higher-Order Beam Theories

Authors: Saeed Kamarian

Abstract:

In this paper, free vibrations of Functionally Graded Sandwich (FGS) beams reinforced by randomly oriented Single-Walled Carbon Nanotubes (SWCNTs) are investigated. The Eshelby–Mori–Tanaka approach based on an equivalent fiber is used to investigate the material properties of the structure. The natural frequencies of the FGS nanocomposite beam are analyzed based on various Higher-order Shear Deformation Beam Theories (HSDBTs) and using an analytical method. The verification study represents the simplicity and accuracy of the method for free vibration analysis of nanocomposite beams. The effects of carbon nanotube volume fraction profiles in the face layers, length to span ratio and thicknesses of face layers on the natural frequency of structure are studied for the different HSDBTs. Results show that by utilizing the FGS type of structures, free vibration characteristics of structures can be improved. A comparison is also provided to show the difference between natural frequency responses of the FGS nanocomposite beam reinforced by aligned and randomly oriented SWCNT.

Keywords: sandwich beam, nanocomposite beam, functionally graded materials, higher-order beam theories, Mori-Tanaka approach

Procedia PDF Downloads 452
20126 Developing a Sustainable Transit Planning Index Using Analytical Hierarchy Process Method for ZEB Implementation in Canada

Authors: Mona Ghafouri-Azar, Sara Diamond, Jeremy Bowes, Grace Yuan, Aimee Burnett, Michelle Wyndham-West, Sara Wagner, Anand Pariyarath

Abstract:

Transportation is the fastest growing source of greenhouse gas emissions worldwide. In Canada, it is responsible for 23% of total CO2emissions from fuel combustion, and emissions from the transportation sector are the second largest source of emissions after the oil and gas sector. Currently, most Canadian public transportation systems rely on buses that operateon fossil fuels.Canada is currently investing billions of dollars to replacediesel buses with electric busesas this isperceived to have a significant impact on climate mitigation. This paper focuses on the possible impacts of zero emission buses (ZEB) on sustainable development, considering three dimensions of sustainability; environmental quality, economic growth, and social development.A sustainable transportation system is one that is safe, affordable, accessible, efficient, and resilient and that contributes minimal emissions of carbon and other pollutants.To enable implementation of these goals, relevant indicators were selected and defined that measure progress towards a sustainable transportation system. These were drawn from Canadian and international examples. Studies compare different European cities in terms of development, sustainability, and infrastructures, by using transport performance indicators. A Normalized Transport Sustainability index measures and compares policies in different urban areas and allows fine-tuning of policies. Analysts use a number ofmethods for sustainable analysis, like cost-benefit analysis (CBA) toassess economic benefit, life-cycle assessment (LCA) to assess social, economic, and environment factors and goals, and multi-criteria decision making (MCDM) analysis which can comparediffering stakeholder preferences.A multi criteria decision making approach is an appropriate methodology to plan and evaluate sustainable transit development and to provide insights and meaningful information for decision makers and transit agencies. It is essential to develop a system thataggregates specific discrete indices to assess the sustainability of transportation systems.Theseprioritize indicators appropriate for the differentCanadian transit system agencies and theirpreferences and requirements. This studywill develop an integrating index that alliesexistingdiscrete indexes to supporta reliable comparison between the current transportation system (diesel buses) and the new ZEB system emerging in Canada. As a first step, theindexes for each category are selected, and the index matrix constructed. Second, the selected indicators arenormalized to remove anyinconsistency between them. Next, the normalized matrix isweighted based on the relative importance of each index to the main domains of sustainability using the analytical hierarchy process (AHP) method. This is accomplished through expert judgement around the relative importance of different attributes with respect to the goals through apairwise comparison matrix. The considerationof multiple environmental, economic, and social factors (including equity and health) is integrated intoa sustainable transit planning index (STPI) which supportsrealistic ZEB implementation in Canada and beyond and is useful to different stakeholders, agencies, and ministries.

Keywords: zero emission buses, sustainability, sustainable transit, transportation, analytical hierarchy process, environment, economy, social

Procedia PDF Downloads 120
20125 Programming without Code: An Approach and Environment to Conditions-On-Data Programming

Authors: Philippe Larvet

Abstract:

This paper presents the concept of an object-based programming language where tests (if... then... else) and control structures (while, repeat, for...) disappear and are replaced by conditions on data. According to the object paradigm, by using this concept, data are still embedded inside objects, as variable-value couples, but object methods are expressed into the form of logical propositions (‘conditions on data’ or COD).For instance : variable1 = value1 AND variable2 > value2 => variable3 = value3. Implementing this approach, a central inference engine turns and examines objects one after another, collecting all CODs of each object. CODs are considered as rules in a rule-based system: the left part of each proposition (left side of the ‘=>‘ sign) is the premise and the right part is the conclusion. So, premises are evaluated and conclusions are fired. Conclusions modify the variable-value couples of the object and the engine goes to examine the next object. The paper develops the principles of writing CODs instead of complex algorithms. Through samples, the paper also presents several hints for implementing a simple mechanism able to process this ‘COD language’. The proposed approach can be used within the context of simulation, process control, industrial systems validation, etc. By writing simple and rigorous conditions on data, instead of using classical and long-to-learn languages, engineers and specialists can easily simulate and validate the functioning of complex systems.

Keywords: conditions on data, logical proposition, programming without code, object-oriented programming, system simulation, system validation

Procedia PDF Downloads 211
20124 Flow Reproduction Using Vortex Particle Methods for Wake Buffeting Analysis of Bluff Structures

Authors: Samir Chawdhury, Guido Morgenthal

Abstract:

The paper presents a novel extension of Vortex Particle Methods (VPM) where the study aims to reproduce a template simulation of complex flow field that is generated from impulsively started flow past an upstream bluff body at certain Reynolds number Re-Vibration of a structural system under upstream wake flow is often considered its governing design criteria. Therefore, the attention is given in this study especially for the reproduction of wake flow simulation. The basic methodology for the implementation of the flow reproduction requires the downstream velocity sampling from the template flow simulation; therefore, at particular distances from the upstream section the instantaneous velocity components are sampled using a series of square sampling-cells arranged vertically where each of the cell contains four velocity sampling points at its corner. Since the grid free Lagrangian VPM algorithm discretises vorticity on particle elements, the method requires transformation of the velocity components into vortex circulation, and finally the simulation of the reproduction of the template flow field by seeding these vortex circulations or particles into a free stream flow. It is noteworthy that the vortex particles have to be released into the free stream exactly at same rate of velocity sampling. Studies have been done, specifically, in terms of different sampling rates and velocity sampling positions to find their effects on flow reproduction quality. The quality assessments are mainly done, using a downstream flow monitoring profile, by comparing the characteristic wind flow profiles using several statistical turbulence measures. Additionally, the comparisons are performed using velocity time histories, snapshots of the flow fields, and the vibration of a downstream bluff section by performing wake buffeting analyses of the section under the original and reproduced wake flows. Convergence study is performed for the validation of the method. The study also describes the possibilities how to achieve flow reproductions with less computational effort.

Keywords: vortex particle method, wake flow, flow reproduction, wake buffeting analysis

Procedia PDF Downloads 304
20123 Assessment of Physical Learning Environments in ECE: Interdisciplinary and Multivocal Innovation for Chilean Kindergartens

Authors: Cynthia Adlerstein

Abstract:

Physical learning environment (PLE) has been considered, after family and educators, as the third teacher. There have been conflicting and converging viewpoints on the role of the physical dimensions of places to learn, in facilitating educational innovation and quality. Despite the different approaches, PLE has been widely recognized as a key factor in the quality of the learning experience , and in the levels of learning achievement in ECE . The conceptual frameworks of the field assume that PLE consists of a complex web of factors that shape the overall conditions for learning, and that much more interdisciplinary and complementary methodologies of research and development are required. Although the relevance of PLE attracts a broad international consensus, in Chile it remains under-researched and weakly regulated by public policy. Gaining deeper contextual understanding and more thoughtfully-designed recommendations require the use of innovative assessment tools that cross cultural and disciplinary boundaries to produce new hybrid approaches and improvements. When considering a PLE-based change process for ECE improvement, a central question is what dimensions, variables and indicators could allow a comprehensive assessment of PLE in Chilean kindergartens? Based on a grounded theory social justice inquiry, we adopted a mixed method design, that enabled a multivocal and interdisciplinary construction of data. By using in-depth interviews, discussion groups, questionnaires, and documental analysis, we elicited the PLE discourses of politicians, early childhood practitioners, experts in architectural design and ergonomics, ECE stakeholders, and 3 to 5 year olds. A constant comparison method enabled the construction of the dimensions, variables and indicators through which PLE assessment is possible. Subsequently, the instrument was applied in a sample of 125 early childhood classrooms, to test reliability (internal consistency) and validity (content and construct). As a result, an interdisciplinary and multivocal tool for assessing physical learning environments was constructed and validated, for Chilean kindergartens. The tool is structured upon 7 dimensions (wellbeing, flexible, empowerment, inclusiveness, symbolically meaningful, pedagogically intentioned, institutional management) 19 variables and 105 indicators that are assessed through observation and registration on a mobile app. The overall reliability of the instrument is .938 while the consistency of each dimension varies between .773 (inclusive) and .946 (symbolically meaningful). The validation process through expert opinion and factorial analysis (chi-square test) has shown that the dimensions of the assessment tool reflect the factors of physical learning environments. The constructed assessment tool for kindergartens highlights the significance of the physical environment in early childhood educational settings. The relevance of the instrument relies in its interdisciplinary approach to PLE and in its capability to guide innovative learning environments, based on educational habitability. Though further analysis are required for concurrent validation and standardization, the tool has been considered by practitioners and ECE stakeholders as an intuitive, accessible and remarkable instrument to arise awareness on PLE and on equitable distribution of learning opportunities.

Keywords: Chilean kindergartens, early childhood education, physical learning environment, third teacher

Procedia PDF Downloads 345
20122 Clinical Validation of an Automated Natural Language Processing Algorithm for Finding COVID-19 Symptoms and Complications in Patient Notes

Authors: Karolina Wieczorek, Sophie Wiliams

Abstract:

Introduction: Patient data is often collected in Electronic Health Record Systems (EHR) for purposes such as providing care as well as reporting data. This information can be re-used to validate data models in clinical trials or in epidemiological studies. Manual validation of automated tools is vital to pick up errors in processing and to provide confidence in the output. Mentioning a disease in a discharge letter does not necessarily mean that a patient suffers from this disease. Many of them discuss a diagnostic process, different tests, or discuss whether a patient has a certain disease. The COVID-19 dataset in this study used natural language processing (NLP), an automated algorithm which extracts information related to COVID-19 symptoms, complications, and medications prescribed within the hospital. Free-text patient clinical patient notes are rich sources of information which contain patient data not captured in a structured form, hence the use of named entity recognition (NER) to capture additional information. Methods: Patient data (discharge summary letters) were exported and screened by an algorithm to pick up relevant terms related to COVID-19. Manual validation of automated tools is vital to pick up errors in processing and to provide confidence in the output. A list of 124 Systematized Nomenclature of Medicine (SNOMED) Clinical Terms has been provided in Excel with corresponding IDs. Two independent medical student researchers were provided with a dictionary of SNOMED list of terms to refer to when screening the notes. They worked on two separate datasets called "A” and "B”, respectively. Notes were screened to check if the correct term had been picked-up by the algorithm to ensure that negated terms were not picked up. Results: Its implementation in the hospital began on March 31, 2020, and the first EHR-derived extract was generated for use in an audit study on June 04, 2020. The dataset has contributed to large, priority clinical trials (including International Severe Acute Respiratory and Emerging Infection Consortium (ISARIC) by bulk upload to REDcap research databases) and local research and audit studies. Successful sharing of EHR-extracted datasets requires communicating the provenance and quality, including completeness and accuracy of this data. The results of the validation of the algorithm were the following: precision (0.907), recall (0.416), and F-score test (0.570). Percentage enhancement with NLP extracted terms compared to regular data extraction alone was low (0.3%) for relatively well-documented data such as previous medical history but higher (16.6%, 29.53%, 30.3%, 45.1%) for complications, presenting illness, chronic procedures, acute procedures respectively. Conclusions: This automated NLP algorithm is shown to be useful in facilitating patient data analysis and has the potential to be used in more large-scale clinical trials to assess potential study exclusion criteria for participants in the development of vaccines.

Keywords: automated, algorithm, NLP, COVID-19

Procedia PDF Downloads 92
20121 Constant Order Predictor Corrector Method for the Solution of Modeled Problems of First Order IVPs of ODEs

Authors: A. A. James, A. O. Adesanya, M. R. Odekunle, D. G. Yakubu

Abstract:

This paper examines the development of one step, five hybrid point method for the solution of first order initial value problems. We adopted the method of collocation and interpolation of power series approximate solution to generate a continuous linear multistep method. The continuous linear multistep method was evaluated at selected grid points to give the discrete linear multistep method. The method was implemented using a constant order predictor of order seven over an overlapping interval. The basic properties of the derived corrector was investigated and found to be zero stable, consistent and convergent. The region of absolute stability was also investigated. The method was tested on some numerical experiments and found to compete favorably with the existing methods.

Keywords: interpolation, approximate solution, collocation, differential system, half step, converges, block method, efficiency

Procedia PDF Downloads 325
20120 Development of a Matlab® Program for the Bi-Dimensional Truss Analysis Using the Stiffness Matrix Method

Authors: Angel G. De Leon Hernandez

Abstract:

A structure is defined as a physical system or, in certain cases, an arrangement of connected elements, capable of bearing certain loads. The structures are presented in every part of the daily life, e.g., in the designing of buildings, vehicles and mechanisms. The main goal of a structure designer is to develop a secure, aesthetic and maintainable system, considering the constraint imposed to every case. With the advances in the technology during the last decades, the capabilities of solving engineering problems have increased enormously. Nowadays the computers, play a critical roll in the structural analysis, pitifully, for university students the vast majority of these software are inaccessible due to the high complexity and cost they represent, even when the software manufacturers offer student versions. This is exactly the reason why the idea of developing a more reachable and easy-to-use computing tool. This program is designed as a tool for the university students enrolled in courser related to the structures analysis and designs, as a complementary instrument to achieve a better understanding of this area and to avoid all the tedious calculations. Also, the program can be useful for graduated engineers in the field of structural design and analysis. A graphical user interphase is included in the program to make it even simpler to operate it and understand the information requested and the obtained results. In the present document are included the theoretical basics in which the program is based to solve the structural analysis, the logical path followed in order to develop the program, the theoretical results, a discussion about the results and the validation of those results.

Keywords: stiffness matrix method, structural analysis, Matlab® applications, programming

Procedia PDF Downloads 111
20119 Knowledge and Organisational Success: Developing a Scale of Knowledge Framework

Authors: Mohammed Almohammedali, David Edgar, Duncan Peter

Abstract:

The aim of this exploratory research is to further understand how organisations can evaluate their activities, which generate knowledge creation, to meet changing stakeholder expectations. A Scale of Knowledge (SoK) Framework is proposed which links knowledge management and organisational activities to changing stakeholder expectations. The framework was informed by the knowledge management literature, as well as empirical work conducted via a single case study of a multi-site hospital organisation in Saudi Arabia. Eight in-depth semi-structured interviews were conducted with managers from across the organisation regarding current and future stakeholder expectations, organisational strategy/activities and knowledge management. Data were analysed using thematic analysis and a hierarchical value map technique to identify activities that can produce further knowledge and consequently impact on how stakeholder expectations are met. The SoK Framework developed may be useful to practitioners as an analytical aid to determine if current organisational activities produce organisational knowledge which helps them meet (increasingly higher levels of) stakeholder expectations. The limitations of the research and avenues for future development of the proposed framework are discussed.

Keywords: knowledge creation, knowledge management, organisational knowledge, analytical aid, stakeholders

Procedia PDF Downloads 423
20118 Geosynthetic Reinforced Unpaved Road: Literature Study and Design Example

Authors: D. Jayalakshmi, S. S. Bhosale

Abstract:

This paper, in its first part, presents the state-of-the-art literature of design approaches for geosynthetic reinforced unpaved roads. The literature starting since 1970 and the critical appraisal of flexible pavement design by Giroud and Han (2004) and Jonathan Fannin (2006) is presented. The design example is illustrated for Indian conditions. The example emphasizes the results computed by Giroud and Han's (2004) design method with the Indian road congress guidelines by IRC SP 72 -2015. The input data considered are related to the subgrade soil condition of Maharashtra State in India. The unified soil classification of the subgrade soil is inorganic clay with high plasticity (CH), which is expansive with a California bearing ratio (CBR) of 2% to 3%. The example exhibits the unreinforced case and geotextile as reinforcement by varying the rut depth from 25 mm to 100 mm. The present result reveals the base thickness for the unreinforced case from the IRC design catalogs is in good agreement with Giroud and Han (2004) approach for a range of 75 mm to 100 mm rut depth. Since Giroud and Han (2004) method is applicable for both reinforced and unreinforced cases, for the same data with appropriate Nc factor, for the same rut depth, the base thickness for the reinforced case has arrived for the Indian condition. From this trial, for the CBR of 2%, the base thickness reduction due to geotextile inclusion is 35%. For the CBR range of 2% to 5% with different stiffness in geosynthetics, the reduction in base course thickness will be evaluated, and the validation will be executed by the full-scale accelerated pavement testing set up at the College of Engineering Pune (COE), India.

Keywords: base thickness, design approach, equation, full scale accelerated pavement set up, Indian condition

Procedia PDF Downloads 183
20117 Gender and Older People: Reframing Gender Analysis through Lifecycle Lens

Authors: Supriya Akerkar

Abstract:

The UN Decade on Healthy Ageing (2021-2030) provides a new opportunity to address ageing and gender issues in different societies. The concept of gender has been used to unpack and analyse the power and constructions of gender relations in different societies. Such analysis has been employed and used to inform policy and practices of governments and non-governmental organisations to further gender equalities in their work. Yet, experiences of older women and men are often left out of such mainstream gender analysis, marginalising their existence and issues. This paper argues that new critical analytical tools are needed to capture the realities and issues of interest to older women and men. In particular, it argues that gender analysis needs to integrate analytical concepts of ageing and lifecycle approach in its framework. The paper develops such a framework by critical interrogation of the gender analysis tools that are currently applied for framing gender issues in international development and humanitarian work. Informed by the realities and experiences of older women and men, developed through a synthesis of available literature, the paper will develop a new framework for gender analysis that can be used by governments and non-government organisations in their work to further gender justice across the life cycle.

Keywords: ageing, gender, older people, social inclusion

Procedia PDF Downloads 233
20116 Role of Cellulose Fibers in Tuning the Microstructure and Crystallographic Phase of α-Fe₂O₃ and α-FeOOH Nanoparticles

Authors: Indu Chauhan, Bhupendra S. Butola, Paritosh Mohanty

Abstract:

It is very well known that properties of material changes as their size approach to nanoscale level due to the high surface area to volume ratio. However, in last few decades, a tenet ‘structure dictates function’ is quickly being adopted by researchers working with nanomaterials. The design and exploitation of nanoparticles with tailored shape and size has become one of the primary goals of materials science researchers to expose the properties of nanostructures. To date, various methods, including soft/hard template/surfactant assisted route hydrothermal reaction, seed mediated growth method, capping molecule-assisted synthesis, polyol process, etc. have been adopted to synthesize the nanostructures with controlled size and shape and monodispersity. However controlling the shape and size of nanoparticles is an ultimate challenge of modern material research. In particular, many efforts have been devoted to rational and skillful control of hierarchical and complex nanostructures. Thus in our research work, role of cellulose in manipulating the nanostructures has been discussed. Nanoparticles of α-Fe₂O₃ (diameter ca. 15 to 130 nm) were immobilized on the cellulose fiber surface by a single step in situ hydrothermal method. However, nanoflakes of α-FeOOH having thickness ca. ~25 nm and length ca. ~250 nm were obtained by the same method in absence of cellulose fibers. A possible nucleation and growth mechanism of the formation of nanostructures on cellulose fibers have been proposed. The covalent bond formation between the cellulose fibers and nanostructures has been discussed with supporting evidence from the spectroscopic and other analytical studies such as Fourier transform infrared spectroscopy and X-ray photoelectron spectroscopy. The role of cellulose in manipulating the nanostructures has been discussed.

Keywords: cellulose fibers, α-Fe₂O₃, α-FeOOH, hydrothermal, nanoflakes, nanoparticles

Procedia PDF Downloads 140
20115 Validation of SWAT Model for Prediction of Water Yield and Water Balance: Case Study of Upstream Catchment of Jebba Dam in Nigeria

Authors: Adeniyi G. Adeogun, Bolaji F. Sule, Adebayo W. Salami, Michael O. Daramola

Abstract:

Estimation of water yield and water balance in a river catchment is critical to the sustainable management of water resources at watershed level in any country. Therefore, in the present study, Soil and Water Assessment Tool (SWAT) interfaced with Geographical Information System (GIS) was applied as a tool to predict water balance and water yield of a catchment area in Nigeria. The catchment area, which was 12,992km2, is located upstream Jebba hydropower dam in North central part of Nigeria. In this study, data on the observed flow were collected and compared with simulated flow using SWAT. The correlation between the two data sets was evaluated using statistical measures, such as, Nasch-Sucliffe Efficiency (NSE) and coefficient of determination (R2). The model output shows a good agreement between the observed flow and simulated flow as indicated by NSE and R2, which were greater than 0.7 for both calibration and validation period. A total of 42,733 mm of water was predicted by the calibrated model as the water yield potential of the basin for a simulation period 1985 to 2010. This interesting performance obtained with SWAT model suggests that SWAT model could be a promising tool to predict water balance and water yield in sustainable management of water resources. In addition, SWAT could be applied to other water resources in other basins in Nigeria as a decision support tool for sustainable water management in Nigeria.

Keywords: GIS, modeling, sensitivity analysis, SWAT, water yield, watershed level

Procedia PDF Downloads 426
20114 Large Eddy Simulation of Hydrogen Deflagration in Open Space and Vented Enclosure

Authors: T. Nozu, K. Hibi, T. Nishiie

Abstract:

This paper discusses the applicability of the numerical model for a damage prediction method of the accidental hydrogen explosion occurring in a hydrogen facility. The numerical model was based on an unstructured finite volume method (FVM) code “NuFD/FrontFlowRed”. For simulating unsteady turbulent combustion of leaked hydrogen gas, a combination of Large Eddy Simulation (LES) and a combustion model were used. The combustion model was based on a two scalar flamelet approach, where a G-equation model and a conserved scalar model expressed a propagation of premixed flame surface and a diffusion combustion process, respectively. For validation of this numerical model, we have simulated the previous two types of hydrogen explosion tests. One is open-space explosion test, and the source was a prismatic 5.27 m3 volume with 30% of hydrogen-air mixture. A reinforced concrete wall was set 4 m away from the front surface of the source. The source was ignited at the bottom center by a spark. The other is vented enclosure explosion test, and the chamber was 4.6 m × 4.6 m × 3.0 m with a vent opening on one side. Vent area of 5.4 m2 was used. Test was performed with ignition at the center of the wall opposite the vent. Hydrogen-air mixtures with hydrogen concentrations close to 18% vol. were used in the tests. The results from the numerical simulations are compared with the previous experimental data for the accuracy of the numerical model, and we have verified that the simulated overpressures and flame time-of-arrival data were in good agreement with the results of the previous two explosion tests.

Keywords: deflagration, large eddy simulation, turbulent combustion, vented enclosure

Procedia PDF Downloads 236
20113 A Low-Cost of Foot Plantar Shoes for Gait Analysis

Authors: Zulkifli Ahmad, Mohd Razlan Azizan, Nasrul Hadi Johari

Abstract:

This paper presents a study on development and conducting of a wearable sensor system for gait analysis measurement. For validation, the method of plantar surface measurement by force plate was prepared. In general gait analysis, force plate generally represents a studies about barefoot in whole steps and do not allow analysis of repeating movement step in normal walking and running. The measurements that were usually perform do not represent the whole daily plantar pressures in the shoe insole and only obtain the ground reaction force. The force plate measurement is usually limited a few step and it is done indoor and obtaining coupling information from both feet during walking is not easily obtained. Nowadays, in order to measure pressure for a large number of steps and obtain pressure in each insole part, it could be done by placing sensors within an insole. With this method, it will provide a method for determine the plantar pressures while standing, walking or running of a shoe wearing subject. Inserting pressure sensors in the insole will provide specific information and therefore the point of the sensor placement will result in obtaining the critical part under the insole. In the wearable shoe sensor project, the device consists left and right shoe insole with ten FSR. Arduino Mega was used as a micro-controller that read the analog input from FSR. The analog inputs were transmitted via bluetooth data transmission that gains the force data in real time on smartphone. Blueterm software which is an android application was used as an interface to read the FSR reading on the shoe wearing subject. The subject consist of two healthy men with different age and weight doing test while standing, walking (1.5 m/s), jogging (5 m/s) and running (9 m/s) on treadmill. The data obtain will be saved on the android device and for making an analysis and comparison graph.

Keywords: gait analysis, plantar pressure, force plate, earable sensor

Procedia PDF Downloads 442
20112 Limits Problem Solving in Engineering Careers: Competences and Errors

Authors: Veronica Diaz Quezada

Abstract:

In this article, the performance and errors are featured and analysed in the limit problems solving of a real-valued function, in correspondence to competency-based education in engineering careers, in the south of Chile. The methodological component is contextualised in a qualitative research, with a descriptive and explorative design, with elaboration, content validation and application of quantitative instruments, consisting of two parallel forms of open answer tests, based on limit application problems. The mathematical competences and errors made by students from five engineering careers from a public University are identified and characterized. Results show better performance only to solve routine-context problem-solving competence, thus they are oriented towards a rational solution or they use a suitable problem-solving method, achieving the correct solution. Regarding errors, most of them are related to techniques and the incorrect use of theorems and definitions of real-valued function limits of real variable.

Keywords: engineering education, errors, limits, mathematics competences, problem solving

Procedia PDF Downloads 143
20111 Artificial Neural Network Approach for GIS-Based Soil Macro-Nutrients Mapping

Authors: Shahrzad Zolfagharnassab, Abdul Rashid Mohamed Shariff, Siti Khairunniza Bejo

Abstract:

Conventional methods for nutrient soil mapping are based on laboratory tests of samples that are obtained from surveys. The time and cost involved in gathering and analyzing soil samples are the reasons that researchers use Predictive Soil Mapping (PSM). PSM can be defined as the development of a numerical or statistical model of the relationship among environmental variables and soil properties, which is then applied to a geographic database to create a predictive map. Kriging is a group of geostatistical techniques to spatially interpolate point values at an unobserved location from observations of values at nearby locations. The main problem with using kriging as an interpolator is that it is excessively data-dependent and requires a large number of closely spaced data points. Hence, there is a need to minimize the number of data points without sacrificing the accuracy of the results. In this paper, an Artificial Neural Networks (ANN) scheme was used to predict macronutrient values at un-sampled points. ANN has become a popular tool for prediction as it eliminates certain difficulties in soil property prediction, such as non-linear relationships and non-normality. Back-propagation multilayer feed-forward network structures were used to predict nitrogen, phosphorous and potassium values in the soil of the study area. A limited number of samples were used in the training, validation and testing phases of ANN (pattern reconstruction structures) to classify soil properties and the trained network was used for prediction. The soil analysis results of samples collected from the soil survey of block C of Sawah Sempadan, Tanjung Karang rice irrigation project at Selangor of Malaysia were used. Soil maps were produced by the Kriging method using 236 samples (or values) that were a combination of actual values (obtained from real samples) and virtual values (neural network predicted values). For each macronutrient element, three types of maps were generated with 118 actual and 118 virtual values, 59 actual and 177 virtual values, and 30 actual and 206 virtual values, respectively. To evaluate the performance of the proposed method, for each macronutrient element, a base map using 236 actual samples and test maps using 118, 59 and 30 actual samples respectively produced by the Kriging method. A set of parameters was defined to measure the similarity of the maps that were generated with the proposed method, termed the sample reduction method. The results show that the maps that were generated through the sample reduction method were more accurate than the corresponding base maps produced through a smaller number of real samples. For example, nitrogen maps that were produced from 118, 59 and 30 real samples have 78%, 62%, 41% similarity, respectively with the base map (236 samples) and the sample reduction method increased similarity to 87%, 77%, 71%, respectively. Hence, this method can reduce the number of real samples and substitute ANN predictive samples to achieve the specified level of accuracy.

Keywords: artificial neural network, kriging, macro nutrient, pattern recognition, precision farming, soil mapping

Procedia PDF Downloads 61
20110 Experimental Characterization of Anti-Icing System and Accretion of Re-Emitted Droplets on Turbojet Engine Blades

Authors: Guillaume Linassier, Morgan Balland, Hugo Pervier, Marie Pervier, David Hammond

Abstract:

Atmospheric icing for turbojet is caused by ingestion of super-cooled water droplets. To prevent operability risks, manufacturer can implement ice protection systems. Thermal systems are commonly used for this purpose, but their activation can cause the formation of a water liquid film, that can freeze downstream the heated surface or even on other components. In the framework of STORM, a European project dedicated to icing physics in turbojet engines, a cascade rig representative of engine inlet blades was built and tested in an icing wind tunnel. This mock-up integrates two rows of blades, the upstream one being anti-iced using an electro-thermal device the downstream one being unheated. Under icing conditions, the anti-icing system is activated and set at power level to observe a liquid film on the surface and droplet re-emission at the trailing edge. These re-emitted droplets will impinge on the downstream row and contribute to ice accretion. A complete experimental database was generated, including the characterization of ice accretion shapes, and the characterization of electro-thermal anti-icing system (power limit for apparition of the runback water or ice accretion). These data will be used for validation of numerical tools for modeling thermal anti-icing systems in the scope of engine application, as well as validation of re-emission droplets model for stator parts.

Keywords: turbomachine, anti-icing, cascade rig, runback water

Procedia PDF Downloads 177