Search results for: mathematical music modelling
578 Load-Enabled Deployment and Sensing Range Optimization for Lifetime Enhancement of WSNs
Authors: Krishan P. Sharma, T. P. Sharma
Abstract:
Wireless sensor nodes are resource constrained battery powered devices usually deployed in hostile and ill-disposed areas to cooperatively monitor physical or environmental conditions. Due to their limited power supply, the major challenge for researchers is to utilize their battery power for enhancing the lifetime of whole network. Communication and sensing are two major sources of energy consumption in sensor networks. In this paper, we propose a deployment strategy for enhancing the average lifetime of a sensor network by effectively utilizing communication and sensing energy to provide full coverage. The proposed scheme is based on the fact that due to heavy relaying load, sensor nodes near to the sink drain energy at much faster rate than other nodes in the network and consequently die much earlier. To cover this imbalance, proposed scheme finds optimal communication and sensing ranges according to effective load at each node and uses a non-uniform deployment strategy where there is a comparatively high density of nodes near to the sink. Probable relaying load factor at particular node is calculated and accordingly optimal communication distance and sensing range for each sensor node is adjusted. Thus, sensor nodes are placed at locations that optimize energy during network operation. Formal mathematical analysis for calculating optimized locations is reported in present work.Keywords: load factor, network lifetime, non-uniform deployment, sensing range
Procedia PDF Downloads 383577 Boundary Layer Flow of a Casson Nanofluid Past a Vertical Exponentially Stretching Cylinder in the Presence of a Transverse Magnetic Field with Internal Heat Generation/Absorption
Authors: G. Sarojamma, K. Vendabai
Abstract:
An analysis is carried out to investigate the effect of magnetic field and heat source on the steady boundary layer flow and heat transfer of a Casson nanofluid over a vertical cylinder stretching exponentially along its radial direction. Using a similarity transformation, the governing mathematical equations, with the boundary conditions are reduced to a system of coupled, non –linear ordinary differential equations. The resulting system is solved numerically by the fourth order Runge – Kutta scheme with shooting technique. The influence of various physical parameters such as Reynolds number, Prandtl number, magnetic field, Brownian motion parameter, thermophoresis parameter, Lewis number and the natural convection parameter are presented graphically and discussed for non – dimensional velocity, temperature and nanoparticle volume fraction. Numerical data for the skin – friction coefficient, local Nusselt number and the local Sherwood number have been tabulated for various parametric conditions. It is found that the local Nusselt number is a decreasing function of Brownian motion parameter Nb and the thermophoresis parameter Nt.Keywords: casson nanofluid, boundary layer flow, internal heat generation/absorption, exponentially stretching cylinder, heat transfer, brownian motion, thermophoresis
Procedia PDF Downloads 389576 Determination of Surface Deformations with Global Navigation Satellite System Time Series
Authors: Ibrahim Tiryakioglu, Mehmet Ali Ugur, Caglar Ozkaymak
Abstract:
The development of GNSS technology has led to increasingly widespread and successful applications of GNSS surveys for monitoring crustal movements. However, multi-period GPS survey solutions have not been applied in monitoring vertical surface deformation. This study uses long-term GNSS time series that are required to determine vertical deformations. In recent years, the surface deformations that are parallel and semi-parallel to Bolvadin fault have occurred in Western Anatolia. These surface deformations have continued to occur in Bolvadin settlement area that is located mostly on alluvium ground. Due to these surface deformations, a number of cracks in the buildings located in the residential areas and breaks in underground water and sewage systems have been observed. In order to determine the amount of vertical surface deformations, two continuous GNSS stations have been established in the region. The stations have been operating since 2015 and 2017, respectively. In this study, GNSS observations from the mentioned two GNSS stations were processed with GAMIT/GLOBK (GNSS Analysis Massachusetts Institute of Technology/GLOBal Kalman) program package to create a coordinate time series. With the time series analyses, the GNSS stations’ behavior models (linear, periodical, etc.), the causes of these behaviors, and mathematical models were determined. The study results from the time series analysis of these two 2 GNSS stations shows approximately 50-80 mm/yr vertical movement.Keywords: Bolvadin fault, GAMIT, GNSS time series, surface deformations
Procedia PDF Downloads 165575 Integration of Two Thermodynamic Cycles by Absorption for Simultaneous Production of Fresh Water and Cooling
Authors: Javier Delgado-Gonzaga, Wilfrido Rivera, David Juárez-Romero
Abstract:
Cooling and water purification are processes that have contributed to the economic and social development of the modern world. However, these processes require a significant amount of energy globally. Nowadays, absorption heat pumps have been studied with great interest since they are capable of producing cooling and/or purifying water from low-temperature energy sources such as industrial waste heat or renewable energy. In addition, absorption heat pumps require negligible amounts of electricity for their operation and generally use working fluids that do not represent a risk to the environment. The objective of this work is to evaluate a system that integrates an absorption heat transformer and an absorption cooling system to produce fresh water and cooling from a low-temperature heat source. Both cycles operate with the working pair LiBr-H2O. The integration is possible through the interaction of the LiBr-H2O solution streams between both cycles and also by recycling heat from the absorption heat transformer to the absorption cooling system. Mathematical models were developed to compare the performance of four different configurations. The results showed that the configuration in which the hottest streams of LiBr-H2O solution preheated the coldest streams in the economizers of both cycles was one that achieved the best performance. The interaction of the solution currents and the heat recycling analyzed in this work serves as a record of the possibilities of integration between absorption cycles for cogeneration.Keywords: absorption heat transformer, absorption cooling system, water desalination, integrated system
Procedia PDF Downloads 78574 Algorithm Development of Individual Lumped Parameter Modelling for Blood Circulatory System: An Optimization Study
Authors: Bao Li, Aike Qiao, Gaoyang Li, Youjun Liu
Abstract:
Background: Lumped parameter model (LPM) is a common numerical model for hemodynamic calculation. LPM uses circuit elements to simulate the human blood circulatory system. Physiological indicators and characteristics can be acquired through the model. However, due to the different physiological indicators of each individual, parameters in LPM should be personalized in order for convincing calculated results, which can reflect the individual physiological information. This study aimed to develop an automatic and effective optimization method to personalize the parameters in LPM of the blood circulatory system, which is of great significance to the numerical simulation of individual hemodynamics. Methods: A closed-loop LPM of the human blood circulatory system that is applicable for most persons were established based on the anatomical structures and physiological parameters. The patient-specific physiological data of 5 volunteers were non-invasively collected as personalized objectives of individual LPM. In this study, the blood pressure and flow rate of heart, brain, and limbs were the main concerns. The collected systolic blood pressure, diastolic blood pressure, cardiac output, and heart rate were set as objective data, and the waveforms of carotid artery flow and ankle pressure were set as objective waveforms. Aiming at the collected data and waveforms, sensitivity analysis of each parameter in LPM was conducted to determine the sensitive parameters that have an obvious influence on the objectives. Simulated annealing was adopted to iteratively optimize the sensitive parameters, and the objective function during optimization was the root mean square error between the collected waveforms and data and simulated waveforms and data. Each parameter in LPM was optimized 500 times. Results: In this study, the sensitive parameters in LPM were optimized according to the collected data of 5 individuals. Results show a slight error between collected and simulated data. The average relative root mean square error of all optimization objectives of 5 samples were 2.21%, 3.59%, 4.75%, 4.24%, and 3.56%, respectively. Conclusions: Slight error demonstrated good effects of optimization. The individual modeling algorithm developed in this study can effectively achieve the individualization of LPM for the blood circulatory system. LPM with individual parameters can output the individual physiological indicators after optimization, which are applicable for the numerical simulation of patient-specific hemodynamics.Keywords: blood circulatory system, individual physiological indicators, lumped parameter model, optimization algorithm
Procedia PDF Downloads 137573 Prediction of Fluid Induced Deformation using Cavity Expansion Theory
Authors: Jithin S. Kumar, Ramesh Kannan Kandasami
Abstract:
Geomaterials are generally porous in nature due to the presence of discrete particles and interconnected voids. The porosity present in these geomaterials play a critical role in many engineering applications such as CO2 sequestration, well bore strengthening, enhanced oil and hydrocarbon recovery, hydraulic fracturing, and subsurface waste storage. These applications involves solid-fluid interactions, which govern the changes in the porosity which in turn affect the permeability and stiffness of the medium. Injecting fluid into the geomaterials results in permeation which exhibits small or negligible deformation of the soil skeleton followed by cavity expansion/ fingering/ fracturing (different forms of instabilities) due to the large deformation especially when the flow rate is greater than the ability of the medium to permeate the fluid. The complexity of this problem increases as the geomaterial behaves like a solid and fluid under certain conditions. Thus it is important to understand this multiphysics problem where in addition to the permeation, the elastic-plastic deformation of the soil skeleton plays a vital role during fluid injection. The phenomenon of permeation and cavity expansion in porous medium has been studied independently through extensive experimental and analytical/ numerical models. The analytical models generally use Darcy's/ diffusion equations to capture the fluid flow during permeation while elastic-plastic (Mohr-Coulomb and Modified Cam-Clay) models were used to predict the solid deformations. Hitherto, the research generally focused on modelling cavity expansion without considering the effect of injected fluid coming into the medium. Very few studies have considered the effect of injected fluid on the deformation of soil skeleton. However, the porosity changes during the fluid injection and coupled elastic-plastic deformation are not clearly understood. In this study, the phenomenon of permeation and instabilities such as cavity and finger/ fracture formation will be quantified extensively by performing experiments using a novel experimental setup in addition to utilizing image processing techniques. This experimental study will describe the fluid flow and soil deformation characteristics under different boundary conditions. Further, a well refined coupled semi-analytical model will be developed to capture the physics involved in quantifying the deformation behaviour of geomaterial during fluid injection.Keywords: solid-fluid interaction, permeation, poroelasticity, plasticity, continuum model
Procedia PDF Downloads 73572 Study on Natural Light Distribution Inside the Room by Using Sudare as an Outside Horizontal Blind in Tropical Country of Indonesia
Authors: Agus Hariyadi, Hiroatsu Fukuda
Abstract:
In tropical country like Indonesia, especially in Jakarta, most of the energy consumption on building is for the cooling system, the second one is from lighting electric consumption. One of the passive design strategy that can be done is optimizing the use of natural light from the sun. In this area, natural light is always available almost every day around the year. Natural light have many effect on building. It can reduce the need of electrical lighting but also increase the external load. Another thing that have to be considered in the use of natural light is the visual comfort from occupant inside the room. To optimize the effectiveness of natural light need some modification of façade design. By using external shading device, it can minimize the external load that introduces into the room, especially from direct solar radiation which is the 80 % of the external energy load that introduces into the building. It also can control the distribution of natural light inside the room and minimize glare in the perimeter zone of the room. One of the horizontal blind that can be used for that purpose is Sudare. It is traditional Japanese blind that have been used long time in Japanese traditional house especially in summer. In its original function, Sudare is used to prevent direct solar radiation but still introducing natural ventilation. It has some physical characteristics that can be utilize to optimize the effectiveness of natural light. In this research, different scale of Sudare will be simulated using EnergyPlus and DAYSIM simulation software. EnergyPlus is a whole building energy simulation program to model both energy consumption—for heating, cooling, ventilation, lighting, and plug and process loads—and water use in buildings, while DAYSIM is a validated, RADIANCE-based daylighting analysis software that models the annual amount of daylight in and around buildings. The modelling will be done in Ladybug and Honeybee plugin. These are two open source plugins for Grasshopper and Rhinoceros 3D that help explore and evaluate environmental performance which will directly be connected to EnergyPlus and DAYSIM engines. Using the same model will maintain the consistency of the same geometry used both in EnergyPlus and DAYSIM. The aims of this research is to find the best configuration of façade design which can reduce the external load from the outside of the building to minimize the need of energy for cooling system but maintain the natural light distribution inside the room to maximize the visual comfort for occupant and minimize the need of electrical energy consumption.Keywords: façade, natural light, blind, energy
Procedia PDF Downloads 345571 Dynamic Risk Identification Using Fuzzy Failure Mode Effect Analysis in Fabric Process Industries: A Research Article as Management Perspective
Authors: A. Sivakumar, S. S. Darun Prakash, P. Navaneethakrishnan
Abstract:
In and around Erode District, it is estimated that more than 1250 chemical and allied textile processing fabric industries are affected, partially closed and shut off for various reasons such as poor management, poor supplier performance, lack of planning for productivity, fluctuation of output, poor investment, waste analysis, labor problems, capital/labor ratio, accumulation of stocks, poor maintenance of resources, deficiencies in the quality of fabric, low capacity utilization, age of plant and equipment, high investment and input but low throughput, poor research and development, lack of energy, workers’ fear of loss of jobs, work force mix and work ethic. The main objective of this work is to analyze the existing conditions in textile fabric sector, validate the break even of Total Productivity (TP), analyze, design and implement fuzzy sets and mathematical programming for improvement of productivity and quality dimensions in the fabric processing industry. It needs to be compatible with the reality of textile and fabric processing industries. The highly risk events from productivity and quality dimension were found by fuzzy systems and results are wrapped up among the textile fabric processing industry.Keywords: break even point, fuzzy crisp data, fuzzy sets, productivity, productivity cycle, total productive maintenance
Procedia PDF Downloads 338570 Sensitivity Analysis of the Heat Exchanger Design in Net Power Oxy-Combustion Cycle for Carbon Capture
Authors: Hirbod Varasteh, Hamidreza Gohari Darabkhani
Abstract:
The global warming and its impact on climate change is one of main challenges for current century. Global warming is mainly due to the emission of greenhouse gases (GHG) and carbon dioxide (CO2) is known to be the major contributor to the GHG emission profile. Whilst the energy sector is the primary source for CO2 emission, Carbon Capture and Storage (CCS) are believed to be the solution for controlling this emission. Oxyfuel combustion (Oxy-combustion) is one of the major technologies for capturing CO2 from power plants. For gas turbines, several Oxy-combustion power cycles (Oxyturbine cycles) have been investigated by means of thermodynamic analysis. NetPower cycle is one of the leading oxyturbine power cycles with almost full carbon capture capability from a natural gas fired power plant. In this manuscript, sensitivity analysis of the heat exchanger design in NetPower cycle is completed by means of process modelling. The heat capacity variation and supercritical CO2 with gaseous admixtures are considered for multi-zone analysis with Aspen Plus software. It is found that the heat exchanger design has a major role to increase the efficiency of NetPower cycle. The pinch-point analysis is done to extract the composite and grand composite curve for the heat exchanger. In this paper, relationship between the cycle efficiency and the minimum approach temperature (∆Tmin) of the heat exchanger has also been evaluated. Increase in ∆Tmin causes a decrease in the temperature of the recycle flue gases (RFG) and an overall decrease in the required power for the recycled gas compressor. The main challenge in the design of heat exchangers in power plants is a tradeoff between the capital and operational costs. To achieve lower ∆Tmin, larger size of heat exchanger is required. This means a higher capital cost but leading to a better heat recovery and lower operational cost. To achieve this, ∆Tmin is selected from the minimum point in the diagrams of capital and operational costs. This study provides an insight into the NetPower Oxy-combustion cycle’s performance analysis and operational condition based on its heat exchanger design.Keywords: carbon capture and storage, oxy-combustion, netpower cycle, oxy turbine cycles, zero emission, heat exchanger design, supercritical carbon dioxide, oxy-fuel power plant, pinch point analysis
Procedia PDF Downloads 204569 Planning Quality and Maintenance Activities in a Closed-Loop Serial Multi-Stage Manufacturing System under Constant Degradation
Authors: Amauri Josafat Gomez Aguilar, Jean Pierre Kenné
Abstract:
This research presents the development of a self-sustainable manufacturing system from a circular economy perspective, structured by a multi-stage serial production system consisting of a series of machines under deterioration in charge of producing a single product and a reverse remanufacturing system constituted by the same productive systems of the first scheme and different tooling, fed by-products collected at the end of their life cycle, and non-conforming elements of the first productive scheme. Since the advanced production manufacturing system is unable to satisfy the customer's quality expectations completely, we propose the development of a mixed integer linear mathematical model focused on the optimal search and assignment of quality stations and preventive maintenance operation to the machines over a time horizon, intending to segregate the correct number of non-conforming parts for reuse in the remanufacturing system and thereby minimizing production, quality, maintenance, and customer non-conformance penalties. Numerical experiments are performed to analyze the solutions found by the model under different scenarios. The results showed that the correct implementation of a closed manufacturing system and allocation of quality inspection and preventive maintenance operations generate better levels of customer satisfaction and an efficient manufacturing system.Keywords: closed loop, mixed integer linear programming, preventive maintenance, quality inspection
Procedia PDF Downloads 85568 Quantum Graph Approach for Energy and Information Transfer through Networks of Cables
Authors: Mubarack Ahmed, Gabriele Gradoni, Stephen C. Creagh, Gregor Tanner
Abstract:
High-frequency cables commonly connect modern devices and sensors. Interestingly, the proportion of electric components is rising fast in an attempt to achieve lighter and greener devices. Modelling the propagation of signals through these cable networks in the presence of parameter uncertainty is a daunting task. In this work, we study the response of high-frequency cable networks using both Transmission Line and Quantum Graph (QG) theories. We have successfully compared the two theories in terms of reflection spectra using measurements on real, lossy cables. We have derived a generalisation of the vertex scattering matrix to include non-uniform networks – networks of cables with different characteristic impedances and propagation constants. The QG model implicitly takes into account the pseudo-chaotic behavior, at the vertices, of the propagating electric signal. We have successfully compared the asymptotic growth of eigenvalues of the Laplacian with the predictions of Weyl law. We investigate the nearest-neighbour level-spacing distribution of the resonances and compare our results with the predictions of Random Matrix Theory (RMT). To achieve this, we will compare our graphs with the generalisation of Wigner distribution for open systems. The problem of scattering from networks of cables can also provide an analogue model for wireless communication in highly reverberant environments. In this context, we provide a preliminary analysis of the statistics of communication capacity for communication across cable networks, whose eventual aim is to enable detailed laboratory testing of information transfer rates using software defined radio. We specialise this analysis in particular for the case of MIMO (Multiple-Input Multiple-Output) protocols. We have successfully validated our QG model with both TL model and laboratory measurements. The growth of Eigenvalues compares well with Weyl’s law and the level-spacing distribution agrees so well RMT predictions. The results we achieved in the MIMO application compares favourably with the prediction of a parallel on-going research (sponsored by NEMF21.)Keywords: eigenvalues, multiple-input multiple-output, quantum graph, random matrix theory, transmission line
Procedia PDF Downloads 173567 A Project Screening System for Energy Enterprise Based on Dempster-Shafer Theory
Authors: Woosik Jang, Seung Heon Han, Seung Won Baek
Abstract:
Natural gas (NG) is an energy resource in a few countries, and most NG producers do business in politically unstable countries. In addition, as 90% of the LNG market is controlled by a small number of international oil companies (IOCs) and national oil companies (NOCs), entry of latecomers into the market is extremely limited. To meet these challenges, project viability needs to be assessed based on limited information from a project screening perspective. However, the early stages of the project have the following difficulties: (1) What are the factors to consider? (2) How many professionals do you need to decide? (3) How to make the best decision with limited information? To address this problem, this study proposes a model for evaluating LNG project viability based on the Dempster-Shafer theory (DST). A total of 11 indicators for analyzing the gas field, reflecting the characteristics of the LNG industry, and 23 indicators for analyzing the market environment, were identified. The proposed model also evaluates the LNG project based on the survey and provides uncertainty of the results based on DST as well as quantified results. Thus, the proposed model is expected to be able to support the decision-making process of the gas field project using quantitative results as a systematic framework, and it was developed as a stand-alone system to improve its usefulness in practice. Consequently, the amount of information and the mathematical approach are expected to improve the quality and opportunity of decision making for LNG projects for enterprises.Keywords: project screen, energy enterprise, decision support system, Dempster-Shafer theory
Procedia PDF Downloads 341566 The Impact of Research Anxiety on Research Orientation and Interest in Research Courses in Social Work Students
Authors: Daniel Gredig, Annabelle Bartelsen-Raemy
Abstract:
Social work professionals should underpin their decisions with scientific knowledge and research findings. Hence, research is used as a framework for social work education and research courses have become a taken-for-granted component of study programmes. However, it has been acknowledged that social work students have negative beliefs and attitudes as well as frequently feelings of fear of research courses. Against this background, the present study aimed to establish the relationship between student’s fear of research courses, their research orientation and interest in research courses. We hypothesized that fear predicts the interest in research courses. Further, we hypothesized that research orientation (perceived importance and attributed usefulness for research for social work practice and perceived unbiased nature of research) was a mediating variable. In the years 2014, 2015 and 2016, we invited students enrolled for a bachelor programme in social work in Switzerland to participate in the study during their introduction day to the school taking place two weeks before their programme started. For data collection, we used an anonymous self-administered on-line questionnaire filled in on site. Data were analysed using descriptive statistics and structural equation modelling (generalized least squares estimates method). The sample included 708 students enrolled in a social work bachelor-programme, 501 being female, 184 male, and 5 intersexual, aged 19–56, having various entitlements to study, and registered for three different types of programme modes (full time programme; part time study with field placements in blocks; part time study involving concurrent field placement). Analysis showed that the interest in research courses was predicted by fear of research courses (β = -0.29) as well as by the perceived importance (β = 0.27), attributed usefulness of research (β = 0.15) and perceived unbiased nature of research (β = 0.08). These variables were predicted, in turn, by fear of research courses (β = -0.10, β = -0.23, and β = -0.13). Moreover, interest was predicted by age (β = 0.13). Fear of research courses was predicted by age (β = -0.10) female gender (β = 0.28) and having completed a general baccalaureate (β = -0.09). (GFI = 0.997, AGFI = 0.988, SRMR = 0.016, CMIN/df = 0.946, adj. R2 = 0.312). Findings evidence a direct as well as a mediated impact of fear on the interest in research courses in entering first-year students in a social work bachelor-programme. It highlights one of the challenges social work education in a research framework has to meet with. It seems, there have been considerable efforts to address the research orientation of students. However, these findings point out that, additionally, research anxiety in terms of fear of research courses should be considered and addressed by teachers when conceptualizing research courses.Keywords: research anxiety, research courses, research interest, research orientation, social work students, teaching
Procedia PDF Downloads 188565 Controlled Release of Glucosamine from Pluronic-Based Hydrogels for the Treatment of Osteoarthritis
Authors: Papon Thamvasupong, Kwanchanok Viravaidya-Pasuwat
Abstract:
Osteoarthritis affects a lot of people worldwide. Local injection of glucosamine is one of the alternative treatment methods to replenish the natural lubrication of cartilage. However, multiple injections can potentially lead to possible bacterial infection. Therefore, a drug delivery system is desired to reduce the frequencies of injections. A hydrogel is one of the delivery systems that can control the release of drugs. Thermo-reversible hydrogels can be beneficial to the drug delivery system especially in the local injection route because this formulation can change from liquid to gel after getting into human body. Once the gel is in the body, it will slowly release the drug in a controlled manner. In this study, various formulations of Pluronic-based hydrogels were synthesized for the controlled release of glucosamine. One of the challenges of the Pluronic controlled release system is its fast dissolution rate. To overcome this problem, alginate and calcium sulfate (CaSO4) were added to the polymer solution. The characteristics of the hydrogels were investigated including the gelation temperature, gelation time, hydrogel dissolution and glucosamine release mechanism. Finally, a mathematical model of glucosamine release from Pluronic-alginate-hyaluronic acid hydrogel was developed. Our results have shown that crosslinking Pluronic gel with alginate did not significantly extend the dissolution rate of the gel. Moreover, the gel dissolution profiles and the glucosamine release mechanisms were best described using the zeroth-order kinetic model, indicating that the release of glucosamine was primarily governed by the gel dissolution.Keywords: controlled release, drug delivery system, glucosamine, pluronic, thermoreversible hydrogel
Procedia PDF Downloads 270564 Investigation of Crack Formation in Ordinary Reinforced Concrete Beams and in Beams Strengthened with Carbon Fiber Sheet: Theory and Experiment
Authors: Anton A. Bykov, Irina O. Glot, Igor N. Shardakov, Alexey P. Shestakov
Abstract:
This paper presents the results of experimental and theoretical investigations of the mechanisms of crack formation in reinforced concrete beams subjected to quasi-static bending. The boundary-value problem has been formulated in the framework of brittle fracture mechanics and has been solved by using the finite-element method. Numerical simulation of the vibrations of an uncracked beam and a beam with cracks of different size serves to determine the pattern of changes in the spectrum of eigenfrequencies observed during crack evolution. Experiments were performed on the sequential quasistatic four-point bending of the beam leading to the formation of cracks in concrete. At each loading stage, the beam was subjected to an impulse load to induce vibrations. Two stages of cracking were detected. At the first stage the conservative process of deformation is realized. The second stage is an active cracking, which is marked by a sharp change in eingenfrequencies. The boundary of a transition from one stage to another is well registered. The vibration behavior was examined for the beams strengthened by carbon-fiber sheet before loading and at the intermediate stage of loading after the grouting of initial cracks. The obtained results show that the vibrodiagnostic approach is an effective tool for monitoring of cracking and for assessing the quality of measures aimed at strengthening concrete structures.Keywords: crack formation, experiment, mathematical modeling, reinforced concrete, vibrodiagnostics
Procedia PDF Downloads 305563 Statistical Time-Series and Neural Architecture of Malaria Patients Records in Lagos, Nigeria
Authors: Akinbo Razak Yinka, Adesanya Kehinde Kazeem, Oladokun Oluwagbenga Peter
Abstract:
Time series data are sequences of observations collected over a period of time. Such data can be used to predict health outcomes, such as disease progression, mortality, hospitalization, etc. The Statistical approach is based on mathematical models that capture the patterns and trends of the data, such as autocorrelation, seasonality, and noise, while Neural methods are based on artificial neural networks, which are computational models that mimic the structure and function of biological neurons. This paper compared both parametric and non-parametric time series models of patients treated for malaria in Maternal and Child Health Centres in Lagos State, Nigeria. The forecast methods considered linear regression, Integrated Moving Average, ARIMA and SARIMA Modeling for the parametric approach, while Multilayer Perceptron (MLP) and Long Short-Term Memory (LSTM) Network were used for the non-parametric model. The performance of each method is evaluated using the Mean Absolute Error (MAE), R-squared (R2) and Root Mean Square Error (RMSE) as criteria to determine the accuracy of each model. The study revealed that the best performance in terms of error was found in MLP, followed by the LSTM and ARIMA models. In addition, the Bootstrap Aggregating technique was used to make robust forecasts when there are uncertainties in the data.Keywords: ARIMA, bootstrap aggregation, MLP, LSTM, SARIMA, time-series analysis
Procedia PDF Downloads 75562 Test Suite Optimization Using an Effective Meta-Heuristic BAT Algorithm
Authors: Anuradha Chug, Sunali Gandhi
Abstract:
Regression Testing is a very expensive and time-consuming process carried out to ensure the validity of modified software. Due to the availability of insufficient resources to re-execute all the test cases in time constrained environment, efforts are going on to generate test data automatically without human efforts. Many search based techniques have been proposed to generate efficient, effective as well as optimized test data, so that the overall cost of the software testing can be minimized. The generated test data should be able to uncover all potential lapses that exist in the software or product. Inspired from the natural behavior of bat for searching her food sources, current study employed a meta-heuristic, search-based bat algorithm for optimizing the test data on the basis certain parameters without compromising their effectiveness. Mathematical functions are also applied that can effectively filter out the redundant test data. As many as 50 Java programs are used to check the effectiveness of proposed test data generation and it has been found that 86% saving in testing efforts can be achieved using bat algorithm while covering 100% of the software code for testing. Bat algorithm was found to be more efficient in terms of simplicity and flexibility when the results were compared with another nature inspired algorithms such as Firefly Algorithm (FA), Hill Climbing Algorithm (HC) and Ant Colony Optimization (ACO). The output of this study would be useful to testers as they can achieve 100% path coverage for testing with minimum number of test cases.Keywords: regression testing, test case selection, test case prioritization, genetic algorithm, bat algorithm
Procedia PDF Downloads 380561 Designing Ecologically and Economically Optimal Electric Vehicle Charging Stations
Authors: Y. Ghiassi-Farrokhfal
Abstract:
The number of electric vehicles (EVs) is increasing worldwide. Replacing gas fueled cars with EVs reduces carbon emission. However, the extensive energy consumption of EVs stresses the energy systems, requiring non-green sources of energy (such as gas turbines) to compensate for the new energy demand caused by EVs in the energy systems. To make EVs even a greener solution for the future energy systems, new EV charging stations are equipped with solar PV panels and batteries. This will help serve the energy demand of EVs through the green energy of solar panels. To ensure energy availability, solar panels are combined with batteries. The energy surplus at any point is stored in batteries and is used when there is not enough solar energy to serve the demand. While EV charging stations equipped with solar panels and batteries are green and ecologically optimal, they might not be financially viable solutions, due to battery prices. To make the system viable, we should size the battery economically and operate the system optimally. This is, in general, a challenging problem because of the stochastic nature of the EV arrivals at the charging station, the available solar energy, and the battery operating system. In this work, we provide a mathematical model for this problem and we compute the return on investment (ROI) of such a system, which is designed to be ecologically and financially optimal. We also quantify the minimum required investment in terms of battery and solar panels along with the operating strategy to ensure that a charging station has enough energy to serve its EV demand at any time.Keywords: solar energy, battery storage, electric vehicle, charging stations
Procedia PDF Downloads 220560 An Analytical Approach to Assess and Compare the Vulnerability Risk of Operating Systems
Authors: Pubudu K. Hitigala Kaluarachchilage, Champike Attanayake, Sasith Rajasooriya, Chris P. Tsokos
Abstract:
Operating system (OS) security is a key component of computer security. Assessing and improving OSs strength to resist against vulnerabilities and attacks is a mandatory requirement given the rate of new vulnerabilities discovered and attacks occurring. Frequency and the number of different kinds of vulnerabilities found in an OS can be considered an index of its information security level. In the present study five mostly used OSs, Microsoft Windows (windows 7, windows 8 and windows 10), Apple’s Mac and Linux are assessed for their discovered vulnerabilities and the risk associated with each. Each discovered and reported vulnerability has an exploitability score assigned in CVSS score of the national vulnerability database. In this study the risk from vulnerabilities in each of the five Operating Systems is compared. Risk Indexes used are developed based on the Markov model to evaluate the risk of each vulnerability. Statistical methodology and underlying mathematical approach is described. Initially, parametric procedures are conducted and measured. There were, however, violations of some statistical assumptions observed. Therefore the need for non-parametric approaches was recognized. 6838 vulnerabilities recorded were considered in the analysis. According to the risk associated with all the vulnerabilities considered, it was found that there is a statistically significant difference among average risk levels for some operating systems, indicating that according to our method some operating systems have been more risk vulnerable than others given the assumptions and limitations. Relevant test results revealing a statistically significant difference in the Risk levels of different OSs are presented.Keywords: cybersecurity, Markov chain, non-parametric analysis, vulnerability, operating system
Procedia PDF Downloads 183559 Evaluation of Microbial Accumulation of Household Wastewater Purified by Advanced Oxidation Process
Authors: Nazlı Çetindağ, Pelin Yılmaz Çetiner, Metin Mert İlgün, Emine Birci, Gizemnur Yıldız Uysal, Özcan Hatipoğlu, Ehsan Tuzcuoğlu, Gökhan Sır
Abstract:
Water scarcity is an unavoidable issue impacting an increasing number of individuals daily, representing a global crisis stemming from swift population growth, urbanization, and excessive resource exploitation. Consequently, solutions that involve the reclamation of wastewater are considered essential. In this context, household wastewater, categorized as greywater, plays a significant role in freshwater used for residential purposes and is attributed to washing. This type of wastewater comprises diverse elements, including organic substances, soaps, detergents, solvents, biological components, and inorganic elements such as certain metal ions and particles. The physical characteristics of wastewater vary depending on its source, whether commercial, domestic, or from a hospital setting. Consequently, the treatment strategy for this wastewater type necessitates comprehensive investigation and appropriate handling. The advanced oxidation process (AOP) emerges as a promising technique associated with the generation of reactive hydroxyl radicals highly effective in oxidizing organic pollutants. This method takes precedence over others like coagulation, flocculation, sedimentation, and filtration due to its avoidance of undesirable by-products. In the current study, the focus was on exploring the feasibility of the AOP for treating actual household wastewater. To achieve this, a laboratory-scale device was designed to effectively target the formed radicals toward organic pollutants, resulting in lower organic compounds in wastewater. Then, the number of microorganisms present in treated wastewater, in addition to the chemical content of the water, was analyzed to determine whether the lab-scale device eliminates microbial accumulation with AOP. This was also an important parameter since microbes can indirectly affect human health and machine hygiene. To do this, water samples were taken from treated and untreated conditions and then inoculated on general purpose agar to track down the total plate count. Analysis showed that AOP might be an option to treat household wastewater and lower microorganism growth.Keywords: usage of household water, advanced oxidation process, water reuse, modelling
Procedia PDF Downloads 50558 Mapping the Turbulence Intensity and Excess Energy Available to Small Wind Systems over 4 Major UK Cities
Authors: Francis C. Emejeamara, Alison S. Tomlin, James Gooding
Abstract:
Due to the highly turbulent nature of urban air flows, and by virtue of the fact that turbines are likely to be located within the roughness sublayer of the urban boundary layer, proposed urban wind installations are faced with major challenges compared to rural installations. The challenge of operating within turbulent winds can however, be counteracted by the development of suitable gust tracking solutions. In order to assess the cost effectiveness of such controls, a detailed understanding of the urban wind resource, including its turbulent characteristics, is required. Estimating the ambient turbulence and total kinetic energy available at different control response times is essential in evaluating the potential performance of wind systems within the urban environment should effective control solutions be employed. However, high resolution wind measurements within the urban roughness sub-layer are uncommon, and detailed CFD modelling approaches are too computationally expensive to apply routinely on a city wide scale. This paper therefore presents an alternative semi-empirical methodology for estimating the excess energy content (EEC) present in the complex and gusty urban wind. An analytical methodology for predicting the total wind energy available at a potential turbine site is proposed by assessing the relationship between turbulence intensities and EEC, for different control response times. The semi-empirical model is then incorporated with an analytical methodology that was initially developed to predict mean wind speeds at various heights within the built environment based on detailed mapping of its aerodynamic characteristics. Based on the current methodology, additional estimates of turbulence intensities and EEC allow a more complete assessment of the available wind resource. The methodology is applied to 4 UK cities with results showing the potential of mapping turbulence intensities and the total wind energy available at different heights within each city. Considering the effect of ambient turbulence and choice of wind system, the wind resource over neighbourhood regions (of 250 m uniform resolution) and building rooftops within the 4 cities were assessed with results highlighting the promise of mapping potential turbine sites within each city.Keywords: excess energy content, small-scale wind, turbulence intensity, urban wind energy, wind resource assessment
Procedia PDF Downloads 474557 Tales of Two Cities: 'Motor City' Detroit and 'King Cotton' Manchester: Transatlantic Transmissions and Transformations, Flows of Communications, Commercial and Cultural Connections
Authors: Dominic Sagar
Abstract:
Manchester ‘King Cotton’, the first truly industrial city of the nineteenth century, passing on the baton to Detroit ‘Motor City’, is the first truly modern city. We are exploring the tales of the two cities, their rise and fall and subsequent post-industrial decline, their transitions and transformations, whilst alongside paralleling their corresponding, commercial, cultural, industrial and even agricultural, artistic and musical transactions and connections. The paper will briefly contextualize how technologies of the industrial age and modern age have been instrumental in the development of these cities and other similar cities including New York. However, the main focus of the study will be the present and more importantly the future, how globalisation and the advancements of digital technologies and industries have shaped the cities developments from AlanTuring and the making of the first programmable computer to the effect of digitalisation and digital initiatives. Manchester now has a thriving creative digital infrastructure of Digilabs, FabLabs, MadLabs and hubs, the study will reference the Smart Project and the Manchester Digital Development Association whilst paralleling similar digital and creative industrial initiatives now starting to happen in Detroit. The paper will explore other topics including the need to allow for zones of experimentation, areas to play, think and create in order develop and instigate new initiatives and ideas of production, carrying on the tradition of influential inventions throughout the history of these key cities. Other topics will be briefly touched on, such as urban farming, citing the Biospheric foundation in Manchester and other similar projects in Detroit. However, the main thread will focus on the music industries and how they are contributing to the regeneration of cities. Musically and artistically, Manchester and Detroit have been closely connected by the flow and transmission of information and transfer of ideas via ‘cars and trains and boats and planes’ through to the new ‘super highway’. From Detroit to Manchester often via New York and Liverpool and back again, these musical and artistic connections and flows have greatly affected and influenced both cities and the advancement of technology are still connecting the cities. In summary two hugely important industrial cities, subsequently both experienced massive decline in fortunes, having had their large industrial hearts ripped out, ravaged leaving dying industrial carcasses and car crashes of despair, dereliction, desolation and post-industrial wastelands vacated by a massive exodus of the cities’ inhabitants. To examine the affinity, similarity and differences between Manchester & Detroit, from their industrial importance to their post-industrial decline and their current transmutations, transformations, transient transgressions, cities in transition; contrasting how they have dealt with these problems and how they can learn from each other. With a view to framing these topics with regard to how various communities have shaped these cities and the creative industries and design [the new cotton/car manufacturing industries] are reinventing post-industrial cities, to speculate on future development of these themes in relation to Globalisation, digitalisation and how cities can function to develop solutions to communal living in cities of the future.Keywords: cultural capital, digital developments, musical initiatives, zones of experimentation
Procedia PDF Downloads 194556 Uncertainty Quantification of Fuel Compositions on Premixed Bio-Syngas Combustion at High-Pressure
Abstract:
Effect of fuel variabilities on premixed combustion of bio-syngas mixtures is of great importance in bio-syngas utilisation. The uncertainties of concentrations of fuel constituents such as H2, CO and CH4 may lead to unpredictable combustion performances, combustion instabilities and hot spots which may deteriorate and damage the combustion hardware. Numerical modelling and simulations can assist in understanding the behaviour of bio-syngas combustion with pre-defined species concentrations, while the evaluation of variabilities of concentrations is expensive. To be more specific, questions such as ‘what is the burning velocity of bio-syngas at specific equivalence ratio?’ have been answered either experimentally or numerically, while questions such as ‘what is the likelihood of burning velocity when precise concentrations of bio-syngas compositions are unknown, but the concentration ranges are pre-described?’ have not yet been answered. Uncertainty quantification (UQ) methods can be used to tackle such questions and assess the effects of fuel compositions. An efficient probabilistic UQ method based on Polynomial Chaos Expansion (PCE) techniques is employed in this study. The method relies on representing random variables (combustion performances) with orthogonal polynomials such as Legendre or Gaussian polynomials. The constructed PCE via Galerkin Projection provides easy access to global sensitivities such as main, joint and total Sobol indices. In this study, impacts of fuel compositions on combustion (adiabatic flame temperature and laminar flame speed) of bio-syngas fuel mixtures are presented invoking this PCE technique at several equivalence ratios. High-pressure effects on bio-syngas combustion instability are obtained using detailed chemical mechanism - the San Diego Mechanism. Guidance on reducing combustion instability from upstream biomass gasification process is provided by quantifying the significant contributions of composition variations to variance of physicochemical properties of bio-syngas combustion. It was found that flame speed is very sensitive to hydrogen variability in bio-syngas, and reducing hydrogen uncertainty from upstream biomass gasification processes can greatly reduce bio-syngas combustion instability. Variation of methane concentration, although thought to be important, has limited impacts on laminar flame instabilities especially for lean combustion. Further studies on the UQ of percentage concentration of hydrogen in bio-syngas can be conducted to guide the safer use of bio-syngas.Keywords: bio-syngas combustion, clean energy utilisation, fuel variability, PCE, targeted uncertainty reduction, uncertainty quantification
Procedia PDF Downloads 275555 The Taste of Macau: An Exploratory Study of Destination Food Image
Authors: Jianlun Zhang, Christine Lim
Abstract:
Local food is one of the most attractive elements to tourists. The role of local cuisine in destination branding is very important because it is the distinctive identity that helps tourists remember the destination. The objectives of this study are: (1) Test the direct relation between the cognitive image of destination food and tourists’ intention to eat local food. (2) Examine the mediating effect of tourists’ desire to try destination food on the relationship between the cognitive image of local food and tourists’ intention to eat destination food. (3) Study the moderating effect of tourists’ perceived difficulties in finding local food on the relationship between tourists’ desire to try destination food and tourists’ intention to eat local food. To achieve the goals of this study, Macanese cuisine is selected as the destination food. Macau is located in Southeastern China and is a former colonial city of Portugal. The taste and texture of Macanese cuisine are unique because it is a fusion of cuisine from many countries and regions of mainland China. As people travel to seek authentically exotic experience, it is important to investigate if the food image of Macau leaves a good impression on tourists and motivate them to try local cuisine. A total of 449 Chinese tourists were involved in this study. To analyze the data collected, partial least square-structural equation modelling (PLS-SEM) technique is employed. Results suggest that the cognitive image of Macanese cuisine has a direct effect on tourists’ intention to eat Macanese cuisine. Tourists’ desire to try Macanese cuisine mediates the cognitive image-intention relationship. Tourists’ perceived difficulty of finding Macanese cuisine moderates the desire-intention relationship. The lower tourists’ perceived difficulty in finding Macanese cuisine is, the stronger the desire-intention relationship it will be. There are several practical implications of this study. First, the government tourism website can develop an authentic storyline about the evolvement of local cuisine, which provides an opportunity for tourists to taste the history of the destination and create a novel experience for them. Second, the government should consider the development of food events, restaurants, and hawker businesses. Third, to lower tourists’ perceived difficulty in finding local cuisine, there should be locations of restaurants and hawker stalls with clear instructions for finding them on the websites of the government tourism office, popular tourism sites, and public transportation stations in the destination. Fourth, in the post-COVID-19 era, travel risk will be a major concern for tourists. Therefore, when promoting local food, the government tourism website should post images that show food safety and hygiene.Keywords: cognitive image of destination food, desire to try destination food, intention to eat food in the destination, perceived difficulties of finding local cuisine, PLS-SEM
Procedia PDF Downloads 189554 Surgical Hip Dislocation of Femoroacetabular Impingement: Survivorship and Functional Outcomes at 10 Years
Authors: L. Hoade, O. O. Onafowokan, K. Anderson, G. E. Bartlett, E. D. Fern, M. R. Norton, R. G. Middleton
Abstract:
Aims: Femoroacetabular impingement (FAI) was first recognised as a potential driver for hip pain at the turn of the last millennium. While there is an increasing trend towards surgical management of FAI by arthroscopic means, open surgical hip dislocation and debridement (SHD) remains the Gold Standard of care in terms of reported outcome measures. (1) Long-term functional and survivorship outcomes of SHD as a treatment for FAI are yet to be sufficiently reported in the literature. This study sets out to help address this imbalance. Methods: We undertook a retrospective review of our institutional database for all patients who underwent SHD for FAI between January 2003 and December 2008. A total of 223 patients (241 hips) were identified and underwent a ten year review with a standardised radiograph and patient-reported outcome measures questionnaire. The primary outcome measure of interest was survivorship, defined as progression to total hip arthroplasty (THA). Negative predictive factors were analysed. Secondary outcome measures of interest were survivorship to further (non-arthroplasty) surgery, functional outcomes as reflected by patient reported outcome measure scores (PROMS) scores, and whether a learning curve could be identified. Results: The final cohort consisted of 131 females and 110 males, with a mean age of 34 years. There was an overall native hip joint survival rate of 85.4% at ten years. Those who underwent a THA were significantly older at initial surgery, had radiographic evidence of preoperative osteoarthritis and pre- and post-operative acetabular undercoverage. In those whom had not progressed to THA, the average Non-arthritic Hip Score and Oxford Hip Score at ten year follow-up were 72.3% and 36/48, respectively, and 84% still deemed their surgery worthwhile. A learning curve was found to exist that was predicated on case selection rather than surgical technique. Conclusion: This is only the second study to evaluate the long-term outcomes (beyond ten years) of SHD for FAI and the first outside the originating centre. Our results suggest that, with correct patient selection, this remains an operation with worthwhile outcomes at ten years. How the results of open surgery compared to those of arthroscopy remains to be answered. While these results precede the advent of collison software modelling tools, this data helps set a benchmark for future comparison of other techniques effectiveness at the ten year mark.Keywords: femoroacetabular impingement, hip pain, surgical hip dislocation, hip debridement
Procedia PDF Downloads 84553 Numerical Modelling and Experiment of a Composite Single-Lap Joint Reinforced by Multifunctional Thermoplastic Composite Fastener
Authors: Wenhao Li, Shijun Guo
Abstract:
Carbon fibre reinforced composites are progressively replacing metal structures in modern civil aircraft. This is because composite materials have large potential of weight saving compared with metal. However, the achievement to date of weight saving in composite structure is far less than the theoretical potential due to many uncertainties in structural integrity and safety concern. Unlike the conventional metallic structure, composite components are bonded together along the joints where structural integrity is a major concern. To ensure the safety, metal fasteners are used to reinforce the composite bonded joints. One of the solutions for a significant weight saving of composite structure is to develop an effective technology of on-board Structural Health Monitoring (SHM) System. By monitoring the real-life stress status of composite structures during service, the safety margin set in the structure design can be reduced with confidence. It provides a means of safeguard to minimize the need for programmed inspections and allow for maintenance to be need-driven, rather than usage-driven. The aim of this paper is to develop smart composite joint. The key technology is a multifunctional thermoplastic composite fastener (MTCF). The MTCF will replace some of the existing metallic fasteners in the most concerned locations distributed over the aircraft composite structures to reinforce the joints and form an on-board SHM network system. Each of the MTCFs will work as a unit of the AU and AE technology. The proposed MTCF technology has been patented and developed by Prof. Guo in Cranfield University, UK in the past a few years. The manufactured MTCF has been successfully employed in the composite SLJ (Single-Lap Joint). In terms of the structure integrity, the hybrid SLJ reinforced by MTCF achieves 19.1% improvement in the ultimate failure strength in comparison to the bonded SLJ. By increasing the diameter or rearranging the lay-up sequence of MTCF, the hybrid SLJ reinforced by MTCF is able to achieve the equivalent ultimate strength as that reinforced by titanium fastener. The predicted ultimate strength in simulation is in good agreement with the test results. In terms of the structural health monitoring, a signal from the MTCF was measured well before the load of mechanical failure. This signal provides a warning of initial crack in the joint which could not be detected by the strain gauge until the final failure.Keywords: composite single-lap joint, crack propagation, multifunctional composite fastener, structural health monitoring
Procedia PDF Downloads 163552 Hydrogen-Fueled Micro-Thermophotovoltaic Power Generator: Flame Regimes and Flame Stability
Authors: Hosein Faramarzpour
Abstract:
This work presents the optimum operational conditions for a hydrogen-based micro-scale power source, using a verified mathematical model including fluid dynamics and reaction kinetics. Thereafter the stable operational flame regime is pursued as a key factor in optimizing the design of micro-combustors. The results show that with increasing velocities, four H2 flame regimes develop in the micro-combustor, namely: 1) periodic ignition-extinction regime, 2) steady symmetric regime, 3) pulsating asymmetric regime, and 4) steady asymmetric regime. The first regime that appears in 0.8 m/s inlet velocity is a periodic ignition-extinction regime which is characterized by counter flows and tulip-shape flames. For flow velocity above 0.2 m/s, the flame shifts downstream, and the combustion regime switches to a steady symmetric flame where temperature increases considerably due to the increased rate of incoming energy. Further elevation in flow velocity up to 1 m/s leads to the pulsating asymmetric flame formation, which is associated with pulses in various flame properties such as temperature and species concentration. Further elevation in flow velocity up to 1 m/s leads to the pulsating asymmetric flame formation, which is associated with pulses in various flame properties such as temperature and species concentration. Ultimately, when the inlet velocity reached 1.2 m/s, the last regime was observed, and a steady asymmetric regime appeared.Keywords: thermophotovoltaic generator, micro combustor, micro power generator, combustion regimes, flame dynamic
Procedia PDF Downloads 101551 Design and Development of Optical Sensor Based Ground Reaction Force Measurement Platform for GAIT and Geriatric Studies
Authors: K. Chethana, A. S. Guru Prasad, S. N. Omkar, B. Vadiraj, S. Asokan
Abstract:
This paper describes an ab-initio design, development and calibration results of an Optical Sensor Ground Reaction Force Measurement Platform (OSGRFP) for gait and geriatric studies. The developed system employs an array of FBG sensors to measure the respective ground reaction forces from all three axes (X, Y and Z), which are perpendicular to each other. The novelty of this work is two folded. One is in its uniqueness to resolve the tri axial resultant forces during the stance in to the respective pure axis loads and the other is the applicability of inherently advantageous FBG sensors which are most suitable for biomechanical instrumentation. To validate the response of the FBG sensors installed in OSGRFP and to measure the cross sensitivity of the force applied in other directions, load sensors with indicators are used. Further in this work, relevant mathematical formulations are presented for extracting respective ground reaction forces from wavelength shifts/strain of FBG sensors on the OSGRFP. The result of this device has implications in understanding the foot function, identifying issues in gait cycle and measuring discrepancies between left and right foot. The device also provides a method to quantify and compare relative postural stability of different subjects under test, which has implications in post surgical rehabilitation, geriatrics and optimizing training protocols for sports personnel.Keywords: balance and stability, gait analysis, FBG applications, optical sensor ground reaction force platform
Procedia PDF Downloads 403550 Project HDMI: A Hybrid-Differentiated Mathematics Instruction for Grade 11 Senior High School Students at Las Piñas City Technical Vocational High School
Authors: Mary Ann Cristine R. Olgado
Abstract:
Diversity in the classroom might make it difficult to promote individualized learning, but differentiated instruction that caters to students' various learning preferences may prove to be beneficial. Hence, this study examined the effectiveness of Hybrid-Differentiated Mathematics Instruction (HDMI) in improving the students’ academic performance in Mathematics. It employed the quasi-experimental research design by using a comparative analysis of the two variables: the experimental and control groups. The learning styles of the students were identified using the Grasha-Riechmann Student Learning Style Scale (GRSLSS), which served as the basis for designing differentiated action plans in Mathematics. In addition, adapted survey questionnaires, pre-tests, and post-tests were used to gather information and were analyzed using descriptive and correlational statistics to find the relationship between variables. The experimental group received differentiated instruction for a month, while the control group received traditional teaching instruction. The study found that Hybrid-Differentiated Mathematics Instruction (HDMI) improved the academic performance of Grade 11-TVL students, with the experimental group performing better than the control group. This program has effectively tailored the teaching methods to meet the diverse learning needs of the students, fostering and enhancing a deeper understanding of mathematical concepts in Statistics & Probability, both within and beyond the classroom.Keywords: differentiated instruction, hybrid, learning styles, academic performance
Procedia PDF Downloads 61549 A Comparative Soft Computing Approach to Supplier Performance Prediction Using GEP and ANN Models: An Automotive Case Study
Authors: Seyed Esmail Seyedi Bariran, Khairul Salleh Mohamed Sahari
Abstract:
In multi-echelon supply chain networks, optimal supplier selection significantly depends on the accuracy of suppliers’ performance prediction. Different methods of multi criteria decision making such as ANN, GA, Fuzzy, AHP, etc have been previously used to predict the supplier performance but the “black-box” characteristic of these methods is yet a major concern to be resolved. Therefore, the primary objective in this paper is to implement an artificial intelligence-based gene expression programming (GEP) model to compare the prediction accuracy with that of ANN. A full factorial design with %95 confidence interval is initially applied to determine the appropriate set of criteria for supplier performance evaluation. A test-train approach is then utilized for the ANN and GEP exclusively. The training results are used to find the optimal network architecture and the testing data will determine the prediction accuracy of each method based on measures of root mean square error (RMSE) and correlation coefficient (R2). The results of a case study conducted in Supplying Automotive Parts Co. (SAPCO) with more than 100 local and foreign supply chain members revealed that, in comparison with ANN, gene expression programming has a significant preference in predicting supplier performance by referring to the respective RMSE and R-squared values. Moreover, using GEP, a mathematical function was also derived to solve the issue of ANN black-box structure in modeling the performance prediction.Keywords: Supplier Performance Prediction, ANN, GEP, Automotive, SAPCO
Procedia PDF Downloads 419