Search results for: generalized linear model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18915

Search results for: generalized linear model

16635 On ‘Freaks’ and the Feminine in Margaret Atwood’s ‘Lusus Naturae’

Authors: Shahd Alshammari

Abstract:

This paper considers one of Margaret Atwood’s short stories ‘Lusus Naturae'. Through a critical lens that makes use of Julia Kristeva’s work on Powers of Horror and abjection, this paper suggests that the monstrous girl is the disabled woman, the abject in society. The monster is used as a metaphor for the unknown, the misunderstood, and the ‘different’ woman. Culturally Relevant Teaching (CRT) is a pedagogy that calls for making course material accessible and relevant to students. Through the study of literary texts, we are able to help create agency inside and outside the classroom. Stories are a necessary part of establishing connections across borders and boundaries. Stories are meant to raise awareness both inside and outside the classroom. The discussion is equally important, and the text is meant to facilitate relevant questions that the students need to consider when it comes to identity. Questions to consider are: what does it mean to be a ‘girl’ today, and what implications and consequences are at hand when you fail to perform this gendered identity? Gender is sometimes a fatal bond in the Middle East, and even more so, is the disability. In the case of our unnamed protagonist, she undergoes a process of un-becoming, a non-linear process of growing up. In a sense, it is a counter-Bildungsroman. The reading of this text emphasizes that a non-linear narrative is sometimes necessary for the female protagonist’s self-awareness and development. Discussion in class facilitates this sense of agency and questioning of gender and disability.

Keywords: disability, gender, literature, pedagogy

Procedia PDF Downloads 653
16634 Development and Adaptation of a LGBM Machine Learning Model, with a Suitable Concept Drift Detection and Adaptation Technique, for Barcelona Household Electric Load Forecasting During Covid-19 Pandemic Periods (Pre-Pandemic and Strict Lockdown)

Authors: Eric Pla Erra, Mariana Jimenez Martinez

Abstract:

While aggregated loads at a community level tend to be easier to predict, individual household load forecasting present more challenges with higher volatility and uncertainty. Furthermore, the drastic changes that our behavior patterns have suffered due to the COVID-19 pandemic have modified our daily electrical consumption curves and, therefore, further complicated the forecasting methods used to predict short-term electric load. Load forecasting is vital for the smooth and optimized planning and operation of our electric grids, but it also plays a crucial role for individual domestic consumers that rely on a HEMS (Home Energy Management Systems) to optimize their energy usage through self-generation, storage, or smart appliances management. An accurate forecasting leads to higher energy savings and overall energy efficiency of the household when paired with a proper HEMS. In order to study how COVID-19 has affected the accuracy of forecasting methods, an evaluation of the performance of a state-of-the-art LGBM (Light Gradient Boosting Model) will be conducted during the transition between pre-pandemic and lockdowns periods, considering day-ahead electric load forecasting. LGBM improves the capabilities of standard Decision Tree models in both speed and reduction of memory consumption, but it still offers a high accuracy. Even though LGBM has complex non-linear modelling capabilities, it has proven to be a competitive method under challenging forecasting scenarios such as short series, heterogeneous series, or data patterns with minimal prior knowledge. An adaptation of the LGBM model – called “resilient LGBM” – will be also tested, incorporating a concept drift detection technique for time series analysis, with the purpose to evaluate its capabilities to improve the model’s accuracy during extreme events such as COVID-19 lockdowns. The results for the LGBM and resilient LGBM will be compared using standard RMSE (Root Mean Squared Error) as the main performance metric. The models’ performance will be evaluated over a set of real households’ hourly electricity consumption data measured before and during the COVID-19 pandemic. All households are located in the city of Barcelona, Spain, and present different consumption profiles. This study is carried out under the ComMit-20 project, financed by AGAUR (Agència de Gestiód’AjutsUniversitaris), which aims to determine the short and long-term impacts of the COVID-19 pandemic on building energy consumption, incrementing the resilience of electrical systems through the use of tools such as HEMS and artificial intelligence.

Keywords: concept drift, forecasting, home energy management system (HEMS), light gradient boosting model (LGBM)

Procedia PDF Downloads 98
16633 On the Effect of Immigration on Destination: Country Corruption

Authors: Eugen Dimant, Tim Krieger, Margarete Redlin

Abstract:

This paper analyzes the impact of migration on destination-country corruption levels. Capitalizing on a comprehensive dataset consisting of annual immigration stocks of OECD coun-tries from 207 countries of origin for the period 1984-2008, we explore different channels through which corruption might migrate. We employ different estimation methods using fixed effects and Tobit regressions in order to validate our findings. What is more, we also address the issue of endogeneity by using the Difference-Generalized Method of Moments (GMM) estimator. Independent of the econometric methodology we consistently find that while general migration has an insignificant effect on the destination country’s corruption level, immigration from corruption-ridden origin countries boosts corruption in the destination country. Our findings provide a more profound understanding of the economic implications associated with migration flows.

Keywords: corruption, migration, impact of migration, destination-country corruption

Procedia PDF Downloads 317
16632 A Model-Reference Sliding Mode for Dual-Stage Actuator Servo Control in HDD

Authors: S. Sonkham, U. Pinsopon, W. Chatlatanagulchai

Abstract:

This paper presents a method of sliding mode control (SMC) designing and developing for the servo system in a dual-stage actuator (DSA) hard disk drive. Mathematical modelling of hard disk drive actuators is obtained, extracted from measuring frequency response of the voice-coil motor (VCM) and PZT micro-actuator separately. Matlab software tools are used for mathematical model estimation and also for controller design and simulation. A model-reference approach for tracking requirement is selected as a proposed technique. The simulation results show that performance of a model-reference SMC controller design in DSA servo control can be satisfied in the tracking error, as well as keeping the positioning of the head within the boundary of +/-5% of track width under the presence of internal and external disturbance. The overall results of model-reference SMC design in DSA are met per requirement specifications and significant reduction in %off track is found when compared to the single-state actuator (SSA).

Keywords: hard disk drive, dual-stage actuator, track following, hdd servo control, sliding mode control, model-reference, tracking control

Procedia PDF Downloads 361
16631 Vibration-Based Data-Driven Model for Road Health Monitoring

Authors: Guru Prakash, Revanth Dugalam

Abstract:

A road’s condition often deteriorates due to harsh loading such as overload due to trucks, and severe environmental conditions such as heavy rain, snow load, and cyclic loading. In absence of proper maintenance planning, this results in potholes, wide cracks, bumps, and increased roughness of roads. In this paper, a data-driven model will be developed to detect these damages using vibration and image signals. The key idea of the proposed methodology is that the road anomaly manifests in these signals, which can be detected by training a machine learning algorithm. The use of various machine learning techniques such as the support vector machine and Radom Forest method will be investigated. The proposed model will first be trained and tested with artificially simulated data, and the model architecture will be finalized by comparing the accuracies of various models. Once a model is fixed, the field study will be performed, and data will be collected. The field data will be used to validate the proposed model and to predict the future road’s health condition. The proposed will help to automate the road condition monitoring process, repair cost estimation, and maintenance planning process.

Keywords: SVM, data-driven, road health monitoring, pot-hole

Procedia PDF Downloads 80
16630 An Integreated Intuitionistic Fuzzy ELECTRE Model for Multi-Criteria Decision-Making

Authors: Babek Erdebilli

Abstract:

The aim of this study is to develop and describe a new methodology for the Multi-Criteria Decision-Making (MCDM) problem using IFE (Elimination Et Choix Traduisant La Realite (ELECTRE) model. The proposed models enable Decision-Makers (DMs) on the assessment and use Intuitionistic Fuzzy Numbers (IFN). A numerical example is provided to demonstrate and clarify the proposed analysis procedure. Also, an empirical experiment is conducted to validation the effectiveness.

Keywords: multi-criteria decision-making, IFE, DM’s, fuzzy electre model

Procedia PDF Downloads 642
16629 Forecasting Model to Predict Dengue Incidence in Malaysia

Authors: W. H. Wan Zakiyatussariroh, A. A. Nasuhar, W. Y. Wan Fairos, Z. A. Nazatul Shahreen

Abstract:

Forecasting dengue incidence in a population can provide useful information to facilitate the planning of the public health intervention. Many studies on dengue cases in Malaysia were conducted but are limited in modeling the outbreak and forecasting incidence. This article attempts to propose the most appropriate time series model to explain the behavior of dengue incidence in Malaysia for the purpose of forecasting future dengue outbreaks. Several seasonal auto-regressive integrated moving average (SARIMA) models were developed to model Malaysia’s number of dengue incidence on weekly data collected from January 2001 to December 2011. SARIMA (2,1,1)(1,1,1)52 model was found to be the most suitable model for Malaysia’s dengue incidence with the least value of Akaike information criteria (AIC) and Bayesian information criteria (BIC) for in-sample fitting. The models further evaluate out-sample forecast accuracy using four different accuracy measures. The results indicate that SARIMA (2,1,1)(1,1,1)52 performed well for both in-sample fitting and out-sample evaluation.

Keywords: time series modeling, Box-Jenkins, SARIMA, forecasting

Procedia PDF Downloads 474
16628 Removal of Cr (VI) from Water through Adsorption Process Using GO/PVA as Nanosorbent

Authors: Syed Hadi Hasan, Devendra Kumar Singh, Viyaj Kumar

Abstract:

Cr (VI) is a known toxic heavy metal and has been considered as a priority pollutant in water. The effluent of various industries including electroplating, anodizing baths, leather tanning, steel industries and chromium based catalyst are the major source of Cr (VI) contamination in the aquatic environment. Cr (VI) show high mobility in the environment and can easily penetrate cell membrane of the living tissues to exert noxious effects. The Cr (VI) contamination in drinking water causes various hazardous health effects to the human health such as cancer, skin and stomach irritation or ulceration, dermatitis, damage to liver, kidney circulation and nerve tissue damage. Herein, an attempt has been done to develop an efficient adsorbent for the removal of Cr (VI) from water. For this purpose nanosorbent composed of polyvinyl alcohol functionalized graphene oxide (GO/PVA) was prepared. Thus, obtained GO/PVA was characterized through FTIR, XRD, SEM, and Raman Spectroscopy. As prepared nanosorbent of GO/PVA was utilized for the removal Cr (VI) in batch mode experiment. The process variables such as contact time, initial Cr (VI) concentration, pH, and temperature were optimized. The maximum 99.8 % removal of Cr (VI) was achieved at initial Cr (VI) concentration 60 mg/L, pH 2, temperature 35 °C and equilibrium was achieved within 50 min. The two widely used isotherm models viz. Langmuir and Freundlich were analyzed using linear correlation coefficient (R2) and it was found that Langmuir model gives best fit with high value of R2 for the data of present adsorption system which indicate the monolayer adsorption of Cr (VI) on the GO/PVA. Kinetic studies were also conducted using pseudo-first order and pseudo-second order models and it was observed that chemosorptive pseudo-second order model described the kinetics of current adsorption system in better way with high value of correlation coefficient. Thermodynamic studies were also conducted and results showed that the adsorption was spontaneous and endothermic in nature.

Keywords: adsorption, GO/PVA, isotherm, kinetics, nanosorbent, thermodynamics

Procedia PDF Downloads 386
16627 Analysis of Diabetes Patients Using Pearson, Cost Optimization, Control Chart Methods

Authors: Devatha Kalyan Kumar, R. Poovarasan

Abstract:

In this paper, we have taken certain important factors and health parameters of diabetes patients especially among children by birth (pediatric congenital) where using the above three metrics methods we are going to assess the importance of each attributes in the dataset and thereby determining the most highly responsible and co-related attribute causing diabetics among young patients. We use cost optimization, control chart and Spearmen methodologies for the real-time application of finding the data efficiency in this diabetes dataset. The Spearmen methodology is the correlation methodologies used in software development process to identify the complexity between the various modules of the software. Identifying the complexity is important because if the complexity is higher, then there is a higher chance of occurrence of the risk in the software. With the use of control; chart mean, variance and standard deviation of data are calculated. With the use of Cost optimization model, we find to optimize the variables. Hence we choose the Spearmen, control chart and cost optimization methods to assess the data efficiency in diabetes datasets.

Keywords: correlation, congenital diabetics, linear relationship, monotonic function, ranking samples, pediatric

Procedia PDF Downloads 252
16626 Practical Design Procedures of 3D Reinforced Concrete Shear Wall-Frame Structure Based on Structural Optimization Method

Authors: H. Nikzad, S. Yoshitomi

Abstract:

This study investigates and develops the structural optimization method. The effect of size constraints on practical solution of reinforced concrete (RC) building structure with shear wall is proposed. Cross-sections of beam and column, and thickness of shear wall are considered as design variables. The objective function to be minimized is total cost of the structure by using a simple and efficient automated MATLAB platform structural optimization methodology. With modification of mathematical formulations, the result is compared with optimal solution without size constraints. The most suitable combination of section sizes is selected as for the final design application based on linear static analysis. The findings of this study show that defining higher value of upper bound of sectional sizes significantly affects optimal solution, and defining of size constraints play a vital role in finding of global and practical solution during optimization procedures. The result and effectiveness of proposed method confirm the ability and efficiency of optimal solutions for 3D RC shear wall-frame structure.

Keywords: structural optimization, linear static analysis, ETABS, MATLAB, RC shear wall-frame structures

Procedia PDF Downloads 371
16625 Creating Renewable Energy Investment Portfolio in Turkey between 2018-2023: An Approach on Multi-Objective Linear Programming Method

Authors: Berker Bayazit, Gulgun Kayakutlu

Abstract:

The World Energy Outlook shows that energy markets will substantially change within a few forthcoming decades. First, determined action plans according to COP21 and aim of CO₂ emission reduction have already impact on policies of countries. Secondly, swiftly changed technological developments in the field of renewable energy will be influential upon medium and long-term energy generation and consumption behaviors of countries. Furthermore, share of electricity on global energy consumption is to be expected as high as 40 percent in 2040. Electrical vehicles, heat pumps, new electronical devices and digital improvements will be outstanding technologies and innovations will be the testimony of the market modifications. In order to meet highly increasing electricity demand caused by technologies, countries have to make new investments in the field of electricity production, transmission and distribution. Specifically, electricity generation mix becomes vital for both prevention of CO₂ emission and reduction of power prices. Majority of the research and development investments are made in the field of electricity generation. Hence, the prime source diversity and source planning of electricity generation are crucial for improving the wealth of citizen life. Approaches considering the CO₂ emission and total cost of generation, are necessary but not sufficient to evaluate and construct the product mix. On the other hand, employment and positive contribution to macroeconomic values are important factors that have to be taken into consideration. This study aims to constitute new investments in renewable energies (solar, wind, geothermal, biogas and hydropower) between 2018-2023 under 4 different goals. Therefore, a multi-objective programming model is proposed to optimize the goals of minimizing the CO₂ emission, investment amount and electricity sales price while maximizing the total employment and positive contribution to current deficit. In order to avoid the user preference among the goals, Dinkelbach’s algorithm and Guzel’s approach have been combined. The achievements are discussed with comparison to the current policies. Our study shows that new policies like huge capacity allotment might be discussible although obligation for local production is positive. The improvements in grid infrastructure and re-design support for the biogas and geothermal can be recommended.

Keywords: energy generation policies, multi-objective linear programming, portfolio planning, renewable energy

Procedia PDF Downloads 240
16624 A Broadband Tri-Cantilever Vibration Energy Harvester with Magnetic Oscillator

Authors: Xiaobo Rui, Zhoumo Zeng, Yibo Li

Abstract:

A novel tri-cantilever energy harvester with magnetic oscillator was presented, which could convert the ambient vibration into electrical energy to power the low-power devices such as wireless sensor networks. The most common way to harvest vibration energy is based on the use of linear resonant devices such as cantilever beam, since this structure creates the highest strain for a given force. The highest efficiency will be achieved when the resonance frequency of the harvester matches the vibration frequency. The limitation of the structure is the narrow effective bandwidth. To overcome this limitation, this article introduces a broadband tri-cantilever harvester with nonlinear stiffness. This energy harvester typically consists of three thin cantilever beams vertically arranged with Neodymium Magnets ( NdFeB)magnetics at its free end and a fixed base at the other end. The three cantilevers have different resonant frequencies by designed in different thicknesses. It is obviously that a similar advantage of multiple resonant frequencies as piezoelectric cantilevers array structure is built. To achieve broadband energy harvesting, magnetic interaction is used to introduce the nonlinear system stiffness to tune the resonant frequency to match the excitation. Since the three cantilever tips are all free and the magnetic force is distance dependent, the resonant frequencies will be complexly changed with the vertical vibration of the free end. Both model and experiment are built. The electromechanically coupled lumped-parameter model is presented. An electromechanical formulation and analytical expressions for the coupled nonlinear vibration response and voltage response are given. The entire structure is fabricated and mechanically attached to a electromagnetic shaker as a vibrating body via the fixed base, in order to couple the vibrations to the cantilever. The cantilevers are bonded with piezoelectric macro-fiber composite (MFC) materials (Model: M8514P2). The size of the cantilevers is 120*20mm2 and the thicknesses are separately 1mm, 0.8mm, 0.6mm. The prototype generator has a measured performance of 160.98 mW effective electrical power and 7.93 DC output voltage via the excitation level of 10m/s2. The 130% increase in the operating bandwidth is achieved. This device is promising to support low-power devices, peer-to-peer wireless nodes, and small-scale wireless sensor networks in ambient vibration environment.

Keywords: tri-cantilever, ambient vibration, energy harvesting, magnetic oscillator

Procedia PDF Downloads 150
16623 Analysis of the Impact of NVivo and EndNote on Academic Research Productivity

Authors: Sujit K. Basak

Abstract:

The aim of this paper is to analyze the impact of literature review software on researchers. The aim of this study was achieved by analyzing models in terms of perceived usefulness, perceived ease of use, and acceptance level. Collected data was analyzed using WarpPLS 4.0 software. This study used two theoretical frameworks namely Technology Acceptance Model and the Training Needs Assessment Model. The study was experimental and was conducted at a public university in South Africa. The results of the study showed that acceptance level has a high impact on research workload and productivity followed by perceived usefulness and perceived ease of use.

Keywords: technology acceptance model, training needs assessment model, literature review software, research productivity

Procedia PDF Downloads 496
16622 A Spatial Approach to Model Mortality Rates

Authors: Yin-Yee Leong, Jack C. Yue, Hsin-Chung Wang

Abstract:

Human longevity has been experiencing its largest increase since the end of World War II, and modeling the mortality rates is therefore often the focus of many studies. Among all mortality models, the Lee–Carter model is the most popular approach since it is fairly easy to use and has good accuracy in predicting mortality rates (e.g., for Japan and the USA). However, empirical studies from several countries have shown that the age parameters of the Lee–Carter model are not constant in time. Many modifications of the Lee–Carter model have been proposed to deal with this problem, including adding an extra cohort effect and adding another period effect. In this study, we propose a spatial modification and use clusters to explain why the age parameters of the Lee–Carter model are not constant. In spatial analysis, clusters are areas with unusually high or low mortality rates than their neighbors, where the “location” of mortality rates is measured by age and time, that is, a 2-dimensional coordinate. We use a popular cluster detection method—Spatial scan statistics, a local statistical test based on the likelihood ratio test to evaluate where there are locations with mortality rates that cannot be described well by the Lee–Carter model. We first use computer simulation to demonstrate that the cluster effect is a possible source causing the problem of the age parameters not being constant. Next, we show that adding the cluster effect can solve the non-constant problem. We also apply the proposed approach to mortality data from Japan, France, the USA, and Taiwan. The empirical results show that our approach has better-fitting results and smaller mean absolute percentage errors than the Lee–Carter model.

Keywords: mortality improvement, Lee–Carter model, spatial statistics, cluster detection

Procedia PDF Downloads 168
16621 An Infinite Mixture Model for Modelling Stutter Ratio in Forensic Data Analysis

Authors: M. A. C. S. Sampath Fernando, James M. Curran, Renate Meyer

Abstract:

Forensic DNA analysis has received much attention over the last three decades, due to its incredible usefulness in human identification. The statistical interpretation of DNA evidence is recognised as one of the most mature fields in forensic science. Peak heights in an Electropherogram (EPG) are approximately proportional to the amount of template DNA in the original sample being tested. A stutter is a minor peak in an EPG, which is not masking as an allele of a potential contributor, and considered as an artefact that is presumed to be arisen due to miscopying or slippage during the PCR. Stutter peaks are mostly analysed in terms of stutter ratio that is calculated relative to the corresponding parent allele height. Analysis of mixture profiles has always been problematic in evidence interpretation, especially with the presence of PCR artefacts like stutters. Unlike binary and semi-continuous models; continuous models assign a probability (as a continuous weight) for each possible genotype combination, and significantly enhances the use of continuous peak height information resulting in more efficient reliable interpretations. Therefore, the presence of a sound methodology to distinguish between stutters and real alleles is essential for the accuracy of the interpretation. Sensibly, any such method has to be able to focus on modelling stutter peaks. Bayesian nonparametric methods provide increased flexibility in applied statistical modelling. Mixture models are frequently employed as fundamental data analysis tools in clustering and classification of data and assume unidentified heterogeneous sources for data. In model-based clustering, each unknown source is reflected by a cluster, and the clusters are modelled using parametric models. Specifying the number of components in finite mixture models, however, is practically difficult even though the calculations are relatively simple. Infinite mixture models, in contrast, do not require the user to specify the number of components. Instead, a Dirichlet process, which is an infinite-dimensional generalization of the Dirichlet distribution, is used to deal with the problem of a number of components. Chinese restaurant process (CRP), Stick-breaking process and Pólya urn scheme are frequently used as Dirichlet priors in Bayesian mixture models. In this study, we illustrate an infinite mixture of simple linear regression models for modelling stutter ratio and introduce some modifications to overcome weaknesses associated with CRP.

Keywords: Chinese restaurant process, Dirichlet prior, infinite mixture model, PCR stutter

Procedia PDF Downloads 321
16620 Impact of VARK Learning Model at Tertiary Level Education

Authors: Munazza A. Mirza, Khawar Khurshid

Abstract:

Individuals are generally associated with different learning styles, which have been explored extensively in recent past. The learning styles refer to the potential of an individual by which s/he can easily comprehend and retain information. Among various learning style models, VARK is the most accepted model which categorizes the learners with respect to their sensory characteristics. Based on the number of preferred learning modes, the learners can be categorized as uni-modal, bi-modal, tri-modal, or quad/multi-modal. Although there is a prevalent belief in the learning styles, however, the model is not being frequently and effectively utilized in the higher education. This research describes the identification model to validate teacher’s didactic practice and student’s performance linkage with the learning styles. The identification model is recommended to check the effective application and evaluation of the various learning styles. The proposed model is a guideline to effectively implement learning styles inventory in order to ensure that it will validate performance linkage with learning styles. If performance is linked with learning styles, this may help eradicate the distrust on learning style theory. For this purpose, a comprehensive study was conducted to compare and understand how VARK inventory model is being used to identify learning preferences and their correlation with learner’s performance. A comparative analysis of the findings of these studies is presented to understand the learning styles of tertiary students in various disciplines. It is concluded with confidence that the learning styles of students cannot be associated with any specific discipline. Furthermore, there is not enough empirical proof to link performance with learning styles.

Keywords: learning style, VARK, sensory preferences, identification model, didactic practices

Procedia PDF Downloads 267
16619 Magnetohydrodynamics Flow and Heat Transfer in a Non-Newtonian Power-Law Fluid due to a Rotating Disk with Velocity Slip and Temperature Jump

Authors: Nur Dayana Khairunnisa Rosli, Seripah Awang Kechil

Abstract:

Swirling flows with velocity slip are important in nature and industrial processes. The present work considers the effects of velocity slip, temperature jump and suction/injection on the flow and heat transfer of power-law fluids due to a rotating disk in the presence of magnetic field. The system of the partial differential equations is highly non-linear. The number of independent variables is reduced by transforming the system into a system of coupled non-linear ordinary differential equations using similarity transformations. The effects of suction/injection, velocity slip and temperature jump on the flow rates are investigated for various cases of shear thinning and shear thickening power law fluids. The thermal and velocity jump strongly reduce the heat transfer rate and skin friction coefficient. Suction decreases the radial and tangential skin friction coefficient and the rate of heat transfer. It is also observed that the effects are more pronounced in the case of shear thinning fluids as compared to shear thickening fluids.

Keywords: heat transfer, power-law fluids, rotating disk, suction or injection, temperature jump, velocity slip

Procedia PDF Downloads 258
16618 Applying the Extreme-Based Teaching Model in Post-Secondary Online Classroom Setting: A Field Experiment

Authors: Leon Pan

Abstract:

The first programming course within post-secondary education has long been recognized as a challenging endeavor for both educators and students alike. Historically, these courses have exhibited high failure rates and a notable number of dropouts. Instructors often lament students' lack of effort in their coursework, and students often express frustration that the teaching methods employed are not effective. Drawing inspiration from the successful principles of Extreme Programming, this study introduces an approach—the Extremes-based teaching model — aimed at enhancing the teaching of introductory programming courses. To empirically determine the effectiveness of the model, a comparison was made between a section taught using the extreme-based model and another utilizing traditional teaching methods. Notably, the extreme-based teaching class required students to work collaboratively on projects while also demanding continuous assessment and performance enhancement within groups. This paper details the application of the extreme-based model within the post-secondary online classroom context and presents the compelling results that emphasize its effectiveness in advancing the teaching and learning experiences. The extreme-based model led to a significant increase of 13.46 points in the weighted total average and a commendable 10% reduction in the failure rate.

Keywords: extreme-based teaching model, innovative pedagogical methods, project-based learning, team-based learning

Procedia PDF Downloads 53
16617 Functional Decomposition Based Effort Estimation Model for Software-Intensive Systems

Authors: Nermin Sökmen

Abstract:

An effort estimation model is needed for software-intensive projects that consist of hardware, embedded software or some combination of the two, as well as high level software solutions. This paper first focuses on functional decomposition techniques to measure functional complexity of a computer system and investigates its impact on system development effort. Later, it examines effects of technical difficulty and design team capability factors in order to construct the best effort estimation model. With using traditional regression analysis technique, the study develops a system development effort estimation model which takes functional complexity, technical difficulty and design team capability factors as input parameters. Finally, the assumptions of the model are tested.

Keywords: functional complexity, functional decomposition, development effort, technical difficulty, design team capability, regression analysis

Procedia PDF Downloads 286
16616 Robustified Asymmetric Logistic Regression Model for Global Fish Stock Assessment

Authors: Osamu Komori, Shinto Eguchi, Hiroshi Okamura, Momoko Ichinokawa

Abstract:

The long time-series data on population assessments are essential for global ecosystem assessment because the temporal change of biomass in such a database reflects the status of global ecosystem properly. However, the available assessment data usually have limited sample sizes and the ratio of populations with low abundance of biomass (collapsed) to those with high abundance (non-collapsed) is highly imbalanced. To allow for the imbalance and uncertainty involved in the ecological data, we propose a binary regression model with mixed effects for inferring ecosystem status through an asymmetric logistic model. In the estimation equation, we observe that the weights for the non-collapsed populations are relatively reduced, which in turn puts more importance on the small number of observations of collapsed populations. Moreover, we extend the asymmetric logistic regression model using propensity score to allow for the sample biases observed in the labeled and unlabeled datasets. It robustified the estimation procedure and improved the model fitting.

Keywords: double robust estimation, ecological binary data, mixed effect logistic regression model, propensity score

Procedia PDF Downloads 260
16615 Investigating the Challenges Faced by English Language Teachers in Implementing Outcome Based Education the Outcome Based Education model in Engineering Universities of Sindh

Authors: Habibullah Pathan

Abstract:

The present study aims to explore problems faced by English Language Teachers (ELT) while implementing the Outcome Based Education (OBE) model in engineering universities of Sindh. OBE is an emerging model initiative of the International Engineering Alliance. Traditional educational systems are teacher-centered or curriculum-centered, in which learners are not able to achieve desired outcomes, but the OBE model enables learners to know the outcomes before the start of the program. OBE is a circular process that begins from the needs and demands of society to stakeholders who ask the experts to produce the alumnus who can fulfill the needs and ends up getting new enrollment in the respective programs who can work according to the demands. In all engineering institutions, engineering courses besides English language courses are taught on the OBE model. English language teachers were interviewed to learn the in-depth of the problems faced by them. The study found that teachers were facing problems including pedagogical, OBE training, assessment, evaluation and administrative support. This study will be a guide for public and private English language teachers to cope with these challenges while teaching the English language on the OBE model. OBE is an emerging model by which the institutions can produce such a product that can meet the demands.

Keywords: problems of ELT teachers, outcome based education (OBE), implementing, assessment

Procedia PDF Downloads 82
16614 On the Application and Comparison of Two Geostatistics Methods in the Parameterisation Step to Calibrate Groundwater Model: Grid-Based Pilot Point and Head-Zonation Based Pilot Point Methods

Authors: Dua K. S. Y. Klaas, Monzur A. Imteaz, Ika Sudiayem, Elkan M. E. Klaas, Eldav C. M. Klaas

Abstract:

Properly selecting the most suitable and effective geostatistics method in the parameterization step of groundwater modeling is critical to attain a satisfactory model. In this paper, two geostatistics methods, i.e., Grid-Based Pilot Point (GB-PP) and Head-Zonation Based Pilot Point (HZB-PP) methods, were applied in an eogenetic karst catchment and compared using as model performances and computation time the criteria. Overall, the results show that appropriate selection of method is substantial in the parameterization of physically-based groundwater models, as it influences both the accuracy and simulation times. It was found that GB-PP method performed comparably superior to HZB-PP method. However, reflecting its model performances, HZB-PP method is promising for further application in groundwater modeling.

Keywords: groundwater model, geostatistics, pilot point, parameterization step

Procedia PDF Downloads 158
16613 The Development of Nursing Model for Pregnant Women to Prevention of Early Postpartum Hemorrhage

Authors: Wadsana Sarakarn, Pimonpan Charoensri, Baliya Chaiyara

Abstract:

Objectives: To study the outcomes of the developed nursing model to prevent early postpartum hemorrhage (PPH). Materials and Methods: The analytical study was conducted in Sunpasitthiprasong Hospital during October 1st, 2015, until May 31st, 2017. After review the prevalence, risk factors, and outcomes of postpartum hemorrhage of the parturient who gave birth in Sunpasitthiprasong Hospital, the nursing model was developed under research regulation of Kemmis&McTaggart using 4 steps of operating procedures: 1) analyzing problem situation and gathering 2) creating the plan 3) noticing and performing 4) reflecting the result of the operation. The nursing model consisted of the screening tools for risk factors associated with PPH, the clinical nursing practice guideline (CNPG), and the collecting bag for measuring postpartum blood loss. Primary outcome was early postpartum hemorrhage. Secondary outcomes were postpartum hysterectomy, maternal mortality, personnel’s practice, knowledge, and satisfaction of the nursing model. The data were analyzed by using content analysis for qualitative data and descriptive statistics for quantitative data. Results: Before using the nursing model, the prevalence of early postpartum hemorrhage was under estimated (2.97%). There were 5 cases of postpartum hysterectomy and 2 cases of maternal death due to postpartum hemorrhage. During the study period, there was 22.7% prevalence of postpartum hemorrhage among 220 pregnant women who were vaginally delivered at Sunpasitthiprasong Hospital. No maternal death or postpartum hysterectomy was reported after using the nursing model. Among 16 registered nurses at the delivery room who evaluated using of the nursing model, they reported the high level of practice, knowledge, and satisfaction Conclusion: The nursing model for the prevention of early PPH is effective to decrease early PPH and other serious complications.

Keywords: the development of a nursing model, prevention of postpartum hemorrhage, pregnant women, postpartum hemorrhage

Procedia PDF Downloads 94
16612 Nonlinear Porous Diffusion Modeling of Ionic Agrochemicals in Astomatous Plant Cuticle Aqueous Pores: A Mechanistic Approach

Authors: Eloise C. Tredenick, Troy W. Farrell, W. Alison Forster, Steven T. P. Psaltis

Abstract:

The agriculture industry requires improved efficacy of sprays being applied to crops. More efficacious sprays provide many environmental and financial benefits. The plant leaf cuticle is known to be the main barrier to diffusion of agrochemicals within the leaf. The importance of a mathematical model to simulate uptake of agrochemicals in plant cuticles has been noted, as the results of each uptake experiments are specific to each formulation of active ingredient and plant species. In this work we develop a mathematical model and numerical simulation for the uptake of ionic agrochemicals through aqueous pores in plant cuticles. We propose a nonlinear porous diffusion model of ionic agrochemicals in isolated cuticles, which provides additions to a simple diffusion model through the incorporation of parameters capable of simulating plant species' variations, evaporation of surface droplet solutions and swelling of the aqueous pores with water. The model could feasibly be adapted to other ionic active ingredients diffusing through other plant species' cuticles. We validate our theoretical results against appropriate experimental data, discuss the key sensitivities in the model and relate theoretical predictions to appropriate physical mechanisms.

Keywords: aqueous pores, ionic active ingredient, mathematical model, plant cuticle, porous diffusion

Procedia PDF Downloads 256
16611 Simulation of Government Management Model to Increase Financial Productivity System Using Govpilot

Authors: Arezou Javadi

Abstract:

The use of algorithmic models dependent on software calculations and simulation of new government management assays with the help of specialized software had increased the productivity and efficiency of the government management system recently. This has caused the management approach to change from the old bitch & fix model, which has low efficiency and less usefulness, to the capable management model with higher efficiency called the partnership with resident model. By using Govpilot TM software, the relationship between people in a system and the government was examined. The method of two tailed interaction was the outsourcing of a goal in a system, which is formed in the order of goals, qualified executive people, optimal executive model, and finally, summarizing additional activities at the different statistical levels. The results showed that the participation of people in a financial implementation system with a statistical potential of P≥5% caused a significant increase in investment and initial capital in the government system with maximum implement project in a smart government.

Keywords: machine learning, financial income, statistical potential, govpilot

Procedia PDF Downloads 82
16610 Simulation of Government Management Model to Increase Financial Productivity System Using Govpilot

Authors: Arezou Javadi

Abstract:

The use of algorithmic models dependent on software calculations and simulation of new government management assays with the help of specialized software had increased the productivity and efficiency of the government management system recently. This has caused the management approach to change from the old bitch & fix model, which has low efficiency and less usefulness, to the capable management model with higher efficiency called the partnership with resident model. By using Govpilot TM software, the relationship between people in a system and the government was examined. The method of two tailed interaction was the outsourcing of a goal in a system, which is formed in the order of goals, qualified executive people, optimal executive model, and finally, summarizing additional activities at the different statistical levels. The results showed that the participation of people in a financial implementation system with a statistical potential of P≥5% caused a significant increase in investment and initial capital in the government system with maximum implement project in a smart government.

Keywords: machine learning, financial income, statistical potential, govpilot

Procedia PDF Downloads 65
16609 A Study of Using Multiple Subproblems in Dantzig-Wolfe Decomposition of Linear Programming

Authors: William Chung

Abstract:

This paper is to study the use of multiple subproblems in Dantzig-Wolfe decomposition of linear programming (DW-LP). Traditionally, the decomposed LP consists of one LP master problem and one LP subproblem. The master problem and the subproblem is solved alternatively by exchanging the dual prices of the master problem and the proposals of the subproblem until the LP is solved. It is well known that convergence is slow with a long tail of near-optimal solutions (asymptotic convergence). Hence, the performance of DW-LP highly depends upon the number of decomposition steps. If the decomposition steps can be greatly reduced, the performance of DW-LP can be improved significantly. To reduce the number of decomposition steps, one of the methods is to increase the number of proposals from the subproblem to the master problem. To do so, we propose to add a quadratic approximation function to the LP subproblem in order to develop a set of approximate-LP subproblems (multiple subproblems). Consequently, in each decomposition step, multiple subproblems are solved for providing multiple proposals to the master problem. The number of decomposition steps can be reduced greatly. Note that each approximate-LP subproblem is nonlinear programming, and solving the LP subproblem must faster than solving the nonlinear multiple subproblems. Hence, using multiple subproblems in DW-LP is the tradeoff between the number of approximate-LP subproblems being formed and the decomposition steps. In this paper, we derive the corresponding algorithms and provide some simple computational results. Some properties of the resulting algorithms are also given.

Keywords: approximate subproblem, Dantzig-Wolfe decomposition, large-scale models, multiple subproblems

Procedia PDF Downloads 155
16608 Model Predictive Control Using Thermal Inputs for Crystal Growth Dynamics

Authors: Takashi Shimizu, Tomoaki Hashimoto

Abstract:

Recently, crystal growth technologies have made progress by the requirement for the high quality of crystal materials. To control the crystal growth dynamics actively by external forces is useuful for reducing composition non-uniformity. In this study, a control method based on model predictive control using thermal inputs is proposed for crystal growth dynamics of semiconductor materials. The control system of crystal growth dynamics considered here is governed by the continuity, momentum, energy, and mass transport equations. To establish the control method for such thermal fluid systems, we adopt model predictive control known as a kind of optimal feedback control in which the control performance over a finite future is optimized with a performance index that has a moving initial time and terminal time. The objective of this study is to establish a model predictive control method for crystal growth dynamics of semiconductor materials.

Keywords: model predictive control, optimal control, process control, crystal growth

Procedia PDF Downloads 350
16607 Observer-Based Leader-Following Consensus of Nonlinear Fractional-Order Multi-Agent Systems

Authors: Ali Afaghi, Sehraneh Ghaemi

Abstract:

The coordination of the multi-agent systems has been one of the interesting topic in recent years, because of its potential applications in many branches of science and engineering such as sensor networks, flocking, underwater vehicles and etc. In the most of the related studies, it is assumed that the dynamics of the multi-agent systems are integer-order and linear and the multi-agent systems with the fractional-order nonlinear dynamics are rarely considered. However many phenomena in nature cannot be described within integer-order and linear characteristics. This paper investigates the leader-following consensus problem for a class of nonlinear fractional-order multi-agent systems based on observer-based cooperative control. In the system, the dynamics of each follower and leader are nonlinear. For a multi-agent system with fixed directed topology firstly, an observer-based consensus protocol is proposed based on the relative observer states of neighboring agents. Secondly, based on the property of the stability theory of fractional-order system, some sufficient conditions are presented for the asymptotical stability of the observer-based fractional-order control systems. The proposed method is applied on a five-agent system with the fractional-order nonlinear dynamics and unavailable states. The simulation example shows that the proposed scenario results in the good performance and can be used in many practical applications.

Keywords: fractional-order multi-agent systems, leader-following consensus, nonlinear dynamics, directed graphs

Procedia PDF Downloads 391
16606 Stability Analysis of Modelling the Effect of Vaccination and Novel Quarantine-Adjusted Incidence on the Spread of Newcastle Disease

Authors: Nurudeen O. Lasisi, Sirajo Abdulrahman, Abdulkareem A. Ibrahim

Abstract:

Newcastle disease is an infection of domestic poultry and other bird species with the virulent Newcastle disease virus (NDV). In this paper, we study the dynamics of the modeling of the Newcastle disease virus (NDV) using a novel quarantine-adjusted incidence. The comparison of Vaccination, linear incident rate and novel quarantine-adjusted incident rate in the models are discussed. The dynamics of the models yield disease-free and endemic equilibrium states.The effective reproduction numbers of the models are computed in order to measure the relative impact of an individual bird or combined intervention for effective disease control. We showed the local and global stability of endemic equilibrium states of the models and we found that the stability of endemic equilibrium states of models are globally asymptotically stable if the effective reproduction numbers of the models equations are greater than a unit.

Keywords: effective reproduction number, Endemic state, Mathematical model, Newcastle disease virus, novel quarantine-adjusted incidence, stability analysis

Procedia PDF Downloads 108