Search results for: Normalized least mean square methods
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4650

Search results for: Normalized least mean square methods

2910 A POX Controller Module to Collect Web Traffic Statistics in SDN Environment

Authors: Wisam H. Muragaa, Kamaruzzaman Seman, Mohd Fadzli Marhusin

Abstract:

Software Defined Networking (SDN) is a new norm of networks. It is designed to facilitate the way of managing, measuring, debugging and controlling the network dynamically, and to make it suitable for the modern applications. Generally, measurement methods can be divided into two categories: Active and passive methods. Active measurement method is employed to inject test packets into the network in order to monitor their behaviour (ping tool as an example). Meanwhile the passive measurement method is used to monitor the traffic for the purpose of deriving measurement values. The measurement methods, both active and passive, are useful for the collection of traffic statistics, and monitoring of the network traffic. Although there has been a work focusing on measuring traffic statistics in SDN environment, it was only meant for measuring packets and bytes rates for non-web traffic. In this study, a feasible method will be designed to measure the number of packets and bytes in a certain time, and facilitate obtaining statistics for both web traffic and non-web traffic. Web traffic refers to HTTP requests that use application layer; while non-web traffic refers to ICMP and TCP requests. Thus, this work is going to be more comprehensive than previous works. With a developed module on POX OpenFlow controller, information will be collected from each active flow in the OpenFlow switch, and presented on Command Line Interface (CLI) and wireshark interface. Obviously, statistics that will be displayed on CLI and on wireshark interfaces include type of protocol, number of bytes and number of packets, among others. Besides, this module will show the number of flows added to the switch whenever traffic is generated from and to hosts in the same statistics list. In order to carry out this work effectively, our Python module will send a statistics request message to the switch requesting its current ports and flows statistics in every five seconds; while the switch will reply with the required information in a message called statistics reply message. Thus, POX controller will be notified and updated with any changes could happen in the entire network in a very short time. Therefore, our aim of this study is to prepare a list for the important statistics elements that are collected from the whole network, to be used for any further researches; particularly, those that are dealing with the detection of the network attacks that cause a sudden rise in the number of packets and bytes like Distributed Denial of Service (DDoS).

Keywords: Mininet, OpenFlow, POX controller, SDN.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2972
2909 Vocational Skills, Recognition of Prior Learning and Technology: The Future of Higher Education

Authors: Shankar Subramanian Iyer

Abstract:

The vocational education, enhanced by technology and Recognition of Prior Learning (RPL) is going to be the main ingredient of the future of education. This is coming from the various issues of the current educational system like cost, time, type of course, type of curriculum, unemployment, to name the major reasons. Most millennials like to perform and learn rather than learning how to perform. This is the essence of vocational education be it any field from cooking, painting, plumbing to modern technologies using computers. Even a more theoretical course like entrepreneurship can be taught as to be an entrepreneur and learn about its nuances. The best way to learn accountancy is actually keeping accounts for a small business or grocer and learn the ropes of accountancy and finance. The purpose of this study is to investigate the relationship between vocational skills, RPL and new technologies with future employability. This study implies that individual's knowledge and skills are essential aspects to be emphasized in future education and to give credit for prior experience for future employability. Virtual reality can be used to stimulate workplace situations for vocational learning for fields like hospitality, medical emergencies, healthcare, draughtsman ship, building inspection, quantity surveying, estimation, to name a few. All disruptions in future education, especially vocational education, are going to be technology driven with the advent of AI, ML, IoT, VR, VI etc. Vocational education not only helps institutes cut costs drastically, but allows all students to have hands-on experiences, rather than to be observers. The earlier experiential learning theory and the recent theory of knowledge and skills-based learning modified and applied to the vocational education and development of skills is the proposed contribution of this paper. Apart from secondary research study on major scholarly articles, books, primary research using interviews, questionnaire surveys have been used to validate and test the reliability of the suggested model using Partial Least Square- Structural Equation Method (PLS-SEM), the factors being assimilated using an existing literature review. Major findings have been that there exists high relationship between the vocational skills, RPL, new technology to the future employability through mediation of future employability skills.

Keywords: Vocational education, vocational skills, competencies, modern technologies, Recognition of Prior Learning, RPL.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 775
2908 Using Scanning Electron Microscope and Computed Tomography for Concrete Diagnostics of Airfield Pavements

Authors: M. Linek

Abstract:

This article presents the comparison of selected evaluation methods regarding microstructure modification of hardened cement concrete intended for airfield pavements. Basic test results were presented for two pavement quality concrete lots. Analysis included standard concrete used for airfield pavements and modern material solutions based on concrete composite modification. In case of basic grain size distribution of concrete cement CEM I 42,5HSR NA, fine aggregate and coarse aggregate fractions in the form of granite chippings, water and admixtures were considered. In case of grain size distribution of modified concrete, the use of modern modifier as substitute of fine aggregate was suggested. Modification influence on internal concrete structure parameters using scanning electron microscope was defined. Obtained images were compared to the results obtained using computed tomography. Opportunity to use this type of equipment for internal concrete structure diagnostics and an attempt of its parameters evaluation was presented. Obtained test results enabled to reach a conclusion that both methods can be applied for pavement quality concrete diagnostics, with particular purpose of airfield pavements.

Keywords: Scanning electron microscope, computed tomography, cement concrete, airfield pavements.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1115
2907 Development of Integrated GIS Interface for Characteristics of Regional Daily Flow

Authors: Ju Young Lee, Jung-Seok Yang, Jaeyoung Choi

Abstract:

The purpose of this paper primarily intends to develop GIS interface for estimating sequences of stream-flows at ungauged stations based on known flows at gauged stations. The integrated GIS interface is composed of three major steps. The first, precipitation characteristics using statistical analysis is the procedure for making multiple linear regression equation to get the long term mean daily flow at ungauged stations. The independent variables in regression equation are mean daily flow and drainage area. Traditionally, mean flow data are generated by using Thissen polygon method. However, method for obtaining mean flow data can be selected by user such as Kriging, IDW (Inverse Distance Weighted), Spline methods as well as other traditional methods. At the second, flow duration curve (FDC) is computing at unguaged station by FDCs in gauged stations. Finally, the mean annual daily flow is computed by spatial interpolation algorithm. The third step is to obtain watershed/topographic characteristics. They are the most important factors which govern stream-flows. In summary, the simulated daily flow time series are compared with observed times series. The results using integrated GIS interface are closely similar and are well fitted each other. Also, the relationship between the topographic/watershed characteristics and stream flow time series is highly correlated.

Keywords: Integrated GIS interface, spatial interpolation algorithm, FDC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1510
2906 Environmental Assessment Methods in Abu Dhabi

Authors: Huda Al Salmi, Hisham Al Kadi, Simone Leao

Abstract:

Abu Dhabi is one of the fastest developed cities in the region. On top of all the current and future environmental challenges, Abu Dhabi aims to be among the top governments in the world in sustainable development. Abu Dhabi plans to create an attractive, livable and sustainable managed urban environment in which all necessary services and infrastructure are provided in a sustainable and timely manner. Abu Dhabi is engaged in a difficult challenge to develop credible environmental indicators that would assess the ambitious environmental targets. The aim of those indicators is to provide reliable guidance to decision makers and the public concerning key factors that determine the state of urban environment and identify major areas for policy intervention. In order to ensure sustainable development in UAE in general, and of Abu Dhabi City in particular, relevant and contextual environmental indicators need to be carefully considered. These indicators provide a gauge at a national government scale of how close countries are to establish environmental policy goals. The environment indicators assist city decision-making in such areas as identification of significant environmental aspects and observation of environmental performance trends. Those can help to find ways of reducing environmental pollution and in improving eco-efficiency. This paper outlines recent strategies implemented in Abu Dhabi that aims to improve the sustainable performance of the city-s built environment. The paper explores the variety of current and possible indicators at different levels and their roles in the development of the city.

Keywords: Abu Dhabi, Sustainable Development, Indicators, Assessment Methods.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2738
2905 Probabilistic Method of Wind Generation Placement for Congestion Management

Authors: S. Z. Moussavi, A. Badri, F. Rastegar Kashkooli

Abstract:

Wind farms (WFs) with high level of penetration are being established in power systems worldwide more rapidly than other renewable resources. The Independent System Operator (ISO), as a policy maker, should propose appropriate places for WF installation in order to maximize the benefits for the investors. There is also a possibility of congestion relief using the new installation of WFs which should be taken into account by the ISO when proposing the locations for WF installation. In this context, efficient wind farm (WF) placement method is proposed in order to reduce burdens on congested lines. Since the wind speed is a random variable and load forecasts also contain uncertainties, probabilistic approaches are used for this type of study. AC probabilistic optimal power flow (P-OPF) is formulated and solved using Monte Carlo Simulations (MCS). In order to reduce computation time, point estimate methods (PEM) are introduced as efficient alternative for time-demanding MCS. Subsequently, WF optimal placement is determined using generation shift distribution factors (GSDF) considering a new parameter entitled, wind availability factor (WAF). In order to obtain more realistic results, N-1 contingency analysis is employed to find the optimal size of WF, by means of line outage distribution factors (LODF). The IEEE 30-bus test system is used to show and compare the accuracy of proposed methodology.

Keywords: Probabilistic optimal power flow, Wind power, Pointestimate methods, Congestion management

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1888
2904 Hypertensive Response to Maximal Exercise Test in Young and Middle Age Hypertensive on Blood Pressure Lowering Medication: Monotherapy vs. Combination Therapy

Authors: James Patrick A. Diaz, Raul E. Ramboyong

Abstract:

Background: Hypertensive response during maximal exercise test provides important information on the level of blood pressure control and evaluation of treatment. Method: A single center retrospective descriptive study was conducted among 117 young (aged 20 to 40) and middle age (aged 40 to 65) hypertensive patients, who underwent treadmill stress test. Currently on maintenance frontline medication either monotherapy (Angiotensin-converting enzyme inhibitor/Angiotensin receptor blocker [ACEi/ARB], Calcium channel blocker [CCB], Diuretic - Hydrochlorthiazide [HCTZ]) or combination therapy (ARB+CCB, ARB+HCTZ), who attained a maximal exercise on treadmill stress test (TMST) with hypertensive response (systolic blood pressure: male >210 mm Hg, female >190 mm Hg, diastolic blood pressure >100 mmHg, or increase of >10 mm Hg at any time during the test), on Bruce and Modified Bruce protocol. Exaggerated blood pressure response during exercise (systolic [SBP] and diastolic [DBP]), peak exercise blood pressure (SBP and DBP), recovery period (SBP and DBP) and test for ischemia and their antihypertensive medication/s were investigated. Analysis of variance and chi-square test were used for statistical analysis. Results: Hypertensive responses on maximal exercise test were seen mostly among female population (P < 0.000) and middle age (P < 0.000) patients. Exaggerated diastolic blood pressure responses were significantly lower in patients who were taking CCB (P < 0.004). A longer recovery period that showed a delayed decline in SBP was observed in patients taking ARB+HCTZ (P < 0.036). There were no significant differences in the level of exaggerated systolic blood pressure response and during peak exercise (both systolic and diastolic) in patients using either monotherapy or combination antihypertensives. Conclusion: Calcium channel blockers provided lower exaggerated diastolic BP response during maximal exercise test in hypertensive middle age patients. Patients on combination therapy using ARB+HCTZ exhibited a longer recovery period of systolic blood pressure.

Keywords: Antihypertensive, exercise test, hypertension, hypertensive response.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 920
2903 Hypoglycemic Activity of Water Soluble Polysaccharides of Yam (Dioscorea hispida Dents) Prepared by Aqueous, Papain, and Tempeh Inoculum Assisted Extractions

Authors: Teti Estiasih, Harijono, Weny Bekti Sunarharum, Atina Rahmawati

Abstract:

This research studied the hypoglycemic effect of water soluble polysaccharide (WSP) extracted from yam (Dioscorea hispida) tuber by three different methods: aqueous extraction, papain assisted extraction, and tempeh inoculums assisted extraction. The two later extraction methods were aimed to remove WSP binding protein to have more pure WSP. The hypoglycemic activities were evaluated by means in vivo test on alloxan induced hyperglycemic rats, glucose response test (GRT), in situ glucose absorption test using everted sac, and short chain fatty acids (SCFAs) analysis. All yam WSP extracts exhibited ability to decrease blood glucose level in hyperglycemia condition as well as inhibited glucose absorption and SCFA formation. The order of hypoglycemic activity was tempeh inoculums assisted- >papain assisted- >aqueous WSP extracts. GRT and in situ glucose absorption test showed that order of inhibition was papain assisted- >tempeh inoculums assisted- >aqueous WSP extracts. Digesta of caecum of yam WSP extracts oral fed rats had more SCFA than control. Tempeh inoculums assisted WSP extract exhibited the most significant hypoglycemic activity.

Keywords: hypoglycemic activity, papain, tempeh inoculums, water soluble polysaccharides, yam (Discorea hispida)

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3054
2902 Comparison between Higher-Order SVD and Third-order Orthogonal Tensor Product Expansion

Authors: Chiharu Okuma, Jun Murakami, Naoki Yamamoto

Abstract:

In digital signal processing it is important to approximate multi-dimensional data by the method called rank reduction, in which we reduce the rank of multi-dimensional data from higher to lower. For 2-dimennsional data, singular value decomposition (SVD) is one of the most known rank reduction techniques. Additional, outer product expansion expanded from SVD was proposed and implemented for multi-dimensional data, which has been widely applied to image processing and pattern recognition. However, the multi-dimensional outer product expansion has behavior of great computation complex and has not orthogonally between the expansion terms. Therefore we have proposed an alterative method, Third-order Orthogonal Tensor Product Expansion short for 3-OTPE. 3-OTPE uses the power method instead of nonlinear optimization method for decreasing at computing time. At the same time the group of B. D. Lathauwer proposed Higher-Order SVD (HOSVD) that is also developed with SVD extensions for multi-dimensional data. 3-OTPE and HOSVD are similarly on the rank reduction of multi-dimensional data. Using these two methods we can obtain computation results respectively, some ones are the same while some ones are slight different. In this paper, we compare 3-OTPE to HOSVD in accuracy of calculation and computing time of resolution, and clarify the difference between these two methods.

Keywords: Singular value decomposition (SVD), higher-order SVD (HOSVD), higher-order tensor, outer product expansion, power method.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1562
2901 Two Phase Frictional Pressure Drop of Carbon Dioxide in Horizontal Micro Tubes

Authors: M. Tarawneh

Abstract:

Two-phase frictional pressure drop data were obtained for condensation of carbon dioxide in single horizontal micro tube of inner diameter ranged from 0.6 mm up to 1.6 mm over mass flow rates from 2.5*10-5 to 17*10-5 kg/s and vapor qualities from 0.0 to 1.0. The inlet condensing pressure is changed from 33.5 to 45 bars. The saturation temperature ranged from -1.5 oC up to 10 oC. These data have then been compared against three (two-phase) frictional pressure drop prediction methods. The first method is by Muller-Steinhagen and Heck (Muller-Steinhagen H, Heck K. A simple friction pressure drop correlation for two-phase flow in pipes. Chem. Eng. Process 1986;20:297–308) and that by Gronnerud R. Investigation of liquid hold-up, flow-resistance and heat transfer in circulation type evaporators, part IV: two-phase flow resistance in boiling refrigerants, Annexe 1972. Then the method used by FriedelL. Improved friction pressures drop in horizontal and vertical two-phase pipe flow. European Two-Phase Flow Group Meeting, Paper E2; 1979 June, Ispra, Italy. The methods are used by M.B Ould Didi et al (2001) “Prediction of two-phase pressure gradients of refrigerant in horizontal tubes". Int.J.of Refrigeration 25(2002) 935- 947. The best available method for annular flow was that of Muller- Steinhagen and Heck. It was observed that the peak in the two-phase frictional pressure gradient is at high vapor qualities.

Keywords: Two-phase flow, frictional pressure drop, horizontalmicro tube, carbon dioxide, condensers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3365
2900 Brain Image Segmentation Using Conditional Random Field Based On Modified Artificial Bee Colony Optimization Algorithm

Authors: B. Thiagarajan, R. Bremananth

Abstract:

Tumor is an uncontrolled growth of tissues in any part of the body. Tumors are of different types and they have different characteristics and treatments. Brain tumor is inherently serious and life-threatening because of its character in the limited space of the intracranial cavity (space formed inside the skull). Locating the tumor within MR (magnetic resonance) image of brain is integral part of the treatment of brain tumor. This segmentation task requires classification of each voxel as either tumor or non-tumor, based on the description of the voxel under consideration. Many studies are going on in the medical field using Markov Random Fields (MRF) in segmentation of MR images. Even though the segmentation process is better, computing the probability and estimation of parameters is difficult. In order to overcome the aforementioned issues, Conditional Random Field (CRF) is used in this paper for segmentation, along with the modified artificial bee colony optimization and modified fuzzy possibility c-means (MFPCM) algorithm. This work is mainly focused to reduce the computational complexities, which are found in existing methods and aimed at getting higher accuracy. The efficiency of this work is evaluated using the parameters such as region non-uniformity, correlation and computation time. The experimental results are compared with the existing methods such as MRF with improved Genetic Algorithm (GA) and MRF-Artificial Bee Colony (MRF-ABC) algorithm.

Keywords: Conditional random field, Magnetic resonance, Markov random field, Modified artificial bee colony.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2948
2899 Investigation of the Effect of Teaching a Thinking and Research Lesson by Cooperative and Traditional Methods on the Creativity of Sixth Grade Students

Authors: Faroogh Khakzad, Marzieh Dehghani, Elahe Hejazi

Abstract:

The present study investigates the effect of teaching a Thinking and Research lesson by cooperative and traditional methods on the creativity of sixth-grade students in Piranshahr province. The statistical society includes all the sixth-grade students of Piranshahr province. The sample of this studytable was selected by available sampling from among male elementary schools of Piranshahr. They were randomly assigned into two groups of cooperative teaching method and traditional teaching method. The design of the study is quasi-experimental with a control group. In this study, to assess students’ creativity, Abedi’s creativity questionnaire was used. Based on Cronbach’s alpha coefficient, the reliability of the factor flow was 0.74, innovation was 0.61, flexibility was 0.63, and expansion was 0.68. To analyze the data, t-test, univariate and multivariate covariance analysis were used for evaluation of the difference of means and the pretest and posttest scores. The findings of the research showed that cooperative teaching method does not significantly increase creativity (p > 0.05). Moreover, cooperative teaching method was found to have significant effect on flow factor (p < 0.05), but in innovation and expansion factors no significant effect was observed (p < 0.05).

Keywords: Cooperative teaching method, traditional teaching method, creativity, flow, innovation, flexibility, expansion, thinking and research lesson.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 692
2898 Improved Computational Efficiency of Machine Learning Algorithms Based on Evaluation Metrics to Control the Spread of Coronavirus in the UK

Authors: Swathi Ganesan, Nalinda Somasiri, Rebecca Jeyavadhanam, Gayathri Karthick

Abstract:

The COVID-19 crisis presents a substantial and critical hazard to worldwide health. Since the occurrence of the disease in late January 2020 in the UK, the number of infected people confirmed to acquire the illness has increased tremendously across the country, and the number of individuals affected is undoubtedly considerably high. The purpose of this research is to figure out a predictive machine learning (ML) archetypal that could forecast the COVID-19 cases within the UK. This study concentrates on the statistical data collected from 31st January 2020 to 31st March 2021 in the United Kingdom. Information on total COVID-19 cases registered, new cases encountered on a daily basis, total death registered, and patients’ death per day due to Coronavirus is collected from World Health Organization (WHO). Data preprocessing is carried out to identify any missing values, outliers, or anomalies in the dataset. The data are split into 8:2 ratio for training and testing purposes to forecast future new COVID-19 cases. Support Vector Machine (SVM), Random Forest (RF), and linear regression (LR) algorithms are chosen to study the model performance in the prediction of new COVID-19 cases. From the evaluation metrics such as r-squared value and mean squared error, the statistical performance of the model in predicting the new COVID-19 cases is evaluated. RF outperformed the other two ML algorithms with a training accuracy of 99.47% and testing accuracy of 98.26% when n = 30. The mean square error obtained for RF is 4.05e11, which is lesser compared to the other predictive models used for this study. From the experimental analysis, RF algorithm can perform more effectively and efficiently in predicting the new COVID-19 cases, which could help the health sector to take relevant control measures for the spread of the virus.

Keywords: COVID-19, machine learning, supervised learning, unsupervised learning, linear regression, support vector machine, random forest.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 172
2897 A Fuzzy-Rough Feature Selection Based on Binary Shuffled Frog Leaping Algorithm

Authors: Javad Rahimipour Anaraki, Saeed Samet, Mahdi Eftekhari, Chang Wook Ahn

Abstract:

Feature selection and attribute reduction are crucial problems, and widely used techniques in the field of machine learning, data mining and pattern recognition to overcome the well-known phenomenon of the Curse of Dimensionality. This paper presents a feature selection method that efficiently carries out attribute reduction, thereby selecting the most informative features of a dataset. It consists of two components: 1) a measure for feature subset evaluation, and 2) a search strategy. For the evaluation measure, we have employed the fuzzy-rough dependency degree (FRFDD) of the lower approximation-based fuzzy-rough feature selection (L-FRFS) due to its effectiveness in feature selection. As for the search strategy, a modified version of a binary shuffled frog leaping algorithm is proposed (B-SFLA). The proposed feature selection method is obtained by hybridizing the B-SFLA with the FRDD. Nine classifiers have been employed to compare the proposed approach with several existing methods over twenty two datasets, including nine high dimensional and large ones, from the UCI repository. The experimental results demonstrate that the B-SFLA approach significantly outperforms other metaheuristic methods in terms of the number of selected features and the classification accuracy.

Keywords: Binary shuffled frog leaping algorithm, feature selection, fuzzy-rough set, minimal reduct.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 731
2896 Using Emotional Learning in Rescue Simulation Environment

Authors: Maziar Ahmad Sharbafi, Caro Lucas, Abolfazel Toroghi Haghighat, Omid AmirGhiasvand, Omid Aghazade

Abstract:

RoboCup Rescue simulation as a large-scale Multi agent system (MAS) is one of the challenging environments for keeping coordination between agents to achieve the objectives despite sensing and communication limitations. The dynamicity of the environment and intensive dependency between actions of different kinds of agents make the problem more complex. This point encouraged us to use learning-based methods to adapt our decision making to different situations. Our approach is utilizing reinforcement leaning. Using learning in rescue simulation is one of the current ways which has been the subject of several researches in recent years. In this paper we present an innovative learning method implemented for Police Force (PF) Agent. This method can cope with the main difficulties that exist in other learning approaches. Different methods used in the literature have been examined. Their drawbacks and possible improvements have led us to the method proposed in this paper which is fast and accurate. The Brain Emotional Learning Based Intelligent Controller (BELBIC) is our solution for learning in this environment. BELBIC is a physiologically motivated approach based on a computational model of amygdale and limbic system. The paper presents the results obtained by the proposed approach, showing the power of BELBIC as a decision making tool in complex and dynamic situation.

Keywords: Emotional learning, rescue, simulation environment, RoboCup, multi-agent system.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1621
2895 Effect of Submaximal Eccentric versus Maximal Isometric Contraction on Delayed Onset Muscle Soreness

Authors: Mohamed M. Ragab, Neveen A. Abdel Raoof, Reham H. Diab

Abstract:

Background: Delayed onset muscle soreness (DOMS) is the most common symptom when ordinary individuals and athletes are exposed to unaccustomed physical activity, especially eccentric contraction which impairs athletic performance, ordinary people work ability and physical functioning. Multitudes of methods have been investigated to reduce DOMS. One of the valuable methods to control DOMS is repeated bout effect (RBE) as a prophylactic method. Purpose: To compare the repeated bout effect of submaximal eccentric with maximal isometric contraction on induced DOMS. Methods: Sixty normal male volunteers were assigned randomly into three equal groups: Group A (first study group): 20 subjects received submaximal eccentric contraction on non-dominant elbow flexors as a prophylactic exercise. Group B (second study group): 20 subjects received maximal isometric contraction on nondominant elbow flexors as a prophylactic exercise. Group C (control group): 20 subjects did not receive any prophylactic exercises. Maximal isometric peak torque of elbow flexors and patient related elbow evaluation (PREE) scale were measured for each subject 3 times before, immediately after, and 48 hours after induction of DOMS. Results: Post-hoc test for maximal isometric peak torque and PREE scale immediately and 48 hours after induction of DOMS revealed that group (A) and group (B) resulted in significant decrease in maximal isometric strength loss and elbow pain and disability rather than control group (C), but submaximal eccentric group (A) was more effective than maximal isometric group (B) as it showed more rapid recovery of functional strength and less degrees of elbow pain and disability. Conclusion: Both submaximal eccentric contraction and maximal isometric contraction were effective in prevention of DOMS but submaximal eccentric contraction produced a greater protective effect against muscle damage induced by maximal eccentric exercise performed 2 days later.

Keywords: Delayed onset muscle soreness, maximal isometric peak torque, patient related elbow evaluation scale, repeated bout effect.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2090
2894 Enhancing Children’s English Vocabulary Acquisition through Digital Storytelling at Happy Kids Kindergarten, Palembang, Indonesia

Authors: Gaya Tridinanti

Abstract:

Enhanching English vocabulary in early childhood is the main problem often faced by teachers. Thus, the purpose of this study was to determine the enhancement of children’s English vocabulary acquisition by using digital storytelling. This type of research was an action research. It consisted of a series of four activities done in repeated cycles: planning, implementation, observation, and reflection. The subject of the study consisted of 30 students of B group (5-6 years old) attending Happy Kids Kindergarten Palembang, Indonesia. This research was conducted in three cycles. The methods used for data collection were observation and documentation. Descriptive qualitative and quantitative methods were also used to analyse the data. The research showed that the digital storytelling learning activities could enhance the children’s English vocabulary acquisition. It is based on the data in which the enhancement in pre-cycle was 37% and 51% in Cycle I. In Cycle II it was 71% and in Cycle III it was 89.3%. The results showed an enhancement of about 14% from the pre-cycle to Cycle I, 20% from Cycle I to Cycle II, and enhancement of about 18.3% from Cycle II to Cycle III. The conclusion of this study suggests that digital storytelling learning method could enhance the English vocabulary acquisition of B group children at the Happy Kids Kindergarten Palembang. Therefore, digital storytelling can be considered as an alternative to improve English language learning in the classroom.

Keywords: Acquisition, enhancing, digital storytelling, English vocabulary.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1655
2893 Developing Improvements to Multi-Hazard Risk Assessments

Authors: A. Fathianpour, M. B. Jelodar, S. Wilkinson

Abstract:

This paper outlines the approaches taken to assess multi-hazard assessments. There is currently confusion in assessing multi-hazard impacts, and so this study aims to determine which of the available options are the most useful. The paper uses an international literature search, and analysis of current multi-hazard assessments and a case study to illustrate the effectiveness of the chosen method. Findings from this study will help those wanting to assess multi-hazards to undertake a straightforward approach. The paper is significant as it helps to interpret the various approaches and concludes with the preferred method. Many people in the world live in hazardous environments and are susceptible to disasters. Unfortunately, when a disaster strikes it is often compounded by additional cascading hazards, thus people would confront more than one hazard simultaneously. Hazards include natural hazards (earthquakes, floods, etc.) or cascading human-made hazards (for example, Natural Hazard Triggering Technological disasters (Natech) such as fire, explosion, toxic release). Multi-hazards have a more destructive impact on urban areas than one hazard alone. In addition, climate change is creating links between different disasters such as causing landslide dams and debris flows leading to more destructive incidents. Much of the prevailing literature deals with only one hazard at a time. However, recently sophisticated multi-hazard assessments have started to appear. Given that multi-hazards occur, it is essential to take multi-hazard risk assessment under consideration. This paper aims to review the multi-hazard assessment methods through articles published to date and categorize the strengths and disadvantages of using these methods in risk assessment. Napier City is selected as a case study to demonstrate the necessity of using multi-hazard risk assessments. In order to assess multi-hazard risk assessments, first, the current multi-hazard risk assessment methods were described. Next, the drawbacks of these multi-hazard risk assessments were outlined. Finally, the improvements to current multi-hazard risk assessments to date were summarised. Generally, the main problem of multi-hazard risk assessment is to make a valid assumption of risk from the interactions of different hazards. Currently, risk assessment studies have started to assess multi-hazard situations, but drawbacks such as uncertainty and lack of data show the necessity for more precise risk assessment. It should be noted that ignoring or partial considering multi-hazards in risk assessment will lead to an overestimate or overlook in resilient and recovery action managements.

Keywords: Cascading hazards, multi-hazard, risk assessment, risk reduction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1096
2892 An Investigation of Final Tests of Translation as Practiced in Iranian Undergraduate English Translation Program

Authors: Hossein Heidari Tabrizi, Azizeh Chalak

Abstract:

The present study examined how translation teachers develop final tests as measures for checking on the quality of students’ academic translation in Iranian context. To achieve this goal, thirty experienced male and female translation teachers from the four types of the universities offering the program were invited to an in-depth 30-minute one-session semi-structured interview. The responses provided showed how much discrepancy exists among the Iranian translation teachers (as developers of final translation tests), who are least informed with the current translation evaluation methods. It was also revealed that the criteria they use for developing such tests and scoring student translations are not theory-driven but are highly subjective, mainly based on their personal experience and intuition. Hence, the quality and accountability of such tests are under serious question. The results also confirmed that the dominant method commonly and currently practiced is the purely essay-type format. To remedy the situation, some suggestions are in order. As part of the solution, to improve the reliability and validity of such tests, the present summative, product-oriented evaluation should be accompanied with some formative, process-oriented methods of evaluation. Training the teachers and helping them get acquainted with modern principles of translation evaluation as well as the existing models, and rating scales does improve the quality of academic translation evaluation.

Keywords: Iranian universities, students’ academic translations, translation final tests, undergraduate translation programs.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2028
2891 Potential Use of Local Materials as Synthesizing One Part Geopolymer Cement

Authors: Areej Almalkawi, Sameer Hamadna, Parviz Soroushian, Nalin Darsana

Abstract:

The work on indigenous binders in this paper focused on the following indigenous raw materials: red clay, red lava and pumice (as primary aluminosilicate precursors), wood ash and gypsum (as supplementary minerals), and sodium sulfate and lime (as alkali activators). The experimental methods used for evaluation of these indigenous raw materials included laser granulometry, x-ray fluorescence (XRF) spectroscopy, and chemical reactivity. Formulations were devised for transforming these raw materials into alkali aluminosilicate-based hydraulic cements. These formulations were processed into hydraulic cements via simple heating and milling actions to render thermal activation, mechanochemical and size reduction effects. The resulting hydraulic cements were subjected to laser granulometry, heat of hydration and reactivity tests. These cements were also used to prepare mortar mixtures, which were evaluated via performance of compressive strength tests. The measured values of strength were correlated with the reactivity, size distribution and microstructural features of raw materials. Some of the indigenous hydraulic cements produced in this reporting period yielded viable levels of compressive strength. The correlation trends established in this work are being evaluated for development of simple and thorough methods of qualifying indigenous raw materials for use in production of indigenous hydraulic cements.

Keywords: One-part geopolymer cement, aluminosilicate precursors, thermal activation, mechanochemical.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 703
2890 Comparison of Multivariate Adaptive Regression Splines and Random Forest Regression in Predicting Forced Expiratory Volume in One Second

Authors: P. V. Pramila, V. Mahesh

Abstract:

Pulmonary Function Tests are important non-invasive diagnostic tests to assess respiratory impairments and provides quantifiable measures of lung function. Spirometry is the most frequently used measure of lung function and plays an essential role in the diagnosis and management of pulmonary diseases. However, the test requires considerable patient effort and cooperation, markedly related to the age of patients resulting in incomplete data sets. This paper presents, a nonlinear model built using Multivariate adaptive regression splines and Random forest regression model to predict the missing spirometric features. Random forest based feature selection is used to enhance both the generalization capability and the model interpretability. In the present study, flow-volume data are recorded for N= 198 subjects. The ranked order of feature importance index calculated by the random forests model shows that the spirometric features FVC, FEF25, PEF, FEF25-75, FEF50 and the demographic parameter height are the important descriptors. A comparison of performance assessment of both models prove that, the prediction ability of MARS with the `top two ranked features namely the FVC and FEF25 is higher, yielding a model fit of R2= 0.96 and R2= 0.99 for normal and abnormal subjects. The Root Mean Square Error analysis of the RF model and the MARS model also shows that the latter is capable of predicting the missing values of FEV1 with a notably lower error value of 0.0191 (normal subjects) and 0.0106 (abnormal subjects) with the aforementioned input features. It is concluded that combining feature selection with a prediction model provides a minimum subset of predominant features to train the model, as well as yielding better prediction performance. This analysis can assist clinicians with a intelligence support system in the medical diagnosis and improvement of clinical care.

Keywords: FEV1, Multivariate Adaptive Regression Splines Pulmonary Function Test, Random Forest.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3737
2889 Fault Detection and Diagnosis of Broken Bar Problem in Induction Motors Base Wavelet Analysis and EMD Method: Case Study of Mobarakeh Steel Company in Iran

Authors: M. Ahmadi, M. Kafil, H. Ebrahimi

Abstract:

Nowadays, induction motors have a significant role in industries. Condition monitoring (CM) of this equipment has gained a remarkable importance during recent years due to huge production losses, substantial imposed costs and increases in vulnerability, risk, and uncertainty levels. Motor current signature analysis (MCSA) is one of the most important techniques in CM. This method can be used for rotor broken bars detection. Signal processing methods such as Fast Fourier transformation (FFT), Wavelet transformation and Empirical Mode Decomposition (EMD) are used for analyzing MCSA output data. In this study, these signal processing methods are used for broken bar problem detection of Mobarakeh steel company induction motors. Based on wavelet transformation method, an index for fault detection, CF, is introduced which is the variation of maximum to the mean of wavelet transformation coefficients. We find that, in the broken bar condition, the amount of CF factor is greater than the healthy condition. Based on EMD method, the energy of intrinsic mode functions (IMF) is calculated and finds that when motor bars become broken the energy of IMFs increases.

Keywords: Broken bar, condition monitoring, diagnostics, empirical mode decomposition, Fourier transform, wavelet transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 801
2888 Capability Investigation of Carbon Sequestration in Two Species (Artemisia sieberi Besser and Stipabarbata Desf) Under Different Treatments of Vegetation Management (Saveh, Iran)

Authors: M. Alizadeh, M. Mahdavi, M.H. Jouri

Abstract:

The rangelands, as one of the largest dynamic biomes in the world, have very capabilities. Regulation of greenhouse gases in the Earth's atmosphere, particularly carbon dioxide as the main these gases, is one of these cases. The attention to rangeland, as cheep and reachable resources to sequestrate the carbon dioxide, increases after the Industrial Revolution. Rangelands comprise the large parts of Iran as a steppic area. Rudshur (Saveh), as area index of steppic area, was selected under three sites include long-term exclosure, medium-term exclosure, and grazable area in order to the capable of carbon dioxide’s sequestration of dominated species. Canopy cover’s percentage of two dominated species (Artemisia sieberi Besser & Stipa barbata Desf) was determined via establishing of random 1 square meter plot. The sampling of above and below ground biomass style was obtained by complete random. After determination of ash percentage in the laboratory; conversion ratio of plant biomass to organic carbon was calculated by ignition method. Results of the paired t-test showed that the amount of carbon sequestration in above ground and underground biomass of Artemisia sieberi Besser & Stipa barbata Desf is different in three regions. It, of course, hasn’t any difference between under and surface ground’s biomass of Artemisia sieberi Besser in long-term exclosure. The independent t-test results indicate differences between underground biomass corresponding each other in the studied sites. Carbon sequestration in the Stipa barbata Desf was totally more than Artemisia sieberi Besser. Altogether, the average sequestration of the long-term exclosure was 5.842gr/m², the medium-term exclosure was 4.115gr/m², and grazable area was 5.975gr/m² so that there isn’t valuable statistical difference in term of total amount of carbon sequestration to three sites.

Keywords: Carbon sequestration, the Industrial Revolution, greenhouse gases, Artemisia sieberi Besser, Stipa barbata Desf, steppic rangelands

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1746
2887 Two Dimensionnal Model for Extraction Packed Column Simulation using Finite Element Method

Authors: N. Outili, A-H. Meniai

Abstract:

Modeling transfer phenomena in several chemical engineering operations leads to the resolution of partial differential equations systems. According to the complexity of the operations mechanisms, the equations present a nonlinear form and analytical solution became difficult, we have then to use numerical methods which are based on approximations in order to transform a differential system to an algebraic one.Finite element method is one of numerical methods which can be used to obtain an accurate solution in many complex cases of chemical engineering.The packed columns find a large application like contactor for liquid-liquid systems such solvent extraction. In the literature, the modeling of this type of equipment received less attention in comparison with the plate columns.A mathematical bidimensionnal model with radial and axial dispersion, simulating packed tower extraction behavior was developed and a partial differential equation was solved using the finite element method by adopting the Galerkine model. We developed a Mathcad program, which can be used for a similar equations and concentration profiles are obtained along the column. The influence of radial dispersion was prooved and it can-t be neglected, the results were compared with experimental concentration at the top of the column in the extraction system: acetone/toluene/water.

Keywords: finite element method, Galerkine method, liquidliquid extraction modelling, packed column simulation, two dimensional model

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1690
2886 A Hybrid Neural Network and Traditional Approach for Forecasting Lumpy Demand

Authors: A. Nasiri Pour, B. Rostami Tabar, A.Rahimzadeh

Abstract:

Accurate demand forecasting is one of the most key issues in inventory management of spare parts. The problem of modeling future consumption becomes especially difficult for lumpy patterns, which characterized by intervals in which there is no demand and, periods with actual demand occurrences with large variation in demand levels. However, many of the forecasting methods may perform poorly when demand for an item is lumpy. In this study based on the characteristic of lumpy demand patterns of spare parts a hybrid forecasting approach has been developed, which use a multi-layered perceptron neural network and a traditional recursive method for forecasting future demands. In the described approach the multi-layered perceptron are adapted to forecast occurrences of non-zero demands, and then a conventional recursive method is used to estimate the quantity of non-zero demands. In order to evaluate the performance of the proposed approach, their forecasts were compared to those obtained by using Syntetos & Boylan approximation, recently employed multi-layered perceptron neural network, generalized regression neural network and elman recurrent neural network in this area. The models were applied to forecast future demand of spare parts of Arak Petrochemical Company in Iran, using 30 types of real data sets. The results indicate that the forecasts obtained by using our proposed mode are superior to those obtained by using other methods.

Keywords: Lumpy Demand, Neural Network, Forecasting, Hybrid Approach.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2680
2885 Bone Mineral Density and Quality, Body Composition of Women in the Postmenopausal Period

Authors: Vladyslav Povoroznyuk, Oksana Ivanyk, Nataliia Dzerovych

Abstract:

In the diagnostics of osteoporosis, the gold standard is considered to be bone mineral density; however, X-ray densitometry is not an accurate indicator of osteoporotic fracture risk under all circumstances. In this regard, the search for new methods that could determine the indicators not only of the mineral density, but of the bone tissue quality, is a logical step for diagnostic optimization. One of these methods is the evaluation of trabecular bone quality. The aim of this study was to examine the quality and mineral density of spine bone tissue, femoral neck, and body composition of women depending on the duration of the postmenopausal period, to determine the correlation of body fat with indicators of bone mineral density and quality. The study examined 179 women in premenopausal and postmenopausal periods. The patients were divided into the following groups: Women in the premenopausal period and women in the postmenopausal period at various stages (early, middle, late postmenopause). A general examination and study of the above parameters were conducted with General Electric X-ray densitometer. The results show that bone quality and mineral density probably deteriorate with advancing of postmenopausal period. Total fat and lean mass ratio is not likely to change with age. In the middle and late postmenopausal periods, the bone tissue mineral density of the spine and femoral neck increases along with total fat mass.

Keywords: Osteoporosis, bone tissue mineral density, bone quality, fat mass, lean mass, postmenopausal osteoporosis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 941
2884 Classification of Potential Biomarkers in Breast Cancer Using Artificial Intelligence Algorithms and Anthropometric Datasets

Authors: Aref Aasi, Sahar Ebrahimi Bajgani, Erfan Aasi

Abstract:

Breast cancer (BC) continues to be the most frequent cancer in females and causes the highest number of cancer-related deaths in women worldwide. Inspired by recent advances in studying the relationship between different patient attributes and features and the disease, in this paper, we have tried to investigate the different classification methods for better diagnosis of BC in the early stages. In this regard, datasets from the University Hospital Centre of Coimbra were chosen, and different machine learning (ML)-based and neural network (NN) classifiers have been studied. For this purpose, we have selected favorable features among the nine provided attributes from the clinical dataset by using a random forest algorithm. This dataset consists of both healthy controls and BC patients, and it was noted that glucose, BMI, resistin, and age have the most importance, respectively. Moreover, we have analyzed these features with various ML-based classifier methods, including Decision Tree (DT), K-Nearest Neighbors (KNN), eXtreme Gradient Boosting (XGBoost), Logistic Regression (LR), Naive Bayes (NB), and Support Vector Machine (SVM) along with NN-based Multi-Layer Perceptron (MLP) classifier. The results revealed that among different techniques, the SVM and MLP classifiers have the most accuracy, with amounts of 96% and 92%, respectively. These results divulged that the adopted procedure could be used effectively for the classification of cancer cells, and also it encourages further experimental investigations with more collected data for other types of cancers.

Keywords: Breast cancer, health diagnosis, Machine Learning, biomarker classification, Neural Network.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 321
2883 Study on the Effect of Pre-Operative Patient Education on Post-Operative Outcomes

Authors: Chaudhary Itisha, Shankar Manu

Abstract:

Patient satisfaction represents a crucial aspect in the evaluation of health care services. Preoperative teaching provides the patient with pertinent information concerning the surgical process and the intended surgical procedure as well as anticipated patient behavior (anxiety, fear), expected sensation, and the probable outcomes. Although patient education is part of Accreditation protocols, it is not uniform at most places. The aim of this study was to try to assess the benefit of preoperative patient education on selected post-operative outcome parameters; mainly, post-operative pain scores, requirement of additional analgesia, return to activity of daily living and overall patient satisfaction, and try to standardize few education protocols. Dependent variables were measured before and after the treatment on a study population of 302 volunteers. Educational intervention was provided by the Investigator in the preoperative period to the study group through personal counseling. An information booklet contained detailed information was also provided. Statistical Analysis was done using Chi square test, Mann Whitney u test and Fischer Exact Test on a total of 302 subjects. P value <0.05 was considered as level of statistical significance and p<0.01 was considered as highly significant. This study suggested that patients who are given a structured, individualized and elaborate preoperative education and counseling have a better ability to cope up with postoperative pain in the immediate post-operative period. However, there was not much difference when the patients have had almost complete recovery. There was no difference in the requirement of additional analgesia among the two groups. There is a positive effect of preoperative counseling on expected return to the activities of daily living and normal work schedule. However, no effect was observed on the activities in the immediate post-operative period. There is no difference in the overall satisfaction score among the two groups of patients. Thus this study concludes that there is a positive benefit as suggested by the results for pre-operative patient education. Although the difference in various parameters studied might not be significant over a long term basis, they definitely point towards the benefits of preoperative patient education. 

Keywords: Patient education, post-operative pain, patient satisfaction, post-operative outcome.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3341
2882 Persian/Arabic Document Segmentation Based On Pyramidal Image Structure

Authors: Seyyed Yasser Hashemi, Khalil Monfaredi

Abstract:

Automatic transformation of paper documents into electronic documents requires document segmentation at the first stage. However, some parameters restrictions such as variations in character font sizes, different text line spacing, and also not uniform document layout structures altogether have made it difficult to design a general-purpose document layout analysis algorithm for many years. Thus in most previously reported methods it is inevitable to include these parameters. This problem becomes excessively acute and severe, especially in Persian/Arabic documents. Since the Persian/Arabic scripts differ considerably from the English scripts, most of the proposed methods for the English scripts do not render good results for the Persian scripts. In this paper, we present a novel parameter-free method for segmenting the Persian/Arabic document images which also works well for English scripts. This method segments the document image into maximal homogeneous regions and identifies them as texts and non-texts based on a pyramidal image structure. In other words the proposed method is capable of document segmentation without considering the character font sizes, text line spacing, and document layout structures. This algorithm is examined for 150 Arabic/Persian and English documents and document segmentation process are done successfully for 96 percent of documents.

Keywords: Persian/Arabic document, document segmentation, Pyramidal Image Structure, skew detection and correction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1765
2881 Continuous Measurement of Spatial Exposure Based on Visual Perception in Three-Dimensional Space

Authors: Nanjiang Chen

Abstract:

In the backdrop of expanding urban landscapes, accurately assessing spatial openness is critical. Traditional visibility analysis methods grapple with discretization errors and inefficiencies, creating a gap in truly capturing the human experience of space. Addressing these gaps, this paper presents a continuous visibility algorithm, providing a potentially valuable approach to measuring urban spaces from a human - centric perspective. This study presents a methodological breakthrough by applying this algorithm to urban visibility analysis. Unlike conventional approaches, this technique allows for a continuous range of visibility assessment, closely mirroring human visual perception. By eliminating the need for predefined subdivisions in ray casting, it offers a more accurate and efficient tool for urban planners and architects. The proposed algorithm not only reduces computational errors but also demonstrates faster processing capabilities, validated through a case study in Beijing's urban setting. Its key distinction lies in its potential to benefit a broad spectrum of stakeholders, ranging from urban developers to public policymakers, aiding in the creation of urban spaces that prioritize visual openness and quality of life. This advancement in urban analysis methods could lead to more inclusive, comfortable, and well-integrated urban environments, enhancing the spatial experience for communities worldwide.

Keywords: Visual openness, spatial continuity, ray-tracing algorithms, urban computation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31