Search results for: group efficiency
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3829

Search results for: group efficiency

2779 Holistic Approach to Teaching Mathematics in Secondary School as a Means of Improving Students’ Comprehension of Study Material

Authors: Natalia Podkhodova, Olga Sheremeteva, Mariia Soldaeva

Abstract:

Creating favourable conditions for students’ comprehension of mathematical content is one of the primary problems in teaching mathematics in secondary school. The fact of comprehension includes the ability to build a working situational model and thus becomes an important means of solving mathematical problems. This paper describes a holistic approach to teaching mathematics designed to address the primary challenges of such teaching; specifically, the challenge of students’ comprehension. Essentially, this approach consists of (1) establishing links between the attributes of the notion: the sense, the meaning, and the term; (2) taking into account the components of student’s subjective experience—value-based emotions, contextual, procedural and communicative—during the educational process; (3) linking together different ways to present mathematical information; (4) identifying and leveraging the relationships between real, perceptual and conceptual (scientific) mathematical spaces by applying real-life situational modelling. The article describes approaches to the practical use of these foundational concepts. Identifying how proposed methods and techniques influence understanding of material used in teaching mathematics was the primary goal. The study included an experiment in which 256 secondary school students took part: 142 in the study group and 114 in the control group. All students in these groups had similar levels of achievement in math and studied math under the same curriculum. In the course of the experiment, comprehension of two topics — “Derivative” and “Trigonometric functions”—was evaluated. Control group participants were taught using traditional methods. Students in the study group were taught using the holistic method: under teacher’s guidance, they carried out assignments designed to establish linkages between notion’s characteristics, to convert information from one mode of presentation to another, as well as assignments that required the ability to operate with all modes of presentation. Identification, accounting for and transformation of subjective experience were associated with methods of stimulating the emotional value component of the studied mathematical content (discussions of lesson titles, assignments aimed to create study dominants, performing theme-related physical exercise ...) The use of techniques that forms inter-subject notions based on linkages between, perceptual real and mathematical conceptual spaces proved to be of special interest to the students. Results of the experiment were analysed by presenting students in each of the groups with a final test in each of the studied topics. The test included assignments that required building real situational models. Statistical analysis was used to aggregate test results. Pierson criterion x2 was used to reveal statistics significance of results (pass-fail the modelling test). Significant difference of results was revealed (p < 0.001), which allowed to conclude that students in the study group showed better comprehension of mathematical information than those in the control group. The total number of completed assignments of each student was analysed as well, with average results calculated for each group. Statistical significance of result differences against the quantitative criterion (number of completed assignments) was determined using Student’s t-test, which showed that students in the study group completed significantly more assignments than those in the control group (p = 0.0001). Authors thus come to the conclusion that suggested increase in the level of comprehension of study material took place as a result of applying implemented methods and techniques.

Keywords: Comprehension of mathematical content, holistic approach to teaching mathematics in secondary school, subjective experience, technology of the formation of inter-subject notions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 544
2778 An Autonomous Collaborative Forecasting System Implementation – The First Step towards Successful CPFR System

Authors: Chi-Fang Huang, Yun-Shiow Chen, Yun-Kung Chung

Abstract:

In the past decade, artificial neural networks (ANNs) have been regarded as an instrument for problem-solving and decision-making; indeed, they have already done with a substantial efficiency and effectiveness improvement in industries and businesses. In this paper, the Back-Propagation neural Networks (BPNs) will be modulated to demonstrate the performance of the collaborative forecasting (CF) function of a Collaborative Planning, Forecasting and Replenishment (CPFR®) system. CPFR functions the balance between the sufficient product supply and the necessary customer demand in a Supply and Demand Chain (SDC). Several classical standard BPN will be grouped, collaborated and exploited for the easy implementation of the proposed modular ANN framework based on the topology of a SDC. Each individual BPN is applied as a modular tool to perform the task of forecasting SKUs (Stock-Keeping Units) levels that are managed and supervised at a POS (point of sale), a wholesaler, and a manufacturer in an SDC. The proposed modular BPN-based CF system will be exemplified and experimentally verified using lots of datasets of the simulated SDC. The experimental results showed that a complex CF problem can be divided into a group of simpler sub-problems based on the single independent trading partners distributed over SDC, and its SKU forecasting accuracy was satisfied when the system forecasted values compared to the original simulated SDC data. The primary task of implementing an autonomous CF involves the study of supervised ANN learning methodology which aims at making “knowledgeable" decision for the best SKU sales plan and stocks management.

Keywords: CPFR, artificial neural networks, global logistics, supply and demand chain.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1965
2777 Domestic Tourist Behaviours of the Tourism Place in Bangkok and Greater Area

Authors: Tanit P ruktara

Abstract:

This research aims to study the preferable tourism and the elements of choosing tourist destination from domestic tourist in Bangkok and the nearby areas in Thailand.The data were collected by using 1249 set of questionnaires, in mid-August 2012. The result illustrates that religious destinations are the most preferable places for the tourist. The average expense per travel is approximately 47 USD a time. Travellers travel based on the advertisement in the television and internet and their decisions is based on the reputation of the destinations. The result on a place dimension demonstrates the neatness and well managed location play a crucial role on tourist destination. Gender, age, marriage status and their origins are affecting their spending and travelling behaviour. The researcher reckon that providing the area of arcade, selling the souvenir and promoting tourism among a young professional group would be an important key follow the income distribution policy, including managing the destination to welcome the family group, which the result is to identified as the highest spending.

Keywords: Domestic Tourist, Tourism, Tourist Behaviours, Tourism Place.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2165
2776 The Effectiveness of ICT-Assisted PBL on College-Level Nano Knowledge and Learning Skills

Authors: Ya-Ting Carolyn Yang, Ping-Han Cheng, Shi-Hui Gilbert Chang, Terry Yuan-Fang Chen, Chih-Chieh Li

Abstract:

Nanotechnology is widely applied in various areas so professionals in the related fields have to know more than nano knowledge. In the study, we focus on adopting ICT-assisted PBL in college general education to foster professionals who possess multiple abilities. The research adopted a pretest and posttest quasi-experimental design. The control group received traditional instruction, and the experimental group received ICT-assisted PBL instruction. Descriptive statistics will be used to describe the means, standard deviations, and adjusted means for the tests between the two groups. Next, analysis of covariance (ANCOVA) will be used to compare the final results of the two research groups after 6 weeks of instruction. Statistics gathered in the end of the research can be used to make contrasts. Therefore, we will see how different teaching strategies can improve students’ understanding about nanotechnology and learning skills.

Keywords: Nanotechnology, science education, project-based learning, information and communication technology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2054
2775 Adopting Artificial Intelligence and Deep Learning Techniques in Cloud Computing for Operational Efficiency

Authors: Sandesh Achar

Abstract:

Artificial intelligence (AI) is being increasingly incorporated into many applications across various sectors such as health, education, security, and agriculture. Recently, there has been rapid development in cloud computing technology, resulting in AI’s implementation into cloud computing to enhance and optimize the technology service rendered. The deployment of AI in cloud-based applications has brought about autonomous computing, whereby systems achieve stated results without human intervention. Despite the amount of research into autonomous computing, work incorporating AI/ML into cloud computing to enhance its performance and resource allocation remains a fundamental challenge. This paper highlights different manifestations, roles, trends, and challenges related to AI-based cloud computing models. This work reviews and highlights investigations and progress in the domain. Future directions are suggested for leveraging AI/ML in next-generation computing for emerging computing paradigms such as cloud environments. Adopting AI-based algorithms and techniques to increase operational efficiency, cost savings, automation, reducing energy consumption and solving complex cloud computing issues are the major findings outlined in this paper.

Keywords: Artificial intelligence, AI, cloud computing, deep learning, machine learning, ML, internet of things, IoT.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 571
2774 Performance of Partially Covered N Number of Photovoltaic Thermal (PVT) - Compound Parabolic Concentrator (CPC) Series Connected Water Heating System

Authors: Rohit Tripathi, Sumit Tiwari, G. N. Tiwari

Abstract:

In present study, an approach is adopted where photovoltaic thermal flat plate collector is integrated with compound parabolic concentrator. Analytical expression of temperature dependent electrical efficiency of N number of partially covered Photovoltaic Thermal (PVT) - Compound Parabolic Concentrator (CPC) water collector connected in series has been derived with the help of basic thermal energy balance equations. Analysis has been carried for winter weather condition at Delhi location, India. Energy and exergy performance of N - partially covered Photovoltaic Thermal (PVT) - Compound Parabolic Concentrator (CPC) Water collector system has been compared for two cases: (i) 25% area of water collector covered by PV module, (ii) 75% area of water collector covered by PV module. It is observed that case (i) has been best suited for thermal performance and case (ii) for electrical energy as well as overall exergy.

Keywords: Compound parabolic concentrator, Energy, Photovoltaic thermal, Temperature dependent electrical efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1598
2773 Computational Analysis of the Scaling Effects on the Performance of an Axial Compressor

Authors: Junting Xiang, Jörg Uwe Schlüter, Fei Duan

Abstract:

The miniaturization of gas turbines promises many advantages. Miniature gas turbines can be used for local power generation or the propulsion of small aircraft, such as UAV and MAV. However, experience shows that the miniaturization of conventional gas turbines, which are optimized at their current large size, leads to a substantial loss of efficiency and performance at smaller scales. This may be due to a number of factors, such as the Reynolds-number effect, the increased heat transfer, and manufacturing tolerances. In the present work, we focus on computational investigations of the Reynolds number effect and the wall heat transfer on the performance of axial compressor during its size change. The NASA stage 35 compressor is selected as the configuration in this study and computational fluid dynamics (CFD) is used to carry out the miniaturization process and simulations. We perform parameter studies on the effect of Reynolds number and wall thermal conditions. Our results indicate a decrease of efficiency, if the compressor is miniaturized based on its original geometry due to the increase of viscous effects. The increased heat transfer through wall has only a small effect and will actually benefit compressor performance based on our study.

Keywords: Axial compressor, CFD, heat transfer, miniature gas turbines, Reynolds number.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3188
2772 Perceptual and Ultrasound Articulatory Training Effects on English L2 Vowels Production by Italian Learners

Authors: I. Sonia d’Apolito, Bianca Sisinni, Mirko Grimaldi, Barbara Gili Fivela

Abstract:

The American English contrast /ɑ-ʌ/ (cop-cup) is difficult to be produced by Italian learners since they realize L2-/ɑ-ʌ/ as L1-/ɔ-a/ respectively, due to differences in phonetic-phonological systems and also in grapheme-to-phoneme conversion rules. In this paper, we try to answer the following research questions: Can a short training improve the production of English /ɑ-ʌ/ by Italian learners? Is a perceptual training better than an articulatory (ultrasound - US) training? Thus, we compare a perceptual training with an US articulatory one to observe: 1) the effects of short trainings on L2-/ɑ-ʌ/ productions; 2) if the US articulatory training improves the pronunciation better than the perceptual training. In this pilot study, 9 Salento-Italian monolingual adults participated: 3 subjects performed a 1-hour perceptual training (ES-P); 3 subjects performed a 1-hour US training (ES-US); and 3 control subjects did not receive any training (CS). Verbal instructions about the phonetic properties of L2-/ɑ-ʌ/ and L1-/ɔ-a/ and their differences (representation on F1-F2 plane) were provided during both trainings. After these instructions, the ES-P group performed an identification training based on the High Variability Phonetic Training procedure, while the ES-US group performed the articulatory training, by means of US video of tongue gestures in L2-/ɑ-ʌ/ production and dynamic view of their own tongue movements and position using a probe under their chin. The acoustic data were analyzed and the first three formants were calculated. Independent t-tests were run to compare: 1) /ɑ-ʌ/ in pre- vs. post-test respectively; /ɑ-ʌ/ in pre- and post-test vs. L1-/a-ɔ/ respectively. Results show that in the pre-test all speakers realize L2-/ɑ-ʌ/ as L1-/ɔ-a/ respectively. Contrary to CS and ES-P groups, the ES-US group in the post-test differentiates the L2 vowels from those produced in the pre-test as well as from the L1 vowels, although only one ES-US subject produces both L2 vowels accurately. The articulatory training seems more effective than the perceptual one since it favors the production of vowels in the correct direction of L2 vowels and differently from the similar L1 vowels.

Keywords: L2 vowel production, perceptual training, articulatory training, ultrasound.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 996
2771 Numerical Investigations on Group Piles’ Lateral Bearing Capacity Considering Interaction of Soil and Structure

Authors: Mahdi Sadeghian, Mahmoud Hassanlourad, Alireza Ardakani, Reza Dinarvand

Abstract:

In this research, the behavior of monopiles, under lateral loads, was investigated with vertical and oblique piles by Finite Element Method. In engineering practice when soil-pile interaction comes to the picture some simplifications are applied to reduce the design time. As a simplified replacement of soil and pile interaction analysis, pile could be replaced by a column. The height of the column would be equal to the free length of the pile plus a portion of the embedded length of it. One of the important factors studied in this study was that columns with an equivalent length (free length plus a part of buried depth) could be used instead of soil and pile modeling. The results of the analysis show that the more internal friction angle of the soil increases, the more the bearing capacity of the soil is achieved. This additional length is 6 to 11 times of the pile diameter in dense soil although in loose sandy soil this range might increase.

Keywords: Lateral bearing capacity, pile group, oblique pile, soil-structure interaction, depth of fixity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1010
2770 Revising the Student Experiment Materials and Practices at the National University of Laos

Authors: Syhalath Xaphakdy, Toshio Nagata, Saykham Phommathat, Pavy Souwannavong, Vilayvanh Srithilat, Phoxay Sengdala, Bounaom Phetarnousone, Boualay Siharath, Xaya Chemcheng

Abstract:

The National University of Laos (NUOL) invited a group of volunteers from the Japan International Cooperation Agency (JICA) to revise the physics experiments to utilize the materials that were already available to students. The intension was to review and revise the materials regularly utilized in physics class. The project had access to limited materials and a small budget for the class in the unit; however, by developing experimental textbooks related to mechanics, electricity, and wave and vibration, the group found a way to apply them in the classroom and enhance the students teaching activities. The aim was to introduce a way to incorporate the materials and practices in the classroom to enhance the students learning and teaching skills, particularly when they graduate and begin working as high school teachers.

Keywords: NUOL, JICA, physics experiment materials, small budget, mechanics, electricity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1217
2769 Treatment of Wastewater from Wet Scrubbers in Secondary Lead Smelters for Recycling and Lead Recovery

Authors: Mahmoud A. Rabah

Abstract:

The present study shows a method to recover lead metal from wastewater of wet scrubber in secondary lead smelter. The wastewater is loaded with 42,000 ppm of insoluble lead compounds (TSP) submicron in diameter. The technical background benefits the use of cationic polyfloc solution to flocculate these colloidal solids before press filtration. The polymer solution is injected in the wastewater stream in a countercurrent flow design. The study demonstrates the effect of polymer dose, temperature, pH, flow velocity of the wastewater and different filtration media on the filtration extent. Results indicated that filtration rate (¦r), quality of purified water, purifying efficiency (¦e) and floc diameter decrease regularly with increase in mass flow rate and velocity up to turbulence of 0.5 m.sec-1. Laminar flow is in favor of flocculation. Polyfloc concentration of 0.75 – 1.25 g/m3 wastewater is convenient. Increasing temperature of the wastewater and pneumatic pressure of filtration enhances ¦r. High pH value deforms floc formation and assists degradation of the filtration fabric. The overall efficiency of the method amounts to 93.2 %. Lead metal was recovered from the filtrate cake using carbon as a reducing agent at 900°C.

Keywords: Wastewater, wet scrubbers, filtration, secondary lead.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3421
2768 Evaluating Alternative Fuel Vehicles from Technical, Environmental and Economic Perspectives: Case of Light-Duty Vehicles in Iran

Authors: Vahid Aryanpur , Ehsan Shafiei

Abstract:

This paper presents an environmental and technoeconomic evaluation of light duty vehicles in Iran. A comprehensive well-to-wheel (WTW) analysis is applied to compare different automotive fuel chains, conventional internal combustion engines and innovative vehicle powertrains. The study examines the competitiveness of 15 various pathways in terms of energy efficiencies, GHG emissions, and levelized cost of different energy carriers. The results indicate that electric vehicles including battery electric vehicles (BEV), fuel cell vehicles (FCV) and plug-in hybrid electric vehicles (PHEV) increase the WTW energy efficiency by 54%, 51% and 46%, respectively, compared to common internal combustion engines powered by gasoline. On the other hand, greenhouse gas (GHG) emissions per kilometer of FCV and BEV would be 48% lower than that of gasoline engines. It is concluded that BEV has the lowest total cost of energy consumption and external cost of emission, followed by internal combustion engines (ICE) fueled by CNG. Conventional internal combustion engines fueled by gasoline, on the other hand, would have the highest costs.

Keywords: Well-to-Wheel analysis, Energy Efficiency, GHG emissions, Levelized cost of energy, Alternative fuel vehicles.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1731
2767 A Case Study on Appearance Based Feature Extraction Techniques and Their Susceptibility to Image Degradations for the Task of Face Recognition

Authors: Vitomir Struc, Nikola Pavesic

Abstract:

Over the past decades, automatic face recognition has become a highly active research area, mainly due to the countless application possibilities in both the private as well as the public sector. Numerous algorithms have been proposed in the literature to cope with the problem of face recognition, nevertheless, a group of methods commonly referred to as appearance based have emerged as the dominant solution to the face recognition problem. Many comparative studies concerned with the performance of appearance based methods have already been presented in the literature, not rarely with inconclusive and often with contradictory results. No consent has been reached within the scientific community regarding the relative ranking of the efficiency of appearance based methods for the face recognition task, let alone regarding their susceptibility to appearance changes induced by various environmental factors. To tackle these open issues, this paper assess the performance of the three dominant appearance based methods: principal component analysis, linear discriminant analysis and independent component analysis, and compares them on equal footing (i.e., with the same preprocessing procedure, with optimized parameters for the best possible performance, etc.) in face verification experiments on the publicly available XM2VTS database. In addition to the comparative analysis on the XM2VTS database, ten degraded versions of the database are also employed in the experiments to evaluate the susceptibility of the appearance based methods on various image degradations which can occur in "real-life" operating conditions. Our experimental results suggest that linear discriminant analysis ensures the most consistent verification rates across the tested databases.

Keywords: Biometrics, face recognition, appearance based methods, image degradations, the XM2VTS database.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2264
2766 The Cardiac Diagnostic Prediction Applied to a Designed Holter

Authors: Leonardo Juan Ramírez López, Javier Oswaldo Rodriguez Velasquez

Abstract:

We have designed a Holter that measures the heart´s activity for over 24 hours, implemented a prediction methodology, and generate alarms as well as indicators to patients and treating physicians. Various diagnostic advances have been developed in clinical cardiology thanks to Holter implementation; however, their interpretation has largely been conditioned to clinical analysis and measurements adjusted to diverse population characteristics, thus turning it into a subjective examination. This, however, requires vast population studies to be validated that, in turn, have not achieved the ultimate goal: mortality prediction. Given this context, our Insight Research Group developed a mathematical methodology that assesses cardiac dynamics through entropy and probability, creating a numerical and geometrical attractor which allows quantifying the normalcy of chronic and acute disease as well as the evolution between such states, and our Tigum Research Group developed a holter device with 12 channels and advanced computer software. This has been shown in different contexts with 100% sensitivity and specificity results.

Keywords: Entropy, mathematical, prediction, cardiac, holter, attractor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 680
2765 Kinetic Modeling of Transesterification of Triacetin Using Synthesized Ion Exchange Resin (SIERs)

Authors: Hafizuddin W. Yussof, Syamsutajri S. Bahri, Adam P. Harvey

Abstract:

Strong anion exchange resins with QN+OH-, have the potential to be developed and employed as heterogeneous catalyst for transesterification, as they are chemically stable to leaching of the functional group. Nine different SIERs (SIER1-9) with QN+OH-were prepared by suspension polymerization of vinylbenzyl chloridedivinylbenzene (VBC-DVB) copolymers in the presence of n-heptane (pore-forming agent). The amine group was successfully grafted into the polymeric resin beads through functionalization with trimethylamine. These SIERs are then used as a catalyst for the transesterification of triacetin with methanol. A set of differential equations that represents the Langmuir-Hinshelwood-Hougen- Watson (LHHW) and Eley-Rideal (ER) models for the transesterification reaction were developed. These kinetic models of LHHW and ER were fitted to the experimental data. Overall, the synthesized ion exchange resin-catalyzed reaction were welldescribed by the Eley-Rideal model compared to LHHW models, with sum of square error (SSE) of 0.742 and 0.996, respectively.

Keywords: Anion exchange resin, Eley-Rideal, Langmuir-Hinshelwood-Hougen-Watson, transesterification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2370
2764 Comparing Test Equating by Item Response Theory and Raw Score Methods with Small Sample Sizes on a Study of the ARTé: Mecenas Learning Game

Authors: Steven W. Carruthers

Abstract:

The purpose of the present research is to equate two test forms as part of a study to evaluate the educational effectiveness of the ARTé: Mecenas art history learning game. The researcher applied Item Response Theory (IRT) procedures to calculate item, test, and mean-sigma equating parameters. With the sample size n=134, test parameters indicated “good” model fit but low Test Information Functions and more acute than expected equating parameters. Therefore, the researcher applied equipercentile equating and linear equating to raw scores and compared the equated form parameters and effect sizes from each method. Item scaling in IRT enables the researcher to select a subset of well-discriminating items. The mean-sigma step produces a mean-slope adjustment from the anchor items, which was used to scale the score on the new form (Form R) to the reference form (Form Q) scale. In equipercentile equating, scores are adjusted to align the proportion of scores in each quintile segment. Linear equating produces a mean-slope adjustment, which was applied to all core items on the new form. The study followed a quasi-experimental design with purposeful sampling of students enrolled in a college level art history course (n=134) and counterbalancing design to distribute both forms on the pre- and posttests. The Experimental Group (n=82) was asked to play ARTé: Mecenas online and complete Level 4 of the game within a two-week period; 37 participants completed Level 4. Over the same period, the Control Group (n=52) did not play the game. The researcher examined between group differences from post-test scores on test Form Q and Form R by full-factorial Two-Way ANOVA. The raw score analysis indicated a 1.29% direct effect of form, which was statistically non-significant but may be practically significant. The researcher repeated the between group differences analysis with all three equating methods. For the IRT mean-sigma adjusted scores, form had a direct effect of 8.39%. Mean-sigma equating with a small sample may have resulted in inaccurate equating parameters. Equipercentile equating aligned test means and standard deviations, but resultant skewness and kurtosis worsened compared to raw score parameters. Form had a 3.18% direct effect. Linear equating produced the lowest Form effect, approaching 0%. Using linearly equated scores, the researcher conducted an ANCOVA to examine the effect size in terms of prior knowledge. The between group effect size for the Control Group versus Experimental Group participants who completed the game was 14.39% with a 4.77% effect size attributed to pre-test score. Playing and completing the game increased art history knowledge, and individuals with low prior knowledge tended to gain more from pre- to post test. Ultimately, researchers should approach test equating based on their theoretical stance on Classical Test Theory and IRT and the respective  assumptions. Regardless of the approach or method, test equating requires a representative sample of sufficient size. With small sample sizes, the application of a range of equating approaches can expose item and test features for review, inform interpretation, and identify paths for improving instruments for future study.

Keywords: Effectiveness, equipercentile equating, IRT, learning games, linear equating, mean-sigma equating.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 994
2763 Investigation of Scour Depth at Bridge Piers using Bri-Stars Model in Iran

Authors: Gh. Saeidifar, F. Raeiszadeh

Abstract:

BRI-STARS (BRIdge Stream Tube model for Alluvial River Simulation) program was used to investigate the scour depth around bridge piers in some of the major river systems in Iran. Model calibration was performed by collecting different field data. Field data are cataloged on three categories, first group of bridges that their rivers bed are formed by fine material, second group of bridges that their rivers bed are formed by sand material, and finally bridges that their rivers bed are formed by gravel or cobble materials. Verification was performed with some field data in Fars Province. Results show that for wide piers, computed scour depth is more than measured one. In gravel bed streams, computed scour depth is greater than measured scour depth, the reason is due to formation of armor layer on bed of channel. Once this layer is eroded, the computed scour depth is close to the measured one.

Keywords: BRI-STARS, local scour, bridge, computer modeling

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1970
2762 Physics of Decision for Polling Place Management: A Case Study from the 2020 USA Presidential Election

Authors: Nafe Moradkhani, Frederick Benaben, Benoit Montreuil, Ali Vatankhah Barenji, Dima Nazzal

Abstract:

In the context of the global pandemic, the practical management of the 2020 presidential election in the USA was a strong concern. To anticipate and prepare for this election accurately, one of the main challenges was to confront: (i) forecasts of voter turnout, (ii) capacities of the facilities and, (iii) potential configuration options of resources. The approach chosen to conduct this anticipative study consists of collecting data about forecasts and using simulation models to work simultaneously on resource allocation and facility configuration of polling places in Fulton County, Georgia’s largest county. This article presents the results of the simulations of such places facing pre-identified potential risks. These results are oriented towards the efficiency of these places according to different criteria (health, trust, comfort). Then a dynamic framework is introduced to describe risks as physical forces perturbing the efficiency of the observed system. Finally, the main benefits and contributions resulting from this simulation campaign are presented.

Keywords: performance, decision support, simulation, artificial intelligence, risk management, election, pandemics, information system

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 598
2761 Variant Polymorphisms of GST and XRCC Genes and the Early Risk of Age Associated Disease in Kazakhstan

Authors: Zeinep A. Berkimbayeva, Almagul T. Mansharipova, Elmira M. Khussainova, Leyla B. Djansugurova

Abstract:

It is believed that DNA damaging toxic metabolites contributes to the development of different pathological conditions. To prevent harmful influence of toxic agents, cells developed number of protecting mechanisms, such as enzymatic reaction of detoxification of reactive metabolites and repair of DNA damage. The aim of the study was to examine the association between polymorphism of GSTT1/GSTM1 and XRCC1/3 genes and coronary artery disease (CAD) incidence. To examine a polymorphism of these genes in CAD susceptibility in patients and controls, PCR based genotyping assay was performed. For GST genes, frequency of GSTM1 null genotype among CAD affected group was significantly increased than in control group (P<0.001). Frequencies of the GSTT1 null and positive alleles are almost equal in both groups (P>0.1). We found that neither XRCC1 Arg399Gln nor XRCC3 Thr241Met were associated with CAD risk. Obtained data suggests that GSTM1 null genotype carriers are more susceptible to CAD development.

Keywords: Cardiovascular disease, DNA reparation, gene polymorphism, risk factors, xenobiotic detoxification.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1915
2760 Response Surface Modeling of Lactic Acid Extraction by Emulsion Liquid Membrane: Box-Behnken Experimental Design

Authors: A. Thakur, P. S. Panesar, M. S. Saini

Abstract:

Extraction of lactic acid by emulsion liquid membrane technology (ELM) using n-trioctyl amine (TOA) in n-heptane as carrier within the organic membrane along with sodium carbonate as acceptor phase was optimized by using response surface methodology (RSM). A three level Box-Behnken design was employed for experimental design, analysis of the results and to depict the combined effect of five independent variables, vizlactic acid concentration in aqueous phase (cl), sodium carbonate concentration in stripping phase (cs), carrier concentration in membrane phase (ψ), treat ratio, and batch extraction time (τ)  with equal volume of organic and external aqueous phase on lactic acid extraction efficiency. The maximum lactic acid extraction efficiency (ηext) of 98.21%from aqueous phase in a batch reactor using ELM was found at the optimized values for test variables, cl, cs, ψ, and τ as 0.06 [M], 0.18 [M], 4.72 (%,v/v), 1.98 (v/v) and 13.36 min respectively. 

Keywords: Emulsion liquid membrane, extraction, lactic acid, n-trioctylamine, response surface methodology.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2307
2759 Cooperative Energy Efficient Routing for Wireless Sensor Networks in Smart Grid Communications

Authors: Ghazi AL-Sukkar, Iyad Jafar, Khalid Darabkh, Raed Al-Zubi, Mohammed Hawa

Abstract:

Smart Grids employ wireless sensor networks for their control and monitoring. Sensors are characterized by limitations in the processing power, energy supply and memory spaces, which require a particular attention on the design of routing and data management algorithms. Since most routing algorithms for sensor networks, focus on finding energy efficient paths to prolong the lifetime of sensor networks, the power of sensors on efficient paths depletes quickly, and consequently sensor networks become incapable of monitoring events from some parts of their target areas. In consequence, the design of routing protocols should consider not only energy efficiency paths, but also energy efficient algorithms in general. In this paper we propose an energy efficient routing protocol for wireless sensor networks without the support of any location information system. The reliability and the efficiency of this protocol have been demonstrated by simulation studies where we compare them to the legacy protocols. Our simulation results show that these algorithms scale well with network size and density.

Keywords: Data-centric storage, Dynamic Address Allocation, Sensor networks, Smart Grid Communications.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1833
2758 Towards Real-Time Classification of Finger Movement Direction Using Encephalography Independent Components

Authors: Mohamed Mounir Tellache, Hiroyuki Kambara, Yasuharu Koike, Makoto Miyakoshi, Natsue Yoshimura

Abstract:

This study explores the practicality of using electroencephalographic (EEG) independent components to predict eight-direction finger movements in pseudo-real-time. Six healthy participants with individual-head MRI images performed finger movements in eight directions with two different arm configurations. The analysis was performed in two stages. The first stage consisted of using independent component analysis (ICA) to separate the signals representing brain activity from non-brain activity signals and to obtain the unmixing matrix. The resulting independent components (ICs) were checked, and those reflecting brain-activity were selected. Finally, the time series of the selected ICs were used to predict eight finger-movement directions using Sparse Logistic Regression (SLR). The second stage consisted of using the previously obtained unmixing matrix, the selected ICs, and the model obtained by applying SLR to classify a different EEG dataset. This method was applied to two different settings, namely the single-participant level and the group-level. For the single-participant level, the EEG dataset used in the first stage and the EEG dataset used in the second stage originated from the same participant. For the group-level, the EEG datasets used in the first stage were constructed by temporally concatenating each combination without repetition of the EEG datasets of five participants out of six, whereas the EEG dataset used in the second stage originated from the remaining participants. The average test classification results across datasets (mean ± S.D.) were 38.62 ± 8.36% for the single-participant, which was significantly higher than the chance level (12.50 ± 0.01%), and 27.26 ± 4.39% for the group-level which was also significantly higher than the chance level (12.49% ± 0.01%). The classification accuracy within [–45°, 45°] of the true direction is 70.03 ± 8.14% for single-participant and 62.63 ± 6.07% for group-level which may be promising for some real-life applications. Clustering and contribution analyses further revealed the brain regions involved in finger movement and the temporal aspect of their contribution to the classification. These results showed the possibility of using the ICA-based method in combination with other methods to build a real-time system to control prostheses.

Keywords: Brain-computer interface, BCI, electroencephalography, EEG, finger motion decoding, independent component analysis, pseudo-real-time motion decoding.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 572
2757 How Efficiency of Password Attack Based on a Keyboard

Authors: Hsien-cheng Chou, Fei-pei Lai, Hung-chang Lee

Abstract:

At present, dictionary attack has been the basic tool for recovering key passwords. In order to avoid dictionary attack, users purposely choose another character strings as passwords. According to statistics, about 14% of users choose keys on a keyboard (Kkey, for short) as passwords. This paper develops a framework system to attack the password chosen from Kkeys and analyzes its efficiency. Within this system, we build up keyboard rules using the adjacent and parallel relationship among Kkeys and then use these Kkey rules to generate password databases by depth-first search method. According to the experiment results, we find the key space of databases derived from these Kkey rules that could be far smaller than the password databases generated within brute-force attack, thus effectively narrowing down the scope of attack research. Taking one general Kkey rule, the combinations in all printable characters (94 types) with Kkey adjacent and parallel relationship, as an example, the derived key space is about 240 smaller than those in brute-force attack. In addition, we demonstrate the method's practicality and value by successfully cracking the access password to UNIX and PC using the password databases created

Keywords: Brute-force attack, dictionary attack, depth-firstsearch, password attack.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3454
2756 Video-On-Demand QoE Evaluation across Different Age-Groups and Its Significance for Network Capacity

Authors: Mujtaba Roshan, John A. Schormans

Abstract:

Quality of Experience (QoE) drives churn in the broadband networks industry, and good QoE plays a large part in the retention of customers. QoE is known to be affected by the Quality of Service (QoS) factors packet loss probability (PLP), delay and delay jitter caused by the network. Earlier results have shown that the relationship between these QoS factors and QoE is non-linear, and may vary from application to application. We use the network emulator Netem as the basis for experimentation, and evaluate how QoE varies as we change the emulated QoS metrics. Focusing on Video-on-Demand, we discovered that the reported QoE may differ widely for users of different age groups, and that the most demanding age group (the youngest) can require an order of magnitude lower PLP to achieve the same QoE than is required by the most widely studied age group of users. We then used a bottleneck TCP model to evaluate the capacity cost of achieving an order of magnitude decrease in PLP, and found it be (almost always) a 3-fold increase in link capacity that was required.

Keywords: Quality of experience, quality of service, packet loss probability, network capacity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 918
2755 The Association of Vitamin B₁₂ with Body Weight-and Fat-Based Indices in Childhood Obesity

Authors: Mustafa M. Donma, Orkide Donma

Abstract:

Vitamin deficiencies are common in obese individuals. Particularly, the status of vitamin B12 and its association with vitamin B9 (folate) and vitamin D is under investigation in recent time. Vitamin B12 is closely related to many vital processes in the body. In clinical studies, its involvement in fat metabolism draws attention from the obesity point of view. Obesity, in its advanced stages and in combination with metabolic syndrome (MetS) findings, may be a life-threatening health problem. Pediatric obesity is particularly important, because it may be a predictor of the severe chronic diseases during adulthood period of the child. Due to its role in fat metabolism, vitamin B12 deficiency may disrupt metabolic pathways of the lipid and energy metabolisms in the body. The association of low B12 levels with obesity degree may be an interesting topic to be investigated. Obesity indices may be helpful at this point. Weight- and fat-based indices are available. Of them, body mass index (BMI) is in the first group. Fat mass index (FMI), fat-free mass index (FFMI) and diagnostic obesity notation model assessment-II (D2I) index lie in the latter group. The aim of this study is to clarify possible associations between vitamin B12 status and obesity indices in pediatric population. The study comprises a total of 122 children. 32 children were included in the normal-body mass index (N-BMI) group. 46 and 44 children constitute groups with morbid obese children without MetS and with MetS, respectively. Informed consent forms and the approval of the institutional ethics committee were obtained. Tables prepared for obesity classification by World Health Organization were used. MetS criteria were defined. Anthropometric and blood pressure measurements were taken. BMI, FMI, FFMI, D2I were calculated. Routine laboratory tests were performed. Vitamin B9, B12, D concentrations were determined. Statistical evaluation of the study data was performed. Vitamin B9 and vitamin D levels were reduced in MetS group compared to children with N-BMI (p > 0.05). Significantly lower values were observed in vitamin B12 concentrations of MetS group (p < 0.01). Upon evaluation of blood pressure as well as triglyceride levels, there exist significant increases in morbid obese children. Significantly decreased concentrations of high-density lipoprotein cholesterol were observed. All of the obesity indices and insulin resistance index exhibit increasing tendency with the severity of obesity. Inverse correlations were calculated between vitamin D and insulin resistance index as well as vitamin B12 and D2I in morbid obese groups. In conclusion, a fat-based index, D2I, was the most prominent body index, which shows strong correlation with vitamin B12 concentrations in the late stage of obesity in children. A negative correlation between these two parameters was a confirmative finding related to the association between vitamin B12 and obesity degree. 

Keywords: Body mass index, children, D2I index, fat mass index, obesity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 666
2754 New Features for Specific JPEG Steganalysis

Authors: Johann Barbier, Eric Filiol, Kichenakoumar Mayoura

Abstract:

We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.

Keywords: Compressed frequency domain, Fisher discriminant, specific JPEG steganalysis.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2142
2753 Rotor Concepts for the Counter Flow Heat Recovery Fan

Authors: Christoph Speer

Abstract:

Decentralized ventilation systems should combine a small and economical design with high aerodynamic and thermal efficiency. The Counter Flow Heat Recovery Fan (CHRF) provides the ability to meet these requirements by using only one cross flow fan with a large number of blades to generate both airflows and which simultaneously acts as a regenerative counter flow heat exchanger. The successful development of the first laboratory prototype has shown the potential of this ventilation system. Occurring condensate on the surfaces of the fan blades during the cold and dry season can be recovered through the characteristic mode of operation. Hence the CHRF provides the possibility to avoid the need for frost protection and condensate drain. Through the implementation of system-specific solutions for flow balancing and summer bypass the required functionality is assured. The scalability of the CHRF concept allows the use in renovation as well as in new buildings from single-room devices through to systems for office buildings. High aerodynamic and thermal efficiency and the lower number of required mechatronic components should enable a reduction in investment as well as operating costs. The rotor is the key component of the system, the requirements and possible implementation variants are presented.

Keywords: CHRF, counter flow heat recovery fan, decentralized ventilation system, renovation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 865
2752 An E-Assessment Website to Implement Hierarchical Aggregate Assessment

Authors: M. Lesage, G. Raîche, M. Riopel, F. Fortin, D. Sebkhi

Abstract:

This paper describes a Web server implementation of the hierarchical aggregate assessment process in the field of education. This process describes itself as a field of teamwork assessment where teams can have multiple levels of hierarchy and supervision. This process is applied everywhere and is part of the management, education, assessment and computer science fields. The E-Assessment website named “Cluster” records in its database the students, the course material, the teams and the hierarchical relationships between the students. For the present research, the hierarchical relationships are team member, team leader and group administrator appointments. The group administrators have the responsibility to supervise team leaders. The experimentation of the application has been performed by high school students in geology courses and Canadian army cadets for navigation patrols in teams. This research extends the work of Nance that uses a hierarchical aggregation process similar as the one implemented in the “Cluster” application. 

Keywords: E-Learning, E-Assessment, Teamwork Assessment, Hierarchical Aggregate Assessment.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1820
2751 An Evaluation of Average Run Length of MaxEWMA and MaxGWMA Control Charts

Authors: S. Phanyaem

Abstract:

Exponentially weighted moving average control chart (EWMA) is a popular chart used for detecting shift in the mean of parameter of distributions in quality control. The objective of this paper is to compare the efficiency of control chart to detect an increases in the mean of a process. In particular, we compared the Maximum Exponentially Weighted Moving Average (MaxEWMA) and Maximum Generally Weighted Moving Average (MaxGWMA) control charts when the observations are Exponential distribution. The criteria for evaluate the performance of control chart is called, the Average Run Length (ARL). The result of comparison show that in the case of process is small sample size, the MaxEWMA control chart is more efficiency to detect shift in the process mean than MaxGWMA control chart. For the case of large sample size, the MaxEWMA control chart is more sensitive to detect small shift in the process mean than MaxGWMA control chart, and when the process is a large shift in mean, the MaxGWMA control chart is more sensitive to detect mean shift than MaxEWMA control chart.

Keywords: Maximum Exponentially Weighted Moving Average, Maximum General Weighted Moving Average, Average Run Length.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2154
2750 Estimated Production Potential Types of Wind Turbines Connected to the Network Using Random Numbers Simulation

Authors: Saeid Nahi, Seyed Mohammad Hossein Nabavi

Abstract:

Nowadays, power systems, energy generation by wind has been very important. Noting that the production of electrical energy by wind turbines on site to several factors (such as wind speed and profile site for the turbines, especially off the wind input speed, wind rated speed and wind output speed disconnect) is dependent. On the other hand, several different types of turbines in the market there. Therefore, selecting a turbine that its capacity could also answer the need for electric consumers the efficiency is high something is important and necessary. In this context, calculating the amount of wind power to help optimize overall network, system operation, in determining the parameters of wind power is very important. In this article, to help calculate the amount of wind power plant, connected to the national network in the region Manjil wind, selecting the best type of turbine and power delivery profile appropriate to the network using Monte Carlo method has been. In this paper, wind speed data from the wind site in Manjil, as minute and during the year has been. Necessary simulations based on Random Numbers Simulation method and repeat, using the software MATLAB and Excel has been done.

Keywords: wind turbine, efficiency, wind turbine work points, Random Numbers, reliability

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1387