Search results for: real number sequences
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14777

Search results for: real number sequences

12737 The Appropriate Number of Test Items That a Classroom-Based Reading Assessment Should Include: A Generalizability Analysis

Authors: Jui-Teng Liao

Abstract:

The selected-response (SR) format has been commonly adopted to assess academic reading in both formal and informal testing (i.e., standardized assessment and classroom assessment) because of its strengths in content validity, construct validity, as well as scoring objectivity and efficiency. When developing a second language (L2) reading test, researchers indicate that the longer the test (e.g., more test items) is, the higher reliability and validity the test is likely to produce. However, previous studies have not provided specific guidelines regarding the optimal length of a test or the most suitable number of test items or reading passages. Additionally, reading tests often include different question types (e.g., factual, vocabulary, inferential) that require varying degrees of reading comprehension and cognitive processes. Therefore, it is important to investigate the impact of question types on the number of items in relation to the score reliability of L2 reading tests. Given the popularity of the SR question format and its impact on assessment results on teaching and learning, it is necessary to investigate the degree to which such a question format can reliably measure learners’ L2 reading comprehension. The present study, therefore, adopted the generalizability (G) theory to investigate the score reliability of the SR format in L2 reading tests focusing on how many test items a reading test should include. Specifically, this study aimed to investigate the interaction between question types and the number of items, providing insights into the appropriate item count for different types of questions. G theory is a comprehensive statistical framework used for estimating the score reliability of tests and validating their results. Data were collected from 108 English as a second language student who completed an English reading test comprising factual, vocabulary, and inferential questions in the SR format. The computer program mGENOVA was utilized to analyze the data using multivariate designs (i.e., scenarios). Based on the results of G theory analyses, the findings indicated that the number of test items had a critical impact on the score reliability of an L2 reading test. Furthermore, the findings revealed that different types of reading questions required varying numbers of test items for reliable assessment of learners’ L2 reading proficiency. Further implications for teaching practice and classroom-based assessments are discussed.

Keywords: second language reading assessment, validity and reliability, Generalizability theory, Academic reading, Question format

Procedia PDF Downloads 67
12736 Deriving Framework for Slum Rehabilitation through Environmental Perspective: Case of Mumbai

Authors: Ashwini Bhosale, Yogesh Patil

Abstract:

Urban areas are extremely complicated environmental settings, where health and well-being of an individual and population are governed by a large number of bio-physical, socio-economical, and inclusive aspects. Although poverty and slums are the prime issues under UN-HABITAT agenda of environmental sustainability, slums, the inevitable part of urban environment, have not accounted for inclusive city planning. Developing nations, where about 60 % of world slum population resides, are increasingly under pressure to uplift the urban poor, particularly slum dwellers. From a point of advantage, these new slum redevelopment projects have succeeded in providing legitimized and more permanent and stable shelter for the low income people, as well as individualized sanitation and water supply. However, they unfortunately follow the “one type fits all" approach and exhibit no response to the climatic design needs on Mumbai. The thesis focuses on the study of environmental perspectives in the context of Daylight, natural ventilation and social aspects in the design process of Slum-Rehabilitation schemes (SRS) – case of Mumbai. It attempts to investigate into Indian approaches about SRS and concludes upon strategies to be incorporated in SRS to improve the overall SRS environment. The main objectives of this work have been to identify and study the spatial configuration and possibilities of daylight and natural ventilation in Slum Rehabilitated buildings. The performance of the proposed method was evaluated by comparison with the daylight luminance simulated by lighting software, namely ECOTECT, and with measurements under real skies whereas for the ventilation study purpose, software named FLOW DESIGN was used.

Keywords: urban environment, slum-rehabilitation, daylight, natural-ventilation, architectural consequences

Procedia PDF Downloads 367
12735 Effect of Integrated Nutrient Management Practice on Cultivation Scented Rice Varieties- a Better Approach for Resource Conservation

Authors: Amit Kumar Patel, M. C. Bhambri, Damini Thawait, Srishti Pandey

Abstract:

The experiment was carried out at Raipur during rainy season of 2012. The experiment revealed that the performance of Dubraj was comparatively better than that of badshah bhog, Vishnu bhog and bisni. The number of grains panicle-1, number of filled grains panicle-1 were comparable in Dubraj and badshah bhog. Among the different nutrient, application of 80:50:40 kg N:P2O5:K2O ha-1(50% Inorganic+50% Organic) gave better performance in all the above characters. It is revealed that the variety Dubraj fertilized with 80:50:40 kg N:P2O5:K2O ha-1(50% Inorganic+50% Organic) gave the good yield attributing characters along with highest yield.

Keywords: scented rice, organic manures, chemical fertilizers, yield, varieties

Procedia PDF Downloads 485
12734 PET Image Resolution Enhancement

Authors: Krzysztof Malczewski

Abstract:

PET is widely applied scanning procedure in medical imaging based research. It delivers measurements of functioning in distinct areas of the human brain while the patient is comfortable, conscious and alert. This article presents the new compression sensing based super-resolution algorithm for improving the image resolution in clinical Positron Emission Tomography (PET) scanners. The issue of motion artifacts is well known in Positron Emission Tomography (PET) studies as its side effect. The PET images are being acquired over a limited period of time. As the patients cannot hold breath during the PET data gathering, spatial blurring and motion artefacts are the usual result. These may lead to wrong diagnosis. It is shown that the presented approach improves PET spatial resolution in cases when Compressed Sensing (CS) sequences are used. Compressed Sensing (CS) aims at signal and images reconstructing from significantly fewer measurements than were traditionally thought necessary. The application of CS to PET has the potential for significant scan time reductions, with visible benefits for patients and health care economics. In this study the goal is to combine super-resolution image enhancement algorithm with CS framework to achieve high resolution PET output. Both methods emphasize on maximizing image sparsity on known sparse transform domain and minimizing fidelity.

Keywords: PET, super-resolution, image reconstruction, pattern recognition

Procedia PDF Downloads 358
12733 Development and Validation of Thermal Stability in Complex System ABDM has two ASIC by NISA and COMSOL Tools

Authors: A. Oukaira, A. Lakhssassi, O. Ettahri

Abstract:

To make a good thermal management in an ABDM (Adapter Board Detector Module) card, we must first control temperature and its gradient from the first step in the design of integrated circuits ASIC of our complex system. In this paper, our main goal is to develop and validate the thermal stability in order to get an idea of the flow of heat around the ASIC in transient and thus address the thermal issues for integrated circuits at the ABDM card. However, we need heat sources simulations for ABDM card to establish its thermal mapping. This led us to perform simulations at each ASIC that will allow us to understand the thermal ABDM map and find real solutions for each one of our complex system that contains 36 ABDM map, taking into account the different layers around ASIC. To do a transient simulation under NISA, we had to build a function of power modulation in time TIMEAMP. The maximum power generated in the ASIC is 0.6 W. We divided the power uniformly in the volume of the ASIC. This power was applied for 5 seconds to visualize the evolution and distribution of heat around the ASIC. The DBC (Dirichlet Boundary conditions) method was applied around the ABDM at 25°C and just after these simulations in NISA tool we will validate them by COMSOL tool, wich is a numerical calculation software for a modular finite element for modeling a wide variety of physical phenomena characterizing a real problem. It will also be a design tool with its ability to handle 3D geometries for complex systems.

Keywords: ABDM, APD, thermal mapping, complex system

Procedia PDF Downloads 252
12732 An Ensemble Learning Method for Applying Particle Swarm Optimization Algorithms to Systems Engineering Problems

Authors: Ken Hampshire, Thomas Mazzuchi, Shahram Sarkani

Abstract:

As a subset of metaheuristics, nature-inspired optimization algorithms such as particle swarm optimization (PSO) have shown promise both in solving intractable problems and in their extensibility to novel problem formulations due to their general approach requiring few assumptions. Unfortunately, single instantiations of algorithms require detailed tuning of parameters and cannot be proven to be best suited to a particular illustrative problem on account of the “no free lunch” (NFL) theorem. Using these algorithms in real-world problems requires exquisite knowledge of the many techniques and is not conducive to reconciling the various approaches to given classes of problems. This research aims to present a unified view of PSO-based approaches from the perspective of relevant systems engineering problems, with the express purpose of then eliciting the best solution for any problem formulation in an ensemble learning bucket of models approach. The central hypothesis of the research is that extending the PSO algorithms found in the literature to real-world optimization problems requires a general ensemble-based method for all problem formulations but a specific implementation and solution for any instance. The main results are a problem-based literature survey and a general method to find more globally optimal solutions for any systems engineering optimization problem.

Keywords: particle swarm optimization, nature-inspired optimization, metaheuristics, systems engineering, ensemble learning

Procedia PDF Downloads 79
12731 Qualitative and Quantitative Research Methodology Theoretical Framework and Descriptive Theory: PhD Construction Management

Authors: Samuel Quashie

Abstract:

PhDs in Construction Management often designs their methods based on those established in social sciences using theoretical models, to collect, gather and analysis data to answer research questions. Work aim is to apply qualitative and quantitative as a data analysis method, and as part of the theoretical framework - descriptive theory. To improve the ability to replicate the contribution to knowledge the research. Using practical triangulation approach, which covers, interviews and observations, literature review and (archival) document studies, project-based case studies, questionnaires surveys and review of integrated systems used in, construction and construction related industries. The clarification of organisational context and management delivery that influences organizational performance and quality of product and measures are achieved. Results illustrate improved reliability in this research approach when interpreting real world phenomena; cumulative results of research can be applied with confidence under similar environments. Assisted validity of the PhD research outcomes and strengthens the confidence to apply cumulative results of research under similar conditions in the Built Environment research systems, which have been criticised for the lack of reliability in approaches when interpreting real world phenomena.

Keywords: case studies, descriptive theory, theoretical framework, qualitative and quantitative research

Procedia PDF Downloads 364
12730 Multi-Objective Evolutionary Computation Based Feature Selection Applied to Behaviour Assessment of Children

Authors: F. Jiménez, R. Jódar, M. Martín, G. Sánchez, G. Sciavicco

Abstract:

Abstract—Attribute or feature selection is one of the basic strategies to improve the performances of data classification tasks, and, at the same time, to reduce the complexity of classifiers, and it is a particularly fundamental one when the number of attributes is relatively high. Its application to unsupervised classification is restricted to a limited number of experiments in the literature. Evolutionary computation has already proven itself to be a very effective choice to consistently reduce the number of attributes towards a better classification rate and a simpler semantic interpretation of the inferred classifiers. We present a feature selection wrapper model composed by a multi-objective evolutionary algorithm, the clustering method Expectation-Maximization (EM), and the classifier C4.5 for the unsupervised classification of data extracted from a psychological test named BASC-II (Behavior Assessment System for Children - II ed.) with two objectives: Maximizing the likelihood of the clustering model and maximizing the accuracy of the obtained classifier. We present a methodology to integrate feature selection for unsupervised classification, model evaluation, decision making (to choose the most satisfactory model according to a a posteriori process in a multi-objective context), and testing. We compare the performance of the classifier obtained by the multi-objective evolutionary algorithms ENORA and NSGA-II, and the best solution is then validated by the psychologists that collected the data.

Keywords: evolutionary computation, feature selection, classification, clustering

Procedia PDF Downloads 353
12729 To Study the Performance of FMS under Different Manufacturing Strategies

Authors: Mohammed Ali

Abstract:

A flexible manufacturing system has been studied under different manufacturing strategies. The aim of this paper is to test the impact of number of pallets and routing flexibility (design strategy) on system performance operating at different sequencing and dispatching rules (control strategies) at unbalanced load condition (planning strategies). A computer simulation model is developed to evaluate the effects of aforementioned strategies on the make-span time, which is taken as the system performance measure. The impact of number of pallets is shown with the different levels of routing flexibility. In this paper, the same manufacturing system is modeled under different combination of sequencing and dispatching rules. The result of the simulation shows that there is definite range of pallets for each level of routing flexibility at which the systems performs satisfactorily.

Keywords: flexible manufacturing system, manufacturing, strategy, makespan

Procedia PDF Downloads 652
12728 A Convolution Neural Network PM-10 Prediction System Based on a Dense Measurement Sensor Network in Poland

Authors: Piotr A. Kowalski, Kasper Sapala, Wiktor Warchalowski

Abstract:

PM10 is a suspended dust that primarily has a negative effect on the respiratory system. PM10 is responsible for attacks of coughing and wheezing, asthma or acute, violent bronchitis. Indirectly, PM10 also negatively affects the rest of the body, including increasing the risk of heart attack and stroke. Unfortunately, Poland is a country that cannot boast of good air quality, in particular, due to large PM concentration levels. Therefore, based on the dense network of Airly sensors, it was decided to deal with the problem of prediction of suspended particulate matter concentration. Due to the very complicated nature of this issue, the Machine Learning approach was used. For this purpose, Convolution Neural Network (CNN) neural networks have been adopted, these currently being the leading information processing methods in the field of computational intelligence. The aim of this research is to show the influence of particular CNN network parameters on the quality of the obtained forecast. The forecast itself is made on the basis of parameters measured by Airly sensors and is carried out for the subsequent day, hour after hour. The evaluation of learning process for the investigated models was mostly based upon the mean square error criterion; however, during the model validation, a number of other methods of quantitative evaluation were taken into account. The presented model of pollution prediction has been verified by way of real weather and air pollution data taken from the Airly sensor network. The dense and distributed network of Airly measurement devices enables access to current and archival data on air pollution, temperature, suspended particulate matter PM1.0, PM2.5, and PM10, CAQI levels, as well as atmospheric pressure and air humidity. In this investigation, PM2.5, and PM10, temperature and wind information, as well as external forecasts of temperature and wind for next 24h served as inputted data. Due to the specificity of the CNN type network, this data is transformed into tensors and then processed. This network consists of an input layer, an output layer, and many hidden layers. In the hidden layers, convolutional and pooling operations are performed. The output of this system is a vector containing 24 elements that contain prediction of PM10 concentration for the upcoming 24 hour period. Over 1000 models based on CNN methodology were tested during the study. During the research, several were selected out that give the best results, and then a comparison was made with the other models based on linear regression. The numerical tests carried out fully confirmed the positive properties of the presented method. These were carried out using real ‘big’ data. Models based on the CNN technique allow prediction of PM10 dust concentration with a much smaller mean square error than currently used methods based on linear regression. What's more, the use of neural networks increased Pearson's correlation coefficient (R²) by about 5 percent compared to the linear model. During the simulation, the R² coefficient was 0.92, 0.76, 0.75, 0.73, and 0.73 for 1st, 6th, 12th, 18th, and 24th hour of prediction respectively.

Keywords: air pollution prediction (forecasting), machine learning, regression task, convolution neural networks

Procedia PDF Downloads 130
12727 Object-Based Flow Physics for Aerodynamic Modelling in Real-Time Environments

Authors: William J. Crowther, Conor Marsh

Abstract:

Object-based flow simulation allows fast computation of arbitrarily complex aerodynamic models made up of simple objects with limited flow interactions. The proposed approach is universally applicable to objects made from arbitrarily scaled ellipsoid primitives at arbitrary aerodynamic attitude and angular rate. The use of a component-based aerodynamic modelling approach increases efficiency by allowing selective inclusion of different physics models at run-time and allows extensibility through the development of new models. Insight into the numerical stability of the model under first order fixed-time step integration schemes is provided by stability analysis of the drag component. The compute cost of model components and functions is evaluated and compared against numerical benchmarks. Model static outputs are verified against theoretical expectations and dynamic behaviour using falling plate data from the literature. The model is applied to a range of case studies to demonstrate the efficacy of its application in extensibility, ease of use, and low computational cost. Dynamically complex multi-body systems can be implemented in a transparent and efficient manner, and we successfully demonstrate large scenes with hundreds of objects interacting with diverse flow fields.

Keywords: aerodynamics, real-time simulation, low-order model, flight dynamics

Procedia PDF Downloads 88
12726 Layout Optimization of a Start-up COVID-19 Testing Kit Manufacturing Facility

Authors: Poojan Vora, Hardik Pancholi, Sanket Tajane, Harsh Shah, Elias Keedy

Abstract:

The global COVID-19 pandemic has affected the industry drastically in many ways. Even though the vaccine is being distributed quickly and despite the decreasing number of positive cases, testing is projected to remain a key aspect of the ‘new normal’. Improving existing plant layout and improving safety within the facility are of great importance in today’s industries because of the need to ensure productivity optimization and reduce safety risks. In practice, it is essential for any manufacturing plant to reduce nonvalue adding steps such as the movement of materials and rearrange similar processes. In the current pandemic situation, optimized layouts will not only increase safety measures but also decrease the fixed cost per unit manufactured. In our case study, we carefully studied the existing layout and the manufacturing steps of a new Texas start-up company that manufactures COVID testing kits. The effects of production rate are incorporated with the computerized relative allocation of facilities technique (CRAFT) algorithm to improve the plant layout and estimate the optimization parameters. Our work reduces the company’s material handling time and increases their daily production. Real data from the company are used in the case study to highlight the importance of colleges in fostering small business needs and improving the collaboration between college researchers and industries by using existing models to advance best practices.

Keywords: computerized relative allocation of facilities technique, facilities planning, optimization, start-up business

Procedia PDF Downloads 127
12725 Inversion of the Spectral Analysis of Surface Waves Dispersion Curves through the Particle Swarm Optimization Algorithm

Authors: A. Cerrato Casado, C. Guigou, P. Jean

Abstract:

In this investigation, the particle swarm optimization (PSO) algorithm is used to perform the inversion of the dispersion curves in the spectral analysis of surface waves (SASW) method. This inverse problem usually presents complicated solution spaces with many local minima that make difficult the convergence to the correct solution. PSO is a metaheuristic method that was originally designed to simulate social behavior but has demonstrated powerful capabilities to solve inverse problems with complex space solution and a high number of variables. The dispersion curve of the synthetic soils is constructed by the vertical flexibility coefficient method, which is especially convenient for soils where the stiffness does not increase gradually with depth. The reason is that these types of soil profiles are not normally dispersive since the dominant mode of Rayleigh waves is usually not coincident with the fundamental mode. Multiple synthetic soil profiles have been tested to show the characteristics of the convergence process and assess the accuracy of the final soil profile. In addition, the inversion procedure is applied to multiple real soils and the final profile compared with the available information. The combination of the vertical flexibility coefficient method to obtain the dispersion curve and the PSO algorithm to carry out the inversion process proves to be a robust procedure that is able to provide good solutions for complex soil profiles even with scarce prior information.

Keywords: dispersion, inverse problem, particle swarm optimization, SASW, soil profile

Procedia PDF Downloads 169
12724 Selection of Optimal Reduced Feature Sets of Brain Signal Analysis Using Heuristically Optimized Deep Autoencoder

Authors: Souvik Phadikar, Nidul Sinha, Rajdeep Ghosh

Abstract:

In brainwaves research using electroencephalogram (EEG) signals, finding the most relevant and effective feature set for identification of activities in the human brain is a big challenge till today because of the random nature of the signals. The feature extraction method is a key issue to solve this problem. Finding those features that prove to give distinctive pictures for different activities and similar for the same activities is very difficult, especially for the number of activities. The performance of a classifier accuracy depends on this quality of feature set. Further, more number of features result in high computational complexity and less number of features compromise with the lower performance. In this paper, a novel idea of the selection of optimal feature set using a heuristically optimized deep autoencoder is presented. Using various feature extraction methods, a vast number of features are extracted from the EEG signals and fed to the autoencoder deep neural network. The autoencoder encodes the input features into a small set of codes. To avoid the gradient vanish problem and normalization of the dataset, a meta-heuristic search algorithm is used to minimize the mean square error (MSE) between encoder input and decoder output. To reduce the feature set into a smaller one, 4 hidden layers are considered in the autoencoder network; hence it is called Heuristically Optimized Deep Autoencoder (HO-DAE). In this method, no features are rejected; all the features are combined into the response of responses of the hidden layer. The results reveal that higher accuracy can be achieved using optimal reduced features. The proposed HO-DAE is also compared with the regular autoencoder to test the performance of both. The performance of the proposed method is validated and compared with the other two methods recently reported in the literature, which reveals that the proposed method is far better than the other two methods in terms of classification accuracy.

Keywords: autoencoder, brainwave signal analysis, electroencephalogram, feature extraction, feature selection, optimization

Procedia PDF Downloads 102
12723 Regional Dynamics of Innovation and Entrepreneurship in the Optics and Photonics Industry

Authors: Mustafa İlhan Akbaş, Özlem Garibay, Ivan Garibay

Abstract:

The economic entities in innovation ecosystems form various industry clusters, in which they compete and cooperate to survive and grow. Within a successful and stable industry cluster, the entities acquire different roles that complement each other in the system. The universities and research centers have been accepted to have a critical role in these systems for the creation and development of innovations. However, the real effect of research institutions on regional economic growth is difficult to assess. In this paper, we present our approach for the identification of the impact of research activities on the regional entrepreneurship for a specific high-tech industry: optics and photonics. The optics and photonics has been defined as an enabling industry, which combines the high-tech photonics technology with the developing optics industry. The recent literature suggests that the growth of optics and photonics firms depends on three important factors: the embedded regional specializations in the labor market, the research and development infrastructure, and a dynamic small firm network capable of absorbing new technologies, products and processes. Therefore, the role of each factor and the dynamics among them must be understood to identify the requirements of the entrepreneurship activities in optics and photonics industry. There are three main contributions of our approach. The recent studies show that the innovation in optics and photonics industry is mostly located around metropolitan areas. There are also studies mentioning the importance of research center locations and universities in the regional development of optics and photonics industry. These studies are mostly limited with the number of patents received within a short period of time or some limited survey results. Therefore the first contribution of our approach is conducting a comprehensive analysis for the state and recent history of the photonics and optics research in the US. For this purpose, both the research centers specialized in optics and photonics and the related research groups in various departments of institutions (e.g. Electrical Engineering, Materials Science) are identified and a geographical study of their locations is presented. The second contribution of the paper is the analysis of regional entrepreneurship activities in optics and photonics in recent years. We use the membership data of the International Society for Optics and Photonics (SPIE) and the regional photonics clusters to identify the optics and photonics companies in the US. Then the profiles and activities of these companies are gathered by extracting and integrating the related data from the National Establishment Time Series (NETS) database, ES-202 database and the data sets from the regional photonics clusters. The number of start-ups, their employee numbers and sales are some examples of the extracted data for the industry. Our third contribution is the utilization of collected data to investigate the impact of research institutions on the regional optics and photonics industry growth and entrepreneurship. In this analysis, the regional and periodical conditions of the overall market are taken into consideration while discovering and quantifying the statistical correlations.

Keywords: entrepreneurship, industrial clusters, optics, photonics, emerging industries, research centers

Procedia PDF Downloads 394
12722 Quality and Quantity in the Strategic Network of Higher Education Institutions

Authors: Juha Kettunen

Abstract:

The study analyzes the quality and the size of the strategic network of higher education institutions and the concept of fitness for purpose in quality assurance. It also analyses the transaction costs of networking that have consequences on the number of members in the network. Empirical evidence is presented from the Consortium on Applied Research and Professional Education, which is a European strategic network of six higher education institutions. The results of the study support the argument that the number of members in the strategic network should be relatively small to provide high-quality results. The practical importance is that networking has been able to promote international research and development projects. The results of this study are important for those who want to design and improve international networks in higher education.

Keywords: higher education, network, research and development, strategic management

Procedia PDF Downloads 328
12721 The Effect of Chemical Degradation of a Nonwoven Filter Media Membrane in Polyester

Authors: Rachid El Aidani, Phuong Nguyen-Tri, Toan Vu-Khanh

Abstract:

The filter media in synthetic fibre is the most geotextile materials used in aerosol and drainage filtration, particularly for buildings soil reinforcement in civil engineering due to its appropriated properties and its low cost. However, the current understanding of the durability and stability of this material in real service conditions, especially under severe long-term conditions are completely limited. This study has examined the effects of the chemical aging of a filter media in polyester non-woven under different temperatures (50, 70 and 80˚C) and pH (2. 7 and 12). The effect of aging conditions on mechanical properties, morphology, permeability, thermal stability and molar weigh changes is investigated. The results showed a significant reduction of mechanical properties in term of tensile strength, puncture force and tearing forces of the filter media after chemical aging due to the chemical degradation. The molar mass and mechanical properties changes in different temperature and pH showed a complex dependence of material properties on environmental conditions. The SEM and AFM characterizations showed a significant impact of the thermal aging on the morphological properties of the fibers. Based on the obtained results, the lifetime of the material in different temperatures was determined by the use of the Arrhenius model. These results provide useful information to better understand phenomena occurring during chemical aging of the filter media and may help to predict the service lifetime of this material in real used conditions.

Keywords: nonwoven membrane, chemical aging, mechanical properties, lifetime, filter media

Procedia PDF Downloads 307
12720 Comparison of Number of Waves Surfed and Duration Using Global Positioning System and Inertial Sensors

Authors: João Madureira, Ricardo Lagido, Inês Sousa, Fraunhofer Portugal

Abstract:

Surf is an increasingly popular sport and its performance evaluation is often qualitative. This work aims at using a smartphone to collect and analyze the GPS and inertial sensors data in order to obtain quantitative metrics of the surfing performance. Two approaches are compared for detection of wave rides, computing the number of waves rode in a surfing session, the starting time of each wave and its duration. The first approach is based on computing the velocity from the Global Positioning System (GPS) signal and finding the velocity thresholds that allow identifying the start and end of each wave ride. The second approach adds information from the Inertial Measurement Unit (IMU) of the smartphone, to the velocity thresholds obtained from the GPS unit, to determine the start and end of each wave ride. The two methods were evaluated using GPS and IMU data from two surfing sessions and validated with similar metrics extracted from video data collected from the beach. The second method, combining GPS and IMU data, was found to be more accurate in determining the number of waves, start time and duration. This paper shows that it is feasible to use smartphones for quantification of performance metrics during surfing. In particular, detection of the waves rode and their duration can be accurately determined using the smartphone GPS and IMU.

Keywords: inertial measurement unit (IMU), global positioning system (GPS), smartphone, surfing performance

Procedia PDF Downloads 389
12719 Trends in Research Regarding International Student Connectedness, A Systematic Review

Authors: Zilola Kozimova

Abstract:

Humans are highly social creatures, and our social surroundings create a large part of our daily experiences. Feeling connected and belonging at school have been studied a lot, especially in the period up to college. The need to feel connected becomes even more vital when people choose to study abroad. The number of published research in the field has increased recently, creating sufficient studies for a systematic literature review. The current study was conducted to find out existing trends and central themes in the field regarding international student connectedness. Using PRISMA 2020 and Shariff et al.’s work as the guidelines, I conducted a systematic literature review of studies regarding international student connectedness in higher education. Three steps of inclusion/exclusion criteria were used to determine the final studies to be included. The results show an increasing trend in the field as the number of related studies drastically rose after 2017. the results showed that there are three phases in the research regarding the connectedness of international students: a rejection period, a sudden increase of interest in the topic, and merging as an essential part of the mental well-being of international students. There is also a change in the themes regarding the topic, as there is a rise in the number of research published regarding international students’ mental health in recent years, connectedness being a sub-topic.

Keywords: international students, connectedness, mental well-being of international students, trends, higher education

Procedia PDF Downloads 101
12718 Educational Sustainability: Teaching the Next Generation of Educators in Medical Simulation

Authors: Thomas Trouton, Sebastian Tanner, Manvir Sandher

Abstract:

The use of simulation in undergraduate and postgraduate medical curricula is ever-growing, is a useful addition to the traditional apprenticeship model of learning within medical education, and better prepares graduates for the team-based approach to healthcare seen in real-life clinical practice. As a learning tool, however, undergraduate medical students often have little understanding of the theory behind the use of medical simulation and have little experience in planning and delivering their own simulated teaching sessions. We designed and implemented a student-selected component (SSC) as part of the undergraduate medical curriculum at the University of Buckingham Medical School to introduce students to the concepts behind the use of medical simulation in education and allow them to plan and deliver their own simulated medical scenario to their peers. The SSC took place over a 2-week period in the 3rd year of the undergraduate course. There was a mix of lectures, seminars and interactive group work sessions, as well as hands-on experience in the simulation suite, to introduce key concepts related to medical simulation, including technical considerations in simulation, human factors, debriefing and troubleshooting scenarios. We evaluated the success of our SSC using “Net Promotor Scores” (NPS) to assess students’ confidence in planning and facilitating a simulation-based teaching session, as well as leading a debrief session. In all three domains, we showed an increase in the confidence of the students. We also showed an increase in confidence in the management of common medical emergencies as a result of the SSC. Overall, the students who chose our SSC had the opportunity to learn new skills in medical education, with a particular focus on the use of simulation-based teaching, and feedback highlighted that a number of students would take these skills forward in their own practice. We demonstrated an increase in confidence in several domains related to the use of medical simulation in education and have hopefully inspired a new generation of medical educators.

Keywords: simulation, SSC, teaching, medical students

Procedia PDF Downloads 108
12717 Widely Diversified Macroeconomies in the Super-Long Run Casts a Doubt on Path-Independent Equilibrium Growth Model

Authors: Ichiro Takahashi

Abstract:

One of the major assumptions of mainstream macroeconomics is the path independence of capital stock. This paper challenges this assumption by employing an agent-based approach. The simulation results showed the existence of multiple "quasi-steady state" equilibria of the capital stock, which may cast serious doubt on the validity of the assumption. The finding would give a better understanding of many phenomena that involve hysteresis, including the causes of poverty. The "market-clearing view" has been widely shared among major schools of macroeconomics. They understand that the capital stock, the labor force, and technology, determine the "full-employment" equilibrium growth path and demand/supply shocks can move the economy away from the path only temporarily: the dichotomy between the short-run business cycles and the long-run equilibrium path. The view then implicitly assumes the long-run capital stock to be independent of how the economy has evolved. In contrast, "Old Keynesians" have recognized fluctuations in output as arising largely from fluctuations in real aggregate demand. It will then be an interesting question to ask if an agent-based macroeconomic model, which is known to have path dependence, can generate multiple full-employment equilibrium trajectories of the capital stock in the super-long run. If the answer is yes, the equilibrium level of capital stock, an important supply-side factor, would no longer be independent of the business cycle phenomenon. This paper attempts to answer the above question by using the agent-based macroeconomic model developed by Takahashi and Okada (2010). The model would serve this purpose well because it has neither population growth nor technology progress. The objective of the paper is twofold: (1) to explore the causes of long-term business cycle, and (2) to examine the super-long behaviors of the capital stock of full-employment economies. (1) The simulated behaviors of the key macroeconomic variables such as output, employment, real wages showed widely diversified macro-economies. They were often remarkably stable but exhibited both short-term and long-term fluctuations. The long-term fluctuations occur through the following two adjustments: the quantity and relative cost adjustments of capital stock. The first one is obvious and assumed by many business cycle theorists. The reduced aggregate demand lowers prices, which raises real wages, thereby decreasing the relative cost of capital stock with respect to labor. (2) The long-term business cycles/fluctuations were synthesized with the hysteresis of real wages, interest rates, and investments. In particular, a sequence of the simulation runs with a super-long simulation period generated a wide range of perfectly stable paths, many of which achieved full employment: all the macroeconomic trajectories, including capital stock, output, and employment, were perfectly horizontal over 100,000 periods. Moreover, the full-employment level of capital stock was influenced by the history of unemployment, which was itself path-dependent. Thus, an experience of severe unemployment in the past kept the real wage low, which discouraged a relatively costly investment in capital stock. Meanwhile, a history of good performance sometimes brought about a low capital stock due to a high-interest rate that was consistent with a strong investment.

Keywords: agent-based macroeconomic model, business cycle, hysteresis, stability

Procedia PDF Downloads 197
12716 Random Subspace Ensemble of CMAC Classifiers

Authors: Somaiyeh Dehghan, Mohammad Reza Kheirkhahan Haghighi

Abstract:

The rapid growth of domains that have data with a large number of features, while the number of samples is limited has caused difficulty in constructing strong classifiers. To reduce the dimensionality of the feature space becomes an essential step in classification task. Random subspace method (or attribute bagging) is an ensemble classifier that consists of several classifiers that each base learner in ensemble has subset of features. In the present paper, we introduce Random Subspace Ensemble of CMAC neural network (RSE-CMAC), each of which has training with subset of features. Then we use this model for classification task. For evaluation performance of our model, we compare it with bagging algorithm on 36 UCI datasets. The results reveal that the new model has better performance.

Keywords: classification, random subspace, ensemble, CMAC neural network

Procedia PDF Downloads 316
12715 Comparative Study of Fatigue and Drowsiness in the Night-Time Passenger Transportation Industry in Japan

Authors: Hiroshi Ikeda

Abstract:

In this research, a questionnaire survey was conducted to measure nap, drowsiness and fatigue of drivers who work long shifts, to discuss about the work environment and health conditions for taxi and bus drivers who work at night time. The questionnaire sheet used for this research was organized into the following categories: tension/tiredness, drowsiness while driving, and the nap situation during night-time work. The number of taxi drivers was 127 and the number of bus drivers was 40. Concerning the results of a comparison of nap hours of taxi and bus drivers, the taxi drivers’ nap hours are overwhelmingly shorter, and also the frequency of drivers who feel drowsiness is higher. The burden on bus drivers does not change because of the system of a two-driver rotation shift. In particular, the working environment of the taxi driver may lead to greater fatigue accumulation than the bus driver’s environment.

Keywords: bus and taxi, drowsiness, fatigue, nap

Procedia PDF Downloads 313
12714 Physiological Effects on Scientist Astronaut Candidates: Hypobaric Training Assessment

Authors: Pedro Llanos, Diego García

Abstract:

This paper is addressed to expanding our understanding of the effects of hypoxia training on our bodies to better model its dynamics and leverage some of its implications and effects on human health. Hypoxia training is a recommended practice for military and civilian pilots that allow them to recognize their early hypoxia signs and symptoms, and Scientist Astronaut Candidates (SACs) who underwent hypobaric hypoxia (HH) exposure as part of a training activity for prospective suborbital flight applications. This observational-analytical study describes physiologic responses and symptoms experienced by a SAC group before, during and after HH exposure and proposes a model for assessing predicted versus observed physiological responses. A group of individuals with diverse Science Technology Engineering Mathematics (STEM) backgrounds conducted a hypobaric training session to an altitude up to 22,000 ft (FL220) or 6,705 meters, where heart rate (HR), breathing rate (BR) and core temperature (Tc) were monitored with the use of a chest strap sensor pre and post HH exposure. A pulse oximeter registered levels of saturation of oxygen (SpO2), number and duration of desaturations during the HH chamber flight. Hypoxia symptoms as described by the SACs during the HH training session were also registered. This data allowed to generate a preliminary predictive model of the oxygen desaturation and O2 pressure curve for each subject, which consists of a sixth-order polynomial fit during exposure, and a fifth or fourth-order polynomial fit during recovery. Data analysis showed that HR and BR showed no significant differences between pre and post HH exposure in most of the SACs, while Tc measures showed slight but consistent decrement changes. All subjects registered SpO2 greater than 94% for the majority of their individual HH exposures, but all of them presented at least one clinically significant desaturation (SpO2 < 85% for more than 5 seconds) and half of the individuals showed SpO2 below 87% for at least 30% of their HH exposure time. Finally, real time collection of HH symptoms presented temperature somatosensory perceptions (SP) for 65% of individuals, and task-focus issues for 52.5% of individuals as the most common HH indications. 95% of the subjects experienced HH onset symptoms below FL180; all participants achieved full recovery of HH symptoms within 1 minute of donning their O2 mask. The current HH study performed on this group of individuals suggests a rapid and fully reversible physiologic response after HH exposure as expected and obtained in previous studies. Our data showed consistent results between predicted versus observed SpO2 curves during HH suggesting a mathematical function that may be used to model HH performance deficiencies. During the HH study, real-time HH symptoms were registered providing evidenced SP and task focusing as the earliest and most common indicators. Finally, an assessment of HH signs of symptoms in a group of heterogeneous, non-pilot individuals showed similar results to previous studies in homogeneous populations of pilots.

Keywords: slow onset hypoxia, hypobaric chamber training, altitude sickness, symptoms and altitude, pressure cabin

Procedia PDF Downloads 105
12713 Histogenesis of the Stomach of Pre-Hatching Quail: A Light and Electron Microscopic Study

Authors: Soha A Soliman, Yasser A Ahmed, Mohamed A Khalaf

Abstract:

Although the enormous literature describing the histology of the stomach of different avian species during the posthatching development, the available literature on the pre-hatching development of quail stomach development is scanty. Thus, the current study was undertaken to provide a careful description of the main histological events during the embryonic development of quail stomach. To achieve this aim, daily histological specimens from the stomach of quail of 4 days post-incubation till the day 17 (few hours before hatching) were examined with light microscopy. The current study showed that the primitive gut tube of the embryonic quail appeared at the 4th day post incubation, and both parts of stomach (proventriculus and gizzard) were similar in structure and composed of endodermal epithelium of pseudostratified type surrounded by undifferentiated mesenchymal tissue. The sequences of the developmental events in the gut tube were preceded in a cranio-caudal pattern. By the 5th day, the endodermal covering of the primitive proventriculus gave rise to sac-like invaginations. The primitive gizzard was distinguished into thick-walled bodies and thin-walled sacs. In the 6th day, the prospective proventricular glandular epithelium became canalized and the muscular layer was developed in the cranial part of the proventriculus, whereas the primitive muscular coat of the gizzard was represented by a layer of condensed mesenchyme. In the 7th day, the proventricular glandular epithelial invaginations increased in depth and number, while, the muscularis mucosa and the muscular layer began to be distinguished. In the 8th day, the myoblasts differentiated into spindle shaped smooth muscle fibers. In the 10th day, branching of the proventricular glands began. The branching continued later on. The surface and the glandular epithelium were transformed into simple columnar type in the 12th day. The epithelial covering of the gizzard gave rise to tubular invaginations lined by simple cuboidal epithelium and the surface epithelium became simple columnar. Canalization of the tubular glands was recognized in the 14th day. In the 15th day, the proventricular surface epithelium invaginated in an concentric manner around a central cavity to form immature secretory units. The central cavity was lined by eosinophilic cells which form the ductal epithelia. The peripheral lamellae were lined by basophilic cells; the undifferentiated oxyntico-peptic cells. Entero-endocrine cells stained positive for silver impregnation in the proventricular glands. The mucosal folding in the gizzard appeared in the 15th day to form the plicae and the sulci. The wall of the proventriculus and gizzard in the 17th day acquired the main histological features of post-hatching birds, but neither the surface nor the ductal epithelium were differentiated to mucous producing cells. The current results shoed be considered in the molecular developmental studies.

Keywords: quail, proventriculus, gizzard, pre-hatching, histology

Procedia PDF Downloads 604
12712 Diversity of Culturable Forms of Microorganisms in Soils with Long-term Exposure to Petroleum Hydrocarbons and Prospects for Bioremediation

Authors: Yessentayeva K. Y., Berzhanova R. Z., Mukasheva T. D.

Abstract:

The purpose of this study was to study the microbial diversity of soils with long-standing hydrocarbon pollution in the S. Balgimbayev field (Kazakhstan), where the transformation of meadow coastal soils technogenic solonchak soils, as well as the assessment of the degradation potential of microorganisms perspective for the use for bioremediation. In the present work autochthonous microorganisms of the surface horizon of soils were investigated. In samples with a low degree of pollution the number of microorganisms, was comparable to the number in the uncontaminated soil and was 103 - 104 CFU/g. and one and two orders of magnitude lower in samples with high oil content. A collection of microorganisms was created using different culture media, which made it possible to isolate isolates that play a key role in different successional stages of biodegradation of petroleum hydrocarbons. The collection included the main bacterial filiiments, Protobacteria, Firmicutes, Actinobacteria and Bacteroidetes. Mycelial fungi andyeast-like fungwere assigned to the Ascomycota division. Studies showed that the percentage of isolates capable of growth in hydrocarbons varied. More than 50 % of the isolates grew on crude oil, a low percentage of less than 10 % of the isolates grew on an anthracene, phenanthrene and naphthalene, more than 20 % of the isolates belonging to different genera Pseudomonas, Bacillus, Rhodococcus, Achromobacter, Gordonia, Microbacterium, and Trichosporon, characterized the growth on two or three different hydrocarbons. The ability to grow using all hydrocarbons, associated with the synthesis of biosurfactants, was detected only in a few isolates.

Keywords: oil, soil, number of bioremediation, biodegradation, microorganisms, hydrocarbons – oxidizing microorganisms

Procedia PDF Downloads 38
12711 Analyzing the Results of Buildings Energy Audit by Using Grey Set Theory

Authors: Tooraj Karimi, Mohammadreza Sadeghi Moghadam

Abstract:

Grey set theory has the advantage of using fewer data to analyze many factors, and it is therefore more appropriate for system study rather than traditional statistical regression which require massive data, normal distribution in the data and few variant factors. So, in this paper grey clustering and entropy of coefficient vector of grey evaluations are used to analyze energy consumption in buildings of the Oil Ministry in Tehran. In fact, this article intends to analyze the results of energy audit reports and defines most favorable characteristics of system, which is energy consumption of buildings, and most favorable factors affecting these characteristics in order to modify and improve them. According to the results of the model, ‘the real Building Load Coefficient’ has been selected as the most important system characteristic and ‘uncontrolled area of the building’ has been diagnosed as the most favorable factor which has the greatest effect on energy consumption of building. Grey clustering in this study has been used for two purposes: First, all the variables of building relate to energy audit cluster in two main groups of indicators and the number of variables is reduced. Second, grey clustering with variable weights has been used to classify all buildings in three categories named ‘no standard deviation’, ‘low standard deviation’ and ‘non- standard’. Entropy of coefficient vector of Grey evaluations is calculated to investigate greyness of results. It shows that among the 38 buildings surveyed in terms of energy consumption, 3 cases are in standard group, 24 cases are in ‘low standard deviation’ group and 11 buildings are completely non-standard. In addition, clustering greyness of 13 buildings is less than 0.5 and average uncertainly of clustering results is 66%.

Keywords: energy audit, grey set theory, grey incidence matrixes, grey clustering, Iran oil ministry

Procedia PDF Downloads 363
12710 Determination of Mechanical Properties of Tomato Fruits: Experimental and Finite Element Analysis

Authors: Mallikarjunachari G., Venkata Ravi M.

Abstract:

The objective of this research work is to evaluate the mechanical properties such as elastic modulus and critical rupture load of tomato fruits. Determination of mechanical properties of tomato fruits is essential in various material handling applications, especially as related to robot harvesting, packaging, and transportation. However, extracting meaningful mechanical properties of tomato fruits are extremely challenging due to its layered structure, i.e., the combination of exocarp, mesocarp, and locular gel tissues. Apart from this layered structure, other physical parameters such as diameter, sphericity, locule number, and, the surface to volume ratio also influence the mechanical properties. In this research work, tomato fruits are cultivated in two different ways, namely organic and inorganic farming. Static compression tests are performed to extract the mechanical properties of tomato fruits. Finite element simulations are done to complement the experimental results. It is observed that the effective modulus decreases as the compression depth increase from 0.5 mm to 10 mm and also a critical load of fracture decreases as the locule number increases from 3 to 5. Significant differences in mechanical properties are observed between organically and inorganically cultivated tomato fruits. The current study significantly helps in the design of material handling systems to avoid damage of tomato fruits.

Keywords: elastic modulus, critical load of fracture, locule number, finite element analysis

Procedia PDF Downloads 106
12709 Reinforcement Effect on Dynamic Properties of Saturated Sand

Authors: R. Ziaie Moayed, M. Alibolandi

Abstract:

Dynamic behavior of soil are evaluated relative to a number of factors including: strain level, density, number of cycles, material type, fine content, geosynthetic inclusion, saturation, and effective stress. This paper investigate the dynamic behavior of saturated reinforced sand under cyclic stress condition. The cyclic triaxial tests are conducted on remolded specimens under various CSR which reinforced by different arrangement of non-woven geotextile. Aforementioned tests simulate field reinforced saturated deposits during earthquake or other cyclic loadings. This analysis revealed that the geotextile arrangement played dominant role on dynamic soil behavior and as geotextile close to top of specimen, the liquefaction resistance increased.

Keywords: dynamic behavior, reinforced sand, triaxial test, non-woven geotextile

Procedia PDF Downloads 220
12708 Classical and Bayesian Inference of the Generalized Log-Logistic Distribution with Applications to Survival Data

Authors: Abdisalam Hassan Muse, Samuel Mwalili, Oscar Ngesa

Abstract:

A generalized log-logistic distribution with variable shapes of the hazard rate was introduced and studied, extending the log-logistic distribution by adding an extra parameter to the classical distribution, leading to greater flexibility in analysing and modeling various data types. The proposed distribution has a large number of well-known lifetime special sub-models such as; Weibull, log-logistic, exponential, and Burr XII distributions. Its basic mathematical and statistical properties were derived. The method of maximum likelihood was adopted for estimating the unknown parameters of the proposed distribution, and a Monte Carlo simulation study is carried out to assess the behavior of the estimators. The importance of this distribution is that its tendency to model both monotone (increasing and decreasing) and non-monotone (unimodal and bathtub shape) or reversed “bathtub” shape hazard rate functions which are quite common in survival and reliability data analysis. Furthermore, the flexibility and usefulness of the proposed distribution are illustrated in a real-life data set and compared to its sub-models; Weibull, log-logistic, and BurrXII distributions and other parametric survival distributions with 3-parmaeters; like the exponentiated Weibull distribution, the 3-parameter lognormal distribution, the 3- parameter gamma distribution, the 3-parameter Weibull distribution, and the 3-parameter log-logistic (also known as shifted log-logistic) distribution. The proposed distribution provided a better fit than all of the competitive distributions based on the goodness-of-fit tests, the log-likelihood, and information criterion values. Finally, Bayesian analysis and performance of Gibbs sampling for the data set are also carried out.

Keywords: hazard rate function, log-logistic distribution, maximum likelihood estimation, generalized log-logistic distribution, survival data, Monte Carlo simulation

Procedia PDF Downloads 185