Search results for: actual exam time usage
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 20584

Search results for: actual exam time usage

18874 Analysing the Benefit of Real-Time Digital Translation for ESL Learners in a Post-secondary Canadian Classroom

Authors: Jordan Shuler

Abstract:

The goal of this study is to determine whether real-time language translation benefits ESL learners by contributing to overall equity in the classroom. Equity will be measured quantitatively through assessment performance and qualitatively through student survey. Two separate sections of students studying the same course will receive identical curriculum: one group, the control, will be taught in English and the other group in English with real-time translation into the students' first languages. The professor will use Microsoft Translator during lectures, in-class discussions, and Q&A time. The college is committed to finding new ways of teaching and learning, as outlined in Strategy 2022. If this research finds a positive relationship between language translation and student academic success, the technology will surely be encouraged for adoption by all George Brown College faculty. With greater acceptance, this technology could influence equity and pedagogy in the larger educational community.

Keywords: ESL learners, equity, innovative teaching strategies, language translation

Procedia PDF Downloads 115
18873 Analysis of Eco-Efficiency and the Determinants of Family Agriculture in Southeast Spain

Authors: Emilio Galdeano-Gómez, Ángeles Godoy-Durán, Juan C. Pérez-Mesa, Laura Piedra-Muñoz

Abstract:

Eco-efficiency is receiving ever-increasing interest as an indicator of sustainability, as it links environmental and economic performances in productive activities. In agriculture, these indicators and their determinants prove relevant due to the close relationships in this activity between the use of natural resources, which is generally limited, and the provision of basic goods to society. In this context, various analyses have focused on eco-efficiency by considering individual family farms as the basic production unit. However, not only must the measure of efficiency be taken into account, but also the existence of a series of factors which constitute socio-economic, political-institutional, and environmental determinants. Said factors have been studied to a lesser extent in the literature. The present work analyzes eco-efficiency at a micro level, focusing on small-scale family farms as the main decision-making units in horticulture in southeast Spain, a sector which represents about 30% of the fresh vegetables produced in the country and about 20% of those consumed in Europe. The objectives of this study are a) to obtain a series of eco-efficiency indicators by estimating several pressure ratios and economic value added in farming, b) to analyze the influence of specific social, economic and environmental variables on the aforementioned eco-efficiency indicators. The present work applies the method of Data Envelopment Analysis (DEA), which calculates different combinations of environmental pressures (water usage, phytosanitary contamination, waste management, etc.) and aggregate economic value. In a second stage, an analysis is conducted on the influence of the socio-economic and environmental characteristics of family farms on the eco-efficiency indicators, as endogeneous variables, through the use of truncated regression and bootstrapping techniques, following Simar-Wilson methodology. The results reveal considerable inefficiency in aspects such as waste management, while there is relatively little inefficiency in water usage and nitrogen balance. On the other hand, characteristics, such as product specialization, the adoption of quality certifications and belonging to a cooperative do have a positive impact on eco-efficiency. These results are deemed to be of interest to agri-food systems structured on small-scale producers, and they may prove useful to policy-makers as regards managing public environmental programs in agriculture.

Keywords: data envelopment analysis, eco-efficiency, family farms, horticulture, socioeconomic features

Procedia PDF Downloads 187
18872 Correlation between Sprint Performance and Vertical Jump Height in Elite Female Football Players

Authors: Svetlana Missina, Anatoliy Shipilov, Alexandr Vavaev

Abstract:

The purpose of the present study was to investigate the relationship between sprint and vertical jump performance in elite female football players. Twenty four professional female football players (age, 18.6±3.1 years; height, 168.3±6.3 cm, body mass 61.6±7.4 kg; mean±SD) were tested for 30-m sprint time, 10-m sprint time and vertical countermovement (CMJ) and squat (SJ) jumps height. Participants performed three countermovement jumps and three squat jumps for maximal height on a force platform. Mean values of three trials were used in statistical analysis. The displacement of center of mass (COM) during flight phase (e.g. jump height) was calculated using the vertical velocity of the COM at the moment of take-off. 30-m and 10-m sprint time were measured using OptoGait optical system. The best of three trials were used for analysis. A significant negative correlation was found between 30-m sprint time and CMJ, SJ height (r = -0.85, r = -0.79 respectively), between 10-m sprint time and CMJ, SJ height (r = -0.73, r = -0.8 respectively), and step frequency was significantly related to CMJ peak power (r = -0.57). Our study indicates that there is strong correlation between sprint and jump performance in elite female football players, thus vertical jump test can be considered as a good sprint and agility predictor in female football.

Keywords: agility, female football players, sprint performance, vertical jump height

Procedia PDF Downloads 462
18871 A Hybrid Genetic Algorithm for Assembly Line Balancing In Automotive Sector

Authors: Qazi Salman Khalid, Muhammad Khalid, Shahid Maqsood

Abstract:

This paper presents a solution for optimizing the cycle time in an assembly line with human-robot collaboration and diverse operators. A genetic algorithm with tailored parameters is used to address the assembly line balancing problem in the automobile sector. A mathematical model is developed, depicting the problem. Currently, the firm runs on the largest candidate rule; however, it causes a lag in orders, which ultimately gets penalized. The results of the study show that the proposed GA is effective in providing efficient solutions and that the cycle time has significantly impacted productivity.

Keywords: line balancing, cycle time, genetic algorithm, productivity

Procedia PDF Downloads 129
18870 Evaluation of Quick Covering Machine for Grain Drying Pavement

Authors: Fatima S. Rodriguez, Victorino T. Taylan, Manolito C. Bulaong, Helen F. Gavino, Vitaliana U. Malamug

Abstract:

In sundrying the quality of the grains are greatly reduced when paddy grains were caught by the rain unsacked and unstored resulting to reduced profit. The objectives of this study were to design and fabricate a quick covering machine for grain drying pavement; to test and evaluate the operating characteristics of the machine according to its deployment speed, recovery speed, deployment time, recovery time, power consumption, aesthetics of laminated sack; and to conduct partial budget and cost curve analysis. The machine was able to cover the grains in a 12.8 m x 22.5 m grain drying pavement at an average time of 17.13 s. It consumed 0.53 W-hr for the deployment and recovery of the cover. The machine entailed an investment cost of $1,344.40 and an annual cost charge of $647.32. Moreover, the savings per year using the quick covering machine was $101.83.

Keywords: quick covering machine, grain drying pavement, laminated polypropylene, recovery time

Procedia PDF Downloads 313
18869 Automatic Content Curation of Visual Heritage

Authors: Delphine Ribes Lemay, Valentine Bernasconi, André Andrade, Lara DéFayes, Mathieu Salzmann, FréDéRic Kaplan, Nicolas Henchoz

Abstract:

Digitization and preservation of large heritage induce high maintenance costs to keep up with the technical standards and ensure sustainable access. Creating impactful usage is instrumental to justify the resources for long-term preservation. The Museum für Gestaltung of Zurich holds one of the biggest poster collections of the world from which 52’000 were digitised. In the process of building a digital installation to valorize the collection, one objective was to develop an algorithm capable of predicting the next poster to show according to the ones already displayed. The work presented here describes the steps to build an algorithm able to automatically create sequences of posters reflecting associations performed by curator and professional designers. The exposed challenge finds similarities with the domain of song playlist algorithms. Recently, artificial intelligence techniques and more specifically, deep-learning algorithms have been used to facilitate their generations. Promising results were found thanks to Recurrent Neural Networks (RNN) trained on manually generated playlist and paired with clusters of extracted features from songs. We used the same principles to create the proposed algorithm but applied to a challenging medium, posters. First, a convolutional autoencoder was trained to extract features of the posters. The 52’000 digital posters were used as a training set. Poster features were then clustered. Next, an RNN learned to predict the next cluster according to the previous ones. RNN training set was composed of poster sequences extracted from a collection of books from the Gestaltung Museum of Zurich dedicated to displaying posters. Finally, within the predicted cluster, the poster with the best proximity compared to the previous poster is selected. The mean square distance between features of posters was used to compute the proximity. To validate the predictive model, we compared sequences of 15 posters produced by our model to randomly and manually generated sequences. Manual sequences were created by a professional graphic designer. We asked 21 participants working as professional graphic designers to sort the sequences from the one with the strongest graphic line to the one with the weakest and to motivate their answer with a short description. The sequences produced by the designer were ranked first 60%, second 25% and third 15% of the time. The sequences produced by our predictive model were ranked first 25%, second 45% and third 30% of the time. The sequences produced randomly were ranked first 15%, second 29%, and third 55% of the time. Compared to designer sequences, and as reported by participants, model and random sequences lacked thematic continuity. According to the results, the proposed model is able to generate better poster sequencing compared to random sampling. Eventually, our algorithm is sometimes able to outperform a professional designer. As a next step, the proposed algorithm should include a possibility to create sequences according to a selected theme. To conclude, this work shows the potentiality of artificial intelligence techniques to learn from existing content and provide a tool to curate large sets of data, with a permanent renewal of the presented content.

Keywords: Artificial Intelligence, Digital Humanities, serendipity, design research

Procedia PDF Downloads 176
18868 The Research of Effectiveness of Animal Protection Act Implementation Reducing Animal Abuse

Authors: Yu Ling Chang

Abstract:

Since the United Nations announced Universal Declaration of Human Rights in 1948, people are paying more and more attention to the value of lives. On the other hand, life education is being vigorously pushed in different countries. Unfortunately, the results have been only moderately successful by reason that the concept is not implemented in everyone’s daily life. Even worse, animal abuse and killing events keep happening. This research is focused on generalizing a conclusion from different countries’ Animal Protection Act and actual execution by case studies, in order to make an approach of whether the number of animal abuse is directly influenced by different laws and regimes or not. It concludes the central notion and spirit of Animal Protection Act in German, Japan, and Taiwan. Providing the reference of specific schemes and analysis based on Taiwanese social culture.

Keywords: animal abuse, Animal Management Act, Animal Protection Act, social culture

Procedia PDF Downloads 204
18867 Virtual Metering and Prediction of Heating, Ventilation, and Air Conditioning Systems Energy Consumption by Using Artificial Intelligence

Authors: Pooria Norouzi, Nicholas Tsang, Adam van der Goes, Joseph Yu, Douglas Zheng, Sirine Maleej

Abstract:

In this study, virtual meters will be designed and used for energy balance measurements of an air handling unit (AHU). The method aims to replace traditional physical sensors in heating, ventilation, and air conditioning (HVAC) systems with simulated virtual meters. Due to the inability to manage and monitor these systems, many HVAC systems have a high level of inefficiency and energy wastage. Virtual meters are implemented and applied in an actual HVAC system, and the result confirms the practicality of mathematical sensors for alternative energy measurement. While most residential buildings and offices are commonly not equipped with advanced sensors, adding, exploiting, and monitoring sensors and measurement devices in the existing systems can cost thousands of dollars. The first purpose of this study is to provide an energy consumption rate based on available sensors and without any physical energy meters. It proves the performance of virtual meters in HVAC systems as reliable measurement devices. To demonstrate this concept, mathematical models are created for AHU-07, located in building NE01 of the British Columbia Institute of Technology (BCIT) Burnaby campus. The models will be created and integrated with the system’s historical data and physical spot measurements. The actual measurements will be investigated to prove the models' accuracy. Based on preliminary analysis, the resulting mathematical models are successful in plotting energy consumption patterns, and it is concluded confidently that the results of the virtual meter will be close to the results that physical meters could achieve. In the second part of this study, the use of virtual meters is further assisted by artificial intelligence (AI) in the HVAC systems of building to improve energy management and efficiency. By the data mining approach, virtual meters’ data is recorded as historical data, and HVAC system energy consumption prediction is also implemented in order to harness great energy savings and manage the demand and supply chain effectively. Energy prediction can lead to energy-saving strategies and considerations that can open a window in predictive control in order to reach lower energy consumption. To solve these challenges, the energy prediction could optimize the HVAC system and automates energy consumption to capture savings. This study also investigates AI solutions possibility for autonomous HVAC efficiency that will allow quick and efficient response to energy consumption and cost spikes in the energy market.

Keywords: virtual meters, HVAC, artificial intelligence, energy consumption prediction

Procedia PDF Downloads 100
18866 Interpretations of Disaster: A Comparative Study on Disaster Film Cycles

Authors: Chi-Ying Yu

Abstract:

In real life, the occurrence of disasters is always dreadful and heartbreaking, yet paradoxically, disaster film is a genre that has been popular at periodic intervals in motion picture history. This study attempts to compare the disaster film cycles of the 1970s, 1990s, and the early 21st century. Two research questions are addressed: First, how this genre has responded to the existing conditions of society in different periods in terms of the disaster proposition? Second, how this genre reflects a certain eternal substance of the human mind in light of its lasting appeal? Through cinematic textual analysis and literature review, this study finds that the emergence of disaster films in the 1970s reflected the turmoil in international relations and domestic politics situation in contemporary American society, and cinema screens showed such disaster stories as shipwrecks, air accidents, and skyscraper blazes due to human negligence. The 1990s saw the fervor of millennial apocalypse legends, and the awakening of environmental consciousness, which, together with the rapid advances in digital technology, once again gave rise to a frenzy of disaster films, with natural disasters and threats from aliens as the major themes of disasters. Since the beginning of the 21st century, the 911 Incident and natural disasters around the world have generated a consciousness of imminent crisis. Cinematic images simulated actual disasters, while aesthetic techniques focused on creating a kind of ‘empathetic’ experience in their exploration of the essence of the disaster experience. At the same time, post-apocalypse films that focus on post-disaster reconstruction have become an even more popular theme. Taking the approach of Jungian/post-Jungian film study, this study also reviews and interprets the commonly exhibited subliminal feelings in the disaster films of the three different periods. The imagination of disaster seems to serve as an underlying state of the human mind.

Keywords: disaster film, Jungian/post-Jungian film studies, stimulation, sublime

Procedia PDF Downloads 254
18865 The Effective Use of the Network in the Distributed Storage

Authors: Mamouni Mohammed Dhiya Eddine

Abstract:

This work aims at studying the exploitation of high-speed networks of clusters for distributed storage. Parallel applications running on clusters require both high-performance communications between nodes and efficient access to the storage system. Many studies on network technologies led to the design of dedicated architectures for clusters with very fast communications between computing nodes. Efficient distributed storage in clusters has been essentially developed by adding parallelization mechanisms so that the server(s) may sustain an increased workload. In this work, we propose to improve the performance of distributed storage systems in clusters by efficiently using the underlying high-performance network to access distant storage systems. The main question we are addressing is: do high-speed networks of clusters fit the requirements of a transparent, efficient and high-performance access to remote storage? We show that storage requirements are very different from those of parallel computation. High-speed networks of clusters were designed to optimize communications between different nodes of a parallel application. We study their utilization in a very different context, storage in clusters, where client-server models are generally used to access remote storage (for instance NFS, PVFS or LUSTRE). Our experimental study based on the usage of the GM programming interface of MYRINET high-speed networks for distributed storage raised several interesting problems. Firstly, the specific memory utilization in the storage access system layers does not easily fit the traditional memory model of high-speed networks. Secondly, client-server models that are used for distributed storage have specific requirements on message control and event processing, which are not handled by existing interfaces. We propose different solutions to solve communication control problems at the filesystem level. We show that a modification of the network programming interface is required. Data transfer issues need an adaptation of the operating system. We detail several propositions for network programming interfaces which make their utilization easier in the context of distributed storage. The integration of a flexible processing of data transfer in the new programming interface MYRINET/MX is finally presented. Performance evaluations show that its usage in the context of both storage and other types of applications is easy and efficient.

Keywords: distributed storage, remote file access, cluster, high-speed network, MYRINET, zero-copy, memory registration, communication control, event notification, application programming interface

Procedia PDF Downloads 213
18864 Measuring Science and Technology Innovation Capacity in Developing Countries: From a National Innovation System

Authors: Haeng A. Seo, Changseok Oh, Seung Jun Yoo

Abstract:

This study attempts to examine the disparities in S&T innovation capacity from 14 developing countries to discuss how to support specific features in national innovation systems. It includes East-Asian, Middle-Asian, Central American and African countries. Here, we particularly focus on five dimensions- resources, activities, network, environment and performance- with 37 indicators. They were derived as structuring components of the relevant diagnostic model, which encompasses the whole process of S&T innovation from the input of resources to the output of economically valuable results. For many developing nations, economic industries remain weaker than actual S&T capabilities, and relevant regulatory authorities may not exist. This paper will be helpful to provide basic evidence and to set directions for better national S&T Innovation capacities and toward national competitiveness.

Keywords: developing countries, measurement, NIS, S&T innovation capacity

Procedia PDF Downloads 278
18863 Quantum Statistical Machine Learning and Quantum Time Series

Authors: Omar Alzeley, Sergey Utev

Abstract:

Minimizing a constrained multivariate function is the fundamental of Machine learning, and these algorithms are at the core of data mining and data visualization techniques. The decision function that maps input points to output points is based on the result of optimization. This optimization is the central of learning theory. One approach to complex systems where the dynamics of the system is inferred by a statistical analysis of the fluctuations in time of some associated observable is time series analysis. The purpose of this paper is a mathematical transition from the autoregressive model of classical time series to the matrix formalization of quantum theory. Firstly, we have proposed a quantum time series model (QTS). Although Hamiltonian technique becomes an established tool to detect a deterministic chaos, other approaches emerge. The quantum probabilistic technique is used to motivate the construction of our QTS model. The QTS model resembles the quantum dynamic model which was applied to financial data. Secondly, various statistical methods, including machine learning algorithms such as the Kalman filter algorithm, are applied to estimate and analyses the unknown parameters of the model. Finally, simulation techniques such as Markov chain Monte Carlo have been used to support our investigations. The proposed model has been examined by using real and simulated data. We establish the relation between quantum statistical machine and quantum time series via random matrix theory. It is interesting to note that the primary focus of the application of QTS in the field of quantum chaos was to find a model that explain chaotic behaviour. Maybe this model will reveal another insight into quantum chaos.

Keywords: machine learning, simulation techniques, quantum probability, tensor product, time series

Procedia PDF Downloads 465
18862 H∞ Sampled-Data Control for Linear Systems Time-Varying Delays: Application to Power System

Authors: Chang-Ho Lee, Seung-Hoon Lee, Myeong-Jin Park, Oh-Min Kwon

Abstract:

This paper investigates improved stability criteria for sampled-data control of linear systems with disturbances and time-varying delays. Based on Lyapunov-Krasovskii stability theory, delay-dependent conditions sufficient to ensure H∞ stability for the system are derived in the form of linear matrix inequalities(LMI). The effectiveness of the proposed method will be shown in numerical examples.

Keywords: sampled-data control system, Lyapunov-Krasovskii functional, time delay-dependent, LMI, H∞ control

Procedia PDF Downloads 317
18861 Big Data Analysis with RHadoop

Authors: Ji Eun Shin, Byung Ho Jung, Dong Hoon Lim

Abstract:

It is almost impossible to store or analyze big data increasing exponentially with traditional technologies. Hadoop is a new technology to make that possible. R programming language is by far the most popular statistical tool for big data analysis based on distributed processing with Hadoop technology. With RHadoop that integrates R and Hadoop environment, we implemented parallel multiple regression analysis with different sizes of actual data. Experimental results showed our RHadoop system was much faster as the number of data nodes increases. We also compared the performance of our RHadoop with lm function and big lm packages available on big memory. The results showed that our RHadoop was faster than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases.

Keywords: big data, Hadoop, parallel regression analysis, R, RHadoop

Procedia PDF Downloads 429
18860 The Data Quality Model for the IoT based Real-time Water Quality Monitoring Sensors

Authors: Rabbia Idrees, Ananda Maiti, Saurabh Garg, Muhammad Bilal Amin

Abstract:

IoT devices are the basic building blocks of IoT network that generate enormous volume of real-time and high-speed data to help organizations and companies to take intelligent decisions. To integrate this enormous data from multisource and transfer it to the appropriate client is the fundamental of IoT development. The handling of this huge quantity of devices along with the huge volume of data is very challenging. The IoT devices are battery-powered and resource-constrained and to provide energy efficient communication, these IoT devices go sleep or online/wakeup periodically and a-periodically depending on the traffic loads to reduce energy consumption. Sometime these devices get disconnected due to device battery depletion. If the node is not available in the network, then the IoT network provides incomplete, missing, and inaccurate data. Moreover, many IoT applications, like vehicle tracking and patient tracking require the IoT devices to be mobile. Due to this mobility, If the distance of the device from the sink node become greater than required, the connection is lost. Due to this disconnection other devices join the network for replacing the broken-down and left devices. This make IoT devices dynamic in nature which brings uncertainty and unreliability in the IoT network and hence produce bad quality of data. Due to this dynamic nature of IoT devices we do not know the actual reason of abnormal data. If data are of poor-quality decisions are likely to be unsound. It is highly important to process data and estimate data quality before bringing it to use in IoT applications. In the past many researchers tried to estimate data quality and provided several Machine Learning (ML), stochastic and statistical methods to perform analysis on stored data in the data processing layer, without focusing the challenges and issues arises from the dynamic nature of IoT devices and how it is impacting data quality. A comprehensive review on determining the impact of dynamic nature of IoT devices on data quality is done in this research and presented a data quality model that can deal with this challenge and produce good quality of data. This research presents the data quality model for the sensors monitoring water quality. DBSCAN clustering and weather sensors are used in this research to make data quality model for the sensors monitoring water quality. An extensive study has been done in this research on finding the relationship between the data of weather sensors and sensors monitoring water quality of the lakes and beaches. The detailed theoretical analysis has been presented in this research mentioning correlation between independent data streams of the two sets of sensors. With the help of the analysis and DBSCAN, a data quality model is prepared. This model encompasses five dimensions of data quality: outliers’ detection and removal, completeness, patterns of missing values and checks the accuracy of the data with the help of cluster’s position. At the end, the statistical analysis has been done on the clusters formed as the result of DBSCAN, and consistency is evaluated through Coefficient of Variation (CoV).

Keywords: clustering, data quality, DBSCAN, and Internet of things (IoT)

Procedia PDF Downloads 130
18859 Application of Robotics to Assemble a Used Fuel Container in the Canadian Used Fuel Packing Plant

Authors: Dimitrie Marinceu

Abstract:

The newest Canadian Used Fuel Container (UFC)- (called also “Mark II”) modifies the design approach for its Assembly Robotic Cell (ARC) in the Canadian Used (Nuclear) Fuel Packing Plant (UFPP). Some of the robotic design solutions are presented in this paper. The design indicates that robots and manipulators are expected to be used in the Canadian UFPP. As normally, the UFPP design will incorporate redundancy of all equipment to allow expedient recovery from any postulated upset conditions. Overall, this paper suggests that robot usage will have a significant positive impact on nuclear safety, quality, productivity, and reliability.

Keywords: used fuel packing plant, robotic assembly cell, used fuel container, deep geological repository

Procedia PDF Downloads 284
18858 Relationship Between Pain Intensity at the Time of the Hamstring Muscle Injury and Hamstring Muscle Lesion Volume Measured by Magnetic Resonance Imaging

Authors: Grange Sylvain, Plancher Ronan, Reurink Guustav, Croisille Pierre, Edouard Pascal

Abstract:

The primary objective of this study was to analyze the potential correlation between the pain experienced at the time of a hamstring muscle injury and the volume of the lesion measured on MRI. The secondary objectives were to analyze a correlation between this pain and the lesion grade as well as the affected hamstring muscle. We performed a retrospective analysis of the data collected in a prospective, multicenter, non-interventional cohort study (HAMMER). Patients with suspected hamstring muscle injury had an MRI after the injury and at the same time were evaluated for their pain intensity experienced at the time of the injury with a Numerical Pain Rating Scale (NPRS) from 0 to 10. A total of 61 patients were included in the present analysis. MRIs were performed in an average of less than 8 days. There was a significant correlation between pain and the injury volume (r=0.287; p=0.025). There was no significant correlation between the pain and the lesion grade (p>0.05), nor between the pain and affected hamstring muscle (p>0.05). Pain at the time of injury appeared to be correlated with the volume of muscle affected. These results confirm the value of a clinical approach in the initial evaluation of hamstring injuries to better select patients eligible for further imaging.

Keywords: hamstring muscle injury, MRI, volume lesion, pain

Procedia PDF Downloads 94
18857 Statistical Comparison of Ensemble Based Storm Surge Forecasting Models

Authors: Amin Salighehdar, Ziwen Ye, Mingzhe Liu, Ionut Florescu, Alan F. Blumberg

Abstract:

Storm surge is an abnormal water level caused by a storm. Accurate prediction of a storm surge is a challenging problem. Researchers developed various ensemble modeling techniques to combine several individual forecasts to produce an overall presumably better forecast. There exist some simple ensemble modeling techniques in literature. For instance, Model Output Statistics (MOS), and running mean-bias removal are widely used techniques in storm surge prediction domain. However, these methods have some drawbacks. For instance, MOS is based on multiple linear regression and it needs a long period of training data. To overcome the shortcomings of these simple methods, researchers propose some advanced methods. For instance, ENSURF (Ensemble SURge Forecast) is a multi-model application for sea level forecast. This application creates a better forecast of sea level using a combination of several instances of the Bayesian Model Averaging (BMA). An ensemble dressing method is based on identifying best member forecast and using it for prediction. Our contribution in this paper can be summarized as follows. First, we investigate whether the ensemble models perform better than any single forecast. Therefore, we need to identify the single best forecast. We present a methodology based on a simple Bayesian selection method to select the best single forecast. Second, we present several new and simple ways to construct ensemble models. We use correlation and standard deviation as weights in combining different forecast models. Third, we use these ensembles and compare with several existing models in literature to forecast storm surge level. We then investigate whether developing a complex ensemble model is indeed needed. To achieve this goal, we use a simple average (one of the simplest and widely used ensemble model) as benchmark. Predicting the peak level of Surge during a storm as well as the precise time at which this peak level takes place is crucial, thus we develop a statistical platform to compare the performance of various ensemble methods. This statistical analysis is based on root mean square error of the ensemble forecast during the testing period and on the magnitude and timing of the forecasted peak surge compared to the actual time and peak. In this work, we analyze four hurricanes: hurricanes Irene and Lee in 2011, hurricane Sandy in 2012, and hurricane Joaquin in 2015. Since hurricane Irene developed at the end of August 2011 and hurricane Lee started just after Irene at the beginning of September 2011, in this study we consider them as a single contiguous hurricane event. The data set used for this study is generated by the New York Harbor Observing and Prediction System (NYHOPS). We find that even the simplest possible way of creating an ensemble produces results superior to any single forecast. We also show that the ensemble models we propose generally have better performance compared to the simple average ensemble technique.

Keywords: Bayesian learning, ensemble model, statistical analysis, storm surge prediction

Procedia PDF Downloads 305
18856 Analysing the Behaviour of Local Hurst Exponent and Lyapunov Exponent for Prediction of Market Crashes

Authors: Shreemoyee Sarkar, Vikhyat Chadha

Abstract:

In this paper, the local fractal properties and chaotic properties of financial time series are investigated by calculating two exponents, the Local Hurst Exponent: LHE and Lyapunov Exponent in a moving time window of a financial series.y. For the purpose of this paper, the Dow Jones Industrial Average (DIJA) and S&P 500, two of the major indices of United States have been considered. The behaviour of the above-mentioned exponents prior to some major crashes (1998 and 2008 crashes in S&P 500 and 2002 and 2008 crashes in DIJA) is discussed. Also, the optimal length of the window for obtaining the best possible results is decided. Based on the outcomes of the above, an attempt is made to predict the crashes and accuracy of such an algorithm is decided.

Keywords: local hurst exponent, lyapunov exponent, market crash prediction, time series chaos, time series local fractal properties

Procedia PDF Downloads 144
18855 Further Development of Offshore Floating Solar and Its Design Requirements

Authors: Madjid Karimirad

Abstract:

Floating solar was not very well-known in the renewable energy field a decade ago; however, there has been tremendous growth internationally with a Compound Annual Growth Rate (CAGR) of nearly 30% in recent years. To reach the goal of global net-zero emission by 2050, all renewable energy sources including solar should be used. Considering that 40% of the world’s population lives within 100 kilometres of the coasts, floating solar in coastal waters is an obvious energy solution. However, this requires more robust floating solar solutions. This paper tries to enlighten the fundamental requirements in the design of floating solar for offshore installations from the hydrodynamic and offshore engineering points of view. In this regard, a closer look at dynamic characteristics, stochastic behaviour and nonlinear phenomena appearing in this kind of structure is a major focus of the current article. Floating solar structures are alternative and very attractive green energy installations with (a) Less strain on land usage for densely populated areas; (b) Natural cooling effect with efficiency gain; and (c) Increased irradiance from the reflectivity of water. Also, floating solar in conjunction with the hydroelectric plants can optimise energy efficiency and improve system reliability. The co-locating of floating solar units with other types such as offshore wind, wave energy, tidal turbines as well as aquaculture (fish farming) can result in better ocean space usage and increase the synergies. Floating solar technology has seen considerable developments in installed capacities in the past decade. Development of design standards and codes of practice for floating solar technologies deployed on both inland water-bodies and offshore is required to ensure robust and reliable systems that do not have detrimental impacts on the hosting water body. Floating solar will account for 17% of all PV energy produced worldwide by 2030. To enhance the development, further research in this area is needed. This paper aims to discuss the main critical design aspects in light of the load and load effects that the floating solar platforms are subjected to. The key considerations in hydrodynamics, aerodynamics and simultaneous effects from the wind and wave load actions will be discussed. The link of dynamic nonlinear loading, limit states and design space considering the environmental conditions is set to enable a better understanding of the design requirements of fast-evolving floating solar technology.

Keywords: floating solar, offshore renewable energy, wind and wave loading, design space

Procedia PDF Downloads 69
18854 Building Information Modeling Implementation for Managing an Extra Large Governmental Building Renovation Project

Authors: Pornpote Nusen, Manop Kaewmoracharoen

Abstract:

In recent years, there was an observable shift in fully developed countries from constructing new buildings to modifying existing buildings. The issue was that although an effective instrument like BIM (Building Information Modeling) was well developed for constructing new buildings, it was not widely used to renovate old buildings. BIM was accepted as an effective means to overcome common managerial problems such as project delay, cost overrun, and poor quality of the project life cycle. It was recently introduced in Thailand and rarely used in a renovation project. Today, in Thailand, BIM is mostly used for creating aesthetic 3D models and quantity takeoff purposes, though it can be an effective tool to use as a project management tool in planning and scheduling. Now the governmental sector in Thailand begins to recognize the uses of using BIM to manage a construction project, but the knowledge about the BIM implementation to governmental construction projects is underdeveloped. Further studies need to be conducted to maximize its advantages for the governmental sector. An educational extra large governmental building of 17,000 square-meters was used in this research. It is currently under construction for a two-year renovation project. BIM models of the building for the exterior and interior areas were created for the whole five floors. Then 4D BIM with combination of 3D BIM plus time was created for planning and scheduling. Three focus groups had been done with executive committee, contractors, and officers of the building to discuss the possibility of usage and usefulness of BIM approach over the traditional process. Several aspects were discussed in the positive sides, especially several foreseen problems, such as the inadequate accessibility of ways, the altered ceiling levels, the impractical construction plan created through a traditional approach, and the lack of constructability information. However, for some parties, the cost of BIM implementation was a concern, though, this study believes, its uses outweigh the cost.

Keywords: building information modeling, extra large building, governmental building renovation, project management, renovation, 4D BIM

Procedia PDF Downloads 143
18853 The Catholic Aristotle: Metaphysics and the Transubstantiation of the Eucharist

Authors: Elizabeth Latham

Abstract:

Aristotle’s definition of substance from Metaphysics is relevant to the Catholic transubstantiation of the Eucharist, the idea that the actual substance of bread and wine is replaced by the substance of the body and blood of Christ. Assuming the physiological nature of the subjects do not change, the Aristotelian view on the nature of substance seems to be incompatible with this belief, since bread and wine have essential qualities different from those of flesh and blood. However, based on a theological view of the essence of the body and blood of Christ as salvation along with the essence of Eucharist, transubstantiation can fit within Aristotle’s brackets. This is one step further than theologians like Aquinas have gone in their similar discussions. Using his work as a method by which to understand the apparently impossible act of transforming food to flesh provides a logical angle on a question of faith.

Keywords: aristotle, catholicism, eucharist, metaphysics transubstantiation

Procedia PDF Downloads 150
18852 Cardiometabolic Risk Factors Responses to Supplemental High Intensity Exercise in Middle School Children

Authors: R. M. Chandler, A. J. Stringer

Abstract:

In adults, short bursts of high-intensity exercise (intensities between 80-95% of maximum heart rates) increase cardiovascular and metabolic function without the time investment of traditional aerobic training. Similar improvements in various health indices are also becoming increasingly evident in children in countries other than the United States. In the United States, physical education programs have become shorter in length and fewer in frequency. With this in the background, it is imperative that health and physical educators delivered well-organized and focused fitness programs that can be tolerated across many different somatotypes. Perhaps the least effective lag-time in a US physical education (PE) class is the first 10 minutes, a time during which children warm up. Replacing a traditional PE warmup with a 10 min high-intensity excise protocol is a time-efficient method to impact health, leaving as much time for other PE material such as skill development, motor behavior development as possible. This supplemented 10 min high-intensity exercise increases cardiovascular function as well as induces favorable body composition changes in as little as six weeks with further enhancement throughout a semester of activity. The supplemental high-intensity exercise did not detract from the PE lesson outcomes.

Keywords: cardiovascular fitness, high intensity interval training, high intensity exercise, pediatric

Procedia PDF Downloads 132
18851 Resource Allocation Modeling and Simulation in Border Security Application

Authors: Kai Jin, Hua Li, Qing Song

Abstract:

Homeland security and border safety is an issue for any country. This paper takes the border security of US as an example to discuss the usage and efficiency of simulation tools in the homeland security application. In this study, available resources and different illegal infiltration parameters are defined, including their individual behavior and objective, in order to develop a model that describes border patrol system. A simulation model is created in Arena. This simulation model is used to study the dynamic activities in the border security. Possible factors that may affect the effectiveness of the border patrol system are proposed. Individual and factorial analysis of these factors is conducted and some suggestions are made.

Keywords: resource optimization, simulation, modeling, border security

Procedia PDF Downloads 509
18850 The Artificial Intelligence Technologies Used in PhotoMath Application

Authors: Tala Toonsi, Marah Alagha, Lina Alnowaiser, Hala Rajab

Abstract:

This report is about the Photomath app, which is an AI application that uses image recognition technology, specifically optical character recognition (OCR) algorithms. The (OCR) algorithm translates the images into a mathematical equation, and the app automatically provides a step-by-step solution. The application supports decimals, basic arithmetic, fractions, linear equations, and multiple functions such as logarithms. Testing was conducted to examine the usage of this app, and results were collected by surveying ten participants. Later, the results were analyzed. This paper seeks to answer the question: To what level the artificial intelligence features are accurate and the speed of process in this app. It is hoped this study will inform about the efficiency of AI in Photomath to the users.

Keywords: photomath, image recognition, app, OCR, artificial intelligence, mathematical equations.

Procedia PDF Downloads 165
18849 Increased Reaction and Movement Times When Text Messaging during Simulated Driving

Authors: Adriana M. Duquette, Derek P. Bornath

Abstract:

Reaction Time (RT) and Movement Time (MT) are important components of everyday life that have an effect on the way in which we move about our environment. These measures become even more crucial when an event can be caused (or avoided) in a fraction of a second, such as the RT and MT required while driving. The purpose of this study was to develop a more simple method of testing RT and MT during simulated driving with or without text messaging, in a university-aged population (n = 170). In the control condition, a randomly-delayed red light stimulus flashed on a computer interface after the participant began pressing the ‘gas’ pedal on a foot switch mat. Simple RT was defined as the time between the presentation of the light stimulus and the initiation of lifting the foot from the switch mat ‘gas’ pedal; while MT was defined as the time after the initiation of lifting the foot, to the initiation of depressing the switch mat ‘brake’ pedal. In the texting condition, upon pressing the ‘gas’ pedal, a ‘text message’ appeared on the computer interface in a dialog box that the participant typed on their cell phone while waiting for the light stimulus to turn red. In both conditions, the sequence was repeated 10 times, and an average RT (seconds) and average MT (seconds) were recorded. Condition significantly (p = .000) impacted overall RTs, as the texting condition (0.47 s) took longer than the no-texting (control) condition (0.34 s). Longer MTs were also recorded during the texting condition (0.28 s) than in the control condition (0.23 s), p = .001. Overall increases in Response Time (RT + MT) of 189 ms during the texting condition would equate to an additional 4.2 meters (to react to the stimulus and begin braking) if the participant had been driving an automobile at 80 km per hour. In conclusion, increasing task complexity due to the dual-task demand of text messaging during simulated driving caused significant increases in RT (41%), MT (23%) and Response Time (34%), thus further strengthening the mounting evidence against text messaging while driving.

Keywords: simulated driving, text messaging, reaction time, movement time

Procedia PDF Downloads 518
18848 Videoconference Technology: An Attractive Vehicle for Challenging and Changing Tutors Practice in Open and Distance Learning Environment

Authors: Ramorola Mmankoko Ziphorah

Abstract:

Videoconference technology represents a recent experiment of technology integration into teaching and learning in South Africa. Increasingly, videoconference technology is commonly used as a substitute for the traditional face-to-face approaches to teaching and learning in helping tutors to reshape and change their teaching practices. Interestingly, though, some studies point out that videoconference technology is commonly used for knowledge dissemination by tutors and not so much for the actual teaching of course content in Open and Distance Learning context. Though videoconference technology has become one of the dominating technologies available among Open and Distance Learning institutions, it is not clear that it has been used as effectively to bridge the learning distance in time, geography, and economy. While tutors are prepared theoretically, in most tutor preparation programs, on the use of videoconference technology, there are still no practical guidelines on how they should go about integrating this technology into their course teaching. Therefore, there is an urgent need to focus on tutor development, specifically on their capacities and skills to use videoconference technology. The assumption is that if tutors become competent in the use of the videoconference technology for course teaching, then their use in Open and Distance Learning environment will become more commonplace. This is the imperative of the 4th Industrial Revolution (4IR) on education generally. Against the current vacuum in the practice of using videoconference technology for course teaching, the current study proposes a qualitative phenomenological approach to investigate the efficacy of videoconferencing as an approach to student learning. Using interviews and observation data from ten participants in Open and Distance Learning institution, the author discusses how dialogue and structure interacted to provide the participating tutors with a rich set of opportunities to deliver course content. The findings to this study highlight various challenges experienced by tutors when using videoconference technology. The study suggests tutor development programs on their capacity and skills and on how to integrate this technology with various teaching strategies in order to enhance student learning. The author argues that it is not merely the existence of the structure, namely the videoconference technology, that provides the opportunity for effective teaching, but that is the interactions, namely, the dialogue amongst tutors and learners that make videoconference technology an attractive vehicle for challenging and changing tutors practice.

Keywords: open distance learning, transactional distance, tutor, videoconference

Procedia PDF Downloads 126
18847 UML Model for Double-Loop Control Self-Adaptive Braking System

Authors: Heung Sun Yoon, Jong Tae Kim

Abstract:

In this paper, we present an activity diagram model for double-loop control self-adaptive braking system. Since activity diagram helps to improve visibility of self-adaption, we can easily find where improvement is needed on double-loop control. Double-loop control is adopted since the design conditions and actual conditions can be different. The system is reconfigured in runtime by using double-loop control. We simulated to verify and validate our model by using MATLAB. We compared single-loop control model with double-loop control model. Simulation results show that double-loop control provides more consistent brake power control than single-loop control.

Keywords: activity diagram, automotive, braking system, double-loop, self-adaptive, UML, vehicle

Procedia PDF Downloads 406
18846 Time Series Analysis the Case of China and USA Trade Examining during Covid-19 Trade Enormity of Abnormal Pricing with the Exchange rate

Authors: Md. Mahadi Hasan Sany, Mumenunnessa Keya, Sharun Khushbu, Sheikh Abujar

Abstract:

Since the beginning of China's economic reform, trade between the U.S. and China has grown rapidly, and has increased since China's accession to the World Trade Organization in 2001. The US imports more than it exports from China, reducing the trade war between China and the U.S. for the 2019 trade deficit, but in 2020, the opposite happens. In international and U.S. trade, Washington launched a full-scale trade war against China in March 2016, which occurred a catastrophic epidemic. The main goal of our study is to measure and predict trade relations between China and the U.S., before and after the arrival of the COVID epidemic. The ML model uses different data as input but has no time dimension that is present in the time series models and is only able to predict the future from previously observed data. The LSTM (a well-known Recurrent Neural Network) model is applied as the best time series model for trading forecasting. We have been able to create a sustainable forecasting system in trade between China and the US by closely monitoring a dataset published by the State Website NZ Tatauranga Aotearoa from January 1, 2015, to April 30, 2021. Throughout the survey, we provided a 180-day forecast that outlined what would happen to trade between China and the US during COVID-19. In addition, we have illustrated that the LSTM model provides outstanding outcome in time series data analysis rather than RFR and SVR (e.g., both ML models). The study looks at how the current Covid outbreak affects China-US trade. As a comparative study, RMSE transmission rate is calculated for LSTM, RFR and SVR. From our time series analysis, it can be said that the LSTM model has given very favorable thoughts in terms of China-US trade on the future export situation.

Keywords: RFR, China-U.S. trade war, SVR, LSTM, deep learning, Covid-19, export value, forecasting, time series analysis

Procedia PDF Downloads 191
18845 Exploring the ‘Many Worlds’ Interpretation in Both a Philosophical and Creative Literary Framework

Authors: Jane Larkin

Abstract:

Combining elements of philosophy, science, and creative writing, this investigation explores how a philosophically structured science-fiction novel can challenge the theory of linearity and singularity of time through the ‘many worlds’ theory. This concept is addressed through the creation of a research exegesis and accompanying creative artefact, designed to be read in conjunction with each other in an explorative, interwoven manner. Research undertaken into scientific concepts, such as the ‘many worlds’ interpretation of quantum mechanics and diverse philosophers and their ideologies on time, is embodied in an original science-fiction narrative titled, It Goes On. The five frames that make up the creative artefact are enhanced not only by five leading philosophers and their philosophies on time but by an appreciation of the research, which comes first in the paper. Research into traditional approaches to storytelling is creatively and innovatively inverted in several ways, thus challenging the singularity and linearity of time. Further nonconventional approaches to literary techniques include an abstract narrator, embodied by time, a concept, and a figure in the text, whose voice and vantage point in relation to death furthers the unreliability of the notion of time. These further challenge individuals’ understanding of complex scientific and philosophical views in a variety of ways. The science-fiction genre is essential when considering the speculative nature of It Goes On, which deals with parallel realities and is a fantastical exploration of human ingenuity in plausible futures. Therefore, this paper documents the research-led methodology used to create It Goes On, the application of the ‘many worlds’ theory within a framed narrative, and the many innovative techniques used to contribute new knowledge in a variety of fields.

Keywords: time, many-worlds theory, Heideggerian philosophy, framed narrative

Procedia PDF Downloads 79