Search results for: the scenario test GR
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10037

Search results for: the scenario test GR

9977 Cognitive Models of Future in Political Texts

Authors: Solopova Olga

Abstract:

The present paper briefly recalls theoretical preconditions for investigating cognitive-discursive models of future in political discourse. The author reviews theories and methods used for strengthening a future focus in this discourse working out two main tools – a model of future and a metaphorical scenario. The paper examines the implications of metaphorical analogies for modeling future in mass media. It argues that metaphor is not merely a rhetorical ornament in the political discourse of media regulation but a conceptual model that legislates and regulates our understanding of future.

Keywords: cognitive approach, future research, political discourse, model, scenario, metaphor

Procedia PDF Downloads 360
9976 Formulation and Test of a Model to explain the Complexity of Road Accident Events in South Africa

Authors: Dimakatso Machetele, Kowiyou Yessoufou

Abstract:

Whilst several studies indicated that road accident events might be more complex than thought, we have a limited scientific understanding of this complexity in South Africa. The present project proposes and tests a more comprehensive metamodel that integrates multiple causality relationships among variables previously linked to road accidents. This was done by fitting a structural equation model (SEM) to the data collected from various sources. The study also fitted the GARCH Model (Generalized Auto-Regressive Conditional Heteroskedasticity) to predict the future of road accidents in the country. The analysis shows that the number of road accidents has been increasing since 1935. The road fatality rate follows a polynomial shape following the equation: y = -0.0114x²+1.2378x-2.2627 (R²=0.76) with y = death rate and x = year. This trend results in an average death rate of 23.14 deaths per 100,000 people. Furthermore, the analysis shows that the number of crashes could be significantly explained by the total number of vehicles (P < 0.001), number of registered vehicles (P < 0.001), number of unregistered vehicles (P = 0.003) and the population of the country (P < 0.001). As opposed to expectation, the number of driver licenses issued and total distance traveled by vehicles do not correlate significantly with the number of crashes (P > 0.05). Furthermore, the analysis reveals that the number of casualties could be linked significantly to the number of registered vehicles (P < 0.001) and total distance traveled by vehicles (P = 0.03). As for the number of fatal crashes, the analysis reveals that the total number of vehicles (P < 0.001), number of registered (P < 0.001) and unregistered vehicles (P < 0.001), the population of the country (P < 0.001) and the total distance traveled by vehicles (P < 0.001) correlate significantly with the number of fatal crashes. However, the number of casualties and again the number of driver licenses do not seem to determine the number of fatal crashes (P > 0.05). Finally, the number of crashes is predicted to be roughly constant overtime at 617,253 accidents for the next 10 years, with the worse scenario suggesting that this number may reach 1 896 667. The number of casualties was also predicted to be roughly constant at 93 531 overtime, although this number may reach 661 531 in the worst-case scenario. However, although the number of fatal crashes may decrease over time, it is forecasted to reach 11 241 fatal crashes within the next 10 years, with the worse scenario estimated at 19 034 within the same period. Finally, the number of fatalities is also predicted to be roughly constant at 14 739 but may also reach 172 784 in the worse scenario. Overall, the present study reveals the complexity of road accidents and allows us to propose several recommendations aimed to reduce the trend of road accidents, casualties, fatal crashes, and death in South Africa.

Keywords: road accidents, South Africa, statistical modelling, trends

Procedia PDF Downloads 130
9975 Case Study Approach Using Scenario Analysis to Analyze Unabsorbed Head Office Overheads

Authors: K. C. Iyer, T. Gupta, Y. M. Bindal

Abstract:

Head office overhead (HOOH) is an indirect cost and is recovered through individual project billings by the contractor. Delay in a project impacts the absorption of HOOH cost allocated to that particular project and thus diminishes the expected profit of the contractor. This unabsorbed HOOH cost is later claimed by contractors as damages. The subjective nature of the available formulae to compute unabsorbed HOOH is the difficulty that contractors and owners face and thus dispute it. The paper attempts to bring together the rationale of various HOOH formulae by gathering contractor’s HOOH cost data on all of its project, using case study approach and comparing variations in values of HOOH using scenario analysis. The case study approach uses project data collected from four construction projects of a contractor in India to calculate unabsorbed HOOH costs from various available formulae. Scenario analysis provides further variations in HOOH values after considering two independent situations mainly scope changes and new projects during the delay period. Interestingly, one of the findings in this study reveals that, in spite of HOOH getting absorbed by additional works available during the period of delay, a few formulae depict an increase in the value of unabsorbed HOOH, neglecting any absorption by the increase in scope. This indicates that these formulae are inappropriate for use in case of a change to the scope of work. Results of this study can help both parties in deciding on an appropriate formula more objectively, considering the events on a project causing the delay and contractor's position in respect of obtaining new projects.

Keywords: absorbed and unabsorbed overheads, head office overheads, scenario analysis, scope variation

Procedia PDF Downloads 137
9974 A Generalized Model for Performance Analysis of Airborne Radar in Clutter Scenario

Authors: Vinod Kumar Jaysaval, Prateek Agarwal

Abstract:

Performance prediction of airborne radar is a challenging and cumbersome task in clutter scenario for different types of targets. A generalized model requires to predict the performance of Radar for air targets as well as ground moving targets. In this paper, we propose a generalized model to bring out the performance of airborne radar for different Pulsed Repetition Frequency (PRF) as well as different type of targets. The model provides a platform to bring out different subsystem parameters for different applications and performance requirements under different types of clutter terrain.

Keywords: airborne radar, blind zone, clutter, probability of detection

Procedia PDF Downloads 437
9973 Effect of Coaching Related Incompetency to Stand Trial on Symptom Validity Test: Robustness, Sensitivity, and Specificity

Authors: Natthawut Arin

Abstract:

In forensic contexts, competency to stand trial assessments are the most common referrals. The defendants may attempt to endorse psychopathology symptoms and feign incompetent. Coaching, which can be teaching them test-taking strategies to avoid detection of psychopathological symptoms feigning. Recently, the Symptom Validity Testings (SVTs) were created to detect feigning. Moreover, the works of the literature showed that the effects of coaching on SVTs may be more robust to the effects of coaching. Thai Symptom Validity Test (SVT-Th) was designed as SVTs which demonstrated adequate psychometric properties and ability to classify between feigners and honest responders. Thus, the current study to examine the utility as the robustness of SVT-Th in the detection of feigned psychopathology. Participants consisted of 120 were recruited from undergraduate courses in psychology, randomly assigned to one of three groups. The SVT-Th was administered to those three scenario-experimental groups: (a) Uncoached group were asked to respond honestly (n=40), (b) Symptom-coached without warning group were asked to feign psychiatric symptoms to gain incompetency to stand trial (n=40), while (c) Test-coached with warning group were asked to feign psychiatric symptoms to avoid test detection but being incompetency to stand trial (n=40). Group differences were analyzed using one-way ANOVAs. The result revealed an uncoached group (M = 4.23, SD.= 5.20) had significantly lower SVT-Th mean scores than those both coached groups (M =185.00, SD.= 72.88 and M = 132.10, SD.= 54.06, respectively). Classification rates were calculated to determine the classification accuracy. Result indicated that SVT-Th had overall classification accuracy rates of 96.67% with acceptable of 95% sensitivity and 100% specificity rates. Overall, the results of the present study indicate that the SVT-Th yielded high adequate indices of accuracy and these findings suggest that the SVT-Th is robustness against coaching.

Keywords: incompetency to stand trial, coaching, robustness, classification accuracy

Procedia PDF Downloads 110
9972 Reaching Students Who “Don’t Like Writing” through Scenario Based Learning

Authors: Shahira Mahmoud Yacout

Abstract:

Writing is an essential skill in many vocational, academic environments, and notably workplaces, yet many students perceive writing as being something tiring and boring or maybe a “waste of time”. Studies in the field of foreign languages related this fact might be due to the lack of connection between what is learned in the university and what students come to encounter in real life situations”. Arabic learners felt they needed more language exposure to the context of their future professions. With this idea in mind, Scenario based learning (SBL) is reported to be an educational approach to motivate, engage and stimulate students’ interest and to achieve the desired writing learning outcomes. In addition, researchers suggested Scenario based learning (SBL)as an instructional approach that develops and enhances students skills through developing higher order thinking skills and active learning. It is a subset of problem-based learning and case-based learning. The approach focuses on authentic rhetorical framing reflecting writing tasks in real life situations. It works successfully when used to simulate real-world practices, providing context that reflects the types of situations professionals respond to in writing. It was claimed that using realistic scenarios customized to the course’s learning objectives as it bridged the gap for students between theory and application. Within this context, it is thought that scenario-based learning is an important approach to enhance the learners’ writing skills and to reflect meaningful learning within authentic contexts. As an Arabicforeign language instructor, it was noticed that students find difficulties in adapting writing styles to authentic writing contexts and addressing different audiences and purposes. This idea is supported by studieswho claimed that AFL students faced difficulties with transferring writing skills to situations outside of the classroom context. In addition, it was observed that some of the Arabic textbooks for teaching Arabic as a foreign language lacked topics that initiated higher order thinking skills and stimulated the learners to understand the setting, and created messages appropriate to different audiences, context, and purposes. The goals of this study are to 1)provide a rational for using scenario-based learning approach to improveAFL learners in writing skills, 2) demonstrate how to design/ implement a scenario-based learning technique aligned with the writing course objectives,3) demonstrate samples of scenario-based approach implemented in AFL writing class, and 4)emphasis the role of peer-review along with the instructor’s feedback, in the process of developing the writing skill. Finally, this presentation highlighted and emphasized the importance of using the scenario-based learning approach in writing as a means to mirror students’ real-life situations and engage them in planning, monitoring, and problem solving. This approach helped in making writing an enjoyable experience and clearly useful to students’ future professional careers.

Keywords: meaningful learning, real life contexts, scenario based learning, writing skill

Procedia PDF Downloads 71
9971 Investigating the Effects of Data Transformations on a Bi-Dimensional Chi-Square Test

Authors: Alexandru George Vaduva, Adriana Vlad, Bogdan Badea

Abstract:

In this research, we conduct a Monte Carlo analysis on a two-dimensional χ2 test, which is used to determine the minimum distance required for independent sampling in the context of chaotic signals. We investigate the impact of transforming initial data sets from any probability distribution to new signals with a uniform distribution using the Spearman rank correlation on the χ2 test. This transformation removes the randomness of the data pairs, and as a result, the observed distribution of χ2 test values differs from the expected distribution. We propose a solution to this problem and evaluate it using another chaotic signal.

Keywords: chaotic signals, logistic map, Pearson’s test, Chi Square test, bivariate distribution, statistical independence

Procedia PDF Downloads 56
9970 An Analysis of Non-Elliptic Curve Based Primality Tests

Authors: William Wong, Zakaria Alomari, Hon Ching Lai, Zhida Li

Abstract:

Modern-day information security depends on implementing Diffie-Hellman, which requires the generation of prime numbers. Because the number of primes is infinite, it is impractical to store prime numbers for use, and therefore, primality tests are indispensable in modern-day information security. A primality test is a test to determine whether a number is prime or composite. There are two types of primality tests, which are deterministic tests and probabilistic tests. Deterministic tests are adopting algorithms that provide a definite answer whether a given number is prime or composite. While in probabilistic tests, a probabilistic result would be provided, there is a degree of uncertainty. In this paper, we review three probabilistic tests: the Fermat Primality Test, the Miller-Rabin Test, and the Baillie-PSW Test, as well as one deterministic test, the Agrawal-Kayal-Saxena (AKS) Test. Furthermore, we do an analysis of these tests. All of the reviews discussed are not based on the Elliptic Curve. The analysis demonstrates that, in the majority of real-world scenarios, the Baillie- PSW test’s favorability stems from its typical operational complexity of O(log 3n) and its capacity to deliver accurate results for numbers below 2^64.

Keywords: primality tests, Fermat’s primality test, Miller-Rabin primality test, Baillie-PSW primality test, AKS primality test

Procedia PDF Downloads 52
9969 The Analysis of Differential Item and Test Functioning between Sexes by Studying on the Scholastic Aptitude Test 2013

Authors: Panwasn Mahalawalert

Abstract:

The purposes of this research were analyzed differential item functioning and differential test functioning of SWUSAT aptitude test classification by sex variable. The data used in this research is the secondary data from Srinakharinwirot University Scholastic Aptitude Test 2013 (SWUSAT). SWUSAT test consists of four subjects. There are verbal ability test, number ability test, reasoning ability test and spatial ability test. The data analysis was analyzed in 2 steps. The first step was analyzing descriptive statistics. In the second step were analyzed differential item functioning (DIF) and differential test functioning (DTF) by using the DIFAS program. The research results were as follows: The results of DIF and DTF analysis for all 10 tests in year 2013. Gender was the characteristic that found DIF all 10 tests. The percentage of item number that found DIF is between 6.67% - 60%. There are 5 tests that most of items favors female group and 2 tests that most of items favors male group. There are 3 tests that the number of items favors female group equal favors male group. For Differential test functioning (DTF), there are 8 tests that have small level.

Keywords: aptitude test, differential item functioning, differential test functioning, educational measurement

Procedia PDF Downloads 370
9968 An Investigation of Differential Item and Test Functioning of Scholastic Aptitude Test 2011 (SWUSAT 2011)

Authors: Ruangdech Sirikit

Abstract:

The purposes of this study were analyzed differential item functioning and differential test functioning of SWUSAT aptitude test classification by sex variable. The data used in this research is the secondary data from Srinakharinwirot University Scholastic Aptitude Test 2011 (SWUSAT 2011) SWUSAT test consists of four subjects. There are verbal ability test, number ability test, reasoning ability test and spatial ability test. The data analysis was carried out in 2 steps. The first step was analyzing descriptive statistics. In the second step were analyzed differential item functioning (DIF) and differential test functioning (DTF) by using the DIFAS program. The research results were as follows: The results of data analysis for all 10 tests in year 2011. Sex was the characteristic that found DIF all 10 tests. The percentage of item number that found DIF was between 10% - 46.67%. There are 4 tests that most of items favors female group. There are 3 tests that most of items favors male group and there are 3 tests that the number of items favors female group equal favors male group. For Differential test functioning (DTF), there are 8 tests that have small DIF effect variance.

Keywords: differential item functioning, differential test functioning, SWUSAT, aptitude test

Procedia PDF Downloads 580
9967 Wally Feelings Test: Validity and Reliability Study

Authors: Gökhan Kayili, Ramazan Ari

Abstract:

In this research, it is aimed to be adapted Wally Feelings Test to Turkish children and performed the reliability and validity analysis of the test. The sampling of the research was composed of three to five year-old 699 Turkish preschoolers who are attending official and private nursery school. The schools selected with simple random sampling method by considering different socio economic conditions and different central district in Konya. In order to determine reliability of Wally Feelings Test, internal consistency coefficients (KR-20), split-half reliability and test- retest reliability analysis have been performed. During validation process construct validity, content/scope validity and concurrent/criterion validity were used. When validity and reliability of the test examined, it is seen that Wally Feelings Test is a valid and reliable instrument to evaluate three to five year old Turkish children’s understanding feeling skills.

Keywords: reliability, validity, wally feelings test, social sciences

Procedia PDF Downloads 504
9966 Forecasting the Sea Level Change in Strait of Hormuz

Authors: Hamid Goharnejad, Amir Hossein Eghbali

Abstract:

Recent investigations have demonstrated the global sea level rise due to climate change impacts. In this study climate changes study the effects of increasing water level in the strait of Hormuz. The probable changes of sea level rise should be investigated to employ the adaption strategies. The climatic output data of a GCM (General Circulation Model) named CGCM3 under climate change scenario of A1b and A2 were used. Among different variables simulated by this model, those of maximum correlation with sea level changes in the study region and least redundancy among themselves were selected for sea level rise prediction by using stepwise regression. One models of Discrete Wavelet artificial Neural Network (DWNN) was developed to explore the relationship between climatic variables and sea level changes. In these models, wavelet was used to disaggregate the time series of input and output data into different components and then ANN was used to relate the disaggregated components of predictors and predictands to each other. The results showed in the Shahid Rajae Station for scenario A1B sea level rise is among 64 to 75 cm and for the A2 Scenario sea level rise is among 90 to 105 cm. Furthermore the result showed a significant increase of sea level at the study region under climate change impacts, which should be incorporated in coastal areas management.

Keywords: climate change scenarios, sea-level rise, strait of Hormuz, forecasting

Procedia PDF Downloads 241
9965 Examining the Relationship between Chi-Square Test Statistics and Skewness of Weibull Distribution: Simulation Study

Authors: Rafida M. Elobaid

Abstract:

Most of the literature on goodness-of-fit test try to provide a theoretical basis for studying empirical distribution functions. Such goodness-of-fit tests are Kolmogorove-Simirnov and Crumer-Von Mises Type tests. However, it is likely that most of literature has not focused in details on the relationship of the values of the test statistics and skewness or kurtosis. The aim of this study is to investigate the behavior of the values of the χ2 test statistic with the variation of the skewness of right skewed distribution. A simulation study is conducted to generate random numbers from Weibull distribution. For a fixed sample sizes, different levels of skewness are considered, and the corresponding values of the χ2 test statistic are calculated. Using different sample sizes, the results show an inverse relationship between the value of χ2 test and the level of skewness for Wiebull distribution, i.e the value of χ2 test statistic decreases as the value of skewness increases. The research results also show that with large values of skewness we are more confident that the data follows the assumed distribution. Nonparametric Kendall τ test is used to confirm these results.

Keywords: goodness-of-fit test, chi-square test, simulation, continuous right skewed distributions

Procedia PDF Downloads 377
9964 End To End Process to Automate Batch Application

Authors: Nagmani Lnu

Abstract:

Often, Quality Engineering refers to testing the applications that either have a User Interface (UI) or an Application Programming Interface (API). We often find mature test practices, standards, and automation regarding UI or API testing. However, another kind is present in almost all types of industries that deal with data in bulk and often get handled through something called a Batch Application. This is primarily an offline application companies develop to process large data sets that often deal with multiple business rules. The challenge gets more prominent when we try to automate batch testing. This paper describes the approaches taken to test a Batch application from a Financial Industry to test the payment settlement process (a critical use case in all kinds of FinTech companies), resulting in 100% test automation in Test Creation and Test execution. One can follow this approach for any other batch use cases to achieve a higher efficiency in their testing process.

Keywords: batch testing, batch test automation, batch test strategy, payments testing, payments settlement testing

Procedia PDF Downloads 21
9963 A Study on the Accelerated Life Cycle Test Method of the Motor for Home Appliances by Using Acceleration Factor

Authors: Youn-Sung Kim, Mi-Sung Kim, Jae-Kun Lee

Abstract:

This paper deals with the accelerated life cycle test method of the motor for home appliances that demand high reliability. Life Cycle of parts in home appliances also should be 10 years because life cycle of the home appliances such as washing machine, refrigerator, TV is at least 10 years. In case of washing machine, the life cycle test method of motor is advanced for 3000 cycle test (1cycle = 2hours). However, 3000 cycle test incurs loss for the time and cost. Objectives of this study are to reduce the life cycle test time and the number of test samples, which could be realized by using acceleration factor for the test time and reduction factor for the number of sample.

Keywords: accelerated life cycle test, motor reliability test, motor for washing machine, BLDC motor

Procedia PDF Downloads 604
9962 Modelling of Reactive Methodologies in Auto-Scaling Time-Sensitive Services With a MAPE-K Architecture

Authors: Óscar Muñoz Garrigós, José Manuel Bernabeu Aubán

Abstract:

Time-sensitive services are the base of the cloud services industry. Keeping low service saturation is essential for controlling response time. All auto-scalable services make use of reactive auto-scaling. However, reactive auto-scaling has few in-depth studies. This presentation shows a model for reactive auto-scaling methodologies with a MAPE-k architecture. Queuing theory can compute different properties of static services but lacks some parameters related to the transition between models. Our model uses queuing theory parameters to relate the transition between models. It associates MAPE-k related times, the sampling frequency, the cooldown period, the number of requests that an instance can handle per unit of time, the number of incoming requests at a time instant, and a function that describes the acceleration in the service's ability to handle more requests. This model is later used as a solution to horizontally auto-scale time-sensitive services composed of microservices, reevaluating the model’s parameters periodically to allocate resources. The solution requires limiting the acceleration of the growth in the number of incoming requests to keep a constrained response time. Business benefits determine such limits. The solution can add a dynamic number of instances and remains valid under different system sizes. The study includes performance recommendations to improve results according to the incoming load shape and business benefits. The exposed methodology is tested in a simulation. The simulator contains a load generator and a service composed of two microservices, where the frontend microservice depends on a backend microservice with a 1:1 request relation ratio. A common request takes 2.3 seconds to be computed by the service and is discarded if it takes more than 7 seconds. Both microservices contain a load balancer that assigns requests to the less loaded instance and preemptively discards requests if they are not finished in time to prevent resource saturation. When load decreases, instances with lower load are kept in the backlog where no more requests are assigned. If the load grows and an instance in the backlog is required, it returns to the running state, but if it finishes the computation of all requests and is no longer required, it is permanently deallocated. A few load patterns are required to represent the worst-case scenario for reactive systems: the following scenarios test response times, resource consumption and business costs. The first scenario is a burst-load scenario. All methodologies will discard requests if the rapidness of the burst is high enough. This scenario focuses on the number of discarded requests and the variance of the response time. The second scenario contains sudden load drops followed by bursts to observe how the methodology behaves when releasing resources that are lately required. The third scenario contains diverse growth accelerations in the number of incoming requests to observe how approaches that add a different number of instances can handle the load with less business cost. The exposed methodology is compared against a multiple threshold CPU methodology allocating/deallocating 10 or 20 instances, outperforming the competitor in all studied metrics.

Keywords: reactive auto-scaling, auto-scaling, microservices, cloud computing

Procedia PDF Downloads 66
9961 Finding the English Competency for Developing Public Health Village Volunteers at Ban Prasukchai in Chumpuang District, Nakhon Ratchasima Province in Thailand

Authors: Kittivate Boonyopakorn

Abstract:

The purposes of this study were to find the English competence of the public health volunteers and to develop the use of their English. The samples for the study were 41 public health village volunteers at Ban Prasukchai, in Thailand. The findings showed that the sum of all scores for the pre-test was 452, while the score for the post-test was 1,080. Therefore, the results of the experiment confirm that the post-test scores (1,080) significantly are higher than the pre-test (452). The mean score (N=41) for the pre-test was 11.02 while the mean score (N=41) for the post-test was 18.10. The standard deviation for the pre-test was 2.734; however, for the post-test it was 1.934. In addition to the experts-evaluated research tools, the results of the evaluation for the structured interviews (IOC) were 1 in value. The evaluation of congruence for the content with learning objectives (IOC) were 0.66 to 1.00 in value. The evaluation of congruence for the pre and post-test with learning objectives (IOC) are 1 in value.

Keywords: finding the English competency, developing public health, village volunteers

Procedia PDF Downloads 415
9960 Phytochemical and Antioxidant Activity Test of Water Fraction Extract of Sisik Naga (Drymoglossum piloselloides) Leaves

Authors: Afifah Nur Aini, Elsa Mega Suryani, Betty Lukiaty

Abstract:

Drymoglossum piloselloides or more commonly known as sisik naga fern is a member of Polipodiaceae Family that is abundant and widely distributed in nature. That being said, there hasn’t been many studies reporting about the benefits of this fern. The aim of this study was to find out the active compounds and antioxidant activity of water fraction extract of sisik naga leaves. The study will be able to optimize the use of this fern in the future. In this study, phytochemical test was done qualitatively by using Mayer, Dragendorff and Wagner reagent for alkaloid test; FeCl3 for phenolic test; Shinoda test for flavonoid; Liebermann-Burchard test for triterprnoid and Forth test for saponin. Antioxidant activity test was done by using 20D spectronic spectrophotometer to determine the percentage of DPPH free radical inhibition. The results showed that water fraction extract of sisik naga leaves contain phenolic and IC50 = 5.44 μg/ml. This means that sisik naga leaves can be used as an antioxidant.

Keywords: antioxidant activity test, dpph, phytochemical test, drymoglossum piloselloides

Procedia PDF Downloads 872
9959 Design of In-House Test Method for Assuring Packing Quality of Bottled Spirits

Authors: S. Ananthakrishnan, U. H. Acharya

Abstract:

Whether shopping in a retail location or via the internet, consumers expect to receive their products intact. When products arrive damaged or over-packaged, the result can be customer dissatisfaction and increased cost for retailers and manufacturers. The packaging performance depends on both the transport situation and the packaging design. During transportation, the packaged products are subjected to the variation in vibration levels from transport vehicles that vary in frequency and acceleration while moving to their destinations. Spirits manufactured by this Company were being transported to various parts of the country by road. There were instances of package breaking and customer complaints. The vibration experienced on a straight road at some speed may not be same as the vibration experienced by the same vehicle on a curve at the same speed. This vibration may negatively affect the product or packing. Hence, it was necessary to conduct a physical road test to understand the effect of vibration in the packaged products. The field transit trial has to be done before the transportations, which results in high investment. The company management was interested in developing an in-house test environment which would adequately represent the transit conditions. With the objective to develop an in-house test condition that can accurately simulate the mechanical loading scenario prevailing during the storage, handling and transportation of the products a brainstorming was done with the concerned people to identify the critical factors affecting vibration rate. Position of corrugated box, the position of bottle and speed of vehicle were identified as factors affecting the vibration rate. Several packing scenarios were identified by Design of Experiment methodology and simulated in the in-house test facility. Each condition was observed for 30 minutes, which was equivalent to 1000 km. The achieved vibration level was considered as the response. The average achieved in the simulated experiments was near to the third quartile (Q3) of the actual data. Thus, we were able to address around three-fourth of the actual phenomenon. Most of the cases in transit could be reproduced. The recommended test condition could generate a vibration level ranging from 9g to 15g as against a maximum of only 7g that was being generated earlier. Thus, the Company was able to test the packaged cartons satisfactorily in the house itself before transporting to the destinations, assuring itself that the breakages of the bottles will not happen.

Keywords: ANOVA, Corrugated box, DOE, Quartile

Procedia PDF Downloads 93
9958 Probabilistic Robustness Assessment of Structures under Sudden Column-Loss Scenario

Authors: Ali Y Al-Attraqchi, P. Rajeev, M. Javad Hashemi, Riadh Al-Mahaidi

Abstract:

This paper presents a probabilistic incremental dynamic analysis (IDA) of a full reinforced concrete building subjected to column loss scenario for the assessment of progressive collapse. The IDA is chosen to explicitly account for uncertainties in loads and system capacity. Fragility curves are developed to predict the probability of progressive collapse given the loss of one or more columns. At a broader scale, it will also provide critical information needed to support the development of a new generation of design codes that attempt to explicitly quantify structural robustness.

Keywords: fire, nonlinear incremental dynamic analysis, progressive collapse, structural engineering

Procedia PDF Downloads 236
9957 Automated Test Data Generation For some types of Algorithm

Authors: Hitesh Tahbildar

Abstract:

The cost of test data generation for a program is computationally very high. In general case, no algorithm to generate test data for all types of algorithms has been found. The cost of generating test data for different types of algorithm is different. Till date, people are emphasizing the need to generate test data for different types of programming constructs rather than different types of algorithms. The test data generation methods have been implemented to find heuristics for different types of algorithms. Some algorithms that includes divide and conquer, backtracking, greedy approach, dynamic programming to find the minimum cost of test data generation have been tested. Our experimental results say that some of these types of algorithm can be used as a necessary condition for selecting heuristics and programming constructs are sufficient condition for selecting our heuristics. Finally we recommend the different heuristics for test data generation to be selected for different types of algorithms.

Keywords: ongest path, saturation point, lmax, kL, kS

Procedia PDF Downloads 368
9956 Methodology to Achieve Non-Cooperative Target Identification Using High Resolution Range Profiles

Authors: Olga Hernán-Vega, Patricia López-Rodríguez, David Escot-Bocanegra, Raúl Fernández-Recio, Ignacio Bravo

Abstract:

Non-Cooperative Target Identification has become a key research domain in the Defense industry since it provides the ability to recognize targets at long distance and under any weather condition. High Resolution Range Profiles, one-dimensional radar images where the reflectivity of a target is projected onto the radar line of sight, are widely used for identification of flying targets. According to that, to face this problem, an approach to Non-Cooperative Target Identification based on the exploitation of Singular Value Decomposition to a matrix of range profiles is presented. Target Identification based on one-dimensional radar images compares a collection of profiles of a given target, namely test set, with the profiles included in a pre-loaded database, namely training set. The classification is improved by using Singular Value Decomposition since it allows to model each aircraft as a subspace and to accomplish recognition in a transformed domain where the main features are easier to extract hence, reducing unwanted information such as noise. Singular Value Decomposition permits to define a signal subspace which contain the highest percentage of the energy, and a noise subspace which will be discarded. This way, only the valuable information of each target is used in the recognition process. The identification algorithm is based on finding the target that minimizes the angle between subspaces and takes place in a transformed domain. Two metrics, F1 and F2, based on Singular Value Decomposition are accomplished in the identification process. In the case of F2, the angle is weighted, since the top vectors set the importance in the contribution to the formation of a target signal, on the contrary F1 simply shows the evolution of the unweighted angle. In order to have a wide database or radar signatures and evaluate the performance, range profiles are obtained through numerical simulation of seven civil aircraft at defined trajectories taken from an actual measurement. Taking into account the nature of the datasets, the main drawback of using simulated profiles instead of actual measured profiles is that the former implies an ideal identification scenario, since measured profiles suffer from noise, clutter and other unwanted information and simulated profiles don't. In this case, the test and training samples have similar nature and usually a similar high signal-to-noise ratio, so as to assess the feasibility of the approach, the addition of noise has been considered before the creation of the test set. The identification results applying the unweighted and weighted metrics are analysed for demonstrating which algorithm provides the best robustness against noise in an actual possible scenario. So as to confirm the validity of the methodology, identification experiments of profiles coming from electromagnetic simulations are conducted, revealing promising results. Considering the dissimilarities between the test and training sets when noise is added, the recognition performance has been improved when weighting is applied. Future experiments with larger sets are expected to be conducted with the aim of finally using actual profiles as test sets in a real hostile situation.

Keywords: HRRP, NCTI, simulated/synthetic database, SVD

Procedia PDF Downloads 328
9955 Developing High-Definition Flood Inundation Maps (HD-Fims) Using Raster Adjustment with Scenario Profiles (RASPTM)

Authors: Robert Jacobsen

Abstract:

Flood inundation maps (FIMs) are an essential tool in communicating flood threat scenarios to the public as well as in floodplain governance. With an increasing demand for online raster FIMs, the FIM State-of-the-Practice (SOP) is rapidly advancing to meet the dual requirements for high-resolution and high-accuracy—or High-Definition. Importantly, today’s technology also enables the resolution of problems of local—neighborhood-scale—bias errors that often occur in FIMs, even with the use of SOP two-dimensional flood modeling. To facilitate the development of HD-FIMs, a new GIS method--Raster Adjustment with Scenario Profiles, RASPTM—is described for adjusting kernel raster FIMs to match refined scenario profiles. With RASPTM, flood professionals can prepare HD-FIMs for a wide range of scenarios with available kernel rasters, including kernel rasters prepared from vector FIMs. The paper provides detailed procedures for RASPTM, along with an example of applying RASPTM to prepare an HD-FIM for the August 2016 Flood in Louisiana using both an SOP kernel raster and a kernel raster derived from an older vector-based flood insurance rate map. The accuracy of the HD-FIMs achieved with the application of RASPTM to the two kernel rasters is evaluated.

Keywords: hydrology, mapping, high-definition, inundation

Procedia PDF Downloads 31
9954 CDIO-Based Teaching Reform for Software Project Management Course

Authors: Liping Li, Wenan Tan, Na Wang

Abstract:

With the rapid development of information technology, project management has gained more and more attention recently. Based on CDIO, this paper proposes some teaching reform ideas for software project management curriculum. We first change from Teacher-centered classroom to Student-centered and adopt project-driven, scenario animation show, teaching rhythms, case study and team work practice to improve students' learning enthusiasm. Results showed these attempts have been well received and very effective; as well, students prefer to learn with this curriculum more than before the reform.

Keywords: CDIO, teaching reform, engineering education, project-driven, scenario animation simulation

Procedia PDF Downloads 403
9953 Relationship between the Yo-Yo Intermittent Recovery Test Level 1 and Anaerobic Performance Tests in Youth Soccer Players

Authors: Turgay Ozgur, Bahar Ozgur, Gurcan Yazici

Abstract:

The aims of the study were to investigate the relationship between the Yo-Yo intermittent recovery test level 1 (YYIR1) and relatively easy to conduct anaerobic power tests such as Sergeant (SJ) and Standing Broad Jump (SBJ), the flexibility Sit&Reach test (S&R) and Hexagon Agility (HA) test in twenty youth soccer players, aged 14 years. Players completed YYIR1 and other performance tests [(SJ), (SBJ] in two consecutive days. The mean YYIR1 distances for the players was 1454 ± 420 m. Peak Anaerobic Power (PAPw) was calculated using SJ (cm) scores. The mean PAPw was 2966,83w. Spearman’s correlation test results revealed that there is a statistically significant negative correlation between HA and YYIR1 tests (r = -0.72, p=0.000) and no significant correlation was found between anaerobic power tests and YYIR1. In conclusion, as a test to measure player’s intermittent aerobic capacity YYIR1 test and anaerobic power test results have not shown significant correlation. Although the YYIR1 test has been used in talent identification, anaerobic qualifications of player’s should be assessed using designated performance tests.

Keywords: yo-yo test, anaerobic power, soccer, sergeant jump test

Procedia PDF Downloads 352
9952 Web Application for Evaluating Tests in Distance Learning Systems

Authors: Bogdan Walek, Vladimir Bradac, Radim Farana

Abstract:

Distance learning systems offer useful methods of learning and usually contain final course test or another form of test. The paper proposes web application for evaluating tests using expert system in distance learning systems. Proposed web application is appropriate for didactic tests or tests with results for subsequent studying follow-up courses. Web application works with test questions and uses expert system and LFLC tool for test evaluation. After test evaluation the results are visualized and shown to student.

Keywords: distance learning, test, uncertainty, fuzzy, expert system, student

Procedia PDF Downloads 460
9951 Directed-Wald Test for Distinguishing Long Memory and Nonlinearity Time Series: Power and Size Simulation

Authors: Heri Kuswanto, Philipp Sibbertsen, Irhamah

Abstract:

A Wald type test to distinguish between long memory and ESTAR nonlinearity has been developed. The test uses a directed-Wald statistic to overcome the problem of restricted parameters under the alternative. The test is derived from a model specification i.e. allows the transition parameter to appear as a nuisance parameter in the transition function. A simulation study has been conducted and it indicates that the approach leads a test with good size and power properties to distinguish between stationary long memory and ESTAR.

Keywords: directed-Wald test, ESTAR, long memory, distinguish

Procedia PDF Downloads 444
9950 Flood Scenarios for Hydrological and Hydrodynamic Modelling

Authors: M. Sharif Imam Ibne Amir, Mohammad Masud Kamal Khan, Mohammad Golam Rasul, Raj H. Sharma, Fatema Akram

Abstract:

Future flood can be predicted using the probable maximum flood (PMF). PMF is calculated using the historical discharge or rainfall data considering the other climatic parameter stationary. However, climate is changing globally and the key climatic variables are temperature, evaporation, rainfall and sea level rise (SLR). To develop scenarios to a basin or catchment scale these important climatic variables should be considered. Nowadays scenario based on climatic variables is more suitable than PMF. Six scenarios were developed for a large Fitzroy basin and presented in this paper.

Keywords: climate change, rainfall, potential evaporation, scenario, sea level rise (SLR), sub-catchment

Procedia PDF Downloads 488
9949 Fostering Students’ Active Learning in Speaking Class through Project-Based Learning

Authors: Rukminingsih Rukmi

Abstract:

This paper addresses the issue of L2 teaching speaking to ESL students by fostering their active learning through project-based learning. Project-based learning was employed in classrooms where teachers support students by giving sufficient guidance and feedback. The students drive the inquiry, engage in research and discovery, and collaborate effectively with teammates to deliver the final work product. The teacher provides the initial direction and acts as a facilitator along the way. This learning approach is considered helpful for fostering students’ active learning. that the steps in implementing of project-based learning that fosters students’ critical thinking in TEFL class are in the following: (1) Discussing the materials about Speaking Class, (2) Working with the group to construct scenario of ways on speaking practice, (3) Practicing the scenario, (4) Recording the speaking practice into video, and (5) Evaluating the video product. This research is aimed to develop a strategy of teaching speaking by implementing project-based learning to improve speaking skill in the second Semester of English Department of STKIP PGRI Jombang. To achieve the purpose, the researcher conducted action research. The data of the study were gathered through the following instruments: test, observation checklists, and questionnaires. The result was indicated by the increase of students’ average speaking scores from 65 in the preliminary study, 73 in the first cycle, and 82 in the second cycle. Besides, the results of the study showed that project-based learning considered to be appropriate strategy to give students the same amount of chance in practicing their speaking skill and to pay attention in creating a learning situation.

Keywords: active learning, project-based learning, speaking ability, L2 teaching speaking

Procedia PDF Downloads 372
9948 Model-Based Software Regression Test Suite Reduction

Authors: Shiwei Deng, Yang Bao

Abstract:

In this paper, we present a model-based regression test suite reducing approach that uses EFSM model dependence analysis and probability-driven greedy algorithm to reduce software regression test suites. The approach automatically identifies the difference between the original model and the modified model as a set of elementary model modifications. The EFSM dependence analysis is performed for each elementary modification to reduce the regression test suite, and then the probability-driven greedy algorithm is adopted to select the minimum set of test cases from the reduced regression test suite that cover all interaction patterns. Our initial experience shows that the approach may significantly reduce the size of regression test suites.

Keywords: dependence analysis, EFSM model, greedy algorithm, regression test

Procedia PDF Downloads 398