Search results for: software testing techniques
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 13032

Search results for: software testing techniques

12612 An Approach to Analyze Testing of Nano On-Chip Networks

Authors: Farnaz Fotovvatikhah, Javad Akbari

Abstract:

Test time of a test architecture is an important factor which depends on the architecture's delay and test patterns. Here a new architecture to store the test results based on network on chip is presented. In addition, simple analytical model is proposed to calculate link test time for built in self-tester (BIST) and external tester (Ext) in multiprocessor systems. The results extracted from the model are verified using FPGA implementation and experimental measurements. Systems consisting 16, 25, and 36 processors are implemented and simulated and test time is calculated. In addition, BIST and Ext are compared in terms of test time at different conditions such as at different number of test patterns and nodes. Using the model the maximum frequency of testing could be calculated and the test structure could be optimized for high speed testing.

Keywords: test, nano on-chip network, JTAG, modelling

Procedia PDF Downloads 454
12611 Factors of Influence in Software Process Improvement: An ISO/IEC 29110 for Very-Small Entities

Authors: N. Wongsai, R. Wetprasit, V. Siddoo

Abstract:

The recently introduced ISO/IEC 29110 standard Lifecycle profile for Very Small Entities (VSE) has been adopted and practiced in many small and medium software companies, including in Thailand’s software industry. Many Thai companies complete their software process improvement (SPI) initiative program and have been certified. There are, however, a number of participants fail to success. This study was concerned with the factors that influence the accomplishment of the standard implementation in various VSE characteristics. In order to achieve this goal, exploring and extracting critical factors from prior studies were carried out and then the obtained factors were validated by the standard experts. Data analysis of comments and recommendations was performed using a qualitative content analysis method. This paper presents the initial set of influence factors in both positive and negative impact the ISO/IEC 29110 implementation with an aim at helping such SPI practitioners with some considerations to manage appropriate adoption approach in order to achieve its implementation.

Keywords: barriers, critical success factors, ISO/IEC 29110, Software Process Improvement, SPI, Very-Small Entity, VSE

Procedia PDF Downloads 293
12610 An Approach to Specify Software Requirements in Semantic Form

Authors: Deepa Vijay, Chellammal Surianarayanan, Gopinath Ganapathy

Abstract:

Requirements of a software project serve as a guideline for the entire project team which enable the team towards producing the right outcome. As requirements are the key in deciding the success of the project, it should be specified in an unambiguous manner. Also, the requirements should be complete and consistent. It should be interpreted in the same way by the entire software project team as the customer interprets. Specifying requirements in textual manner is common in software development. This leads to poor understanding of the requirements which results in more errors and degraded quality. There are some literatures which focus on semantic way of specifying functional requirement which ensure the consistency and completeness of requirements. Alternately in the work, a method is proposed to map the syntactic requirements with corresponding semantics in the form of ontologies. This improves the understanding of requirements, prevents errors and improves quality.

Keywords: functional requirement, ontology, requirements management, semantics

Procedia PDF Downloads 346
12609 Applying Failure Modes and Effect Analysis Concept in a Global Software Development Process

Authors: Camilo Souza, Lidia Melo, Fernanda Terra, Francisco Caio, Marcelo Reis

Abstract:

SIDIA is a research and development (R&D) institute that belongs to Samsung’s global software development process. The SIDIA’s Model Team (MT) is a part of Samsung’s Mobile Division Area, which is responsible for the development of Android releases embedded in Samsung mobile devices. Basically, in this software development process, the kickoff occurs in some strategic countries (e.g., South Korea) where some software requirements are applied and the initial software tests are performed. When the software achieves a more mature level, a new branch is derived, and the development continues in subsidiaries from other strategic countries (e.g., SIDIA-Brazil). However, even in the newly created branches, there are several interactions between developers from different nationalities in order to fix bugs reported during test activities, apply some specific requirements from partners and develop new features as well. Despite the GSD strategy contributes to improving software development, some challenges are also introduced as well. In this paper, we share the initial results about the application of the failure modes and effect analysis (FMEA) concept in the software development process followed by the SIDIA’s model team. The main goal was to identify and mitigate the process potential failures through the application of recommended actions. The initial results show that the application of the FMEA concept allows us to identify the potential failures in our GSD process as well as to propose corrective actions to mitigate them. Finally, FMEA encouraged members of different teams to take actions that contribute to improving our GSD process.

Keywords: global software development, potential failures, FMEA, recommended actions

Procedia PDF Downloads 204
12608 Using Design Sprint For Software Engineering Undergraduate Student Projects: A Method Paper

Authors: Sobhani U. Pilapitiya, Tharanga Peiris

Abstract:

Software Engineering curriculums generally consist of industry-based practices such as project-based learning (PBL) which mainly focuses on efficient and innovative product development. These approaches can be tailored and used in project-based modules in software engineering curriculums. However, there are very limited attempts in the area especially related to the Sri Lankan context. This paper describes a tailored pedagogical approach and its results of using design sprint which can be used for project-based modules in SE curriculums. A controlled group of second-year software engineering students was selected for the study. The study results indicate that 100% of students agreed that the Design Sprint approach is effective in group-based projects and 83% of students stated that it minimized the re-work compared to traditional project approaches. The tailored process was effective, easy to implement and produced desired results at the end of the session while providing students an enjoyable experience.

Keywords: design sprint, PBL, software engineering, curriculum

Procedia PDF Downloads 180
12607 Applying Personel Resilence and Emotional Agitation in Occupational, Health and Safety Education and Training

Authors: M. Jayandran

Abstract:

Continual professional development is an important concept for safety professionals to strengthen the knowledge base and to achieve the required qualifications or international memberships in a given time. But the main problems which have observed among most of the safety aspirants are as follows: lack of focus, inferiority complex, superiority complex, lack of interest and lethargy, family and off job stress, health issues, usage of drugs and alcohol, and absenteeism. A HSE trainer should be an expert in soft skills and other stress, emotional handling techniques, so as to manage the above aspirants during training. To do this practice, a trainer has to brainstorm himself of few of the soft skills like personnel resilience, mnemonic techniques, mind healing, and subconscious suggestion techniques by integrating with an emotional intelligence quotient of the aspirants. By adopting these techniques, a trainer can successfully deliver the course and influence the different types of audience to achieve success in training.

Keywords: personnel resilience, mnemonic techniques, mind healing, sub conscious suggestion techniques

Procedia PDF Downloads 280
12606 Survey on Malware Detection

Authors: Doaa Wael, Naswa Abdelbaky

Abstract:

Malware is malicious software that is built to cause destructive actions and damage information systems and networks. Malware infections increase rapidly, and types of malware have become more sophisticated, which makes the malware detection process more difficult. On the other side, the Internet of Things IoT technology is vulnerable to malware attacks. These IoT devices are always connected to the internet and lack security. This makes them easy for hackers to access. These malware attacks are becoming the go-to attack for hackers. Thus, in order to deal with this challenge, new malware detection techniques are needed. Currently, building a blockchain solution that allows IoT devices to download any file from the internet and to verify/approve whether it is malicious or not is the need of the hour. In recent years, blockchain technology has stood as a solution to everything due to its features like decentralization, persistence, and anonymity. Moreover, using blockchain technology overcomes some difficulties in malware detection and improves the malware detection ratio over-than the techniques that do not utilize blockchain technology. In this paper, we study malware detection models which are based on blockchain technology. Furthermore, we elaborate on the effect of blockchain technology in malware detection, especially in the android environment.

Keywords: malware analysis, blockchain, malware attacks, malware detection approaches

Procedia PDF Downloads 48
12605 Software Vulnerability Markets: Discoverers and Buyers

Authors: Abdullah M. Algarni, Yashwant K. Malaiya

Abstract:

Some of the key aspects of vulnerability-discovery, dissemination, and disclosure-have received some attention recently. However, the role of interaction among the vulnerability discoverers and vulnerability acquirers has not yet been adequately addressed. Our study suggests that a major percentage of discoverers, a majority in some cases, are unaffiliated with the software developers and thus are free to disseminate the vulnerabilities they discover in any way they like. As a result, multiple vulnerability markets have emerged. In some of these markets, the exchange is regulated, but in others, there is little or no regulation. In recent vulnerability discovery literature, the vulnerability discoverers have remained anonymous individuals. Although there has been an attempt to model the level of their efforts, information regarding their identities, modes of operation, and what they are doing with the discovered vulnerabilities has not been explored. Reports of buying and selling of the vulnerabilities are now appearing in the press; however, the existence of such markets requires validation, and the natures of the markets need to be analysed. To address this need, we have attempted to collect detailed information. We have identified the most prolific vulnerability discoverers throughout the past decade and examined their motivation and methods. A large percentage of these discoverers are located in Eastern and Western Europe and in the Far East. We have contacted several of them in order to collect first-hand information regarding their techniques, motivations, and involvement in the vulnerability markets. We examine why many of the discoverers appear to retire after a highly successful vulnerability-finding career. The paper identifies the actual vulnerability markets, rather than the hypothetical ideal markets that are often examined. The emergence of worldwide government agencies as vulnerability buyers has significant implications. We discuss potential factors that can impact the risk to society and the need for detailed exploration.

Keywords: risk management, software security, vulnerability discoverers, vulnerability markets

Procedia PDF Downloads 228
12604 Expert-Driving-Criteria Based on Fuzzy Logic Approach for Intelligent Driving Diagnosis

Authors: Andrés C. Cuervo Pinilla, Christian G. Quintero M., Chinthaka Premachandra

Abstract:

This paper considers people’s driving skills diagnosis under real driving conditions. In that sense, this research presents an approach that uses GPS signals which have a direct correlation with driving maneuvers. Besides, it is presented a novel expert-driving-criteria approximation using fuzzy logic which seeks to analyze GPS signals in order to issue an intelligent driving diagnosis. Based on above, this works presents in the first section the intelligent driving diagnosis system approach in terms of its own characteristics properties, explaining in detail significant considerations about how an expert-driving-criteria approximation must be developed. In the next section, the implementation of our developed system based on the proposed fuzzy logic approach is explained. Here, a proposed set of rules which corresponds to a quantitative abstraction of some traffics laws and driving secure techniques seeking to approach an expert-driving- criteria approximation is presented. Experimental testing has been performed in real driving conditions. The testing results show that the intelligent driving diagnosis system qualifies driver’s performance quantitatively with a high degree of reliability.

Keywords: driver support systems, intelligent transportation systems, fuzzy logic, real time data processing

Procedia PDF Downloads 488
12603 Cross Site Scripting (XSS) Attack and Automatic Detection Technology Research

Authors: Tao Feng, Wei-Wei Zhang, Chang-Ming Ding

Abstract:

Cross-site scripting (XSS) is one of the most popular WEB Attacking methods at present, and also one of the most risky web attacks. Because of the population of JavaScript, the scene of the cross site scripting attack is also gradually expanded. However, since the web application developers tend to only focus on functional testing and lack the awareness of the XSS, which has made the on-line web projects exist many XSS vulnerabilities. In this paper, different various techniques of XSS attack are analyzed, and a method automatically to detect it is proposed. It is easy to check the results of vulnerability detection when running it as a plug-in.

Keywords: XSS, no target attack platform, automatic detection,XSS detection

Procedia PDF Downloads 378
12602 Open Source Software in Higher Education: Oman SQU Case Study

Authors: Amal S. Al-Badi, Ali H. Al-Badi

Abstract:

Many organizations are opting to adopt Open Source Software (OSS) as it is the current trend to rely on each other rather than on companies (Software vendors). It is a clear shift from organizations to individuals, the concept being to rely on collective participation rather than companies/vendors. The main objectives of this research are 1) to identify the current level of OSS usage in Sultan Qaboos University; 2) to identify the potential benefits of using OSS in educational institutes; 3) to identify the OSS applications that are most likely to be used within an educational institute; 4) to identify the existing and potential barriers to the successful adoption of OSS in education. To achieve these objectives a two-stage research method was conducted. First a rigorous literature review of previously published material was performed (interpretive/descriptive approach), and then a set of interviews were conducted with the IT professionals at Sultan Qaboos University in Oman in order to explore the extent and nature of their usage of OSS.

Keywords: open source software, social software, e-learning 2.0, Web 2.0, connectivism, personal learning environment (PLE), OpenCourseWare

Procedia PDF Downloads 282
12601 Stabilization of Expansive Soils by Additions Binders Hydraulic Lime and Cement

Authors: Kherafa Abdennasser

Abstract:

A literature review was conducted to gather as much information. Concerns the phenomenon of swelling clays, as well as a presentation of some bibliographic findings on factors affecting the swelling potential. Citing the various techniques of stabilization of clays as well as a presentation of some literature results on the stabilization of swelling. Then a characterization of the materials was carried out at basic bibliographic study. These are standard mechanical geotechnical testing. Simple practical, economical and efficient to minimize the phenomenon swelling.

Keywords: stabilization, expansive soils, cement, lime, oedometer

Procedia PDF Downloads 286
12600 Computer-Aided Teaching of Transformers for Undergraduates

Authors: Rajesh Kumar, Roopali Dogra, Puneet Aggarwal

Abstract:

In the era of technological advancement, use of computer technology has become inevitable. Hence it has become the need of the hour to integrate software methods in engineering curriculum as a part to boost pedagogy techniques. Simulations software is a great help to graduates of disciplines such as electrical engineering. Since electrical engineering deals with high voltages and heavy instruments, extra care must be taken while operating with them. The viable solution would be to have appropriate control. The appropriate control could be well designed if engineers have knowledge of kind of waveforms associated with the system. Though these waveforms can be plotted manually, but it consumes a lot of time. Hence aid of simulation helps to understand steady state of system and resulting in better performance. In this paper computer, aided teaching of transformer is carried out using MATLAB/Simulink. The test carried out on a transformer includes open circuit test and short circuit respectively. The respective parameters of transformer are then calculated using the values obtained from open circuit and short circuit test respectively using Simulink.

Keywords: computer aided teaching, open circuit test, short circuit test, simulink, transformer

Procedia PDF Downloads 344
12599 Improving the Run Times of Existing and Historical Demand Models Using Simple Python Scripting

Authors: Abhijeet Ostawal, Parmjit Lall

Abstract:

The run times for a large strategic model that we were managing had become too long leading to delays in project delivery, increased costs and loss in productivity. Software developers are continuously working towards developing more efficient tools by changing their algorithms and processes. The issue faced by our team was how do you apply the latest technologies on validated existing models which are based on much older versions of software that do not have the latest software capabilities. The multi-model transport model that we had could only be run in sequential assignment order. Recent upgrades to the software now allowed the assignment to be run in parallel, a concept called parallelization. Parallelization is a Python script working only within the latest version of the software. A full model transfer to the latest version was not possible due to time, budget and the potential changes in trip assignment. This article is to show the method to adapt and update the Python script in such a way that it can be used in older software versions by calling the latest version and then recalling the old version for assignment model without affecting the results. Through a process of trial-and-error run time savings of up to 30-40% have been achieved. Assignment results were maintained within the older version and through this learning process we’ve applied this methodology to other even older versions of the software resulting in huge time savings, more productivity and efficiency for both client and consultant.

Keywords: model run time, demand model, parallelisation, python scripting

Procedia PDF Downloads 93
12598 A Comparative Study of Malware Detection Techniques Using Machine Learning Methods

Authors: Cristina Vatamanu, Doina Cosovan, Dragos Gavrilut, Henri Luchian

Abstract:

In the past few years, the amount of malicious software increased exponentially and, therefore, machine learning algorithms became instrumental in identifying clean and malware files through semi-automated classification. When working with very large datasets, the major challenge is to reach both a very high malware detection rate and a very low false positive rate. Another challenge is to minimize the time needed for the machine learning algorithm to do so. This paper presents a comparative study between different machine learning techniques such as linear classifiers, ensembles, decision trees or various hybrids thereof. The training dataset consists of approximately 2 million clean files and 200.000 infected files, which is a realistic quantitative mixture. The paper investigates the above mentioned methods with respect to both their performance (detection rate and false positive rate) and their practicability.

Keywords: ensembles, false positives, feature selection, one side class algorithm

Procedia PDF Downloads 263
12597 A Dynamic Software Product Line Approach to Self-Adaptive Genetic Algorithms

Authors: Abdelghani Alidra, Mohamed Tahar Kimour

Abstract:

Genetic algorithm must adapt themselves at design time to cope with the search problem specific requirements and at runtime to balance exploration and convergence objectives. In a previous article, we have shown that modeling and implementing Genetic Algorithms (GA) using the software product line (SPL) paradigm is very appreciable because they constitute a product family sharing a common base of code. In the present article we propose to extend the use of the feature model of the genetic algorithms family to model the potential states of the GA in what is called a Dynamic Software Product Line. The objective of this paper is the systematic generation of a reconfigurable architecture that supports the dynamic of the GA and which is easily deduced from the feature model. The resultant GA is able to perform dynamic reconfiguration autonomously to fasten the convergence process while producing better solutions. Another important advantage of our approach is the exploitation of recent advances in the domain of dynamic SPLs to enhance the performance of the GAs.

Keywords: self-adaptive genetic algorithms, software engineering, dynamic software product lines, reconfigurable architecture

Procedia PDF Downloads 261
12596 Promoting Couple HIV Testing among Migrants for HIV Prevention: Learnings from Integrated Counselling and Testing Centre (ICTC) in Odisha, India

Authors: Sunil Mekale, Debasish Chowdhury, Sanchita Patnaik, Amitav Das, Ashok Agarwal

Abstract:

Background: Odisha is a low HIV prevalence state in India (ANC-HIV positivity of 0.42% as per HIV sentinel surveillance 2010-2011); however, it is an important source migration state with 3.2% of male migrants reporting to be PLHIV. USAID Public Health Foundation of India -PIPPSE project is piloting a source-destination corridor programme between Odisha and Gujarat. In Odisha, the focus has been on developing a comprehensive strategy to reach out to the out migrants and their spouses in the place of their origin based on their availability. The project has made concerted attempts to identify vulnerable districts with high out migration and high positivity rate. Description: 48 out of 97 ICTCs were selected from nine top high out migration districts through multistage sampling. A retrospective descriptive analysis of HIV positive male migrants and their spouses for two years (April 2013-March 2015) was conducted. A total of 3,645 HIV positive records were analysed. Findings: Among 34.2% detected HIV positive in the ICTCs, 23.3% were male migrants and 11% were spouses of male migrants; almost 50% of total ICTC attendees. More than 70% of the PLHIV male migrants and their spouses were less than 45 years old. Conclusions: Couple HIV testing approach may be considered for male migrants and their spouses. ICTC data analysis could guide in identifying the locations with high HIV positivity among male migrants and their spouses.

Keywords: HIV testing, migrants, spouse of migrants, Integrated Counselling and Testing Centre (ICTC)

Procedia PDF Downloads 358
12595 Graphical User Interface for Presting Matlab Work for Reduction of Chromatic Disperion Using Digital Signal Processing for Optical Communication

Authors: Muhammad Faiz Liew Abdullah, Bhagwan Das, Nor Shahida, Abdul Fattah Chandio

Abstract:

This study presents the designed features of Graphical User Interface (GUI) for chromatic dispersion (CD) reduction using digital signal processing (DSP) techniques. GUI is specially designed for windows platform. The obtained simulation results from matlab are presented via this GUI. After importing results from matlab in GUI, It will present your work on any windows7 and onwards versions platforms without matlab software. First part of the GUI contains the research methodology block diagram and in the second part, output for each stage is shown in separate reserved area for the result display. Each stage of methodology has the captions to display the results. This GUI will be very helpful during presentations instead of making slides this GUI will present all your work easily in the absence of other software’s such as Matlab, Labview, MS PowerPoint. GUI is designed using C programming in MS Visio Studio.

Keywords: Matlab simulation results, C programming, MS VISIO studio, chromatic dispersion

Procedia PDF Downloads 434
12594 SVID: Structured Vulnerability Intelligence for Building Deliberated Vulnerable Environment

Authors: Wenqing Fan, Yixuan Cheng, Wei Huang

Abstract:

The diversity and complexity of modern IT systems make it almost impossible for internal teams to find vulnerabilities in all software before the software is officially released. The emergence of threat intelligence and vulnerability reporting policy has greatly reduced the burden on software vendors and organizations to find vulnerabilities. However, to prove the existence of the reported vulnerability, it is necessary but difficult for security incident response team to build a deliberated vulnerable environment from the vulnerability report with limited and incomplete information. This paper presents a structured, standardized, machine-oriented vulnerability intelligence format, that can be used to automate the orchestration of Deliberated Vulnerable Environment (DVE). This paper highlights the important role of software configuration and proof of vulnerable specifications in vulnerability intelligence, and proposes a triad model, which is called DIR (Dependency Configuration, Installation Configuration, Runtime Configuration), to define software configuration. Finally, this paper has also implemented a prototype system to demonstrate that the orchestration of DVE can be automated with the intelligence.

Keywords: DIR triad model, DVE, vulnerability intelligence, vulnerability recurrence

Procedia PDF Downloads 100
12593 Non-Invasive Evaluation of Patients After Percutaneous Coronary Revascularization. The Role of Cardiac Imaging

Authors: Abdou Elhendy

Abstract:

Numerous study have shown the efficacy of the percutaneous intervention (PCI) and coronary stenting in improving left ventricular function and relieving exertional angina. Furthermore, PCI remains the main line of therapy in acute myocardial infarction. Improvement of procedural techniques and new devices have resulted in an increased number of PCI in those with difficult and extensive lesions, multivessel disease as well as total occlusion. Immediate and late outcome may be compromised by acute thrombosis or the development of fibro-intimal hyperplasia. In addition, progression of coronary artery disease proximal or distal to the stent as well as in non-stented arteries is not uncommon. As a result, complications can occur, such as acute myocardial infarction, worsened heart failure or recurrence of angina. In a stent, restenosis can occur without symptoms or with atypical complaints rendering the clinical diagnosis difficult. Routine invasive angiography is not appropriate as a follow up tool due to associated risk and cost and the limited functional assessment. Exercise and pharmacologic stress testing are increasingly used to evaluate the myocardial function, perfusion and adequacy of revascularization. Information obtained by these techniques provide important clues regarding presence and severity of compromise in myocardial blood flow. Stress echocardiography can be performed in conjunction with exercise or dobutamine infusion. The diagnostic accuracy has been moderate, but the results provide excellent prognostic stratification. Adding myocardial contrast agents can improve imaging quality and allows assessment of both function and perfusion. Stress radionuclide myocardial perfusion imaging is an alternative to evaluate these patients. The extent and severity of wall motion and perfusion abnormalities observed during exercise or pharmacologic stress are predictors of survival and risk of cardiac events. According to current guidelines, stress echocardiography and radionuclide imaging are considered to have appropriate indication among patients after PCI who have cardiac symptoms and those who underwent incomplete revascularization. Stress testing is not recommended in asymptomatic patients, particularly early after revascularization, Coronary CT angiography is increasingly used and provides high sensitive for the diagnosis of coronary artery stenosis. Average sensitivity and specificity for the diagnosis of in stent stenosis in pooled data are 79% and 81%, respectively. Limitations include blooming artifacts and low feasibility in patients with small stents or thick struts. Anatomical and functional cardiac imaging modalities are corner stone for the assessment of patients after PCI and provide salient diagnostic and prognostic information. Current imaging techniques cans serve as gate keeper for coronary angiography, thus limiting the risk of invasive procedures to those who are likely to benefit from subsequent revascularization. The determination of which modality to apply requires careful identification of merits and limitation of each technique as well as the unique characteristic of each individual patient.

Keywords: coronary artery disease, stress testing, cardiac imaging, restenosis

Procedia PDF Downloads 138
12592 Software-Defined Networks in Utility Power Networks

Authors: Ava Salmanpour, Hanieh Saeedi, Payam Rouhi, Elahe Hamzeil, Shima Alimohammadi, Siamak Hossein Khalaj, Mohammad Asadian

Abstract:

Software-defined network (SDN) is a network architecture designed to control network using software application in a central manner. This ability enables remote control of the whole network regardless of the network technology. In fact, in this architecture network intelligence is separated from physical infrastructure, it means that required network components can be implemented virtually using software applications. Today, power networks are characterized by a high range of complexity with a large number of intelligent devices, processing both huge amounts of data and important information. Therefore, reliable and secure communication networks are required. SDNs are the best choice to meet this issue. In this paper, SDN networks capabilities and characteristics will be reviewed and different basic controllers will be compared. The importance of using SDNs to escalate efficiency and reliability in utility power networks is going to be discussed and the comparison between the SDN-based power networks and traditional networks will be explained.

Keywords: software-defined network, SDNs, utility network, open flow, communication, gas and electricity, controller

Procedia PDF Downloads 84
12591 Physical Activity and Cognitive Functioning Relationship in Children

Authors: Comfort Mokgothu

Abstract:

This study investigated the relation between processing information and fitness level of active (fit) and sedentary (unfit) children drawn from rural and urban areas in Botswana. It was hypothesized that fit children would display faster simple reaction time (SRT), choice reaction times (CRT) and movement times (SMT). 60, third grade children (7.0 – 9.0 years) were initially selected and based upon fitness testing, 45 participated in the study (15 each of fit urban, unfit urban, fit rural). All children completed anthropometric measures, skinfold testing and submaximal cycle ergometer testing. The cognitive testing included SRT, CRT, SMT and Choice Movement Time (CMT) and memory sequence length. Results indicated that the rural fit group exhibited faster SMT than the urban fit and unfit groups. For CRT, both fit groups were faster than the unfit group. Collectively, the study shows that the relationship that exists between physical fitness and cognitive function amongst the elderly can tentatively be extended to the pediatric population. Physical fitness could be a factor in the speed at which we process information, including decision making, even in children.

Keywords: decision making, fitness, information processing, reaction time, cognition movement time

Procedia PDF Downloads 121
12590 Trends in Language Testing in Primary Schools in River State, Nigeria

Authors: Okoh Chinasa, Asimuonye Augusta

Abstract:

This study investigated the trends in language testing in Primary Schools in Rivers State. English language past question papers were collected from four (4) Primary Schools in Onelga Local Government Area and Ahoada East Local Government Area. Four research questions guided the study. The study is aimed at finding out the appropriateness of test formats used for language testing and the language skills tested. The past question papers collected which served as the instrument were analyzed based on given criteria developed by the researchers in line with documentary frequency studies, a type of survey study. The study revealed that some of the four language skills were not adequately assessed and that the termly question papers were developed by a central examination body. From the past questions, it was observed that an imbalance exists in the test format used. The paper recommended that all the language skills should be tested using correct test formats to ensure that pupils were given a fair chance to show what they know and can do in English language and for teachers to be able to use the test results for effective decision making.

Keywords: discrete test, integrative test, testing approach, test format

Procedia PDF Downloads 394
12589 Design Systems and the Need for a Usability Method: Assessing the Fitness of Components and Interaction Patterns in Design Systems Using Atmosphere Methodology

Authors: Patrik Johansson, Selina Mardh

Abstract:

The present study proposes a usability test method, Atmosphere, to assess the fitness of components and interaction patterns of design systems. The method covers the user’s perception of the components of the system, the efficiency of the logic of the interaction patterns, perceived ease of use as well as the user’s understanding of the intended outcome of interactions. These aspects are assessed by combining measures of first impression, visual affordance and expectancy. The method was applied to a design system developed for the design of an electronic health record system. The study was conducted involving 15 healthcare personnel. It could be concluded that the Atmosphere method provides tangible data that enable human-computer interaction practitioners to analyze and categorize components and patterns based on perceived usability, success rate of identifying interactive components and success rate of understanding components and interaction patterns intended outcome.

Keywords: atomic design, atmosphere methodology, design system, expectancy testing, first impression testing, usability testing, visual affordance testing

Procedia PDF Downloads 149
12588 Examining the Role of Farmer-Centered Participatory Action Learning in Building Sustainable Communities in Rural Haiti

Authors: Charles St. Geste, Michael Neumann, Catherine Twohig

Abstract:

Our primary aim is to examine farmer-centered participatory action learning as a tool to improve agricultural production, build resilience to climate shocks and, more broadly, advance community-driven solutions for sustainable development in rural communities across Haiti. For over six years, sixty plus farmers from Deslandes, Haiti, organized in three traditional work groups called konbits, have designed and tested low-input agroecology techniques as part of the Konbit Vanyan Kapab Pwoje Agroekoloji. The project utilizes a participatory action learning approach, emphasizing social inclusion, building on local knowledge, experiential learning, active farmer participation in trial design and evaluation, and cross-community sharing. Mixed methods were used to evaluate changes in knowledge and adoption of agroecology techniques, confidence in advancing agroecology locally, and innovation among Konbit Vanyan Kapab farmers. While skill and knowledge in application of agroecology techniques varied among individual farmers, a majority of farmers successfully adopted techniques outside of the trial farms. The use of agroecology techniques on trial and individual farms has doubled crop production in many cases. Farm income has also increased, and farmers report less damage to crops and property caused by extreme weather events. Furthermore, participatory action strategies have led to greater local self-determination and greater capacity for sustainable community development. With increased self-confidence and the knowledge and skills acquired from participating in the project, farmers prioritized sharing their successful techniques with other farmers and have developed a farmer-to-farmer training program that incorporates participatory action learning. Using adult education methods, farmers, trained as agroecology educators, are currently providing training in sustainable farming practices to farmers from five villages in three departments across Haiti. Konbit Vanyan Kapab farmers have also begun testing production of value-added food products, including a dried soup mix and tea. Key factors for success include: opportunities for farmers to actively participate in all phases of the project, group diversity, resources for application of agroecology techniques, focus on group processes and overcoming local barriers to inclusive decision-making.

Keywords: agroecology, participatory action learning, rural Haiti, sustainable community development

Procedia PDF Downloads 126
12587 Analysis of the Annual Proficiency Testing Procedure for Intermediate Reference Laboratories Conducted by the National Reference Laboratory from 2013 to 2017

Authors: Reena K., Mamatha H. G., Somshekarayya, P. Kumar

Abstract:

Objectives: The annual proficiency testing of intermediate reference laboratories is conducted by the National Reference Laboratory (NRL) to assess the efficiency of the laboratories to correctly identify Mycobacterium tuberculosis and to determine its drug susceptibility pattern. The proficiency testing results from 2013 to 2017 were analyzed to determine laboratories that were consistent in reporting quality results and those that had difficulty in doing so. Methods: A panel of twenty cultures were sent out to each of these laboratories. The laboratories were expected to grow the cultures in their own laboratories, set up drug susceptibly testing by all the methods they were certified for and report the results within the stipulated time period. The turnaround time for reporting results, specificity, sensitivity positive and negative predictive values and efficiency of the laboratory in identifying the cultures were analyzed. Results: Most of the laboratories had reported their results within the stipulated time period. However, there was enormous delay in reporting results from few of the laboratories. This was mainly due to improper functioning of the biosafety level III laboratory. Only 40% of the laboratories had 100% efficiency in solid culture using Lowenstein Jensen medium. This was expected as a solid culture, and drug susceptibility testing is not used for diagnosing drug resistance. Rapid molecular methods such as Line probe assay and Genexpert are used to determine drug resistance. Automated liquid culture system such as the Mycobacterial growth indicator tube is used to determine prognosis of the patient while on treatment. It was observed that 90% of the laboratories had achieved 100% in the liquid culture method. Almost all laboratories had achieved 100% efficiency in the line probe assay method which is the method of choice for determining drug-resistant tuberculosis. Conclusion: Since the liquid culture and line probe assay technologies are routinely used for the detection of drug-resistant tuberculosis the laboratories exhibited higher level of efficiency as compared to solid culture and drug susceptibility testing which are rarely used. The infrastructure of the laboratory should be maintained properly so that samples can be processed safely and results could be declared on time.

Keywords: annual proficiency testing, drug susceptibility testing, intermediate reference laboratory, national reference laboratory

Procedia PDF Downloads 165
12586 Reliability of Dry Tissues Sampled from Exhumed Bodies in DNA Analysis

Authors: V. Agostini, S. Gino, S. Inturri, A. Piccinini

Abstract:

In cases of corpse identification or parental testing performed on exhumed alleged dead father, usually, we seek and acquire organic samples as bones and/or bone fragments, teeth, nails and muscle’s fragments. The DNA analysis of these cadaveric matrices usually leads to identifying success, but it often happens that the results of the typing are not satisfactory with highly degraded, partial or even non-interpretable genetic profiles. To aggravate the interpretative panorama deriving from the analysis of such 'classical' organic matrices, we must add a long and laborious treatment of the sample that starts from the mechanical fragmentation up to the protracted decalcification phase. These steps greatly increase the chance of sample contamination. In the present work, instead, we want to report the use of 'unusual' cadaveric matrices, demonstrating that their forensic genetics analysis can lead to better results in less time and with lower costs of reagents. We report six case reports, result of on-field experience, in which eyeswabs and cartilage were sampled and analyzed, allowing to obtain clear single genetic profiles, useful for identification purposes. In all cases we used the standard DNA tissue extraction protocols (as reported on the user manuals of the manufacturers such as QIAGEN or Invitrogen- Thermo Fisher Scientific), thus bypassing the long and difficult phases of mechanical fragmentation and decalcification of bones' samples. PCR was carried out using PowerPlex® Fusion System kit (Promega), and capillary electrophoresis was carried out on an ABI PRISM® 310 Genetic Analyzer (Applied Biosystems®), with GeneMapper ID v3.2.1 (Applied Biosystems®) software. The software Familias (version 3.1.3) was employed for kinship analysis. The genetic results achieved have proved to be much better than the analysis of bones or nails, both from the qualitative and quantitative point of view and from the point of view of costs and timing. This way, by using the standard procedure of DNA extraction from tissue, it is possible to obtain, in a shorter time and with maximum efficiency, an excellent genetic profile, which proves to be useful and can be easily decoded for later paternity tests and/or identification of human remains.

Keywords: DNA, eye swabs and cartilage, identification human remains, paternity testing

Procedia PDF Downloads 89
12585 Computer-Aided Diagnosis of Polycystic Kidney Disease Using ANN

Authors: G. Anjan Babu, G. Sumana, M. Rajasekhar

Abstract:

Many inherited diseases and non-hereditary disorders are common in the development of renal cystic diseases. Polycystic kidney disease (PKD) is a disorder developed within the kidneys in which grouping of cysts filled with water like fluid. PKD is responsible for 5-10% of end-stage renal failure treated by dialysis or transplantation. New experimental models, application of molecular biology techniques have provided new insights into the pathogenesis of PKD. Researchers are showing keen interest for developing an automated system by applying computer aided techniques for the diagnosis of diseases. In this paper a multi-layered feed forward neural network with one hidden layer is constructed, trained and tested by applying back propagation learning rule for the diagnosis of PKD based on physical symptoms and test results of urinanalysis collected from the individual patients. The data collected from 50 patients are used to train and test the network. Among these samples, 75% of the data used for training and remaining 25% of the data are used for testing purpose. Furthermore, this trained network is used to implement for new samples. The output results in normality and abnormality of the patient.

Keywords: dialysis, hereditary, transplantation, polycystic, pathogenesis

Procedia PDF Downloads 356
12584 Electrical Decomposition of Time Series of Power Consumption

Authors: Noura Al Akkari, Aurélie Foucquier, Sylvain Lespinats

Abstract:

Load monitoring is a management process for energy consumption towards energy savings and energy efficiency. Non Intrusive Load Monitoring (NILM) is one method of load monitoring used for disaggregation purposes. NILM is a technique for identifying individual appliances based on the analysis of the whole residence data retrieved from the main power meter of the house. Our NILM framework starts with data acquisition, followed by data preprocessing, then event detection, feature extraction, then general appliance modeling and identification at the final stage. The event detection stage is a core component of NILM process since event detection techniques lead to the extraction of appliance features. Appliance features are required for the accurate identification of the household devices. In this research work, we aim at developing a new event detection methodology with accurate load disaggregation to extract appliance features. Time-domain features extracted are used for tuning general appliance models for appliance identification and classification steps. We use unsupervised algorithms such as Dynamic Time Warping (DTW). The proposed method relies on detecting areas of operation of each residential appliance based on the power demand. Then, detecting the time at which each selected appliance changes its states. In order to fit with practical existing smart meters capabilities, we work on low sampling data with a frequency of (1/60) Hz. The data is simulated on Load Profile Generator software (LPG), which was not previously taken into consideration for NILM purposes in the literature. LPG is a numerical software that uses behaviour simulation of people inside the house to generate residential energy consumption data. The proposed event detection method targets low consumption loads that are difficult to detect. Also, it facilitates the extraction of specific features used for general appliance modeling. In addition to this, the identification process includes unsupervised techniques such as DTW. To our best knowledge, there exist few unsupervised techniques employed with low sampling data in comparison to the many supervised techniques used for such cases. We extract a power interval at which falls the operation of the selected appliance along with a time vector for the values delimiting the state transitions of the appliance. After this, appliance signatures are formed from extracted power, geometrical and statistical features. Afterwards, those formed signatures are used to tune general model types for appliances identification using unsupervised algorithms. This method is evaluated using both simulated data on LPG and real-time Reference Energy Disaggregation Dataset (REDD). For that, we compute performance metrics using confusion matrix based metrics, considering accuracy, precision, recall and error-rate. The performance analysis of our methodology is then compared with other detection techniques previously used in the literature review, such as detection techniques based on statistical variations and abrupt changes (Variance Sliding Window and Cumulative Sum).

Keywords: electrical disaggregation, DTW, general appliance modeling, event detection

Procedia PDF Downloads 49
12583 Vibration Measurements of Single-Lap Cantilevered SPR Beams

Authors: Xiaocong He

Abstract:

Self-pierce riveting (SPR) is a new high-speed mechanical fastening technique which is suitable for point joining dissimilar sheet materials, as well as coated and pre-painted sheet materials. Mechanical structures assembled by SPR are expected to possess a high damping capacity. In this study, experimental measurement techniques were proposed for the prediction of vibration behavior of single-lap cantilevered SPR beams. The dynamic test software and the data acquisition hardware were used in the experimental measurement of the dynamic response of the single-lap cantilevered SPR beams. Free and forced vibration behavior of the single-lap cantilevered SPR beams was measured using the LMS CADA-X experimental modal analysis software and the LMS-DIFA Scadas II data acquisition hardware. The frequency response functions of the SPR beams of different rivet number were compared. The main goal of the paper is to provide a basic measuring method for further research on vibration based non-destructive damage detection in single-lap cantilevered SPR beams.

Keywords: self-piercing riveting, dynamic response, experimental measurement, frequency response functions

Procedia PDF Downloads 406