Search results for: unit testing
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5151

Search results for: unit testing

5001 Testing Method of Soil Failure Pattern of Sand Type as an Effort to Minimize the Impact of the Earthquake

Authors: Luthfi Assholam Solamat

Abstract:

Nowadays many people do not know the soil failure pattern as an important part in planning the under structure caused by the loading occurs. This is because the soil is located under the foundation, so it cannot be seen directly. Based on this study, the idea occurs to do a study for testing the soil failure pattern, especially the type of sand soil under the foundation. The necessity of doing this to the design of building structures on the land which is the initial part of the foundation structure that met with waves/vibrations during an earthquake. If the underground structure is not strong it is feared the building thereon more vulnerable to the risk of building damage. This research focuses on the search of soil failure pattern, which the most applicable in the field with the loading periodic re-testing of a particular time with the help of the integrated video visual observations performed. The results could be useful for planning under the structure in an effort to try the upper structure is minimal risk of the earthquake.

Keywords: soil failure pattern, earthquake, under structure, sand soil testing method

Procedia PDF Downloads 358
5000 JavaScript Object Notation Data against eXtensible Markup Language Data in Software Applications a Software Testing Approach

Authors: Theertha Chandroth

Abstract:

This paper presents a comparative study on how to check JSON (JavaScript Object Notation) data against XML (eXtensible Markup Language) data from a software testing point of view. JSON and XML are widely used data interchange formats, each with its unique syntax and structure. The objective is to explore various techniques and methodologies for validating comparison and integration between JSON data to XML and vice versa. By understanding the process of checking JSON data against XML data, testers, developers and data practitioners can ensure accurate data representation, seamless data interchange, and effective data validation.

Keywords: XML, JSON, data comparison, integration testing, Python, SQL

Procedia PDF Downloads 140
4999 Evaluation of the Quality of Care for Premature Babies in the Neonatology Unit of the Centre Hospitalier Universitaire de Kamenge

Authors: Kankurize Josiane, Nizigama Mediatrice

Abstract:

Introduction: Burundi records a still high infant mortality rate. Despite efforts to reduce it, prematurity is still the leading cause of death in the neonatal period. The objective of this study was to assess the quality of care for premature babies hospitalized in the neonatology unit of the Centre Hospitalier Universitaire de Kamenge. Method: This was a descriptive and evaluative prospective carried out in the neonatology unit of the CHUK (Centre Hospitalier Universitaire de Kamenge) from December 1, 2016, to May 31, 2017, including 70 premature babies, 65 mothers of premature babies and 15 providers including a pediatrician and 14 nurses. Using a tool developed by the World Health Organization and adapted to the local context by national experts, the quality of care for premature babies was assessed. Results: Prematurity accounted for 44.05% of hospitalizations in neonatology at the University Hospital of Kamenge. The assessment of the quality of care for premature babies was of low quality, with an average global score of 2/5 (50%), indicating that there is a considerable need for improvement to reach the standards. Conclusion: Efforts must be made to have infrastructures, materials, and human resources sufficient in quality and quantity so that the neonatology unit of the CHUK can be efficient and optimize the care of premature babies.

Keywords: quality of care, evaluation, premature, standards

Procedia PDF Downloads 60
4998 Infrared Detection Device for Accurate Scanning 3D Objects

Authors: Evgeny A. Rybakov, Dmitry P. Starikov

Abstract:

This article contains information about creating special unit for scanning 3D objects different nature, different materials, for example plastic, plaster, cardboard, wood, metal and etc. The main part of the unit is infrared transducer, which is sends the wave to the object and receive back wave for calculating distance. After that, microcontroller send to PC data, and computer program create model for printing from the plastic, gypsum, brass, etc.

Keywords: clutch, infrared, microcontroller, plastic, shaft, stage

Procedia PDF Downloads 443
4997 Efficient Control of Some Dynamic States of Wheeled Robots

Authors: Boguslaw Schreyer

Abstract:

In some types of wheeled robots it is important to secure starting acceleration and deceleration maxima while at the same time maintaining transversal stability. In this paper torque distribution between the front and rear wheels as well as the timing of torque application have been calculated. Both secure an optimum traction coefficient. This paper also identifies required input signals to a control unit, which controls the torque values and timing. Using a three dimensional, two mass model of a robot developed by the author a computer simulation was performed confirming the calculations presented in this paper. These calculations were also implemented and confirmed during military robot testing.

Keywords: robot dynamics, torque distribution, traction coefficient, wheeled robots

Procedia PDF Downloads 312
4996 Test Suite Optimization Using an Effective Meta-Heuristic BAT Algorithm

Authors: Anuradha Chug, Sunali Gandhi

Abstract:

Regression Testing is a very expensive and time-consuming process carried out to ensure the validity of modified software. Due to the availability of insufficient resources to re-execute all the test cases in time constrained environment, efforts are going on to generate test data automatically without human efforts. Many search based techniques have been proposed to generate efficient, effective as well as optimized test data, so that the overall cost of the software testing can be minimized. The generated test data should be able to uncover all potential lapses that exist in the software or product. Inspired from the natural behavior of bat for searching her food sources, current study employed a meta-heuristic, search-based bat algorithm for optimizing the test data on the basis certain parameters without compromising their effectiveness. Mathematical functions are also applied that can effectively filter out the redundant test data. As many as 50 Java programs are used to check the effectiveness of proposed test data generation and it has been found that 86% saving in testing efforts can be achieved using bat algorithm while covering 100% of the software code for testing. Bat algorithm was found to be more efficient in terms of simplicity and flexibility when the results were compared with another nature inspired algorithms such as Firefly Algorithm (FA), Hill Climbing Algorithm (HC) and Ant Colony Optimization (ACO). The output of this study would be useful to testers as they can achieve 100% path coverage for testing with minimum number of test cases.

Keywords: regression testing, test case selection, test case prioritization, genetic algorithm, bat algorithm

Procedia PDF Downloads 380
4995 New Desiccant Solar Unit for Air Conditioning and Desalination: Study of the Compartments of Desalination and Water Condensation

Authors: Zied Guidara, Alexander Morgenstern, Aref Maalej

Abstract:

In this paper, a new desiccant solar unit for air conditioning and desalination is presented first. Secondly, a dynamic modelling study of the desiccant wheel is developed. After that, a simulation study and an experimental investigation of the behaviour of the desiccant wheel are developed. The experimental investigation is done in the chamber of commerce in Freiburg-Germany. Indeed, the variations of calculated and measured temperatures and specific humidity of dehumidified and rejected air are presented where a good agreement is found when comparing the model predictions with experimental data under the considered range of operating conditions. Finally, the study of the compartments of desalination and water condensation shows that the unit can produce an acceptable quantity of water at the same time of the air conditioning operation.

Keywords: air conditioning, desalination, condensation, design, desiccant wheel, modelling, experimental investigation

Procedia PDF Downloads 385
4994 Predictive Analytics for Theory Building

Authors: Ho-Won Jung, Donghun Lee, Hyung-Jin Kim

Abstract:

Predictive analytics (data analysis) uses a subset of measurements (the features, predictor, or independent variable) to predict another measurement (the outcome, target, or dependent variable) on a single person or unit. It applies empirical methods in statistics, operations research, and machine learning to predict the future, or otherwise unknown events or outcome on a single or person or unit, based on patterns in data. Most analyses of metabolic syndrome are not predictive analytics but statistical explanatory studies that build a proposed model (theory building) and then validate metabolic syndrome predictors hypothesized (theory testing). A proposed theoretical model forms with causal hypotheses that specify how and why certain empirical phenomena occur. Predictive analytics and explanatory modeling have their own territories in analysis. However, predictive analytics can perform vital roles in explanatory studies, i.e., scientific activities such as theory building, theory testing, and relevance assessment. In the context, this study is to demonstrate how to use our predictive analytics to support theory building (i.e., hypothesis generation). For the purpose, this study utilized a big data predictive analytics platform TM based on a co-occurrence graph. The co-occurrence graph is depicted with nodes (e.g., items in a basket) and arcs (direct connections between two nodes), where items in a basket are fully connected. A cluster is a collection of fully connected items, where the specific group of items has co-occurred in several rows in a data set. Clusters can be ranked using importance metrics, such as node size (number of items), frequency, surprise (observed frequency vs. expected), among others. The size of a graph can be represented by the numbers of nodes and arcs. Since the size of a co-occurrence graph does not depend directly on the number of observations (transactions), huge amounts of transactions can be represented and processed efficiently. For a demonstration, a total of 13,254 metabolic syndrome training data is plugged into the analytics platform to generate rules (potential hypotheses). Each observation includes 31 predictors, for example, associated with sociodemographic, habits, and activities. Some are intentionally included to get predictive analytics insights on variable selection such as cancer examination, house type, and vaccination. The platform automatically generates plausible hypotheses (rules) without statistical modeling. Then the rules are validated with an external testing dataset including 4,090 observations. Results as a kind of inductive reasoning show potential hypotheses extracted as a set of association rules. Most statistical models generate just one estimated equation. On the other hand, a set of rules (many estimated equations from a statistical perspective) in this study may imply heterogeneity in a population (i.e., different subpopulations with unique features are aggregated). Next step of theory development, i.e., theory testing, statistically tests whether a proposed theoretical model is a plausible explanation of a phenomenon interested in. If hypotheses generated are tested statistically with several thousand observations, most of the variables will become significant as the p-values approach zero. Thus, theory validation needs statistical methods utilizing a part of observations such as bootstrap resampling with an appropriate sample size.

Keywords: explanatory modeling, metabolic syndrome, predictive analytics, theory building

Procedia PDF Downloads 276
4993 Time Travel Testing: A Mechanism for Improving Renewal Experience

Authors: Aritra Majumdar

Abstract:

While organizations strive to expand their new customer base, retaining existing relationships is a key aspect of improving overall profitability and also showcasing how successful an organization is in holding on to its customers. It is an experimentally proven fact that the lion’s share of profit always comes from existing customers. Hence seamless management of renewal journeys across different channels goes a long way in improving trust in the brand. From a quality assurance standpoint, time travel testing provides an approach to both business and technology teams to enhance the customer experience when they look to extend their partnership with the organization for a defined phase of time. This whitepaper will focus on key pillars of time travel testing: time travel planning, time travel data preparation, and enterprise automation. Along with that, it will call out some of the best practices and common accelerator implementation ideas which are generic across verticals like healthcare, insurance, etc. In this abstract document, a high-level snapshot of these pillars will be provided. Time Travel Planning: The first step of setting up a time travel testing roadmap is appropriate planning. Planning will include identifying the impacted systems that need to be time traveled backward or forward depending on the business requirement, aligning time travel with other releases, frequency of time travel testing, preparedness for handling renewal issues in production after time travel testing is done and most importantly planning for test automation testing during time travel testing. Time Travel Data Preparation: One of the most complex areas in time travel testing is test data coverage. Aligning test data to cover required customer segments and narrowing it down to multiple offer sequencing based on defined parameters are keys for successful time travel testing. Another aspect is the availability of sufficient data for similar combinations to support activities like defect retesting, regression testing, post-production testing (if required), etc. This section will talk about the necessary steps for suitable data coverage and sufficient data availability from a time travel testing perspective. Enterprise Automation: Time travel testing is never restricted to a single application. The workflow needs to be validated in the downstream applications to ensure consistency across the board. Along with that, the correctness of offers across different digital channels needs to be checked in order to ensure a smooth customer experience. This section will talk about the focus areas of enterprise automation and how automation testing can be leveraged to improve the overall quality without compromising on the project schedule. Along with the above-mentioned items, the white paper will elaborate on the best practices that need to be followed during time travel testing and some ideas pertaining to accelerator implementation. To sum it up, this paper will be written based on the real-time experience author had on time travel testing. While actual customer names and program-related details will not be disclosed, the paper will highlight the key learnings which will help other teams to implement time travel testing successfully.

Keywords: time travel planning, time travel data preparation, enterprise automation, best practices, accelerator implementation ideas

Procedia PDF Downloads 159
4992 Characterization of Martensitic Stainless Steel Japanese Grade AISI 420A

Authors: T. Z. Butt, T. A. Tabish, K. Anjum, H. Hafeez

Abstract:

A study of martensitic stainless steel surgical grade AISI 420A produced in Japan was carried out in this research work. The sample was already annealed at about 898˚C. The sample were subjected to chemical analysis, hardness, tensile and metallographic tests. These tests were performed on as received annealed and heat treated samples. In the annealed condition the sample showed 0HRC. However, on tensile testing, in annealed condition the sample showed maximum elongation. The heat treatment is carried out in vacuum furnace within temperature range 980-1035°C. The quenching of samples was carried out using liquid nitrogen. After hardening, the samples were subjected to tempering, which was carried out in vacuum tempering furnace at a temperature of 220˚C. The hardened samples were subjected to hardness and tensile testing. In hardness testing, the samples showed maximum hardness values. In tensile testing the sample showed minimum elongation. The sample in annealed state showed coarse plates of martensite structure. Therefore, the studied steels can be used as biomaterials.

Keywords: biomaterials, martensitic steel, microsrtucture, tensile testing, hardening, tempering, bioinstrumentation

Procedia PDF Downloads 277
4991 Analysis of the Interventions Performed in Pediatric Cardiology Unit Based on Nursing Interventions Classification (NIC-6th): A Pilot Study

Authors: Ji Wen Sun, Nan Ping Shen, Yi Bei Wu

Abstract:

This study used Nursing Interventions Classification (NIC-6th) to identify the interventions performed in a pediatric cardiology unit, and then to analysis its frequency, time and difficulty, so as to give a brief review on what our nurses have done. The research team selected a 35 beds pediatric cardiology unit, and drawn all the nursing interventions in the nursing record from our hospital information system (HIS) from 1 October 2015 to 30 November 2015, using NIC-6th to do the matching and then counting their frequencies. Then giving each intervention its own time and difficulty code according to NIC-6th. The results showed that nurses in pediatric cardiology unit performed totally 43 interventions from 5394 statements, and most of them were in RN(basic) education level needed and less than 15 minutes time needed. There still had some interventions just needed by a nursing assistant but done by nurses, which should call for nurse managers to think about the suitable staffing. Thus, counting the summary of the product of frequency, time and difficulty for each intervention of each nurse can know one's performance. Acknowledgement Clinical Management Optimization Project of Shanghai Shen Kang Hospital Development Center (SHDC2014615); Hundred-Talent Program of Construction of Nursing Plateau Discipline (hlgy16073qnhb).

Keywords: nursing interventions, nursing interventions classification, nursing record, pediatric cardiology

Procedia PDF Downloads 364
4990 Nurse-Identified Barriers and Facilitators to Delivering End-of-Life Care in a Cardiac Intensive Care Unit: A Qualitative Study

Authors: Elena Ivany, Leanne Aitken

Abstract:

Little is known about the delivery of end-of-life care in cardiac intensive care unit (CICU) settings. The aims of this study were to highlight the nurse-identified barriers and facilitators to delivering end-of-life care in the CICU, and to identify whether any of the barriers and/or facilitators are specific to the CICU setting. This was an exploratory qualitative study utilizing semi-structured individual interviews as the data collection method and inductive thematic analysis to structure the data. Six CICU nurses took part in the study. Five key themes were identified, each theme including both barriers and facilitators. The five key themes are as follows: patient-centered care, emotional challenges, reaching concordance, nursing contribution and the surgical intensive care unit.

Keywords: end-of-life, cardiovascular disease, cardiac surgery, critical care

Procedia PDF Downloads 265
4989 On the Resilience of Operational Technology Devices in Penetration Tests

Authors: Marko Schuba, Florian Kessels, Niklas Reitz

Abstract:

Operational technology (OT) controls physical processes in critical infrastructures and economically important industries. With the convergence of OT with classical information technology (IT), rising cybercrime worldwide and the increasingly difficult geopolitical situation, the risks of OT infrastructures being attacked are growing. Classical penetration testing, in which testers take on the role of an attacker, has so far found little acceptance in the OT sector - the risk that a penetration test could do more harm than good seems too great. This paper examines the resilience of various OT systems using typical penetration test tools. It is shown that such a test certainly involves risks, but is also feasible in OT if a cautious approach is taken. Therefore, OT penetration testing should be considered as a tool to improve the cyber security of critical infrastructures.

Keywords: penetration testing, OT, ICS, OT security

Procedia PDF Downloads 15
4988 Navigating Cyber Attacks with Quantum Computing: Leveraging Vulnerabilities and Forensics for Advanced Penetration Testing in Cybersecurity

Authors: Sayor Ajfar Aaron, Ashif Newaz, Sajjat Hossain Abir, Mushfiqur Rahman

Abstract:

This paper examines the transformative potential of quantum computing in the field of cybersecurity, with a focus on advanced penetration testing and forensics. It explores how quantum technologies can be leveraged to identify and exploit vulnerabilities more efficiently than traditional methods and how they can enhance the forensic analysis of cyber-attacks. Through theoretical analysis and practical simulations, this study highlights the enhanced capabilities of quantum algorithms in detecting and responding to sophisticated cyber threats, providing a pathway for developing more resilient cybersecurity infrastructures.

Keywords: cybersecurity, cyber forensics, penetration testing, quantum computing

Procedia PDF Downloads 67
4987 Business Continuity Risk Review for a Large Petrochemical Complex

Authors: Michel A. Thomet

Abstract:

A discrete-event simulation model was used to perform a Reliability-Availability-Maintainability (RAM) study of a large petrochemical complex which included sixteen process units, and seven feeds and intermediate streams. All the feeds and intermediate streams have associated storage tanks, so that if a processing unit fails and shuts down, the downstream units can keep producing their outputs. This also helps the upstream units which do not have to reduce their outputs, but can store their excess production until the failed unit restart. Each process unit and each pipe section carrying the feeds and intermediate streams has a probability of failure with an associated distribution and a Mean Time Between Failure (MTBF), as well as a distribution of the time to restore and a Mean Time To Restore (MTTR). The utilities supporting the process units can also fail and have their own distributions with specific MTBF and MTTR. The model runs are for ten years or more and the runs are repeated several times to obtain statistically relevant results. One of the main results is the On-Stream factor (OSF) of each process unit (percent of hours in a year when the unit is running in nominal conditions). One of the objectives of the study was to investigate if the storage capacity of each of the feeds and the intermediate stream was adequate. This was done by increasing the storage capacities in several steps and through running the simulation to see if the OSF were improved and by how much. Other objectives were to see if the failure of the utilities were an important factor in the overall OSF, and what could be done to reduce their failure rates through redundant equipment.

Keywords: business continuity, on-stream factor, petrochemical, RAM study, simulation, MTBF

Procedia PDF Downloads 219
4986 Reducing the Incidence of Hyperphosphatemia in Patients Receiving Dialysis

Authors: Tsai Su Hui

Abstract:

Background: Hyperphosphatemia in patients receiving dialysis can cause hyperparathyroidism, which can lead to renal osteodystrophy, cardiovascular disease and mortality. Data showed that 26% of patients receiving dialysis had blood phosphate levels of >6.0 mg/dl at this unit from January to March 2017, higher than the Taiwan Society of Nephrology evaluation criteria of < 20%. After analysis, possible reasons included: 1. Incomprehensive education for nurse and lack of relevant training. 2. Insufficient assistive aids for nursing health education instruction. 3. Patients were unsure which foods are high or low in phosphate. 4. Patients did not have habits of taking medicine with them and how to correctly administer the medication. Purpose: To reduce the percentage of patients receiving dialysis with blood phosphate levels of >6.0 mg/dl to less than 20% at this unit. Method: (1) Improve understanding of hyperphosphatemia and food for patients receiving dialysis and their families, (2) Acquire more nursing instruction assistive aids and improve knowledge of hyperphosphatemia for nurse. Results: After implementing the project, the percentage of patients receiving dialysis with blood phosphate levels of >6.0 mg/dl decreased from 26.0% to 18.8% at this unit. By implementing the project, the professional skills of nurse improved, blood phosphate levels of patients receiving dialysis were reduced, and the quality of care for patients receiving dialysis at this unit was enhanced.

Keywords: hemodialysis, hyperphosphatemia, incidence, reducing

Procedia PDF Downloads 126
4985 Determination of Physical Properties of Crude Oil Distillates by Near-Infrared Spectroscopy and Multivariate Calibration

Authors: Ayten Ekin Meşe, Selahattin Şentürk, Melike Duvanoğlu

Abstract:

Petroleum refineries are a highly complex process industry with continuous production and high operating costs. Physical separation of crude oil starts with the crude oil distillation unit, continues with various conversion and purification units, and passes through many stages until obtaining the final product. To meet the desired product specification, process parameters are strictly followed. To be able to ensure the quality of distillates, routine analyses are performed in quality control laboratories based on appropriate international standards such as American Society for Testing and Materials (ASTM) standard methods and European Standard (EN) methods. The cut point of distillates in the crude distillation unit is very crucial for the efficiency of the upcoming processes. In order to maximize the process efficiency, the determination of the quality of distillates should be as fast as possible, reliable, and cost-effective. In this sense, an alternative study was carried out on the crude oil distillation unit that serves the entire refinery process. In this work, studies were conducted with three different crude oil distillates which are Light Straight Run Naphtha (LSRN), Heavy Straight Run Naphtha (HSRN), and Kerosene. These products are named after separation by the number of carbons it contains. LSRN consists of five to six carbon-containing hydrocarbons, HSRN consist of six to ten, and kerosene consists of sixteen to twenty-two carbon-containing hydrocarbons. Physical properties of three different crude distillation unit products (LSRN, HSRN, and Kerosene) were determined using Near-Infrared Spectroscopy with multivariate calibration. The absorbance spectra of the petroleum samples were obtained in the range from 10000 cm⁻¹ to 4000 cm⁻¹, employing a quartz transmittance flow through cell with a 2 mm light path and a resolution of 2 cm⁻¹. A total of 400 samples were collected for each petroleum sample for almost four years. Several different crude oil grades were processed during sample collection times. Extended Multiplicative Signal Correction (EMSC) and Savitzky-Golay (SG) preprocessing techniques were applied to FT-NIR spectra of samples to eliminate baseline shifts and suppress unwanted variation. Two different multivariate calibration approaches (Partial Least Squares Regression, PLS and Genetic Inverse Least Squares, GILS) and an ensemble model were applied to preprocessed FT-NIR spectra. Predictive performance of each multivariate calibration technique and preprocessing techniques were compared, and the best models were chosen according to the reproducibility of ASTM reference methods. This work demonstrates the developed models can be used for routine analysis instead of conventional analytical methods with over 90% accuracy.

Keywords: crude distillation unit, multivariate calibration, near infrared spectroscopy, data preprocessing, refinery

Procedia PDF Downloads 129
4984 Rule-Based Expert System for Headache Diagnosis and Medication Recommendation

Authors: Noura Al-Ajmi, Mohammed A. Almulla

Abstract:

With the increased utilization of technology devices around the world, healthcare and medical diagnosis are critical issues that people worry about these days. Doctors are doing their best to avoid any medical errors while diagnosing diseases and prescribing the wrong medication. Subsequently, artificial intelligence applications that can be installed on mobile devices such as rule-based expert systems facilitate the task of assisting doctors in several ways. Due to their many advantages, the usage of expert systems has increased recently in health sciences. This work presents a backward rule-based expert system that can be used for a headache diagnosis and medication recommendation system. The structure of the system consists of three main modules, namely the input unit, the processing unit, and the output unit.

Keywords: headache diagnosis system, prescription recommender system, expert system, backward rule-based system

Procedia PDF Downloads 215
4983 Correlates of Peer Influence and Resistance to HIV/AIDS Counselling and Testing among Students in Tertiary Institutions in Kano State, Nigeria

Authors: A. S. Haruna, M. U. Tambawal, A. A. Salawu

Abstract:

The psychological impact of peer influence on its individual group members, can make them resist HIV/AIDS counselling and testing. This study investigated the correlate of peer influence and resistance to HIV/AIDS counselling and testing among students in tertiary institutions in Kano state, Nigeria. To achieve this, three null hypotheses were postulated and tested. Cross-Sectional Survey Design was employed in which 1512 sample was selected from a student population of 104,841.Simple Random Sampling was used in the selection. A self-developed 20-item scale called Peer Influence and Psychological Resistance Inventory (PIPRI) was used for data collection. Pearson Product Moment Correlation (PPMCC) via test-retest method was applied to estimate a reliability coefficient of 0.86 for the scale. Data obtained was analyzed using t-test and PPMCC at 0.05 level of confidence. Results reveal 26.3% (397) of the respondents being influenced by their peer group, while 39.8% showed resistance. Also, the t-tests and PPMCC statistics were greater than their respective critical values. This shows that there was a significant gender difference in peer influence and a difference between peer influence and resistance to HIV/AIDS counselling and testing. However, a positive relationship between peer influence and resistance to HIV/AIDS counselling and testing was shown. A major recommendation offered suggests the use of reinforcement and social support for positive attitudes and maintenance of safe behaviour among students who patronize HIV/AIDS counselling.

Keywords: peer group influence, HIV/AIDS counselling and testing, psychological resistance, students

Procedia PDF Downloads 390
4982 New Scheme of Control and Air Supply in a Low-Power Hot Water Boiler

Authors: М. Zh. Khazimov, А. B. Kudasheva

Abstract:

The article presents the state of solid fuel reserves and their share in the world energy sector. The air pollution caused by the operation of heating devices using solid fuels is a significant problem. In order to improve the air quality, heating device producers take constant measures to improve their products. However, the emission results achieved during an initial test of heating devices in the laboratory may be much worse during operation in real operating conditions. The ways of increasing the efficiency of the boiler by improving its design for combustion in full-layer mode are shown. The results of the testing of the improved КВТС-0.2 hot water boiler is presented and the technical and economic indicators are determined, which indicate an increase in the efficiency of the boiler.

Keywords: boiler unit, grate, furnace, coal, ash

Procedia PDF Downloads 70
4981 The Development and Testing of Greenhouse Comprehensive Environment Control System

Authors: Mohammed Alrefaie, Yaser Miaji

Abstract:

Greenhouses provide a convenient means to grow plants in the best environment. They achieve this by trapping heat from the sunlight and using artificial means to enhance the environment of the greenhouse. This includes controlling factors such as air flow, light intensity and amount of water among others that can have a big impact on plant growth. The aim of the greenhouse is to give maximum yield from plants possible. This report details the development and testing of greenhouse environment control system that can regulate light intensity, airflow and power supply inside the greenhouse. The details of the module development to control these three factors along with results of testing are presented.

Keywords: greenhouse, control system, light intensity, comprehensive environment

Procedia PDF Downloads 482
4980 Use of Chlorophyll Meters to Assess In-Season Wheat Nitrogen Fertilizer Requirements in the Southern San Joaquin Valley

Authors: Brian Marsh

Abstract:

Nitrogen fertilizer is the most used and often the most mismanaged nutrient input. Nitrogen management has tremendous implications on crop productivity, quality and environmental stewardship. Sufficient nitrogen is needed to optimum yield and quality. Soil and in-season plant tissue testing for nitrogen status are a time consuming and expensive process. Real time sensing of plant nitrogen status can be a useful tool in managing nitrogen inputs. The objectives of this project were to assess the reliability of remotely sensed non-destructive plant nitrogen measurements compared to wet chemistry data from sampled plant tissue, develop in-season nitrogen recommendations based on remotely sensed data for improved nitrogen use efficiency and assess the potential for determining yield and quality from remotely sensed data. Very good correlations were observed between early-season remotely sensed crop nitrogen status and plant nitrogen concentrations and subsequent in-season fertilizer recommendations. The transmittance/absorbance type meters gave the most accurate readings. Early in-season fertilizer recommendation would be to apply 40 kg nitrogen per hectare plus 16 kg nitrogen per hectare for each unit difference measured with the SPAD meter between the crop and reference area or 25 kg plus 13 kg per hectare for each unit difference measured with the CCM 200. Once the crop was sufficiently fertilized meter readings became inconclusive and were of no benefit for determining nitrogen status, silage yield and quality and grain yield and protein.

Keywords: wheat, nitrogen fertilization, chlorophyll meter

Procedia PDF Downloads 393
4979 Studying the Relationship Between Washback Effects of IELTS Test on Iranian Language Teachers, Teaching Strategies and Candidates

Authors: Afsaneh Jasmine Majidi

Abstract:

Language testing is an important part of language teaching experience and language learning process as it presents assessment strategies for teachers to evaluate the efficiency of teaching and for learners to examine their outcomes. However, language testing is demanding and challenging because it should provide the opportunity for proper and objective decision. In addition to all the efforts test designers put to design valid and reliable tests, there are some other determining factors which are even more complex and complicated. These factors affect the educational system, individuals, and society, and the impact of the tests vary according to the scope of the test. Seemingly, the impact of a simple classroom assessment is not the same as that of high stake tests such as International English Language Testing System (IELTS). As the importance of the test increases, it affects wider domain. Accordingly, the impacts of high stake tests are reflected not only in teaching, learning strategies but also in society. Testing experts use the term ‘washback’ or ‘impact’ to define the different effects of a test on teaching, learning, and community. This paper first looks at the theoretical background of ‘washback’ and ‘impact’ in language testing by reviewing of relevant literature in the field and then investigates washback effects of IELTS test of on Iranian IELTS teachers and students. The study found significant relationship between the washback effect of IELTS test and teaching strategies of Iranian IELTS teachers as well as performance of Iranian IELTS candidates and their community.

Keywords: high stake tests, IELTS, Iranian Candidates, language testing, test impact, washback

Procedia PDF Downloads 327
4978 Electricity Market Categorization for Smart Grid Market Testing

Authors: Rebeca Ramirez Acosta, Sebastian Lenhoff

Abstract:

Decision makers worldwide need to determine if the implementation of a new market mechanism will contribute to the sustainability and resilience of the power system. Due to smart grid technologies, new products in the distribution and transmission system can be traded; however, the impact of changing a market rule will differ between several regions. To test systematically those impacts, a market categorization has been compiled and organized in a smart grid market testing toolbox. This toolbox maps all actual energy products and sets the basis for running a co-simulation test with the new rule to be implemented. It will help to measure the impact of the new rule, based on the sustainable and resilience indicators.

Keywords: co-simulation, electricity market, smart grid market, market testing

Procedia PDF Downloads 189
4977 Structural Evaluation of Airfield Pavement Using Finite Element Analysis Based Methodology

Authors: Richard Ji

Abstract:

Nondestructive deflection testing has been accepted widely as a cost-effective tool for evaluating the structural condition of airfield pavements. Backcalculation of pavement layer moduli can be used to characterize the pavement existing condition in order to compute the load bearing capacity of pavement. This paper presents an improved best-fit backcalculation methodology based on deflection predictions obtained using finite element method (FEM). The best-fit approach is based on minimizing the squared error between falling weight deflectometer (FWD) measured deflections and FEM predicted deflections. Then, concrete elastic modulus and modulus of subgrade reaction were back-calculated using Heavy Weight Deflectometer (HWD) deflections collected at the National Airport Pavement Testing Facility (NAPTF) test site. It is an alternative and more versatile method in considering concrete slab geometry and HWD testing locations compared to methods currently available.

Keywords: nondestructive testing, pavement moduli backcalculation, finite element method, concrete pavements

Procedia PDF Downloads 166
4976 Political Determinants of Sovereign Spread: The Great East-West Divide

Authors: Maruska Vizek, Josip Glaurdic, Marina Tkalec, Goran Vuksic

Abstract:

We empirically explore whether and how taxation affects bilateral real exchange rates in the euro area – relative unit labor costs and relative consumer price indices. We find that employers’ social security contributions and the value added tax changes have the expected effects put forward in the fiscal devaluation literature and simulations. Increases in employers’ contributions appreciate the relative unit labor costs in the short- and the long-run, while value added tax hike appreciates the relative consumer prices. Somewhat surprisingly, for personal income tax increases, we find a short-run depreciating impact on the relative unit labor costs, while increases in employees’ contributions depreciate both measures of real exchange rates in the short-run.

Keywords: sovereign bonds, European Union, developing countries, political determinants

Procedia PDF Downloads 311
4975 Implant Guided Surgery and Immediate Loading

Authors: Omid Tavakol, Mahnaz Gholami

Abstract:

Introduction : In this oral presentation the main goal is discussing immediate loading in dental implants , from treatment planning and surgical guide designing to delivery , follow up and occlusal consideration . Methods and materials : first of all systematic reviews about immediate loading will be considered . besides , a comparison will be made between immediate loading and conventional loading in terms of success rate and complications . After that different methods , prosthetic options and materials best used in immediate loading will be explained. Particularly multi unit abutments and their mechanism of function will be explained .Digital impressions and designing the temporaries is the next topic we are to explicate .Next issue is the differences between single unit , multiple unit and full arch implantation in immediate loading .Following we are going to describe methods for tissue engineering and papilla formation after extraction . Last slides are about a full mouth rehabilitation via immediate loading technique from surgical designing to follow up .At the end we would talk about potential complications , how to prevent from occurrence and what to do if we face up with .

Keywords: guided surgery, digital implantology, immediate loading, digital dentistry

Procedia PDF Downloads 44
4974 Evaluation of JCI Accreditation for Medical Technology in Saudi Arabian Hospitals: A Study Case of PSMMC

Authors: Hamad Albadr

Abstract:

Joint Commission International (JCI) accreditation process intent to improve the safety and quality of care in the international community through the provision of education, publications, consultation, and evaluation services. These standards apply to the entire organization as well as to each department, unit, or service within the organization. Medical Technology that contains both medical equipment and devices, is an essential part of health care. Appropriate management of equipment maintenance for ensuring medical technology safe, the equipment life is maximized, and the total costs are minimized. JCI medical technology evaluation and accreditation use standards, intents, and measurable elements. The paper focuses on evaluation of JCI standards for medical technology in Saudi Arabian hospitals: a Study Case of PSMMC that define the performance expectation, structures, or functions that must be in place for a hospital to be accredited by JCI through measurable elements that indicate a score during the survey process that identify the requirements for full compliance with the standard specially through Facility Management and Safety (FMS) section that require the hospital establishes and implements a program for inspecting, testing, and maintaining medical technology and documenting the results, to ensure that medical technology is available for use and functioning properly, the hospital performs and documents; an inventory of medical technology; regular inspections of medical technology; testing of medical technology according to its use and manufacturers’ requirements; and performance of preventive maintenance.

Keywords: joint commission international (JCI) accreditation, medical technology, Saudi Arabia, Saudi Arabian hospitals

Procedia PDF Downloads 557
4973 Reducing The Frequency of Flooding Accompanied by Low pH Wastewater In 100/200 Unit of Phosphate Fertilizer 1 Plant by Implementing The 3R Program (Reduce, Reuse and Recycle)

Authors: Pradipta Risang Ratna Sambawa, Driya Herseta, Mahendra Fajri Nugraha

Abstract:

In 2020, PT Petrokimia Gresik implemented a program to increase the ROP (Run Of Pile) production rate at the Phosphate Fertilizer 1 plant, causing an increase in scrubbing water consumption in the 100/200 area unit. This increase in water consumption causes a higher discharge of wastewater, which can further cause local flooding, especially during the rainy season. The 100/200 area of the Phosphate Fertilizer 1 plant is close to the warehouse and is often a passing area for trucks transporting raw materials. This causes the pH in the wastewater to become acidic (the worst point is up to pH 1). The problem of flooding and exposure to acidic wastewater in the 100/200 area of Phosphate Fertilizer Plant 1 was then resolved by PT Petrokimia Gresik through wastewater optimization steps called the 3R program (Reduce, Reuse, and Recycle). The 3R (Reduce, reuse, and recycle) program consists of an air consumption reduction program by considering the liquid/gas ratio in scrubbing unit of 100/200 Phosphate Fertilizer 1 plant, creating a wastewater interconnection line so that wastewater from unit 100/200 can be used as scrubbing water in the Phonska 1, Phonska 2, Phonska 3 and unit 300 Phosphate Fertilizer 1 plant and increasing scrubbing effectiveness through scrubbing effectiveness simulations. Through a series of wastewater optimization programs, PT Petrokimia Gresik has succeeded in reducing NaOH consumption for neutralization up to 2,880 kg/day or equivalent in saving up to 314,359.76 dollars/year and reducing process water consumption up to 600 m3/day or equivalent in saving up to 63,739.62 dollars/year.

Keywords: fertilizer, phosphate fertilizer, wastewater, wastewater treatment, water management

Procedia PDF Downloads 26
4972 A Multi-Release Software Reliability Growth Models Incorporating Imperfect Debugging and Change-Point under the Simulated Testing Environment and Software Release Time

Authors: Sujit Kumar Pradhan, Anil Kumar, Vijay Kumar

Abstract:

The testing process of the software during the software development time is a crucial step as it makes the software more efficient and dependable. To estimate software’s reliability through the mean value function, many software reliability growth models (SRGMs) were developed under the assumption that operating and testing environments are the same. Practically, it is not true because when the software works in a natural field environment, the reliability of the software differs. This article discussed an SRGM comprising change-point and imperfect debugging in a simulated testing environment. Later on, we extended it in a multi-release direction. Initially, the software was released to the market with few features. According to the market’s demand, the software company upgraded the current version by adding new features as time passed. Therefore, we have proposed a generalized multi-release SRGM where change-point and imperfect debugging concepts have been addressed in a simulated testing environment. The failure-increasing rate concept has been adopted to determine the change point for each software release. Based on nine goodness-of-fit criteria, the proposed model is validated on two real datasets. The results demonstrate that the proposed model fits the datasets better. We have also discussed the optimal release time of the software through a cost model by assuming that the testing and debugging costs are time-dependent.

Keywords: software reliability growth models, non-homogeneous Poisson process, multi-release software, mean value function, change-point, environmental factors

Procedia PDF Downloads 74