Search results for: sensory evaluation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 6716

Search results for: sensory evaluation

1736 National Plans for Recovery and Resilience between National Recovery and EU Cohesion Objectives: Insights from European Countries

Authors: Arbolino Roberta, Boffardi Raffaele

Abstract:

Achieving the highest effectiveness for the National Plans for Recovery and Resilience (NPRR) while strengthening the objectives of cohesion and reduction of intra-EU unbalances is only possible by means of strategic, coordinated, and coherent policy planning. Therefore, the present research aims at assessing and quantifying the potential impact of NPRRs across the twenty-seven European Member States in terms of economic convergence, considering disaggregated data on industrial, construction, and service sectors. The first step of the research involves a performance analysis of the main macroeconomic indicators describing the trends of twenty-seven EU economies before the pandemic outbreak. Subsequently, in order to define the potential effect of the resources allocated, we perform an impact analysis of previous similar EU investment policies, estimating national-level sectoral elasticity associated with the expenditure of the 2007-2013 and 2014-2020 Cohesion programmes funds. These coefficients are then exploited to construct adjustment scenarios. Finally, convergence analysis is performed on the data used for constructing scenarios in order to understand whether the expenditure of funds might be useful to foster economic convergence besides driving recovery. The results of our analysis show that the allocation of resources largely mirrors the aims of the policy framework underlying the NPRR, thus reporting the largest investments in both those sectors most affected by the economic shock (services) and those considered fundamental for the digital and green transition. Notwithstanding an overall positive effect, large differences exist among European countries, while no convergence process seems to be activated or fostered by these interventions.

Keywords: NPRR, policy evaluation, cohesion policy, scenario Nalsysi

Procedia PDF Downloads 60
1735 Experimental Evaluation of Electrocoagulation for Hardness Removal of Bore Well Water

Authors: Pooja Kumbhare

Abstract:

Water is an important resource for the survival of life. The inadequate availability of surface water makes people depend on ground water for fulfilling their needs. However, ground water is generally too hard to satisfy the requirements for domestic as well as industrial applications. Removal of hardness involves various techniques such as lime soda process, ion exchange, reverse osmosis, nano-filtration, distillation, and, evaporation, etc. These techniques have individual problems such as high annual operating cost, sediment formation on membrane, sludge disposal problem, etc. Electrocoagulation (EC) is being explored as modern and cost-effective technology to cope up with the growing demand of high water quality at the consumer end. In general, earlier studies on electrocoagulation for hardness removal are found to deploy batch processes. As batch processes are always inappropriate to deal with large volume of water to be treated, it is essential to develop continuous flow EC process. So, in the present study, an attempt is made to investigate continuous flow EC process for decreasing excessive hardness of bore-well water. The experimental study has been conducted using 12 aluminum electrodes (25cm*10cm, 1cm thick) provided in EC reactor with volume of 8 L. Bore well water sample, collected from a local bore-well (i.e. at – Vishrambag, Sangli; Maharashtra) having average initial hardness of 680 mg/l (Range: 650 – 700 mg/l), was used for the study. Continuous flow electrocoagulation experiments were carried out by varying operating parameters specifically reaction time (Range: 10 – 60 min), voltage (Range: 5 – 20 V), current (Range: 1 – 5A). Based on the experimental study, it is found that hardness removal to the desired extent could be achieved even for continuous flow EC reactor, so the use of it is found promising.

Keywords: hardness, continuous flow EC process, aluminum electrode, optimal operating parameters

Procedia PDF Downloads 161
1734 Discussion as a Means to Improve Peer Assessment Accuracy

Authors: Jung Ae Park, Jooyong Park

Abstract:

Writing is an important learning activity that cultivates higher level thinking. Effective and immediate feedback is necessary to help improve students' writing skills. Peer assessment can be an effective method in writing tasks because it makes it possible for students not only to receive quick feedback on their writing but also to get a chance to examine different perspectives on the same topic. Peer assessment can be practiced frequently and has the advantage of immediate feedback. However, there is controversy about the accuracy of peer assessment. In this study, we tried to demonstrate experimentally how the accuracy of peer assessment could be improved. Participants (n=76) were randomly assigned to groups of 4 members. All the participant graded two sets of 4 essays on the same topic. They graded the first set twice, and the second set or the posttest once. After the first grading of the first set, each group in the experimental condition 1 (discussion group), were asked to discuss the results of the peer assessment and then to grade the essays again. Each group in the experimental condition 2 (reading group), were asked to read the assessment on each essay by an expert and then to grade the essays again. In the control group, the participants were asked to grade the 4 essays twice in different orders. Afterwards, all the participants graded the second set of 4 essays. The mean score from 4 participants was calculated for each essay. The accuracy of the peer assessment was measured by Pearson correlation with the scores of the expert. The results were analyzed by two-way repeated measure ANOVA. The main effect of grading was observed: Grading accuracy got better as the number of grading experience increased. Analysis of posttest accuracy revealed that the score variations within a group of 4 participants decreased in both discussion and reading conditions but not in the control condition. These results suggest that having students discuss their grading together can be an efficient means to improve peer assessment accuracy. By discussing, students can learn from others about what to consider in grading and whether their grading is too strict or lenient. Further research is needed to examine the exact cause of the grading accuracy.

Keywords: peer assessment, evaluation accuracy, discussion, score variations

Procedia PDF Downloads 251
1733 Assessment of the Electrical, Mechanical, and Thermal Nociceptive Thresholds for Stimulation and Pain Measurements at the Bovine Hind Limb

Authors: Samaneh Yavari, Christiane Pferrer, Elisabeth Engelke, Alexander Starke, Juergen Rehage

Abstract:

Background: Three nociceptive thresholds of thermal, electrical, and mechanical thresholds commonly use to evaluate the local anesthesia in many species, for instance, cow, horse, cat, dog, rabbit, and so on. Due to the lack of investigations to evaluate and/or validate such those nociceptive thresholds, our plan was the comparison of two-foot local anesthesia methods of Intravenous Regional Anesthesia (IVRA) and our modified four-point Nerve Block Anesthesia (NBA). Materials and Methods: Eight healthy nonpregnant nondairy Holstein Frisian cows in a cross-over study design were selected for this study. All cows divided into two different groups to receive two local anesthesia techniques of IVRA and our modified four-point NBA. Three thermal, electrical, and mechanical force and pinpricks were applied to evaluate the quality of local anesthesia methods before and after local anesthesia application. Results: The statistical evaluation demonstrated that our four-point NBA has a qualification to select as a standard foot local anesthesia. However, the recorded results of our study revealed no significant difference between two groups of local anesthesia techniques of IVRA and modified four-point NBA related to quality and duration of anesthesia stimulated by electrical, mechanical and thermal nociceptive stimuli. Conclusion and discussion: All three nociceptive threshold stimuli of electrical, mechanical and heat nociceptive thresholds can be applied to measure and evaluate the efficacy of foot local anesthesia of dairy cows. However, our study revealed no superiority of those three nociceptive methods to evaluate the duration and quality of bovine foot local anesthesia methods. Veterinarians to investigate the duration and quality of their selected anesthesia method can use any of those heat, mechanical, and electrical methods.

Keywords: mechanical, thermal, electrical threshold, IVRA, NBA, hind limb, dairy cow

Procedia PDF Downloads 229
1732 Clinical Prediction Rules for Using Open Kinetic Chain Exercise in Treatment of Knee Osteoarthritis

Authors: Mohamed Aly, Aliaa Rehan Youssef, Emad Sawerees, Mounir Guirgis

Abstract:

Relevance: Osteoarthritis (OA) is the most common degenerative disease seen in all populations. It causes disability and substantial socioeconomic burden. Evidence supports that exercise are the most effective conservative treatment for patients with OA. Therapists experience and clinical judgment play major role in exercise prescription and scientific evidence for this regard is lacking. The development of clinical prediction rules to identify patients who are most likely benefit from exercise may help solving this dilemma. Purpose: This study investigated whether body mass index and functional ability at baseline can predict patients’ response to a selected exercise program. Approach: Fifty-six patients, aged 35 to 65 years, completed an exercise program consisting of open kinetic chain strengthening and passive stretching exercises. The program was given for 3 sessions per week, 45 minutes per session, for 6 weeks Evaluation: At baseline and post treatment, pain severity was assessed using the numerical pain rating scale, whereas functional ability was being assessed by step test (ST), time up and go test (TUG) and 50 feet time walk test (50 FTW). After completing the program, global rate of change (GROC) score of greater than 4 was used to categorize patients as successful and non-successful. Thirty-eight patients (68%) had successful response to the intervention. Logistic regression showed that BMI and 50 FTW test were the only significant predictors. Based on the results, patients with BMI less than 34.71 kg/m2 and 50 FTW test less than 25.64 sec are 68% to 89% more likely to benefit from the exercise program. Conclusions: Clinicians should consider the described strengthening and flexibility exercise program for patents with BMI less than 34.7 Kg/m2 and 50 FTW faster than 25.6 seconds. The validity of these predictors should be investigated for other exercise.

Keywords: clinical prediction rule, knee osteoarthritis, physical therapy exercises, validity

Procedia PDF Downloads 392
1731 Implications of Optimisation Algorithm on the Forecast Performance of Artificial Neural Network for Streamflow Modelling

Authors: Martins Y. Otache, John J. Musa, Abayomi I. Kuti, Mustapha Mohammed

Abstract:

The performance of an artificial neural network (ANN) is contingent on a host of factors, for instance, the network optimisation scheme. In view of this, the study examined the general implications of the ANN training optimisation algorithm on its forecast performance. To this end, the Bayesian regularisation (Br), Levenberg-Marquardt (LM), and the adaptive learning gradient descent: GDM (with momentum) algorithms were employed under different ANN structural configurations: (1) single-hidden layer, and (2) double-hidden layer feedforward back propagation network. Results obtained revealed generally that the gradient descent with momentum (GDM) optimisation algorithm, with its adaptive learning capability, used a relatively shorter time in both training and validation phases as compared to the Levenberg- Marquardt (LM) and Bayesian Regularisation (Br) algorithms though learning may not be consummated; i.e., in all instances considering also the prediction of extreme flow conditions for 1-day and 5-day ahead, respectively especially using the ANN model. In specific statistical terms on the average, model performance efficiency using the coefficient of efficiency (CE) statistic were Br: 98%, 94%; LM: 98 %, 95 %, and GDM: 96 %, 96% respectively for training and validation phases. However, on the basis of relative error distribution statistics (MAE, MAPE, and MSRE), GDM performed better than the others overall. Based on the findings, it is imperative to state that the adoption of ANN for real-time forecasting should employ training algorithms that do not have computational overhead like the case of LM that requires the computation of the Hessian matrix, protracted time, and sensitivity to initial conditions; to this end, Br and other forms of the gradient descent with momentum should be adopted considering overall time expenditure and quality of the forecast as well as mitigation of network overfitting. On the whole, it is recommended that evaluation should consider implications of (i) data quality and quantity and (ii) transfer functions on the overall network forecast performance.

Keywords: streamflow, neural network, optimisation, algorithm

Procedia PDF Downloads 133
1730 Studies on Climatic and Soil Site Suitability of Major Grapes-Growing Soils of Eastern and Southern Dry Zones of Karnataka

Authors: Harsha B. R., Anil Kumar K. S.

Abstract:

Climate and soils are the two most dynamic entities among the factors affecting growth and grapes productivity. Studying of prevailing climate over the years in a region provides sufficient information related to management practices to be carried out in vineyards. Evaluating the suitability of vineyard soils under different climatic conditions serves as the yardstick to analyse the performance of grapevines. This study was formulated to study the climate and evaluate the site-suitability of soils in vineyards of southern Karnataka, which has registered its superiority in the quality production of wine. Ten soil profiles were excavated for suitability evaluation of soils, and six taluks were studied for climatic analysis. In almost all the regions studied, recharge starts at the end of the May or June months, peaking in either September or October months. Soil Starts drying from mid of December months in the taluks studied. Bangalore North (Rajanukunte) soils were highly suited for grapes cultivation with no or slight limitations. Bangalore North (GKVK Farm) was moderately suited with slight to moderate limitations of slope and available nitrogen content. Moderate suitability was observed in the rest of the profiles studied in Eastern dry zone soils with the slight to moderate limitations of either organic carbon or available nitrogen or both in the Eastern dry zone. Magadi (Southern dry zone) soils were moderately suitable with slight to moderate limitations of graveliness, available nitrogen, organic carbon, and exchangeable sodium percentage. Sustainable performance of vineyards in terms of yield can be achieved in these taluks by managing the constraints existing in soils.

Keywords: climatic analysis, dry zone, water recharge, growing period, suitability, sustainability

Procedia PDF Downloads 106
1729 A Novel Heuristic for Analysis of Large Datasets by Selecting Wrapper-Based Features

Authors: Bushra Zafar, Usman Qamar

Abstract:

Large data sample size and dimensions render the effectiveness of conventional data mining methodologies. A data mining technique are important tools for collection of knowledgeable information from variety of databases and provides supervised learning in the form of classification to design models to describe vital data classes while structure of the classifier is based on class attribute. Classification efficiency and accuracy are often influenced to great extent by noisy and undesirable features in real application data sets. The inherent natures of data set greatly masks its quality analysis and leave us with quite few practical approaches to use. To our knowledge first time, we present a new approach for investigation of structure and quality of datasets by providing a targeted analysis of localization of noisy and irrelevant features of data sets. Machine learning is based primarily on feature selection as pre-processing step which offers us to select few features from number of features as a subset by reducing the space according to certain evaluation criterion. The primary objective of this study is to trim down the scope of the given data sample by searching a small set of important features which may results into good classification performance. For this purpose, a heuristic for wrapper-based feature selection using genetic algorithm and for discriminative feature selection an external classifier are used. Selection of feature based on its number of occurrence in the chosen chromosomes. Sample dataset has been used to demonstrate proposed idea effectively. A proposed method has improved average accuracy of different datasets is about 95%. Experimental results illustrate that proposed algorithm increases the accuracy of prediction of different diseases.

Keywords: data mining, generic algorithm, KNN algorithms, wrapper based feature selection

Procedia PDF Downloads 301
1728 A Feasibility Study of Waste (d) Potential: Synergistic Effect Evaluation by Co-digesting Organic Wastes and Kinetics of Biogas Production

Authors: Kunwar Paritosh, Sanjay Mathur, Monika Yadav, Paras Gandhi, Subodh Kumar, Nidhi Pareek, Vivekanand Vivekanand

Abstract:

A significant fraction of energy is wasted every year managing the biodegradable organic waste inadequately as development and sustainability are the inherent enemies. The management of these waste is indispensable to boost its optimum utilization by converting it to renewable energy resource (here biogas) through anaerobic digestion and to mitigate greenhouse gas emission. Food and yard wastes may prove to be appropriate and potential feedstocks for anaerobic co-digestion for biogas production. The present study has been performed to explore the synergistic effect of co-digesting food waste and yard trimmings from MNIT campus for enhanced biogas production in different ratios in batch tests (37±10C, 90 rpm, 45 days). The results were overwhelming and showed that blending two different organic waste in proper ratio improved the biogas generation considerably, with the highest biogas yield (2044±24 mLg-1VS) that was achieved at 75:25 of food waste to yard waste ratio on volatile solids (VS) basis. The yield was 1.7 and 2.2 folds higher than the mono-digestion of food or yard waste (1172±34, 1016±36mLg-1VS) respectively. The increase in biogas production may be credited to optimum C/N ratio resulting in higher yield. Also Adding TiO2 nanoparticles showed virtually no effect on biogas production as sometimes nanoparticles enhance biogas production. ICP-MS, FTIR analysis was carried out to gain an insight of feedstocks. Modified Gompertz and logistics models were applied for the kinetic study of biogas production where modified Gompertz model showed goodness-of-fit (R2=0.9978) with the experimental results.

Keywords: anaerobic co-digestion, biogas, kinetics, nanoparticle, organic waste

Procedia PDF Downloads 363
1727 Adding Business Value in Enterprise Applications through Quality Matrices Using Agile

Authors: Afshan Saad, Muhammad Saad, Shah Muhammad Emaduddin

Abstract:

Nowadays the business condition is so quick paced that enhancing ourselves consistently has turned into a huge factor for the presence of an undertaking. We can check this for structural building and significantly more so in the quick-paced universe of data innovation and programming designing. The lithe philosophies, similar to Scrum, have a devoted advance in the process that objectives the enhancement of the improvement procedure and programming items. Pivotal to process enhancement is to pick up data that grants you to assess the condition of the procedure and its items. From the status data, you can design activities for the upgrade and furthermore assess the accomplishment of those activities. This investigation builds a model that measures the product nature of the improvement procedure. The product quality is dependent on the useful and auxiliary nature of the product items, besides the nature of the advancement procedure is likewise vital to enhance programming quality. Utilitarian quality covers the adherence to client prerequisites, while the auxiliary quality tends to the structure of the product item's source code with reference to its practicality. The procedure quality is identified with the consistency and expectedness of the improvement procedure. The product quality model is connected in a business setting by social occasion the information for the product measurements in the model. To assess the product quality model, we investigate the information and present it to the general population engaged with the light-footed programming improvement process. The outcomes from the application and the client input recommend that the model empowers a reasonable evaluation of the product quality and that it very well may be utilized to help the persistent enhancement of the advancement procedure and programming items.

Keywords: Agile SDLC Tools, Agile Software development, business value, enterprise applications, IBM, IBM Rational Team Concert, RTC, software quality, software metrics

Procedia PDF Downloads 150
1726 Research on Health Emergency Management Based on the Bibliometrics

Authors: Meng-Na Dai, Bao-Fang Wen, Gao-Pei Zhu, Chen-Xi Zhang, Jing Sun, Chang-Hai Tang, Zhi-Qiang Feng, Wen-Qiang Yin

Abstract:

Based on the analysis of literature in the health emergency management in China with recent 10 years, this paper discusses the Chinese current research hotspots, development trends and shortcomings in this field, and provides references for scholars to conduct follow-up research. CNKI(China National Knowledge Infrastructure), Weipu, and Wanfang were the databases of this literature. The key words during the database search were health, emergency, and management with the time from 2009 to 2018. The duplicate, non-academic, and unrelated documents were excluded. 901 articles were included in the literature review database. The main indicators of abstraction were, the number of articles published every year, authors, institutions, periodicals, etc. There are some research findings through the analysis of the literature. Overall, the number of literature in the health emergency management in China has shown a fluctuating downward trend in recent 10 years. Specifically, there is a lack of close cooperation between authors, which has not constituted the core team among them yet. Meanwhile, in this field, the number of high-level periodicals and quality literature is scarce. In addition, there are a lot of research hotspots, such as emergency management system, mechanism research, capacity evaluation index system research, plans and capacity-building research, etc. In the future, we should increase the scientific research funding of the health emergency management, encourage collaborative innovation among authors in multi-disciplinary fields, and create high-quality and high-impact journals in this field. The states should encourage scholars in this field to carry out more academic cooperation and communication with the whole world and improve the research in breadth and depth. Generally speaking, the research in health emergency management in China is still insufficient and needs to be improved.

Keywords: health emergency management, research situation, bibliometrics, literature

Procedia PDF Downloads 118
1725 Clustering-Based Threshold Model for Condition Rating of Concrete Bridge Decks

Authors: M. Alsharqawi, T. Zayed, S. Abu Dabous

Abstract:

To ensure safety and serviceability of bridge infrastructure, accurate condition assessment and rating methods are needed to provide basis for bridge Maintenance, Repair and Replacement (MRR) decisions. In North America, the common practices to assess condition of bridges are through visual inspection. These practices are limited to detect surface defects and external flaws. Further, the thresholds that define the severity of bridge deterioration are selected arbitrarily. The current research discusses the main deteriorations and defects identified during visual inspection and Non-Destructive Evaluation (NDE). NDE techniques are becoming popular in augmenting the visual examination during inspection to detect subsurface defects. Quality inspection data and accurate condition assessment and rating are the basis for determining appropriate MRR decisions. Thus, in this paper, a novel method for bridge condition assessment using the Quality Function Deployment (QFD) theory is utilized. The QFD model is designed to provide an integrated condition by evaluating both the surface and subsurface defects for concrete bridges. Moreover, an integrated condition rating index with four thresholds is developed based on the QFD condition assessment model and using K-means clustering technique. Twenty case studies are analyzed by applying the QFD model and implementing the developed rating index. The results from the analyzed case studies show that the proposed threshold model produces robust MRR recommendations consistent with decisions and recommendations made by bridge managers on these projects. The proposed method is expected to advance the state of the art of bridges condition assessment and rating.

Keywords: concrete bridge decks, condition assessment and rating, quality function deployment, k-means clustering technique

Procedia PDF Downloads 199
1724 Emotional Intelligence Training: Helping Non-Native Pre-Service EFL Teachers to Overcome Speaking Anxiety: The Case of Pre-Service Teachers of English, Algeria

Authors: Khiari Nor El Houda, Hiouani Amira Sarra

Abstract:

Many EFL students with high capacities are hidden because they suffer from speaking anxiety (SA). Most of them find public speaking much demanding. They feel unable to communicate, they fear to make mistakes and they fear negative evaluation or being called on. With the growing number of the learners who suffer from foreign language speaking anxiety (FLSA), it is becoming increasingly difficult to ignore its harmful outcomes on their performance and success, especially during their first contact with the pupils, as they will be teaching in the near future. Different researchers suggested different ways to minimize the negative effects of FLSA. The present study sheds light on emotional intelligence skills training as an effective strategy not only to influence public speaking success but also to help pre-service EFL teachers lessen their speaking anxiety and eventually to prepare them for their professional career. A quasi-experiment was used in order to examine the research hypothesis. We worked with two groups of third-year EFL students at Oum El Bouaghi University. The Foreign Language Classroom Anxiety Scale (FLCAS) and the Emotional Quotient Inventory (EQ-i) were used to collect data about the participants’ FLSA and EI levels. The analysis of the data has yielded that the assumption that there is a negative correlation between EI and FLSA was statistically validated by the Pearson Correlation Test, concluding that, the more emotionally intelligent the individual is the less anxious s/he will be. In addition, the lack of amelioration in the results of the control group and the noteworthy improvement in the experimental group results led us to conclude that EI skills training was an effective strategy in minimizing the FLSA level and therefore, we confirmed our research hypothesis.

Keywords: emotional intelligence, emotional intelligence skills training, EQ-I, FLCAS, foreign language speaking anxiety, pre-service EFL teachers

Procedia PDF Downloads 111
1723 Efficacy of the Use of Different Teaching Approaches of Math Teachers

Authors: Nilda San Miguel, Elymar Pascual

Abstract:

The main focus of this study is exploring the effective approaches in teaching Mathematics that is being applied in public schools, s.y. 2018-2019. This research was written as connected output to the district-wide School Learning Action Cell (DISLAC) on Math teaching approaches which was recently conducted in Victoria, Laguna. Fifty-four math teachers coming from 17 schools in Victoria became the respondents of this study. Qualitative method of doing research was applied. Teachers’ responses to the following concerns were gathered, analyzed and interpreted: (1) evaluation of the recently conducted DISLAC, (2) status of the use of different approaches, (3) perception on the effective use of approaches, (4) preference of approach to explore in classroom sessions, (5) factors affecting the choice of approach, (6) difficulties encountered, (7) and perceived benefit to learners. Results showed that the conduct of DISLAC was very highly satisfactory (mean 4.41). Teachers looked at collaborative approach as very highly effective (mean 4.74). Fifty-two percent of the teachers is using collaborative approach, 17% constructivist, 11% integrative, 11% inquiry-based, and 9% reflective. Reflective approach was chosen to be explored by most of the respondents (29%) in future sessions. The difficulties encountered by teachers in using the different approaches are: (1) learners’ difficulty in following instructions, (2) lack of focus, (3) lack of willingness and cooperation, (4) teachers’ lack of mastery in using different approaches, and (5) lack of time of doing visual aids because of time mismanagement. Teachers deemed the use of various teaching approaches can help the learners to have (1) mastery of competency, (2) increased communication, (3) improved confidence, (4) facility in comprehension, and (5) better academic output. The result obtained from this study can be used as an input for SLACs. Recommendations at the end of the study were given to school/district heads and future researchers.

Keywords: approaches, collaborative, constructivism, inquiry-based, integrative, reflective

Procedia PDF Downloads 258
1722 Emotional Labour and Employee Performance Appraisal: The Missing Link in Some Hotels in South East Nigeria

Authors: Polycarp Igbojekwe

Abstract:

The main objective of this study was to determine if emotional labour has become a criterion in performance appraisal, job description, selection, and training schemes in the hotel industry in Nigeria. Our main assumption was that majority of hotel organizations have not built emotional labour into their human resources management schemes. Data were gathered by the use of structured questionnaires designed in Likert format, and interviews. The focus group was managers of the selected hotels. Analyses revealed that majority of the hotels have not built emotional labour into their human resources schemes particularly in the 1, 2, and 3-star hotels. It was observed that service employees of 1, 2, and 3-star hotels have not been adequately trained to perform emotional labour; a critical factor in quality service delivery. Managers of 1, 2, and 3-star hotels have not given serious thought to emotional labour as a critical factor in quality service delivery. The study revealed that suitability of an individual’s characteristics is not being considered as a criterion for selection and performance appraisal for service employees. The implication of this is that, person-job-fit is not seriously considered. It was observed that there has been a disconnect between required emotional competency, its recognition, evaluation, and training. Based on the findings of this study, it is concluded that selection, training, job description and performance appraisal instruments in use in hotels in Nigeria are inadequate. Human resource implications of the findings in this study are presented. It is recommended that hotel organizations should re-design and plan the emotional content and context of their human resources practices to reflect the emotional demands of front line jobs in the hotel industry and the crucial role emotional labour plays during service encounters.

Keywords: emotional labour, employee selection, job description, performance appraisal, person-job-fit, employee compensation

Procedia PDF Downloads 176
1721 Quality Assurance for the Climate Data Store

Authors: Judith Klostermann, Miguel Segura, Wilma Jans, Dragana Bojovic, Isadora Christel Jimenez, Francisco Doblas-Reyees, Judit Snethlage

Abstract:

The Climate Data Store (CDS), developed by the Copernicus Climate Change Service (C3S) implemented by the European Centre for Medium-Range Weather Forecasts (ECMWF) on behalf of the European Union, is intended to become a key instrument for exploring climate data. The CDS contains both raw and processed data to provide information to the users about the past, present and future climate of the earth. It allows for easy and free access to climate data and indicators, presenting an important asset for scientists and stakeholders on the path for achieving a more sustainable future. The C3S Evaluation and Quality Control (EQC) is assessing the quality of the CDS by undertaking a comprehensive user requirement assessment to measure the users’ satisfaction. Recommendations will be developed for the improvement and expansion of the CDS datasets and products. User requirements will be identified on the fitness of the datasets, the toolbox, and the overall CDS service. The EQC function of the CDS will help C3S to make the service more robust: integrated by validated data that follows high-quality standards while being user-friendly. This function will be closely developed with the users of the service. Through their feedback, suggestions, and contributions, the CDS can become more accessible and meet the requirements for a diverse range of users. Stakeholders and their active engagement are thus an important aspect of CDS development. This will be achieved with direct interactions with users such as meetings, interviews or workshops as well as different feedback mechanisms like surveys or helpdesk services at the CDS. The results provided by the users will be categorized as a function of CDS products so that their specific interests will be monitored and linked to the right product. Through this procedure, we will identify the requirements and criteria for data and products in order to build the correspondent recommendations for the improvement and expansion of the CDS datasets and products.

Keywords: climate data store, Copernicus, quality, user engagement

Procedia PDF Downloads 126
1720 Impact of Massive Weight Loss Body Contouring Surgery in the Patient’s Quality of Life

Authors: Maria Albuquerque, Miguel Matias, Ângelo Sá, Juliana Sousa, Maria Manuel Mouzinho

Abstract:

Obesity is a frequent disease in Portugal. The surgical treatment is very effective and has an indication when there is a failure of the medical treatment. Although massive weight loss is associated with considerable health gains, these patients are characterized by a variable degree of dermolipodistrophy. In some cases, there is even the development of physical symptoms such as intertriginous, and some degree of psychological distress is present. In almost all cases, a desire for a better body contour, which inhibits some aspects of social life, is a fact. A prospective study was made to access the impact of body contouring surgery in the quality of life of patients who underwent a massive weight lost correction surgical procedure at Centro Hospitalar de Lisboa Central between January 2020 and December 2021. The patients were submitted to the Body Q subjective questionnaire adapted for the Portuguese language and accessed for the following categories: Anguish with Appearance, Contempt with Body Image, Satisfaction with the Abdomen, and Overall Satisfaction with the Body. The questionnaire was repeated at the 6 months mark. A total of 80 patients were sampled. The sex distribution was 79 female and 1 male. The median BMI index before surgery was inferior to 28%. The pre operatory questionnaire showed high scores for Anguish with Appearance and low scores for the body image self-evaluation. Overall, there was an improvement of at least 50% in all the evaluated scores. Additionally, a correlation was found between abdominoplasty and the contempt with body image and satisfaction with the abdomen (p-value <0.05). Massive weight loss is associated with important body deformities that have a significant impact on the patient’s personal and social life. Body contouring surgery is then vital for these patients as it implicates major aesthetic and functional benefits.

Keywords: abdominoplasty, cruroplasty, obesity, massive weight loss

Procedia PDF Downloads 133
1719 Evaluation of Possible Application of Cold Energy in Liquefied Natural Gas Complexes

Authors: А. I. Dovgyalo, S. O. Nekrasova, D. V. Sarmin, A. A. Shimanov, D. A. Uglanov

Abstract:

Usually liquefied natural gas (LNG) gasification is performed due to atmospheric heat. In order to produce a liquefied gas a sufficient amount of energy is to be consumed (about 1 kW∙h for 1 kg of LNG). This study offers a number of solutions, allowing using a cold energy of LNG. In this paper it is evaluated the application turbines installed behind the evaporator in LNG complex due to its work additional energy can be obtained and then converted into electricity. At the LNG consumption of G=1000kg/h the expansion work capacity of about 10 kW can be reached. Herewith-open Rankine cycle is realized, where a low capacity cryo-pump (about 500W) performs its normal function, providing the cycle pressure. Additionally discussed an application of Stirling engine within the LNG complex also gives a possibility to realize cold energy. Considering the fact, that efficiency coefficient of Stirling engine reaches 50 %, LNG consumption of G=1000 kg/h may result in getting a capacity of about 142 kW of such a thermal machine. The capacity of the pump, required to compensate pressure losses when LNG passes through the hydraulic channel, will make 500 W. Apart from the above-mentioned converters, it can be proposed to use thermoelectric generating packages (TGP), which are widely used now. At present, the modern thermoelectric generator line provides availability of electric capacity with coefficient of efficiency up to 15%. In the proposed complex, it is suggested to install the thermoelectric generator on the evaporator surface is such a way, that the cold end is contacted with the evaporator’s surface, and the hot one – with the atmosphere. At the LNG consumption of G=1000 kgг/h and specified coefficient of efficiency the capacity of the heat flow Qh will make about 32 kW. The derivable net electric power will be P=4,2 kW, and the number of packages will amount to about 104 pieces. The carried out calculations demonstrate the research perceptiveness in this field of propulsion plant development, as well as allow realizing the energy saving potential with the use of liquefied natural gas and other cryogenics technologies.

Keywords: cold energy, gasification, liquefied natural gas, electricity

Procedia PDF Downloads 256
1718 Comparison between High Resolution Ultrasonography and Magnetic Resonance Imaging in Assessment of Musculoskeletal Disorders Causing Ankle Pain

Authors: Engy S. El-Kayal, Mohamed M. S. Arafa

Abstract:

There are various causes of ankle pain including traumatic and non-traumatic causes. Various imaging techniques are available for assessment of AP. MRI is considered to be the imaging modality of choice for ankle joint evaluation with an advantage of its high spatial resolution, multiplanar capability, hence its ability to visualize small complex anatomical structures around the ankle. However, the high costs and the relatively limited availability of MRI systems, as well as the relatively long duration of the examination all are considered disadvantages of MRI examination. Therefore there is a need for a more rapid and less expensive examination modality with good diagnostic accuracy to fulfill this gap. HRU has become increasingly important in the assessment of ankle disorders, with advantages of being fast, reliable, of low cost and readily available. US can visualize detailed anatomical structures and assess tendinous and ligamentous integrity. The aim of this study was to compare the diagnostic accuracy of HRU with MRI in the assessment of patients with AP. We included forty patients complaining of AP. All patients were subjected to real-time HRU and MRI of the affected ankle. Results of both techniques were compared to surgical and arthroscopic findings. All patients were examined according to a defined protocol that includes imaging the tendon tears or tendinitis, muscle tears, masses, or fluid collection, ligament sprain or tears, inflammation or fluid effusion within the joint or bursa, bone and cartilage lesions, erosions and osteophytes. Analysis of the results showed that the mean age of patients was 38 years. The study comprised of 24 women (60%) and 16 men (40%). The accuracy of HRU in detecting causes of AP was 85%, while the accuracy of MRI in the detection of causes of AP was 87.5%. In conclusions: HRU and MRI are two complementary tools of investigation with the former will be used as a primary tool of investigation and the latter will be used to confirm the diagnosis and the extent of the lesion especially when surgical interference is planned.

Keywords: ankle pain (AP), high-resolution ultrasound (HRU), magnetic resonance imaging (MRI) ultrasonography (US)

Procedia PDF Downloads 174
1717 Adaptive Reuse of Lost Urban Space

Authors: Rana Sameeh

Abstract:

The city is the greatest symbol of human civilization and has been built for safety and comfort. However, uncontrolled urban growth caused some anonymous and unsightly images of the cities such as unused or abandoned spaces. When social interaction is missed in a public space it means the public space is lost since public spaces reflect the social life and interaction of people. Accordingly; this space became one of the most meaningless parts of the cities and has broken the continuity of the urban fabric. Lost urban spaces are the leftover unstructured landscape within the urban fabric. They are generally the unrecognized urban areas that are in need of redesign, since they have a great value that can add to their surrounding urban context. The research significance lies within the importance of urban open spaces, their value and their impact on the urban fabric. The research also addresses the reuse and reclamation of lost urban spaces in order to increase the percentage of green areas along the urban fabric, provide urban open spaces, develop a sustainable approach towards urban landscape and enhance the quality of the public open space and user experience. In addition, the reuse of lost space will give it the identity and function it lacks while also providing places for presence, spending time and observing. Creating continuity in a broken urban fabric represents an exploratory process in the relationship between infrastructure and the urban fabric and seeks to establish an architectural solution to leftover space within the city. In doing so, the research establishes a framework (criteria) for adaptive reuse of lost urban space throughout inductive and deductive methodology, analytical methodology; by analyzing some relevant examples and similar cases of lost spaces and finally through field methodology; by applying the achieved criteria on a case study in Alexandria and carrying on SWOT analysis and evaluation of the potentials of this case study.

Keywords: adaptive reuse, lost urban space, quality of public open space, urban fabric

Procedia PDF Downloads 617
1716 Nectariferous Plant Genetic Resources for Apicultural Entrepreneurship in Nigeria: Prerequisite for Conservation, Sustainable Management and Policy

Authors: C. V. Nnamani, O. L. Adedeji

Abstract:

The contemporary global economic meltdown has devastating effect on the Nigerian’s economy and its frantic search for alternative source of national revenue aside from oil and gas has become imperative for economic emancipation for Nigerians. Apicultural entrepreneurship could provide a source of livelihood if the basic knowledge of those plant genetic resources needed by bees is made available. A palynological evaluation of those palynotaxa which honey bees forage for pollen and nectar was carried out after standard acetolysis method. Results showed that the honey samples were highly diversified and rich in honey plants. A total of 9544.3 honey pollen, consisting of 39 honey plants belonging to 21 plant families and distributed within 38 genera were identified excluding 238 unidentified pollen grains. Data from the analysis equally revealed that Elaeis guineensis Jacq, Anacardium occidentale L, Diospyros mespiliformis Hochist xe ADC, Alchornea cordifolia Muell, Arg, Daniella oliveri (Rolfe) Hutch & Dalz, Irvingia wombolu Okafor ex Baill, Treculia africana Decne, Nauclea latifolia Smith and Crossopteryx febrifuga Afzil ex Benth were the predominant honey plants. It provided a guide to the optimal utilization of floral resources by honeybees in these regions, showing the opportunity and amazing potentials for apiculture entrepreneurship of these palytaxa. Most of these plants are rare, threatened and endangered. It calls for urgent conservation techniques and step by all players. Critical awareness creation to ensure farmers knowledge of these palynotaxa to ensure proper understanding and attendance boost from them as economic empowerment is needed.

Keywords: palynotaxa, acetolysis, enterprise, livelihood, Nigeria

Procedia PDF Downloads 271
1715 Research on the Feasibility of Evaluating Low-Temperature Cracking Performance of Asphalt Mixture Using Fracture Energy

Authors: Tao Yang, Yongli Zhao

Abstract:

Low-temperature cracking is one of the major challenges for asphalt pavement in the cold region. Fracture energy could determine from various test methods, which is a commonly used parameter to evaluate the low-temperature cracking resistance of asphalt mixture. However, the feasibility of evaluating the low-temperature cracking performance of asphalt mixture using fracture energy is not investigated comprehensively. This paper aims to verify whether fracture energy is an appropriate parameter to evaluate the low-temperature cracking performance. To achieve this goal, this paper compared the test results of thermal stress restrained specimen test (TSRST) and semi-circular bending test (SCB) of asphalt mixture with different types of aggregate, TSRST and indirect tensile test (IDT) of asphalt mixture with different additives, and single-edge notched beam test (SENB) and TSRST of asphalt mixture with different asphalt. Finally, the correlation between in-suit cracking performance and fracture energy was surveyed. The experimental results showed the evaluation result of critical cracking temperature and fracture energy are not always consistent; the in-suit cracking performance is also not correlated well with fracture energy. These results indicated that it is not feasible to evaluate low-temperature performance by fracture energy. Then, the composition of fracture energy of TSRST, SCB, disk-shaped compact tension test (DCT), three-point bending test (3PB) and IDT was analyzed. The result showed: the area of thermal stress versus temperature curve is the multiple of fracture energy and could be used to represent fracture energy of TSRST, as the multiple is nearly equal among different asphalt mixtures for a specific specimen; the fracture energy, determined from TSRST, SCB, DCT, 3PB, SENB and IDT, is mainly the surface energy that forms the fracture face; fracture energy is inappropriate to evaluate the low-temperature cracking performance of asphalt mixture, as the relaxation/viscous performance is not considered; if the fracture energy was used, it is recommended to combine this parameter with an index characterizing the relaxation or creep performance of asphalt mixture.

Keywords: asphalt pavement, cold region, critical cracking temperature, fracture energy, low-temperature cracking

Procedia PDF Downloads 159
1714 From Creativity to Innovation: Tracking Rejected Ideas

Authors: Lisete Barlach, Guilherme Ary Plonski

Abstract:

Innovative ideas are not always synonymous with business opportunities. Any idea can be creative and not recognized as a potential project in which money and time will be invested, among other resources. Even in firms that promote and enhance innovation, there are two 'check-points', the first corresponding to the acknowledgment of the idea as creative and the second, its consideration as a business opportunity. Both the recognition of new business opportunities or new ideas involve cognitive and psychological frameworks which provide individuals with a basis for noticing connections between seemingly independent events or trends as if they were 'connecting the dots'. It also involves prototypes-representing the most typical member of a certain category–functioning as 'templates' for this recognition. There is a general assumption that these kinds of evaluation processes develop through experience, explaining why expertise plays a central role in this process: the more experienced a professional, the easier for him (her) to identify new opportunities in business. But, paradoxically, an increase in expertise can lead to the inflexibility of thought due to automation of procedures. And, besides this, other cognitive biases can also be present, because new ideas or business opportunities generally depend on heuristics, rather than on established algorithms. The paper presents a literature review about the Einstellung effect by tracking famous cases of rejected ideas, extracted from historical records. It also presents the results of empirical research, with data upon rejected ideas gathered from two different environments: projects rejected during first semester of 2017 at a large incubator center in Sao Paulo and ideas proposed by employees that were rejected by a well-known business company, at its Brazilian headquarter. There is an implicit assumption that Einstellung effect tends to be more and more present in contemporaneity, due to time pressure upon decision-making and idea generation process. The analysis discusses desirability, viability, and feasibility as elements that affect decision-making.

Keywords: cognitive biases, Einstellung effect, recognition of business opportunities, rejected ideas

Procedia PDF Downloads 173
1713 Creating Database and Building 3D Geological Models: A Case Study on Bac Ai Pumped Storage Hydropower Project

Authors: Nguyen Chi Quang, Nguyen Duong Tri Nguyen

Abstract:

This article is the first step to research and outline the structure of the geotechnical database in the geological survey of a power project; in the context of this report creating the database that has been carried out for the Bac Ai pumped storage hydropower project. For the purpose of providing a method of organizing and storing geological and topographic survey data and experimental results in a spatial database, the RockWorks software is used to bring optimal efficiency in the process of exploiting, using, and analyzing data in service of the design work in the power engineering consulting. Three-dimensional (3D) geotechnical models are created from the survey data: such as stratigraphy, lithology, porosity, etc. The results of the 3D geotechnical model in the case of Bac Ai pumped storage hydropower project include six closely stacked stratigraphic formations by Horizons method, whereas modeling of engineering geological parameters is performed by geostatistical methods. The accuracy and reliability assessments are tested through error statistics, empirical evaluation, and expert methods. The three-dimensional model analysis allows better visualization of volumetric calculations, excavation and backfilling of the lake area, tunneling of power pipelines, and calculation of on-site construction material reserves. In general, the application of engineering geological modeling makes the design work more intuitive and comprehensive, helping construction designers better identify and offer the most optimal design solutions for the project. The database always ensures the update and synchronization, as well as enables 3D modeling of geological and topographic data to integrate with the designed data according to the building information modeling. This is also the base platform for BIM & GIS integration.

Keywords: database, engineering geology, 3D Model, RockWorks, Bac Ai pumped storage hydropower project

Procedia PDF Downloads 141
1712 Textile Wastewater Ecotoxicity Abatement after Aerobic Granular Sludge Treatment and Advanced Oxidation Process

Authors: Ana M. T. Mata, Alexiane Ligneul

Abstract:

Textile effluents are usually heavily loaded with organic carbon and color compounds, the latter being azo dyes in an estimated 70% of the case effluent posing a major challenge in environmental protection. In this study, the ecotoxicity of simulated textile effluent after biological treatment with anaerobic and aerobic phase (aerobic granular sludge, AGS) and after advanced oxidation processes (AOP) namely ozonation and UV irradiation as post-treatment, were tested to evaluate the fitness of this treatments for ecotoxicity abatement. AGS treatment achieved an 80% removal in both COD and color. AOP was applied with the intention to mineralize the metabolites resulting from biodecolorization of the azo dye Acid Red 14, especially the stable aromatic amine (4-amino-1-naphthalenesulfonic acid, 4A1NS). The ecotoxicity evaluation was based on growth inhibition of the algae Pseudokirchneriella subcapitata following OECD TG 201 except regarding the medium, MBL medium was used instead. Five replicate control cultures and samples were performed with an average STD of 2.7% regarding specific algae growth rate determination. It was found that untreated textile effluent holds an inhibition of specific growth rate of 82%. AGS treatment by itself is able to lower ecotoxicity to 53%. This is probably due to the high color removal of the treatment. AOP post-treatment with Ozone and UV irradiation improves the ecotoxicity abatment to 49 and 43% inhibition respectively, less significantly than previously thought. Since over 85% of 4A1NS was removed by either of the AOP (followed by HPLC), an individual ecotoxicity test of 4A1NS was performed showing that 4A1NS does not inhibit algae growth (0% inhibition). It was concluded that AGS treatment is able by itself to achieve a significant ecotoxicity abatement of textile effluent. The cost-benefit of AOP as a post-treatment have to be better accessed since their application resulted in an improvement of only 10% regarding ecotoxicity effluent removal. It was also found that the 4A1NS amine had no apparent effect on ecotoxicity. Further studies will be conducted to study where ecotoxicity is coming from after AGS biological treatment and how to eliminate it.

Keywords: textile wastewate, ecotoxicity, aerobic granular sludge, AOP

Procedia PDF Downloads 143
1711 Comparison of Different Artificial Intelligence-Based Protein Secondary Structure Prediction Methods

Authors: Jamerson Felipe Pereira Lima, Jeane Cecília Bezerra de Melo

Abstract:

The difficulty and cost related to obtaining of protein tertiary structure information through experimental methods, such as X-ray crystallography or NMR spectroscopy, helped raising the development of computational methods to do so. An approach used in these last is prediction of tridimensional structure based in the residue chain, however, this has been proved an NP-hard problem, due to the complexity of this process, explained by the Levinthal paradox. An alternative solution is the prediction of intermediary structures, such as the secondary structure of the protein. Artificial Intelligence methods, such as Bayesian statistics, artificial neural networks (ANN), support vector machines (SVM), among others, were used to predict protein secondary structure. Due to its good results, artificial neural networks have been used as a standard method to predict protein secondary structure. Recent published methods that use this technique, in general, achieved a Q3 accuracy between 75% and 83%, whereas the theoretical accuracy limit for protein prediction is 88%. Alternatively, to achieve better results, support vector machines prediction methods have been developed. The statistical evaluation of methods that use different AI techniques, such as ANNs and SVMs, for example, is not a trivial problem, since different training sets, validation techniques, as well as other variables can influence the behavior of a prediction method. In this study, we propose a prediction method based on artificial neural networks, which is then compared with a selected SVM method. The chosen SVM protein secondary structure prediction method is the one proposed by Huang in his work Extracting Physico chemical Features to Predict Protein Secondary Structure (2013). The developed ANN method has the same training and testing process that was used by Huang to validate his method, which comprises the use of the CB513 protein data set and three-fold cross-validation, so that the comparative analysis of the results can be made comparing directly the statistical results of each method.

Keywords: artificial neural networks, protein secondary structure, protein structure prediction, support vector machines

Procedia PDF Downloads 593
1710 Nanomaterial Based Electrochemical Sensors for Endocrine Disrupting Compounds

Authors: Gaurav Bhanjana, Ganga Ram Chaudhary, Sandeep Kumar, Neeraj Dilbaghi

Abstract:

Main sources of endocrine disrupting compounds in the ecosystem are hormones, pesticides, phthalates, flame retardants, dioxins, personal-care products, coplanar polychlorinated biphenyls (PCBs), bisphenol A, and parabens. These endocrine disrupting compounds are responsible for learning disabilities, brain development problems, deformations of the body, cancer, reproductive abnormalities in females and decreased sperm count in human males. Although discharge of these chemical compounds into the environment cannot be stopped, yet their amount can be retarded through proper evaluation and detection techniques. The available techniques for determination of these endocrine disrupting compounds mainly include high performance liquid chromatography (HPLC), mass spectroscopy (MS) and gas chromatography-mass spectrometry (GC–MS). These techniques are accurate and reliable but have certain limitations like need of skilled personnel, time consuming, interference and requirement of pretreatment steps. Moreover, these techniques are laboratory bound and sample is required in large amount for analysis. In view of above facts, new methods for detection of endocrine disrupting compounds should be devised that promise high specificity, ultra sensitivity, cost effective, efficient and easy-to-operate procedure. Nowadays, electrochemical sensors/biosensors modified with nanomaterials are gaining high attention among researchers. Bioelement present in this system makes the developed sensors selective towards analyte of interest. Nanomaterials provide large surface area, high electron communication feature, enhanced catalytic activity and possibilities of chemical modifications. In most of the cases, nanomaterials also serve as an electron mediator or electrocatalyst for some analytes.

Keywords: electrochemical, endocrine disruptors, microscopy, nanoparticles, sensors

Procedia PDF Downloads 255
1709 Performance Evaluation and Plugging Characteristics of Controllable Self-Aggregating Colloidal Particle Profile Control Agent

Authors: Zhiguo Yang, Xiangan Yue, Minglu Shao, Yue Yang, Rongjie Yan

Abstract:

It is difficult to realize deep profile control because of the small pore-throats and easy water channeling in low-permeability heterogeneous reservoir, and the traditional polymer microspheres have the contradiction between injection and plugging. In order to solve this contradiction, the controllable self-aggregating colloidal particles (CSA) containing amide groups on the surface of microspheres was prepared based on emulsion polymerization of styrene and acrylamide. The dispersed solution of CSA colloidal particles, whose particle size is much smaller than the diameter of pore-throats, was injected into the reservoir. When the microspheres migrated to the deep part of reservoir, , these CSA colloidal particles could automatically self-aggregate into large particle clusters under the action of the shielding agent and the control agent, so as to realize the plugging of the water channels. In this paper, the morphology, temperature resistance and self-aggregation properties of CSA microspheres were studied by transmission electron microscopy (TEM) and bottle test. The results showed that CSA microspheres exhibited heterogeneous core-shell structure, good dispersion, and outstanding thermal stability. The microspheres remain regular and uniform spheres at 100℃ after aging for 35 days. With the increase of the concentration of the cations, the self-aggregation time of CSA was gradually shortened, and the influence of bivalent cations was greater than that of monovalent cations. Core flooding experiments showed that CSA polymer microspheres have good injection properties, CSA particle clusters can effective plug the water channels and migrate to the deep part of the reservoir for profile control.

Keywords: heterogeneous reservoir, deep profile control, emulsion polymerization, colloidal particles, plugging characteristic

Procedia PDF Downloads 204
1708 Evaluation and Analysis of ZigBee-Based Wireless Sensor Network: Home Monitoring as Case Study

Authors: Omojokun G. Aju, Adedayo O. Sule

Abstract:

ZigBee wireless sensor and control network is one of the most popularly deployed wireless technologies in recent years. This is because ZigBee is an open standard lightweight, low-cost, low-speed, low-power protocol that allows true operability between systems. It is built on existing IEEE 802.15.4 protocol and therefore combines the IEEE 802.15.4 features and newly added features to meet required functionalities thereby finding applications in wide variety of wireless networked systems. ZigBee‘s current focus is on embedded applications of general-purpose, inexpensive, self-organising networks which requires low to medium data rates, high number of nodes and very low power consumption such as home/industrial automation, embedded sensing, medical data collection, smart lighting, safety and security sensor networks, and monitoring systems. Although the ZigBee design specification includes security features to protect data communication confidentiality and integrity, however, when simplicity and low-cost are the goals, security is normally traded-off. A lot of researches have been carried out on ZigBee technology in which emphasis has mainly been placed on ZigBee network performance characteristics such as energy efficiency, throughput, robustness, packet delay and delivery ratio in different scenarios and applications. This paper investigate and analyse the data accuracy, network implementation difficulties and security challenges of ZigBee network applications in star-based and mesh-based topologies with emphases on its home monitoring application using the ZigBee ProBee ZE-10 development boards for the network setup. The paper also expose some factors that need to be considered when designing ZigBee network applications and suggest ways in which ZigBee network can be designed to provide more resilient to network attacks.

Keywords: home monitoring, IEEE 802.14.5, topology, wireless security, wireless sensor network (WSN), ZigBee

Procedia PDF Downloads 357
1707 A Damage Level Assessment Model for Extra High Voltage Transmission Towers

Authors: Huan-Chieh Chiu, Hung-Shuo Wu, Chien-Hao Wang, Yu-Cheng Yang, Ching-Ya Tseng, Joe-Air Jiang

Abstract:

Power failure resulting from tower collapse due to violent seismic events might bring enormous and inestimable losses. The Chi-Chi earthquake, for example, strongly struck Taiwan and caused huge damage to the power system on September 21, 1999. Nearly 10% of extra high voltage (EHV) transmission towers were damaged in the earthquake. Therefore, seismic hazards of EHV transmission towers should be monitored and evaluated. The ultimate goal of this study is to establish a damage level assessment model for EHV transmission towers. The data of earthquakes provided by Taiwan Central Weather Bureau serve as a reference and then lay the foundation for earthquake simulations and analyses afterward. Some parameters related to the damage level of each point of an EHV tower are simulated and analyzed by the data from monitoring stations once an earthquake occurs. Through the Fourier transform, the seismic wave is then analyzed and transformed into different wave frequencies, and the data would be shown through a response spectrum. With this method, the seismic frequency which damages EHV towers the most is clearly identified. An estimation model is built to determine the damage level caused by a future seismic event. Finally, instead of relying on visual observation done by inspectors, the proposed model can provide a power company with the damage information of a transmission tower. Using the model, manpower required by visual observation can be reduced, and the accuracy of the damage level estimation can be substantially improved. Such a model is greatly useful for health and construction monitoring because of the advantages of long-term evaluation of structural characteristics and long-term damage detection.

Keywords: damage level monitoring, drift ratio, fragility curve, smart grid, transmission tower

Procedia PDF Downloads 284