Search results for: machine performance
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 14491

Search results for: machine performance

8431 The Impact of Music on Social Identity Formation and Intergroup Relations in American-Born Korean Skaters in 2018 Winter Olympics

Authors: Sehwan Kim, Jepkorir Rose Chepyator Thomson

Abstract:

Music provides opportunities to affirm social identities and facilitate the internalization of one’s identity. The purpose of this study was to examine the role of music in breaking down boundaries between the in-group and out-of-group sport participants. Social identity theory was used to guide an understanding of two American-born South Korean skaters—Yura Min and Alexander Gamelin—who used a Korean representative traditional folk song, Arirang, at the 2018 Winter Olympics. This was an interpretive case study that focused on 2018 Winter Olympic participants whose performance and use of music was understood through the lenses of Koreans. Semi-structured interviews were conducted with 15 Korean audiences who watched two American-born South Korean skaters’ performances. Data analysis involved the determination of themes in the data collected. The findings of this study are as follows: First Koreans viewed the skaters as the out-group based on ethnic appearances and stereotypes. Second, Koreans’ inter-group bias against the skaters was meditated after Koreans watched the skaters as they used Arirang song in performance. Implications for this study include the importance of music as an instrument of unity across diverse populations, including intergroup relations. Music can also offer ways to understand people’s cultures and bridge gaps between age and gender across categories of naturalization.

Keywords: impact of music, intergroup relations, naturalized athletes, social identity theory

Procedia PDF Downloads 196
8430 Modeling of an Insulin Mircopump

Authors: Ahmed Slami, Med El Amine Brixi Nigassa, Nassima Labdelli, Sofiane Soulimane, Arnaud Pothier

Abstract:

Many people suffer from diabetes, a disease marked by abnormal levels of sugar in the blood; 285 million people have diabetes, 6.6% of the world adult population (in 2010), according to the International Diabetes Federation. Insulin medicament is invented to be injected into the body. Generally, the injection requires the patient to do it manually. However, in many cases he will be unable to inject the drug, saw that among the side effects of hyperglycemia is the weakness of the whole body. The researchers designed a medical device that injects insulin too autonomously by using micro-pumps. Many micro-pumps of concepts have been investigated during the last two decades for injecting molecules in blood or in the body. However, all these micro-pumps are intended for slow infusion of drug (injection of few microliters by minute). Now, the challenge is to develop micro-pumps for fast injections (1 microliter in 10 seconds) with accuracy of the order of microliter. Recently, studies have shown that only piezoelectric actuators can achieve this performance, knowing that few systems at the microscopic level were presented. These reasons lead us to design new smart microsystems injection drugs. Therefore, many technological advances are still to achieve the improvement of materials to their uses, while going through their characterization and modeling action mechanisms themselves. Moreover, it remains to study the integration of the piezoelectric micro-pump in the microfluidic platform features to explore and evaluate the performance of these new micro devices. In this work, we propose a new micro-pump model based on piezoelectric actuation with a new design. Here, we use a finite element model with Comsol software. Our device is composed of two pumping chambers, two diaphragms and two actuators (piezoelectric disks). The latter parts will apply a mechanical force on the membrane in a periodic manner. The membrane deformation allows the fluid pumping, the suction and discharge of the liquid. In this study, we present the modeling results as function as device geometry properties, films thickness, and materials properties. Here, we demonstrate that we can achieve fast injection. The results of these simulations will provide quantitative performance of our micro-pumps. Concern the spatial actuation, fluid rate and allows optimization of the fabrication process in terms of materials and integration steps.

Keywords: COMSOL software, piezoelectric, micro-pump, microfluidic

Procedia PDF Downloads 333
8429 The Reflection Framework to Enhance the User Experience for Cultural Heritage Spaces’ Websites in Post-Pandemic Times

Authors: Duyen Lam, Thuong Hoang, Atul Sajjanhar, Feifei Chen

Abstract:

With the emerging interactive technology applications helping users connect progressively with cultural artefacts in new approaches, the cultural heritage sector gains significantly. The interactive apps’ issues can be tested via several techniques, including usability surveys and usability evaluations. The severe usability problems for museums’ interactive technologies commonly involve interactions, control, and navigation processes. This study confirms the low quality of being immersive for audio guides in navigating the exhibition and involving experience in the virtual environment, which are the most vital features of new interactive technologies such as AR and VR. In addition, our usability surveys and heuristic evaluations disclosed many usability issues of these interactive technologies relating to interaction functions. Additionally, we use the Wayback Machine to examine what interactive apps/technologies were deployed on these websites during the physical visits limited due to the COVID-19 pandemic lockdown. Based on those inputs, we propose the reflection framework to enhance the UX in the cultural heritage domain with detailed guidelines.

Keywords: framework, user experience, cultural heritage, interactive technology, museum, COVID-19 pandemic, usability survey, heuristic evaluation, guidelines

Procedia PDF Downloads 43
8428 Roles Currently Played by Educational Middle Leaders

Authors: Elaine Marta Pereira Aaltonen

Abstract:

Effective school leadership materialised in educational settings through the high standard professional performance of senior and middle leaders, has increasingly become an education policy priority around the world due to a wide recognition that schools need knowledgeable, skilled, and committed leaders, along with great teachers, in order to ensure outstanding education at all levels of schooling. The scope of this paper is the work of middle leaders, whose direct influence on teachers and classroom teaching, thus, on student learning outcomes, is a key component for successful school systems. It particularly aims at sharing some of the findings obtained through an academic study recently carried out by the same researcher, which was focused on enhancing understanding about aspects related to the professional performance of educational middle leaders, applied to the context of the lower elementary school division of a private mainstream school located in Brazil. The master´s dissertation findings included identifying the roles performed by a team of educational middle leaders throughout the year of 2021, as well as gaining insights on their perceptions about the roles performed, both through an electronic questionnaire and individual face-to-face interviews. Not only the roles of the middle leaders who participated in the research have been identified through the qualitative case study undertaken, but additional research finding lying within the sphere of the categorisation of such roles, based upon coherent domains of practice, has possibly been made. Hence, the main purpose of this paper is to outline the findings concerning the current roles played by educational middle leaders.

Keywords: roles, middle leaders, educational leadership, school leadership, and management

Procedia PDF Downloads 92
8427 Experimental Investigation of Gas Bubble Behaviours in a Domestic Heat Pump Water Heating System

Authors: J. B. Qin, X. H. Jiang, Y. T. Ge

Abstract:

The growing awareness of global warming potential has internationally aroused interest and demand in reducing greenhouse gas emissions produced by human activity. Much national energy in the UK had been consumed in the residential sector mainly for space heating and domestic hot water production. Currently, gas boilers are mostly applied in the domestic water heating which contribute significantly to excessive CO2 emissions and consumption of primary energy resources. The issues can be solved by popularizing heat pump systems that are attributable to higher performance efficiency than those of traditional gas boilers. Even so, the heat pump system performance can be further enhanced if the dissolved gases in its hot water circuit can be efficiently discharged.  To achieve this target, the bubble behaviors in the heat pump water heating system need to be extensively investigated. In this paper, by varying different experimental conditions, the effects of various heat pump hot water side parameters on gas microbubble diameters were measured and analyzed. Correspondingly, the effect of each parameter has been investigated. These include varied system pressures, water flow rates, saturation ratios and heat outputs. The results measurement showed that the water flow rate is the most significant parameter to influence on gas microbubble productions. The research outcomes can significantly contribute to the understanding of gas bubble behaviors at domestic heat pump water heating systems and thus the efficient way for the discharging of the associated dissolved gases.  

Keywords: heat pump water heating system, microbubble formation, dissolved gases in water, effectiveness

Procedia PDF Downloads 249
8426 Electroencephalogram Based Approach for Mental Stress Detection during Gameplay with Level Prediction

Authors: Priyadarsini Samal, Rajesh Singla

Abstract:

Many mobile games come with the benefits of entertainment by introducing stress to the human brain. In recognizing this mental stress, the brain-computer interface (BCI) plays an important role. It has various neuroimaging approaches which help in analyzing the brain signals. Electroencephalogram (EEG) is the most commonly used method among them as it is non-invasive, portable, and economical. Here, this paper investigates the pattern in brain signals when introduced with mental stress. Two healthy volunteers played a game whose aim was to search hidden words from the grid, and the levels were chosen randomly. The EEG signals during gameplay were recorded to investigate the impacts of stress with the changing levels from easy to medium to hard. A total of 16 features of EEG were analyzed for this experiment which includes power band features with relative powers, event-related desynchronization, along statistical features. Support vector machine was used as the classifier, which resulted in an accuracy of 93.9% for three-level stress analysis; for two levels, the accuracy of 92% and 98% are achieved. In addition to that, another game that was similar in nature was played by the volunteers. A suitable regression model was designed for prediction where the feature sets of the first and second game were used for testing and training purposes, respectively, and an accuracy of 73% was found.

Keywords: brain computer interface, electroencephalogram, regression model, stress, word search

Procedia PDF Downloads 174
8425 Electrocardiogram-Based Heartbeat Classification Using Convolutional Neural Networks

Authors: Jacqueline Rose T. Alipo-on, Francesca Isabelle F. Escobar, Myles Joshua T. Tan, Hezerul Abdul Karim, Nouar Al Dahoul

Abstract:

Electrocardiogram (ECG) signal analysis and processing are crucial in the diagnosis of cardiovascular diseases, which are considered one of the leading causes of mortality worldwide. However, the traditional rule-based analysis of large volumes of ECG data is time-consuming, labor-intensive, and prone to human errors. With the advancement of the programming paradigm, algorithms such as machine learning have been increasingly used to perform an analysis of ECG signals. In this paper, various deep learning algorithms were adapted to classify five classes of heartbeat types. The dataset used in this work is the synthetic MIT-BIH Arrhythmia dataset produced from generative adversarial networks (GANs). Various deep learning models such as ResNet-50 convolutional neural network (CNN), 1-D CNN, and long short-term memory (LSTM) were evaluated and compared. ResNet-50 was found to outperform other models in terms of recall and F1 score using a five-fold average score of 98.88% and 98.87%, respectively. 1-D CNN, on the other hand, was found to have the highest average precision of 98.93%.

Keywords: heartbeat classification, convolutional neural network, electrocardiogram signals, generative adversarial networks, long short-term memory, ResNet-50

Procedia PDF Downloads 107
8424 An Investigation of Surface Texturing by Ultrasonic Impingement of Micro-Particles

Authors: Nagalingam Arun Prasanth, Ahmed Syed Adnan, S. H. Yeo

Abstract:

Surface topography plays a significant role in the functional performance of engineered parts. It is important to have a control on the surface geometry and understanding on the surface details to get the desired performance. Hence, in the current research contribution, a non-contact micro-texturing technique has been explored and developed. The technique involves ultrasonic excitation of a tool as a prime source of surface texturing for aluminum alloy workpieces. The specimen surface is polished first and is then immersed in a liquid bath containing 10% weight concentration of Ti6Al4V grade 5 spherical powders. A submerged slurry jet is used to recirculate the spherical powders under the ultrasonic horn which is excited at an ultrasonic frequency and amplitude of 40 kHz and 70 µm respectively. The distance between the horn and workpiece surface was remained fixed at 200 µm using a precision control stage. Texturing effects were investigated for different process timings of 1, 3 and 5 s. Thereafter, the specimens were cleaned in an ultrasonic bath for 5 mins to remove loose debris on the surface. The developed surfaces are characterized by optical and contact surface profiler. The optical microscopic images show a texture of circular spots on the workpiece surface indented by titanium spherical balls. Waviness patterns obtained from contact surface profiler supports the texturing effect produced from the proposed technique. Furthermore, water droplet tests were performed to show the efficacy of the proposed technique to develop hydrophilic surfaces and to quantify the texturing effect produced.

Keywords: surface texturing, surface modification, topography, ultrasonic

Procedia PDF Downloads 209
8423 In and Out-Of-Sample Performance of Non Simmetric Models in International Price Differential Forecasting in a Commodity Country Framework

Authors: Nicola Rubino

Abstract:

This paper presents an analysis of a group of commodity exporting countries' nominal exchange rate movements in relationship to the US dollar. Using a series of Unrestricted Self-exciting Threshold Autoregressive models (SETAR), we model and evaluate sixteen national CPI price differentials relative to the US dollar CPI. Out-of-sample forecast accuracy is evaluated through calculation of mean absolute error measures on the basis of two-hundred and fifty-three months rolling window forecasts and extended to three additional models, namely a logistic smooth transition regression (LSTAR), an additive non linear autoregressive model (AAR) and a simple linear Neural Network model (NNET). Our preliminary results confirm presence of some form of TAR non linearity in the majority of the countries analyzed, with a relatively higher goodness of fit, with respect to the linear AR(1) benchmark, in five countries out of sixteen considered. Although no model appears to statistically prevail over the other, our final out-of-sample forecast exercise shows that SETAR models tend to have quite poor relative forecasting performance, especially when compared to alternative non-linear specifications. Finally, by analyzing the implied half-lives of the > coefficients, our results confirms the presence, in the spirit of arbitrage band adjustment, of band convergence with an inner unit root behaviour in five of the sixteen countries analyzed.

Keywords: transition regression model, real exchange rate, nonlinearities, price differentials, PPP, commodity points

Procedia PDF Downloads 268
8422 The Proposal for a Framework to Face Opacity and Discrimination ‘Sins’ Caused by Consumer Creditworthiness Machines in the EU

Authors: Diogo José Morgado Rebelo, Francisco António Carneiro Pacheco de Andrade, Paulo Jorge Freitas de Oliveira Novais

Abstract:

Not everything in AI-power consumer credit scoring turns out to be a wonder. When using AI in Creditworthiness Assessment (CWA), opacity and unfairness ‘sins’ must be considered to the task be deemed Responsible. AI software is not always 100% accurate, which can lead to misclassification. Discrimination of some groups can be exponentiated. A hetero personalized identity can be imposed on the individual(s) affected. Also, autonomous CWA sometimes lacks transparency when using black box models. However, for this intended purpose, human analysts ‘on-the-loop’ might not be the best remedy consumers are looking for in credit. This study seeks to explore the legality of implementing a Multi-Agent System (MAS) framework in consumer CWA to ensure compliance with the regulation outlined in Article 14(4) of the Proposal for an Artificial Intelligence Act (AIA), dated 21 April 2021 (as per the last corrigendum by the European Parliament on 19 April 2024), Especially with the adoption of Art. 18(8)(9) of the EU Directive 2023/2225, of 18 October, which will go into effect on 20 November 2026, there should be more emphasis on the need for hybrid oversight in AI-driven scoring to ensure fairness and transparency. In fact, the range of EU regulations on AI-based consumer credit will soon impact the AI lending industry locally and globally, as shown by the broad territorial scope of AIA’s Art. 2. Consequently, engineering the law of consumer’s CWA is imperative. Generally, the proposed MAS framework consists of several layers arranged in a specific sequence, as follows: firstly, the Data Layer gathers legitimate predictor sets from traditional sources; then, the Decision Support System Layer, whose Neural Network model is trained using k-fold Cross Validation, provides recommendations based on the feeder data; the eXplainability (XAI) multi-structure comprises Three-Step-Agents; and, lastly, the Oversight Layer has a 'Bottom Stop' for analysts to intervene in a timely manner. From the analysis, one can assure a vital component of this software is the XAY layer. It appears as a transparent curtain covering the AI’s decision-making process, enabling comprehension, reflection, and further feasible oversight. Local Interpretable Model-agnostic Explanations (LIME) might act as a pillar by offering counterfactual insights. SHapley Additive exPlanation (SHAP), another agent in the XAI layer, could address potential discrimination issues, identifying the contribution of each feature to the prediction. Alternatively, for thin or no file consumers, the Suggestion Agent can promote financial inclusion. It uses lawful alternative sources such as the share of wallet, among others, to search for more advantageous solutions to incomplete evaluation appraisals based on genetic programming. Overall, this research aspires to bring the concept of Machine-Centered Anthropocentrism to the table of EU policymaking. It acknowledges that, when put into service, credit analysts no longer exert full control over the data-driven entities programmers have given ‘birth’ to. With similar explanatory agents under supervision, AI itself can become self-accountable, prioritizing human concerns and values. AI decisions should not be vilified inherently. The issue lies in how they are integrated into decision-making and whether they align with non-discrimination principles and transparency rules.

Keywords: creditworthiness assessment, hybrid oversight, machine-centered anthropocentrism, EU policymaking

Procedia PDF Downloads 21
8421 Railway Composite Flooring Design: Numerical Simulation and Experimental Studies

Authors: O. Lopez, F. Pedro, A. Tadeu, J. Antonio, A. Coelho

Abstract:

The future of the railway industry lies in the innovation of lighter, more efficient and more sustainable trains. Weight optimizations in railway vehicles allow reducing power consumption and CO₂ emissions, increasing the efficiency of the engines and the maximum speed reached. Additionally, they reduce wear of wheels and rails, increase the space available for passengers, etc. Among the various systems that integrate railway interiors, the flooring system is one which has greater impact both on passenger safety and comfort, as well as on the weight of the interior systems. Due to the high weight saving potential, relative high mechanical resistance, good acoustic and thermal performance, ease of modular design, cost-effectiveness and long life, the use of new sustainable composite materials and panels provide the latest innovations for competitive solutions in the development of flooring systems. However, one of the main drawbacks of the flooring systems is their relatively poor resistance to point loads. Point loads in railway interiors can be caused by passengers or by components fixed to the flooring system, such as seats and restraint systems, handrails, etc. In this way, they can originate higher fatigue solicitations under service loads or zones with high stress concentrations under exceptional loads (higher longitudinal, transverse and vertical accelerations), thus reducing its useful life. Therefore, to verify all the mechanical and functional requirements of the flooring systems, many physical prototypes would be created during the design phase, with all of the high costs associated with it. Nowadays, the use of virtual prototyping methods by computer-aided design (CAD) and computer-aided engineering (CAE) softwares allow validating a product before committing to making physical test prototypes. The scope of this work was to current computer tools and integrate the processes of innovation, development, and manufacturing to reduce the time from design to finished product and optimise the development of the product for higher levels of performance and reliability. In this case, the mechanical response of several sandwich panels with different cores, polystyrene foams, and composite corks, were assessed, to optimise the weight and the mechanical performance of a flooring solution for railways. Sandwich panels with aluminum face sheets were tested to characterise its mechanical performance and determine the polystyrene foam and cork properties when used as inner cores. Then, a railway flooring solution was fully modelled (including the elastomer pads to provide the required vibration isolation from the car body) and perform structural simulations using FEM analysis to comply all the technical product specifications for the supply of a flooring system. Zones with high stress concentrations are studied and tested. The influence of vibration modes on the comfort level and stability is discussed. The information obtained with the computer tools was then completed with several mechanical tests performed on some solutions, and on specific components. The results of the numerical simulations and experimental campaign carried out are presented in this paper. This research work was performed as part of the POCI-01-0247-FEDER-003474 (coMMUTe) Project funded by Portugal 2020 through COMPETE 2020.

Keywords: cork agglomerate core, mechanical performance, numerical simulation, railway flooring system

Procedia PDF Downloads 165
8420 Stability-Indicating High-Performance Thin-Layer Chromatography Method for Estimation of Naftopidil

Authors: P. S. Jain, K. D. Bobade, S. J. Surana

Abstract:

A simple, selective, precise and Stability-indicating High-performance thin-layer chromatographic method for analysis of Naftopidil both in a bulk and in pharmaceutical formulation has been developed and validated. The method employed, HPTLC aluminium plates precoated with silica gel as the stationary phase. The solvent system consisted of hexane: ethyl acetate: glacial acetic acid (4:4:2 v/v). The system was found to give compact spot for Naftopidil (Rf value of 0.43±0.02). Densitometric analysis of Naftopidil was carried out in the absorbance mode at 253 nm. The linear regression analysis data for the calibration plots showed good linear relationship with r2=0.999±0.0001 with respect to peak area in the concentration range 200-1200 ng per spot. The method was validated for precision, recovery and robustness. The limits of detection and quantification were 20.35 and 61.68 ng per spot, respectively. Naftopidil was subjected to acid and alkali hydrolysis, oxidation and thermal degradation. The drug undergoes degradation under acidic, basic, oxidation and thermal conditions. This indicates that the drug is susceptible to acid, base, oxidation and thermal conditions. The degraded product was well resolved from the pure drug with significantly different Rf value. Statistical analysis proves that the method is repeatable, selective and accurate for the estimation of investigated drug. The proposed developed HPTLC method can be applied for identification and quantitative determination of Naftopidil in bulk drug and pharmaceutical formulation.

Keywords: naftopidil, HPTLC, validation, stability, degradation

Procedia PDF Downloads 387
8419 Difference between 'HDR Ir-192 and Co-60 Sources' for High Dose Rate Brachytherapy Machine

Authors: Md Serajul Islam

Abstract:

High Dose Rate (HDR) Brachytherapy is used for cancer patients. In our country’s prospect, we are using only cervices and breast cancer treatment by using HDR. The air kerma rate in air at a reference distance of less than a meter from the source is the recommended quantity for the specification of gamma ray source Ir-192 in brachytherapy. The absorbed dose for the patients is directly proportional to the air kerma rate. Therefore the air kerma rate should be determined before the first use of the source on patients by qualified medical physicist who is independent from the source manufacturer. The air kerma rate will then be applied in the calculation of the dose delivered to patients in their planning systems. In practice, high dose rate (HDR) Ir-192 afterloader machines are mostly used in brachytherapy treatment. Currently, HDR-Co-60 increasingly comes into operation too. The essential advantage of the use of Co-60 sources is its longer half-life compared to Ir-192. The use of HDRCo-60 afterloading machines is also quite interesting for developing countries. This work describes the dosimetry at HDR afterloading machines according to the protocols IAEA-TECDOC-1274 (2002) with the nuclides Ir-192 and Co-60. We have used 3 different measurement methods (with a ring chamber, with a solid phantom and in free air and with a well chamber) in dependence of each of the protocols. We have shown that the standard deviations of the measured air kerma rate for the Co-60 source are generally larger than those of the Ir-192 source. The measurements with the well chamber had the lowest deviation from the certificate value. In all protocols and methods, the deviations stood for both nuclides by a maximum of about 1% for Ir-192 and 2.5% for Co-60-Sources respectively.

Keywords: Ir-192 source, cancer, patients, cheap treatment cost

Procedia PDF Downloads 219
8418 Structural Performance Evaluation of Concrete Beams Reinforced with Recycled and Virgin Plastic Fibres

Authors: Vighnesh Daas, David B. Tann, Mahmood Datoo

Abstract:

The incorporation of recycled plastic fibres in concrete as reinforcement is a potential sustainable alternative for replacement of ordinary steel bars. It provides a scope for waste reduction and re-use of plastics in the construction industry on a large scale. Structural use of fibre reinforced concrete is limited to short span members and low reliability classes. In this study, recycled carpet fibres made of 95% polypropylene with length of 45mm were used for experimental investigations. The performance of recycled polypropylene fibres under structural loading has been compared with commercially available virgin fibres at low volume fractions of less than 1%. A series of 100 mm cubes and 125x200x2000 mm beams were used to conduct strength tests in bending and compression to measure the influence of type and volume of fibres on the structural behaviour of fibre reinforced concrete beams. The workability of the concrete mix decreased as a function of fibre content and resulted in a modification of the mix design. The beams failed in a pseudo-ductile manner with an enhanced bending capacity. The specimens showed significant improvement in the post-cracking behaviour and load carrying ability as compared to conventional reinforced concrete members. This was associated to the binding properties of the fibres in the concrete matrix. With the inclusion of fibres at low volumes of 0-0.5%, there was reduction in crack sizes and deflection. This study indicates that the inclusion of recycled polypropylene fibres at low volumes augments the structural behaviour of concrete as compared to conventional reinforced concrete as well as virgin fibre reinforced concrete.

Keywords: fibre reinforced concrete, polypropylene, recycled, strength

Procedia PDF Downloads 230
8417 Text Localization in Fixed-Layout Documents Using Convolutional Networks in a Coarse-to-Fine Manner

Authors: Beier Zhu, Rui Zhang, Qi Song

Abstract:

Text contained within fixed-layout documents can be of great semantic value and so requires a high localization accuracy, such as ID cards, invoices, cheques, and passports. Recently, algorithms based on deep convolutional networks achieve high performance on text detection tasks. However, for text localization in fixed-layout documents, such algorithms detect word bounding boxes individually, which ignores the layout information. This paper presents a novel architecture built on convolutional neural networks (CNNs). A global text localization network and a regional bounding-box regression network are introduced to tackle the problem in a coarse-to-fine manner. The text localization network simultaneously locates word bounding points, which takes the layout information into account. The bounding-box regression network inputs the features pooled from arbitrarily sized RoIs and refine the localizations. These two networks share their convolutional features and are trained jointly. A typical type of fixed-layout documents: ID cards, is selected to evaluate the effectiveness of the proposed system. These networks are trained on data cropped from nature scene images, and synthetic data produced by a synthetic text generation engine. Experiments show that our approach locates high accuracy word bounding boxes and achieves state-of-the-art performance.

Keywords: bounding box regression, convolutional networks, fixed-layout documents, text localization

Procedia PDF Downloads 180
8416 Bhumastra “Unmanned Ground Vehicle”

Authors: Vivek Krishna, Nikhil Jain, A. Mary Posonia A., Albert Mayan J

Abstract:

Terrorism and insurgency are significant global issues that require constant attention and effort from governments and scientists worldwide. To combat these threats, nations invest billions of dollars in developing new defensive technologies to protect civilians. Breakthroughs in vehicle automation have led to the use of sophisticated machines for many dangerous and critical anti-terrorist activities. Our concept of an "Unmanned Ground Vehicle" can carry out tasks such as border security, surveillance, mine detection, and active combat independently or in tandem with human control. The robot's movement can be wirelessly controlled by a person in a distant location or can travel to a pre-programmed destination autonomously in situations where personal control is not feasible. Our defence system comprises two units: the control unit that regulates mobility and the motion tracking unit. The remote operator robot uses the camera's live visual feed to manually operate both units, and the rover can automatically detect movement. The rover is operated by manpower who controls it using a joystick or mouse, and a wireless modem enables a soldier in a combat zone to control the rover via an additional controller feature.

Keywords: robotics, computer vision, Machine learning, Artificial intelligence, future of AI

Procedia PDF Downloads 98
8415 A Sociolinguistic Approach to the Translation of Children’s Literature: Exploring Identity Issues in the American English Translation of Manolito Gafotas

Authors: Owen Harrington-Fernandez, Pilar Alderete-Diez

Abstract:

Up until recently, translation studies treated children’s literature as something of a marginal preoccupation, but the recent attention that this text type has attracted suggests that it may be fertile ground for research. This paper contributes to this new research avenue by applying a sociolinguistic theoretical framework to explore issues around the intersubjective co-construction of identity in the American English translation of the Spanish children’s story, Manolito Gafotas. The application of Bucholtz and Hall’s framework achieves two objectives: (1) it identifies shifts in the translation of the main character’s behaviour as culturally and morally motivated manipulations, and (2) it demonstrates how the context of translation becomes the very censorship machine that delegitimises the identity of the main character, and, concomitantly, the identity of the implied reader(s). If we take identity to be an intersubjective phenomenon, then it logicall follows that expurgating the identity of the main character necessarily shifts the identity of the implied reader(s) also. It is a double censorship of identity carried out under the auspices of an intellectual colonisation of a Spanish text. After reporting on the results of the analysis, the paper ends by raising the question of censorship in translation, and, more specifically, in children’s literature, in order to promote debate around this topic.

Keywords: censorship, identity, sociolinguistics, translation

Procedia PDF Downloads 245
8414 Bioefficacy of Catharanthus roseus on Reproductive Performance of Red Cotton Bug, Dysdercus koenigii (Heteroptera: Pyrrhocoriedae)

Authors: Sunil Kayesth, Kamal Kumar Gupta

Abstract:

Influence of hexane extract of Catharanthus roseus leaves on reproductive fitness of Dysdercus koenigii was investigated by evaluating mating behaviour, oviposition behaviour and fertility of the treated insects. The volatiles of the plants were extracted in hexane by ‘cold extraction method’. The insects were treated with the extracts by ‘dry film residual method’. Our studies indicated that the treated male showed altered courtship behaviour, less number of mounting attempts, took more time to mate, less percent successful mating, and more disrupted mating. Similarly, the treated female exhibited either mating refusal or neutral behaviour towards courting males. The maximum disruption in the mating was observed in a cross T♂ X T♀, where males and females were treated with Catharanthus extract. The Dysdercus treated with Catharanthus extracts also showed marked reduction in their reproductive success. The treated females laid lesser number of egg batches and eggs in their life span. Catharanthus extract was effective in alteration of the oviposition behaviour. The eggs laid by the mated females were fertile indicating insemination of the mated females. However, the percent hatchability of the eggs laid by the treated females was less than control. The GC-MS analysis of the extract revealed the presence of juvenile hormone mimics, and the intermediates of juvenile hormone biosynthesis. Therefore, some of these compounds individually or synergistically alter reproductive behaviour of Dysdercus.

Keywords: Catharanthus roseus, Dysdercus koenigii, GC-MS analysis, reproductive performance

Procedia PDF Downloads 241
8413 Impacts of Nomophobia on Daily Performance: Validity, Reliability and Prevalence Estimates among Undergraduate Dental Students in Bhubaneswar, India

Authors: Ramesh Nagarajappa, Upasana Mohapatra

Abstract:

Considered a modern phobia, Nomophobia (NO MObile PHOne PhoBIA) is a term that describes the irrational fear or anxiety of being unable to access one’s own mobile phone. Objectives: To develop and validate the nomophobia questionnaire, administering it to a sample of adolescents representing undergraduate dental students. To assess the prevalence of Nomophobia, determine the usage pattern of mobile phones, and evaluate the impact due to lack of access to mobile phones among undergraduate dental students. Methodology: A cross-sectional study was conducted on 302 undergraduate students at Bhubaneswar through a self-administered questionnaire via Google Forms consisting of 19 items evaluating the pattern and anxiety related to usage of mobile phones. Responses were recorded on a 5-point Likert scale. Kruskal Wallis, Mann-Whitney U, and Chi-square tests were used for statistical analysis. Results: Test-Retest reliability showed kappa of k=0.86 and Internal consistency Chronbach’s-Alpha to be α=0.82. Prevalence of nomophobia (score ≥ 58) was 32.1%, and students at risk of being nomophobic (score 39-57) was 61.9%. It was highest in males (32.6%) and amongst the interns (41.9%) and lowest (25.5%) amongst the second-year students. Participants felt nervous/insecure if their phones were away from them because of the fear that somebody might have accessed their data (3.07±1.93) and or tried to contact them (3.09±1.13), which were not statistically significant (p>0.05). Conclusions: Effect of mobile phone on dental students and the fear of not having their phones with them is increasing elaborately, that needs to be controlled, which if not achieved, would negatively hamper their academic performance and their being in the society.

Keywords: addiction, dental students, mobile phone, nomophobia

Procedia PDF Downloads 132
8412 A Hybrid Feature Selection and Deep Learning Algorithm for Cancer Disease Classification

Authors: Niousha Bagheri Khulenjani, Mohammad Saniee Abadeh

Abstract:

Learning from very big datasets is a significant problem for most present data mining and machine learning algorithms. MicroRNA (miRNA) is one of the important big genomic and non-coding datasets presenting the genome sequences. In this paper, a hybrid method for the classification of the miRNA data is proposed. Due to the variety of cancers and high number of genes, analyzing the miRNA dataset has been a challenging problem for researchers. The number of features corresponding to the number of samples is high and the data suffer from being imbalanced. The feature selection method has been used to select features having more ability to distinguish classes and eliminating obscures features. Afterward, a Convolutional Neural Network (CNN) classifier for classification of cancer types is utilized, which employs a Genetic Algorithm to highlight optimized hyper-parameters of CNN. In order to make the process of classification by CNN faster, Graphics Processing Unit (GPU) is recommended for calculating the mathematic equation in a parallel way. The proposed method is tested on a real-world dataset with 8,129 patients, 29 different types of tumors, and 1,046 miRNA biomarkers, taken from The Cancer Genome Atlas (TCGA) database.

Keywords: cancer classification, feature selection, deep learning, genetic algorithm

Procedia PDF Downloads 100
8411 Evaluation of the Appropriateness of Common Oxidants for Ruthenium (II) Chemiluminescence in a Microfluidic Detection Device Coupled to Microbore High Performance Liquid Chromatography for the Analysis of Drugs in Formulations and Biological Fluids

Authors: Afsal Mohammed Kadavilpparampu, Haider A. J. Al Lawati, Fakhr Eldin O. Suliman, Salma M. Z. Al Kindy

Abstract:

In this work, we evaluated the appropriateness of various oxidants that can be used potentially with Ru(bipy)32+ CL system while performing CL detection in a microfluidic device using eight common active pharmaceutical ingredients- ciprofloxacin, hydrochlorothiazide, norfloxacin, buspirone, fexofenadine, cetirizine, codeine, and dextromethorphan. This is because, microfludics have very small channel volume and the residence time is also very short. Hence, a highly efficient oxidant is required for on-chip CL detection to obtain analytically acceptable CL emission. Three common oxidants were evaluated, lead dioxide, cerium ammonium sulphate and ammonium peroxydisulphate. Results obtained showed that ammonium peroxydisulphate is the most appropriate oxidant which can be used in microfluidic setup and all the tested analyte give strong CL emission while using this oxidant. We also found that Ru(bipy)33+ generated off-line by oxidizing [Ru(bipy)3]Cl2.6H2O in acetonitrile under acidic condition with lead dioxide was stable for more than 72 hrs. A highly sensitive microbore HPLC- CL method using ammonium peroxydisulphate as an oxidant in a microfluidic on-chip CL detection has been developed for the analyses of fixed-dose combinations of pseudoephedrine (PSE), fexofenadine (FEX) and cetirizine (CIT) in biological fluids and pharmaceutical formulations with minimum sample pre-treatment.

Keywords: oxidants, microbore High Performance Liquid Chromatography, chemiluminescence, microfluidics

Procedia PDF Downloads 428
8410 Effect of Architecture and Operating Conditions of Vehicle on Bulb Lifetime in Automotive

Authors: Hatice Özbek, Caner Çil, Ahmet Rodoplu

Abstract:

Automotive lighting is the leading function in the configuration of vehicle architecture. Especially headlights and taillights from external lighting functions are among the structures that determine the stylistic character of the vehicle. At the same time, the fact that lighting functions are related to many other functions brings along difficulties in design. Customers expect maximum quality from the vehicle. In these circumstances, it is necessary to make designs that aim to keep the performance of bulbs with limited working lives at the highest level. With this study, the factors that influence the working lives of filament lamps were examined and bulb explosions that can occur sooner than anticipated in the future were prevented while the vehicle was still in the design phase by determining the relations with electrical, dynamical and static variables. Especially the filaments of the bulbs used in the front lighting of the vehicle are deformed in a shorter time due to the high voltage requirement. In addition to this, rear lighting lamps vibrate as a result of the tailgate opening and closing and cause the filaments to be exposed to high stress. With this study, the findings that cause bulb explosions were evaluated. Among the most important findings: 1. The structure of the cables to the lighting functions of the vehicle and the effect of the voltage values are drawn; 2. The effect of the vibration to bulb throughout the life of the vehicle; 3 The effect of the loads carried to bulb while the vehicle doors are opened and closed. At the end of the study, the maximum performance was established in the bulb lifetimes with the optimum changes made in the vehicle architecture based on the findings obtained.

Keywords: vehicle architecture, automotive lighting functions, filament lamps, bulb lifetime

Procedia PDF Downloads 142
8409 Floating Building Potential for Adaptation to Rising Sea Levels: Development of a Performance Based Building Design Framework

Authors: Livia Calcagni

Abstract:

Most of the largest cities in the world are located in areas that are vulnerable to coastal erosion and flooding, both linked to climate change and rising sea levels (RSL). Nevertheless, more and more people are moving to these vulnerable areas as cities keep growing. Architects, engineers and policy makers are called to rethink the way we live and to provide timely and adequate responses not only by investigating measures to improve the urban fabric, but also by developing strategies capable of planning change, exploring unusual and resilient frontiers of living, such as floating architecture. Since the beginning of the 21st century we have seen a dynamic growth of water-based architecture. At the same time, the shortage of land available for urban development also led to reclaim the seabed or to build floating structures. In light of these considerations, time is ripe to consider floating architecture not only as a full-fledged building typology but especially as a full-fledged adaptation solution for RSL. Currently, there is no global international legal framework for urban development on water and there is no structured performance based building design (PBBD) approach for floating architecture in most countries, let alone national regulatory systems. Thus, the research intends to identify the technological, morphological, functional, economic, managerial requirements that must be considered in a the development of the PBBD framework conceived as a meta-design tool. As it is expected that floating urban development is mostly likely to take place as extension of coastal areas, the needs and design criteria are definitely more similar to those of the urban environment than of the offshore industry. Therefor, the identification and categorization of parameters takes the urban-architectural guidelines and regulations as the starting point, taking the missing aspects, such as hydrodynamics, from the offshore and shipping regulatory frameworks. This study is carried out through an evidence-based assessment of performance guidelines and regulatory systems that are effective in different countries around the world addressing on-land and on-water architecture as well as offshore and shipping industries. It involves evidence-based research and logical argumentation methods. Overall, this paper highlights how inhabiting water is not only a viable response to the problem of RSL, thus a resilient frontier for urban development, but also a response to energy insecurity, clean water and food shortages, environmental concerns and urbanization, in line with Blue Economy principles and the Agenda 2030. Moreover, the discipline of architecture is presented as a fertile field for investigating solutions to cope with climate change and its effects on life safety and quality. Future research involves the development of a decision support system as an information tool to guide the user through the decision-making process, emphasizing the logical interaction between the different potential choices, based on the PBBD.

Keywords: adaptation measures, floating architecture, performance based building design, resilient architecture, rising sea levels

Procedia PDF Downloads 74
8408 Fault Diagnosis and Fault-Tolerant Control of Bilinear-Systems: Application to Heating, Ventilation, and Air Conditioning Systems in Multi-Zone Buildings

Authors: Abderrhamane Jarou, Dominique Sauter, Christophe Aubrun

Abstract:

Over the past decade, the growing demand for energy efficiency in buildings has attracted the attention of the control community. Failures in HVAC (heating, ventilation and air conditioning) systems in buildings can have a significant impact on the desired and expected energy performance of buildings and on the user's comfort as well. FTC is a recent technology area that studies the adaptation of control algorithms to faulty operating conditions of a system. The application of Fault-Tolerant Control (FTC) in HVAC systems has gained attention in the last two decades. The objective is to maintain the variations in system performance due to faults within an acceptable range with respect to the desired nominal behavior. This paper considers the so-called active approach, which is based on fault and identification scheme combined with a control reconfiguration algorithm that consists in determining a new set of control parameters so that the reconfigured performance is "as close as possible, "in some sense, to the nominal performance. Thermal models of buildings and their HVAC systems are described by non-linear (usually bi-linear) equations. Most of the works carried out so far in FDI (fault diagnosis and isolation) or FTC consider a linearized model of the studied system. However, this model is only valid in a reduced range of variation. This study presents a new fault diagnosis (FD) algorithm based on a bilinear observer for the detection and accurate estimation of the magnitude of the HVAC system failure. The main contribution of the proposed FD algorithm is that instead of using specific linearized models, the algorithm inherits the structure of the actual bilinear model of the building thermal dynamics. As an immediate consequence, the algorithm is applicable to a wide range of unpredictable operating conditions, i.e., weather dynamics, outdoor air temperature, zone occupancy profile. A bilinear fault detection observer is proposed for a bilinear system with unknown inputs. The residual vector in the observer design is decoupled from the unknown inputs and, under certain conditions, is made sensitive to all faults. Sufficient conditions are given for the existence of the observer and results are given for the explicit computation of observer design matrices. Dedicated observer schemes (DOS) are considered for sensor FDI while unknown input bilinear observers are considered for actuator or system components FDI. The proposed strategy for FTC works as follows: At a first level, FDI algorithms are implemented, making it also possible to estimate the magnitude of the fault. Once the fault is detected, the fault estimation is then used to feed the second level and reconfigure the control low so that that expected performances are recovered. This paper is organized as follows. A general structure for fault-tolerant control of buildings is first presented and the building model under consideration is introduced. Then, the observer-based design for Fault Diagnosis of bilinear systems is studied. The FTC approach is developed in Section IV. Finally, a simulation example is given in Section V to illustrate the proposed method.

Keywords: bilinear systems, fault diagnosis, fault-tolerant control, multi-zones building

Procedia PDF Downloads 159
8407 Effects of Live Yeast Supplementation to Reduce Oxidative Stress and Increase Lactation Performance of Dairy Cattle during the Summer Season

Authors: Ahmad Nawid Mirzad, Akira Goto, Takuto Endo, Hitoshi Ano, Hiromu Katamoto, Takenori Yamauchi

Abstract:

The objective of this study was to evaluate the effects of live yeast supplementation on oxidative stress biomarker and antioxidant vitamin levels as well as lactation performance in Holstein Friesian cows during the summer season in Fukuoka prefecture. Sixteen lactating cows weighing 707.50 ± 13.09 kg (Mean ± SE) were used and randomly assigned to either supplemented (n = 8) or control (n = 8) group. The cows in supplemented group were administered with live yeast product at 10 g/d per cow from middle of July to middle of September for eight weeks. In treatment group, serum levels of derivatives of reactive oxygen metabolites (d-ROMs) were lower at week six. In addition, serum levels of glucose and retinol were higher at week eight and those of α-tocopherol were higher at week 2 in treatment group. During study period daily average milk yield decreased in both groups. Daily average milk yield 63 days after the onset of supplementation in treatment and control groups were 23.5 and 22.2 kg, respectively. The reduction rate of milk yield in treatment group tended to be lower (17.6 vs. 20.0%). These results suggest that live yeast supplementation may reduce oxidative stress and improve energy metabolism in lactating dairy cows during the summer season.

Keywords: cow, live yeast, milk, oxidative stress, summer season

Procedia PDF Downloads 145
8406 A Construction Management Tool: Determining a Project Schedule Typical Behaviors Using Cluster Analysis

Authors: Natalia Rudeli, Elisabeth Viles, Adrian Santilli

Abstract:

Delays in the construction industry are a global phenomenon. Many construction projects experience extensive delays exceeding the initially estimated completion time. The main purpose of this study is to identify construction projects typical behaviors in order to develop a prognosis and management tool. Being able to know a construction projects schedule tendency will enable evidence-based decision-making to allow resolutions to be made before delays occur. This study presents an innovative approach that uses Cluster Analysis Method to support predictions during Earned Value Analyses. A clustering analysis was used to predict future scheduling, Earned Value Management (EVM), and Earned Schedule (ES) principal Indexes behaviors in construction projects. The analysis was made using a database with 90 different construction projects. It was validated with additional data extracted from literature and with another 15 contrasting projects. For all projects, planned and executed schedules were collected and the EVM and ES principal indexes were calculated. A complete linkage classification method was used. In this way, the cluster analysis made considers that the distance (or similarity) between two clusters must be measured by its most disparate elements, i.e. that the distance is given by the maximum span among its components. Finally, through the use of EVM and ES Indexes and Tukey and Fisher Pairwise Comparisons, the statistical dissimilarity was verified and four clusters were obtained. It can be said that construction projects show an average delay of 35% of its planned completion time. Furthermore, four typical behaviors were found and for each of the obtained clusters, the interim milestones and the necessary rhythms of construction were identified. In general, detected typical behaviors are: (1) Projects that perform a 5% of work advance in the first two tenths and maintain a constant rhythm until completion (greater than 10% for each remaining tenth), being able to finish on the initially estimated time. (2) Projects that start with an adequate construction rate but suffer minor delays culminating with a total delay of almost 27% of the planned time. (3) Projects which start with a performance below the planned rate and end up with an average delay of 64%, and (4) projects that begin with a poor performance, suffer great delays and end up with an average delay of a 120% of the planned completion time. The obtained clusters compose a tool to identify the behavior of new construction projects by comparing their current work performance to the validated database, thus allowing the correction of initial estimations towards more accurate completion schedules.

Keywords: cluster analysis, construction management, earned value, schedule

Procedia PDF Downloads 247
8405 Exploring Students’ Self-Evaluation on Their Learning Outcomes through an Integrated Cumulative Grade Point Average Reporting Mechanism

Authors: Suriyani Ariffin, Nor Aziah Alias, Khairil Iskandar Othman, Haslinda Yusoff

Abstract:

An Integrated Cumulative Grade Point Average (iCGPA) is a mechanism and strategy to ensure the curriculum of an academic programme is constructively aligned to the expected learning outcomes and student performance based on the attainment of those learning outcomes that is reported objectively in a spider web. Much effort and time has been spent to develop a viable mechanism and trains academics to utilize the platform for reporting. The question is: How well do learners conceive the idea of their achievement via iCGPA and whether quality learner attributes have been nurtured through the iCGPA mechanism? This paper presents the architecture of an integrated CGPA mechanism purported to address a holistic evaluation from the evaluation of courses learning outcomes to aligned programme learning outcomes attainment. The paper then discusses the students’ understanding of the mechanism and evaluation of their achievement from the generated spider web. A set of questionnaires were distributed to a group of students with iCGPA reporting and frequency analysis was used to compare the perspectives of students on their performance. In addition, the questionnaire also explored how they conceive the idea of an integrated, holistic reporting and how it generates their motivation to improve. The iCGPA group was found to be receptive to what they have achieved throughout their study period. They agreed that the achievement level generated from their spider web allows them to develop intervention and enhance the programme learning outcomes before they graduate.

Keywords: learning outcomes attainment, iCGPA, programme learning outcomes, spider web, iCGPA reporting skills

Procedia PDF Downloads 194
8404 The Clustering of Multiple Sclerosis Subgroups through L2 Norm Multifractal Denoising Technique

Authors: Yeliz Karaca, Rana Karabudak

Abstract:

Multifractal Denoising techniques are used in the identification of significant attributes by removing the noise of the dataset. Magnetic resonance (MR) image technique is the most sensitive method so as to identify chronic disorders of the nervous system such as Multiple Sclerosis. MRI and Expanded Disability Status Scale (EDSS) data belonging to 120 individuals who have one of the subgroups of MS (Relapsing Remitting MS (RRMS), Secondary Progressive MS (SPMS), Primary Progressive MS (PPMS)) as well as 19 healthy individuals in the control group have been used in this study. The study is comprised of the following stages: (i) L2 Norm Multifractal Denoising technique, one of the multifractal technique, has been used with the application on the MS data (MRI and EDSS). In this way, the new dataset has been obtained. (ii) The new MS dataset obtained from the MS dataset and L2 Multifractal Denoising technique has been applied to the K-Means and Fuzzy C Means clustering algorithms which are among the unsupervised methods. Thus, the clustering performances have been compared. (iii) In the identification of significant attributes in the MS dataset through the Multifractal denoising (L2 Norm) technique using K-Means and FCM algorithms on the MS subgroups and control group of healthy individuals, excellent performance outcome has been yielded. According to the clustering results based on the MS subgroups obtained in the study, successful clustering results have been obtained in the K-Means and FCM algorithms by applying the L2 norm of multifractal denoising technique for the MS dataset. Clustering performance has been more successful with the MS Dataset (L2_Norm MS Data Set) K-Means and FCM in which significant attributes are obtained by applying L2 Norm Denoising technique.

Keywords: clinical decision support, clustering algorithms, multiple sclerosis, multifractal techniques

Procedia PDF Downloads 152
8403 A Simulation-Optimization Approach to Control Production, Subcontracting and Maintenance Decisions for a Deteriorating Production System

Authors: Héctor Rivera-Gómez, Eva Selene Hernández-Gress, Oscar Montaño-Arango, Jose Ramon Corona-Armenta

Abstract:

This research studies the joint production, maintenance and subcontracting control policy for an unreliable deteriorating manufacturing system. Production activities are controlled by a derivation of the Hedging Point Policy, and given that the system is subject to deterioration, it reduces progressively its capacity to satisfy product demand. Multiple deterioration effects are considered, reflected mainly in the quality of the parts produced and the reliability of the machine. Subcontracting is available as support to satisfy product demand; also overhaul maintenance can be conducted to reduce the effects of deterioration. The main objective of the research is to determine simultaneously the production, maintenance and subcontracting rate which minimize the total incurred cost. A stochastic dynamic programming model is developed and solved through a simulation-based approach composed of statistical analysis and optimization with the response surface methodology. The obtained results highlight the strong interactions between production, deterioration and quality which justify the development of an integrated model. A numerical example and a sensitivity analysis are presented to validate our results.

Keywords: subcontracting, optimal control, deterioration, simulation, production planning

Procedia PDF Downloads 569
8402 Assessment of Soil Erosion Risk Using Soil and Water Assessment Tools Model: Case of Siliana Watershed, Northwest Tunisia

Authors: Sana Dridi, Jalel Aouissi, Rafla Attia, Taoufik Hermassi, Thouraya Sahli

Abstract:

Soil erosion is an increasing issue in Mediterranean countries. In Tunisia, the capacity of dam reservoirs continues to decrease as a consequence of soil erosion. This study aims to predict sediment yield to enrich soil management practices using Soil and Water Assessment Tools model (SWAT) in the Siliana watershed (1041.6 km²), located in the northwest of Tunisia. A database was constructed using remote sensing and Geographical Information System. Climatic and flow data were collected from water resources directorates in Tunisia. The SWAT model was built to simulate hydrological processes and sediment transport. A sensitivity analysis, calibration, and validation were performed using SWAT-CUP software. The model calibration of stream flow simulations shows a good performance with NSE and R² values of 0.77 and 0.79, respectively. The model validation shows a very good performance with values of NSE and R² for 0.8 and 0.88, respectively. After calibration and validation of stream flow simulation, the model was used to simulate the soil erosion and sediment load transport. The spatial distributions of soil loss rate for determining the critical sediment source areas show that 63 % of the study area has a low soil loss rate less than 7 t ha⁻¹y⁻¹. The annual average soil loss rate simulated with the SWAT model in the Siliana watershed is 4.62 t ha⁻¹y⁻¹.

Keywords: water erosion, SWAT model, streamflow, SWATCUP, sediment yield

Procedia PDF Downloads 89