Search results for: mitigation techniques
5394 Basic Calibration and Normalization Techniques for Time Domain Reflectometry Measurements
Authors: Shagufta Tabassum
Abstract:
The study of dielectric properties in a binary mixture of liquids is very useful to understand the liquid structure, molecular interaction, dynamics, and kinematics of the mixture. Time-domain reflectometry (TDR) is a powerful tool for studying the cooperation and molecular dynamics of the H-bonded system. In this paper, we discuss the basic calibration and normalization procedure for time-domain reflectometry measurements. Our approach is to explain the different types of error occur during TDR measurements and how these errors can be eliminated or minimized.Keywords: time domain reflectometry measurement techinque, cable and connector loss, oscilloscope loss, and normalization technique
Procedia PDF Downloads 2105393 Realization and Characterization of TiN Coating and Metal Working Application
Authors: Nadjette Belhamra, Abdelouahed Chala, Ibrahim Guasmi
Abstract:
Titanium nitride coatings have been extensively used in industry, such as in cutting tools. TiN coating were deposited by chemical vapour deposition (CVD) on carbide insert at a temperature between 850°C and 1100°C, which often exceeds the hardening treatment temperature of the metals. The objective of this work is to realize, to characterize of TiN coating and to apply it in the turning of steel 42CrMo4 under lubrification. Various experimental techniques were employed for the microstructural characterization of the coatings, e. g., X-ray diffraction (XRD), scanning electron microscope (SEM) model JOEL JSM-5900 LV, equipped with energy dispersive X-ray (EDX). The results show that TiN-coated demonstrate a good wear resistance.Keywords: hard coating TiN, carbide inserts, machining, turning, wear
Procedia PDF Downloads 5605392 Laser Cooling of Internal Degrees of Freedom of Molecules: Cesium Case
Authors: R. Horchani
Abstract:
Optical pumping technique with laser fields combined with photo-association of ultra-cold atoms leads to control on demand the vibrational and/or the rotational population of molecules. Here, we review the basic concepts and main steps should be followed, including the excitation schemes and detection techniques we use to achieve the ro-vibrational cooling of Cs2 molecules. We also discuss the extension of this technique to other molecules. In addition, we present a theoretical model used to support the experiment. These simulations can be widely used for the preparation of various experiments since they allow the optimization of several important experimental parameters.Keywords: cold molecule, photo-association, optical pumping, vibrational and rotational cooling
Procedia PDF Downloads 3065391 Remote Sensing Application in Environmental Researches: Case Study of Iran Mangrove Forests Quantitative Assessment
Authors: Neda Orak, Mostafa Zarei
Abstract:
Environmental assessment is an important session in environment management. Since various methods and techniques have been produces and implemented. Remote sensing (RS) is widely used in many scientific and research fields such as geology, cartography, geography, agriculture, forestry, land use planning, environment, etc. It can show earth surface objects cyclical changes. Also, it can show earth phenomena limits on basis of electromagnetic reflectance changes and deviations records. The research has been done on mangrove forests assessment by RS techniques. Mangrove forests quantitative analysis in Basatin and Bidkhoon estuaries was the aim of this research. It has been done by Landsat satellite images from 1975- 2013 and match to ground control points. This part of mangroves are the last distribution in northern hemisphere. It can provide a good background to improve better management on this important ecosystem. Landsat has provided valuable images to earth changes detection to researchers. This research has used MSS, TM, +ETM, OLI sensors from 1975, 1990, 2000, 2003-2013. Changes had been studied after essential corrections such as fix errors, bands combination, georeferencing on 2012 images as basic image, by maximum likelihood and IPVI Index. It was done by supervised classification. 2004 google earth image and ground points by GPS (2010-2012) was used to compare satellite images obtained changes. Results showed mangrove area in bidkhoon was 1119072 m2 by GPS and 1231200 m2 by maximum likelihood supervised classification and 1317600 m2 by IPVI in 2012. Basatin areas is respectively: 466644 m2, 88200 m2, 63000 m2. Final results show forests have been declined naturally. It is due to human activities in Basatin. The defect was offset by planting in many years. Although the trend has been declining in recent years again. So, it mentioned satellite images have high ability to estimation all environmental processes. This research showed high correlation between images and indexes such as IPVI and NDVI with ground control points.Keywords: IPVI index, Landsat sensor, maximum likelihood supervised classification, Nayband National Park
Procedia PDF Downloads 2945390 Developing a Machine Learning-based Cost Prediction Model for Construction Projects using Particle Swarm Optimization
Authors: Soheila Sadeghi
Abstract:
Accurate cost prediction is essential for effective project management and decision-making in the construction industry. This study aims to develop a cost prediction model for construction projects using Machine Learning techniques and Particle Swarm Optimization (PSO). The research utilizes a comprehensive dataset containing project cost estimates, actual costs, resource details, and project performance metrics from a road reconstruction project. The methodology involves data preprocessing, feature selection, and the development of an Artificial Neural Network (ANN) model optimized using PSO. The study investigates the impact of various input features, including cost estimates, resource allocation, and project progress, on the accuracy of cost predictions. The performance of the optimized ANN model is evaluated using metrics such as Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), and R-squared. The results demonstrate the effectiveness of the proposed approach in predicting project costs, outperforming traditional benchmark models. The feature selection process identifies the most influential variables contributing to cost variations, providing valuable insights for project managers. However, this study has several limitations. Firstly, the model's performance may be influenced by the quality and quantity of the dataset used. A larger and more diverse dataset covering different types of construction projects would enhance the model's generalizability. Secondly, the study focuses on a specific optimization technique (PSO) and a single Machine Learning algorithm (ANN). Exploring other optimization methods and comparing the performance of various ML algorithms could provide a more comprehensive understanding of the cost prediction problem. Future research should focus on several key areas. Firstly, expanding the dataset to include a wider range of construction projects, such as residential buildings, commercial complexes, and infrastructure projects, would improve the model's applicability. Secondly, investigating the integration of additional data sources, such as economic indicators, weather data, and supplier information, could enhance the predictive power of the model. Thirdly, exploring the potential of ensemble learning techniques, which combine multiple ML algorithms, may further improve cost prediction accuracy. Additionally, developing user-friendly interfaces and tools to facilitate the adoption of the proposed cost prediction model in real-world construction projects would be a valuable contribution to the industry. The findings of this study have significant implications for construction project management, enabling proactive cost estimation, resource allocation, budget planning, and risk assessment, ultimately leading to improved project performance and cost control. This research contributes to the advancement of cost prediction techniques in the construction industry and highlights the potential of Machine Learning and PSO in addressing this critical challenge. However, further research is needed to address the limitations and explore the identified future research directions to fully realize the potential of ML-based cost prediction models in the construction domain.Keywords: cost prediction, construction projects, machine learning, artificial neural networks, particle swarm optimization, project management, feature selection, road reconstruction
Procedia PDF Downloads 655389 A Multilevel Approach for Stroke Prediction Combining Risk Factors and Retinal Images
Authors: Jeena R. S., Sukesh Kumar A.
Abstract:
Stroke is one of the major reasons of adult disability and morbidity in many of the developing countries like India. Early diagnosis of stroke is essential for timely prevention and cure. Various conventional statistical methods and computational intelligent models have been developed for predicting the risk and outcome of stroke. This research work focuses on a multilevel approach for predicting the occurrence of stroke based on various risk factors and invasive techniques like retinal imaging. This risk prediction model can aid in clinical decision making and help patients to have an improved and reliable risk prediction.Keywords: prediction, retinal imaging, risk factors, stroke
Procedia PDF Downloads 3075388 Induction Machine Bearing Failure Detection Using Advanced Signal Processing Methods
Authors: Abdelghani Chahmi
Abstract:
This article examines the detection and localization of faults in electrical systems, particularly those using asynchronous machines. First, the process of failure will be characterized, relevant symptoms will be defined and based on those processes and symptoms, a model of those malfunctions will be obtained. Second, the development of the diagnosis of the machine will be shown. As studies of malfunctions in electrical systems could only rely on a small amount of experimental data, it has been essential to provide ourselves with simulation tools which allowed us to characterize the faulty behavior. Fault detection uses signal processing techniques in known operating phases.Keywords: induction motor, modeling, bearing damage, airgap eccentricity, torque variation
Procedia PDF Downloads 1415387 Preventive Impact of Regional Analgesia on Chronic Neuropathic Pain After General Surgery
Authors: Beloulou Mohamed Lamine, Fedili Benamar, Meliani Walid, Chaid Dalila, Lamara Abdelhak
Abstract:
Introduction: Post-surgical chronic pain (PSCP) is a pathological condition with a rather complex etiopathogenesis that extensively involves sensitization processes and neuronal damage. The neuropathic component of these pains is almost always present, with variable expression depending on the type of surgery. Objective: To assess the presumed beneficial effect of Regional Anesthesia-Analgesia Techniques (RAAT) on the development of post-surgical chronic neuropathic pain (PSCNP) in various surgical procedures. Patients and Methods: A comparative study involving 510 patients distributed across five surgical models (mastectomy, thoracotomy, hernioplasty, cholecystectomy, and major abdominal-pelvic surgery) and randomized into two groups: Group A (240) receiving conventional postoperative analgesia and Group B (270) receiving balanced analgesia, including the implementation of a Regional Anesthesia-Analgesia Technique (RAAT). These patients were longitudinally followed over a 6-month period, with postsurgical chronic neuropathic pain (PSCNP) defined by a Neuropathic Pain Score DN2≥ 3. Comparative measurements through univariate and multivariable analyses were performed to identify associations between the development of PSCNP and certain predictive factors, including the presumed preventive impact (protective effect) of RAAT. Results: At the 6th month post-surgery, 419 patients were analyzed (Group A= 196 and Group B= 223). The incidence of PSCNP was 32.2% (n=135). Among these patients with chronic pain, the prevalence of neuropathic pain was 37.8% (95% CI: [29.6; 46.5]), with n=51/135. It was significantly lower in Group B compared to Group A, with respective percentages of 31.4% vs. 48.8% (p-value = 0.035). The most significant differences were observed in breast and thoracopulmonary surgeries. In a multiple regression analysis, two predictors of PSCNP were identified: the presence of preoperative pain at the surgical site as a risk factor (OR: 3.198; 95% CI [1.326; 7.714]) and RAAT as a protective factor (OR: 0.408; 95% CI [0.173; 0.961]). Conclusion: The neuropathic component of PSCNP can be observed in different types of surgeries. Regional analgesia included in a multimodal approach to postoperative pain management has proven to be effective for acute pain and seems to have a preventive impact on the development of PSCNP and its neuropathic nature, particularly in surgeries that are more prone to chronicization.Keywords: post-surgical chronic pain, post-surgical chronic neuropathic pain, regional anesthesia-analgesia techniques, neuropathic pain score DN2, preventive impact
Procedia PDF Downloads 815386 The Road Ahead: Merging Human Cyber Security Expertise with Generative AI
Authors: Brennan Lodge
Abstract:
Amidst a complex regulatory landscape, Retrieval Augmented Generation (RAG) emerges as a transformative tool for Governance Risk and Compliance (GRC) officers. This paper details the application of RAG in synthesizing Large Language Models (LLMs) with external knowledge bases, offering GRC professionals an advanced means to adapt to rapid changes in compliance requirements. While the development for standalone LLM’s (Large Language Models) is exciting, such models do have their downsides. LLM’s cannot easily expand or revise their memory, and they can’t straightforwardly provide insight into their predictions, and may produce “hallucinations.” Leveraging a pre-trained seq2seq transformer and a dense vector index of domain-specific data, this approach integrates real-time data retrieval into the generative process, enabling gap analysis and the dynamic generation of compliance and risk management content. We delve into the mechanics of RAG, focusing on its dual structure that pairs parametric knowledge contained within the transformer model with non-parametric data extracted from an updatable corpus. This hybrid model enhances decision-making through context-rich insights, drawing from the most current and relevant information, thereby enabling GRC officers to maintain a proactive compliance stance. Our methodology aligns with the latest advances in neural network fine-tuning, providing a granular, token-level application of retrieved information to inform and generate compliance narratives. By employing RAG, we exhibit a scalable solution that can adapt to novel regulatory challenges and cybersecurity threats, offering GRC officers a robust, predictive tool that augments their expertise. The granular application of RAG’s dual structure not only improves compliance and risk management protocols but also informs the development of compliance narratives with pinpoint accuracy. It underscores AI’s emerging role in strategic risk mitigation and proactive policy formation, positioning GRC officers to anticipate and navigate the complexities of regulatory evolution confidently.Keywords: cybersecurity, gen AI, retrieval augmented generation, cybersecurity defense strategies
Procedia PDF Downloads 1015385 Regulatory Measures on Effective Nuclear Security and Safeguards System in Nigeria
Authors: Nnodi Chinweikpe Akelachi, Adebayo Oladini Kachollom Ifeoma
Abstract:
Insecurity and the possession of nuclear weapons for non-peaceful purposes constitute a major threat to global peace and security, and this undermines the capacity for sustainable development. In Nigeria, the threat of terrorism is a challenge to national stability. For over a decade, Nigeria has been faced with insecurity ranging from Boko-Haram terrorist groups, kidnapping and banditry. The threat exhibited by this non-state actor poses a huge challenge to nuclear and radiological high risks facilities in Nigeria. This challenge has resulted in the regulatory authority and International stakeholders formulating policies for a good mitigation strategy. This strategy is enshrined in formulated laws, regulations and guides like the repealed Nuclear Safety and Radiation Protection Act 19 of 1995 (Nuclear safety, Physical Security and Safeguards Bill), the Nigerian Physical Protection of Nuclear Material and Nuclear Facilities, and Nigerian Nuclear Safeguards Regulations of 2021. All this will help Nigeria’s effort to meet its national nuclear security and safeguards obligations. To further enhance the implementation of nuclear security and safeguards system, Nigeria has signed the Non-Proliferation Treaty (NPT) in 1970, the Comprehensive Safeguards Agreement (INFCIRC/358) in 1988, Additional Protocol in 2007 as well as the Convention on Physical Protection of Nuclear Material and its amendment in 2005. In view of the evolving threats by non-state actors in Nigeria, physical protection security upgrades are being implemented in nuclear and all high-risk radiological facilities through the support of the United States Department of Energy (US-DOE). Also, the IAEA has helped strengthen nuclear security and safeguard systems through the provision of technical assistance and capacity development. Efforts are being made to address some of the challenges identified in the cause of implementing the measures for effective nuclear security and safeguards systems in Nigeria. However, there are eminent challenges in the implementation of the measures within the security and systems in Nigeria. These challenges need to be addressed for an effective security and safeguard regime in Nigeria. This paper seeks to address the challenges encountered in implementing the regulatory and stakeholder measures for effective security and safeguards regime in Nigeria, amongst others.Keywords: nuclear regulatory body, nuclear facilities and activities, international stakeholders, security and safeguards measures
Procedia PDF Downloads 1185384 A Corpus-Based Study of Evaluative Language in Leading Articles in British Broadsheet and Tabloid Newspapers
Authors: Fatimah AlSaiari
Abstract:
In recent years, newspapers in the United Kingdom have been no longer just a means of sharing news about what happens in the world; they are also used to influence target readers by having them become more up-to-date, well-informed, entertained, exasperated, delighted, and infuriated. To achieve these objectives and maintain influence on public opinion, journalists use a particular language in which they can convey emotions and opinions, organize their discourse, and establish solidarity with their audience. This type of language has been widely analyzed under different labels, such as evaluation, appraisal, and stance. There is a considerable amount of linguistic and non-linguistic research devoted to analyzing this type of interpersonal language in journalistic discourse, and most of these studies were carried out to challenge the traditional assumptions of the objectivity and impartiality of news reporting. However, very little research has been undertaken on evaluative language in newspaper institutional editorials, and there is hardly any systematic or exhaustive analysis of this type of language in British tabloid and broadsheet newspapers. This study will attempt to provide new insights into the nature of authorial and non-authorial evaluation in leading articles in popular and quality British newspapers, along with their targets, sources, and discourse functions. The study will also attempt to develop a framework of evaluation that can be applied to evaluative lexical items in newspaper opinion texts. The framework is both theory-driven (i.e., it builds on and modifies previous frameworks of evaluation such as appraisal theory and parameter-based approach) and data-driven (i.e., it elicits the evaluative categories from the analysis of the corpus, which helps in the development of the current framework). To achieve this aim, a corpus of 140 leading articles were selected. The findings revealed that the tabloids tended to express their stance through explicitness, dramatization, frequent reference to social actors’ emotions and beliefs, and exaggeration in negativity, while the broadsheets preferred to express their stance through mitigation ambiguity and implicitness. conceptual themes and propositions were more preferable targets for expressing stance in the broadsheets while human behavior and characters were preferable targets for the tabloids.Keywords: appraisal theory, evaluative language, British newspapers, broadsheets & tabloids, evaluative adjectives
Procedia PDF Downloads 2955383 The Green Synthesis AgNPs from Basil Leaf Extract
Authors: Wanida Wonsawat
Abstract:
Bioreduction of silver nanoparticles (AgNPs) from silver ions (Ag+) using water extract of Thai basil leaf was successfully carried out. The basil leaf extract provided a reducing agent and stabilizing agent for a synthesis of metal nanoparticles. Silver nanoparticles received from cut and uncut basil leaf was compared. The resulting silver nanoparticles are characterized by UV-Vis spectroscopy. The maximum intensities of silver nanoparticle from cut and uncut basil leaf were 410 and 420, respectively. The techniques involved are simple, eco-friendly and rapid.Keywords: basil leaves, silver nanoparticles, green synthesis, plant extract
Procedia PDF Downloads 5945382 Understanding Complexity at Pre-Construction Stage in Project Planning of Construction Projects
Authors: Mehran Barani Shikhrobat, Roger Flanagan
Abstract:
The construction planning and scheduling based on using the current tools and techniques is resulted deterministic in nature (Gantt chart, CPM) or applying a very little probability of completion (PERT) for each task. However, every project embodies assumptions and influences and should start with a complete set of clearly defined goals and constraints that remain constant throughout the duration of the project. Construction planners continue to apply the traditional methods and tools of “hard” project management that were developed for “ideal projects,” neglecting the potential influence of complexity on the design and construction process. The aim of this research is to investigate the emergence and growth of complexity in project planning and to provide a model to consider the influence of complexity on the total project duration at the post-contract award pre-construction stage of a project. The literature review showed that complexity originates from different sources of environment, technical, and workflow interactions. They can be divided into two categories of complexity factors, first, project tasks, and second, project organisation management. Project tasks may originate from performance, lack of resources, or environmental changes for a specific task. Complexity factors that relate to organisation and management refer to workflow and interdependence of different parts. The literature review highlighted the ineffectiveness of traditional tools and techniques in planning for complexity. However, this research focus on understanding the fundamental causes of the complexity of construction projects were investigated through a questionnaire with industry experts. The results were used to develop a model that considers the core complexity factors and their interactions. System dynamics were used to investigate the model to consider the influence of complexity on project planning. Feedback from experts revealed 20 major complexity factors that impact project planning. The factors are divided into five categories known as core complexity factors. To understand the weight of each factor in comparison, the Analytical Hierarchy Process (AHP) analysis method is used. The comparison showed that externalities are ranked as the biggest influence across the complexity factors. The research underlines that there are many internal and external factors that impact project activities and the project overall. This research shows the importance of considering the influence of complexity on the project master plan undertaken at the post-contract award pre-construction phase of a project.Keywords: project planning, project complexity measurement, planning uncertainty management, project risk management, strategic project scheduling
Procedia PDF Downloads 1475381 Optimization of Robot Motion Planning Using Biogeography Based Optimization (Bbo)
Authors: Jaber Nikpouri, Arsalan Amralizadeh
Abstract:
In robotics manipulators, the trajectory should be optimum, thus the torque of the robot can be minimized in order to save power. This paper includes an optimal path planning scheme for a robotic manipulator. Recently, techniques based on metaheuristics of natural computing, mainly evolutionary algorithms (EA), have been successfully applied to a large number of robotic applications. In this paper, the improved BBO algorithm is used to minimize the objective function in the presence of different obstacles. The simulation represents that the proposed optimal path planning method has satisfactory performance.Keywords: biogeography-based optimization, path planning, obstacle detection, robotic manipulator
Procedia PDF Downloads 3085380 Structural Analysis and Modelling in an Evolving Iron Ore Operation
Authors: Sameh Shahin, Nannang Arrys
Abstract:
Optimizing pit slope stability and reducing strip ratio of a mining operation are two key tasks in geotechnical engineering. With a growing demand for minerals and an increasing cost associated with extraction, companies are constantly re-evaluating the viability of mineral deposits and challenging their geological understanding. Within Rio Tinto Iron Ore, the Structural Geology (SG) team investigate and collect critical data, such as point based orientations, mapping and geological inferences from adjacent pits to re-model deposits where previous interpretations have failed to account for structurally controlled slope failures. Utilizing innovative data collection methods and data-driven investigation, SG aims to address the root causes of slope instability. Committing to a resource grid drill campaign as the primary source of data collection will often bias data collection to a specific orientation and significantly reduce the capability to identify and qualify complexity. Consequently, these limitations make it difficult to construct a realistic and coherent structural model that identifies adverse structural domains. Without the consideration of complexity and the capability of capturing these structural domains, mining operations run the risk of inadequately designed slopes that may fail and potentially harm people. Regional structural trends have been considered in conjunction with surface and in-pit mapping data to model multi-batter fold structures that were absent from previous iterations of the structural model. The risk is evident in newly identified dip-slope and rock-mass controlled sectors of the geotechnical design rather than a ubiquitous dip-slope sector across the pit. The reward is two-fold: 1) providing sectors of rock-mass controlled design in previously interpreted structurally controlled domains and 2) the opportunity to optimize the slope angle for mineral recovery and reduced strip ratio. Furthermore, a resulting high confidence model with structures and geometries that can account for historic slope instabilities in structurally controlled domains where design assumptions failed.Keywords: structural geology, geotechnical design, optimization, slope stability, risk mitigation
Procedia PDF Downloads 525379 Development of a Data-Driven Method for Diagnosing the State of Health of Battery Cells, Based on the Use of an Electrochemical Aging Model, with a View to Their Use in Second Life
Authors: Desplanches Maxime
Abstract:
Accurate estimation of the remaining useful life of lithium-ion batteries for electronic devices is crucial. Data-driven methodologies encounter challenges related to data volume and acquisition protocols, particularly in capturing a comprehensive range of aging indicators. To address these limitations, we propose a hybrid approach that integrates an electrochemical model with state-of-the-art data analysis techniques, yielding a comprehensive database. Our methodology involves infusing an aging phenomenon into a Newman model, leading to the creation of an extensive database capturing various aging states based on non-destructive parameters. This database serves as a robust foundation for subsequent analysis. Leveraging advanced data analysis techniques, notably principal component analysis and t-Distributed Stochastic Neighbor Embedding, we extract pivotal information from the data. This information is harnessed to construct a regression function using either random forest or support vector machine algorithms. The resulting predictor demonstrates a 5% error margin in estimating remaining battery life, providing actionable insights for optimizing usage. Furthermore, the database was built from the Newman model calibrated for aging and performance using data from a European project called Teesmat. The model was then initialized numerous times with different aging values, for instance, with varying thicknesses of SEI (Solid Electrolyte Interphase). This comprehensive approach ensures a thorough exploration of battery aging dynamics, enhancing the accuracy and reliability of our predictive model. Of particular importance is our reliance on the database generated through the integration of the electrochemical model. This database serves as a crucial asset in advancing our understanding of aging states. Beyond its capability for precise remaining life predictions, this database-driven approach offers valuable insights for optimizing battery usage and adapting the predictor to various scenarios. This underscores the practical significance of our method in facilitating better decision-making regarding lithium-ion battery management.Keywords: Li-ion battery, aging, diagnostics, data analysis, prediction, machine learning, electrochemical model, regression
Procedia PDF Downloads 745378 Voltage Stability Assessment and Enhancement Using STATCOM -A Case Study
Authors: Puneet Chawla, Balwinder Singh
Abstract:
Recently, increased attention has been devoted to the voltage instability phenomenon in power systems. Many techniques have been proposed in the literature for evaluating and predicting voltage stability using steady state analysis methods. In this paper, P-V and Q-V curves have been generated for a 57 bus Patiala Rajpura circle of India. The power-flow program is developed in MATLAB using Newton-Raphson method. Using Q-V curves, the weakest bus of the power system and the maximum reactive power change permissible on that bus is calculated. STATCOMs are placed on the weakest bus to improve the voltage and hence voltage stability and also the power transmission capability of the line.Keywords: voltage stability, reactive power, power flow, weakest bus, STATCOM
Procedia PDF Downloads 5195377 Model Predictive Controller for Pasteurization Process
Authors: Tesfaye Alamirew Dessie
Abstract:
Our study focuses on developing a Model Predictive Controller (MPC) and evaluating it against a traditional PID for a pasteurization process. Utilizing system identification from the experimental data, the dynamics of the pasteurization process were calculated. Using best fit with data validation, residual, and stability analysis, the quality of several model architectures was evaluated. The validation data fit the auto-regressive with exogenous input (ARX322) model of the pasteurization process by roughly 80.37 percent. The ARX322 model structure was used to create MPC and PID control techniques. After comparing controller performance based on settling time, overshoot percentage, and stability analysis, it was found that MPC controllers outperform PID for those parameters.Keywords: MPC, PID, ARX, pasteurization
Procedia PDF Downloads 1675376 Estimation of Relative Subsidence of Collapsible Soils Using Electromagnetic Measurements
Authors: Henok Hailemariam, Frank Wuttke
Abstract:
Collapsible soils are weak soils that appear to be stable in their natural state, normally dry condition, but rapidly deform under saturation (wetting), thus generating large and unexpected settlements which often yield disastrous consequences for structures unwittingly built on such deposits. In this study, a prediction model for the relative subsidence of stressed collapsible soils based on dielectric permittivity measurement is presented. Unlike most existing methods for soil subsidence prediction, this model does not require moisture content as an input parameter, thus providing the opportunity to obtain accurate estimation of the relative subsidence of collapsible soils using dielectric measurement only. The prediction model is developed based on an existing relative subsidence prediction model (which is dependent on soil moisture condition) and an advanced theoretical frequency and temperature-dependent electromagnetic mixing equation (which effectively removes the moisture content dependence of the original relative subsidence prediction model). For large scale sub-surface soil exploration purposes, the spatial sub-surface soil dielectric data over wide areas and high depths of weak (collapsible) soil deposits can be obtained using non-destructive high frequency electromagnetic (HF-EM) measurement techniques such as ground penetrating radar (GPR). For laboratory or small scale in-situ measurements, techniques such as an open-ended coaxial line with widely applicable time domain reflectometry (TDR) or vector network analysers (VNAs) are usually employed to obtain the soil dielectric data. By using soil dielectric data obtained from small or large scale non-destructive HF-EM investigations, the new model can effectively predict the relative subsidence of weak soils without the need to extract samples for moisture content measurement. Some of the resulting benefits are the preservation of the undisturbed nature of the soil as well as a reduction in the investigation costs and analysis time in the identification of weak (problematic) soils. The accuracy of prediction of the presented model is assessed by conducting relative subsidence tests on a collapsible soil at various initial soil conditions and a good match between the model prediction and experimental results is obtained.Keywords: collapsible soil, dielectric permittivity, moisture content, relative subsidence
Procedia PDF Downloads 3675375 Analysis of Histogram Asymmetry for Waste Recognition
Authors: Janusz Bobulski, Kamila Pasternak
Abstract:
Despite many years of effort and research, the problem of waste management is still current. So far, no fully effective waste management system has been developed. Many programs and projects improve statistics on the percentage of waste recycled every year. In these efforts, it is worth using modern Computer Vision techniques supported by artificial intelligence. In the article, we present a method of identifying plastic waste based on the asymmetry analysis of the histogram of the image containing the waste. The method is simple but effective (94%), which allows it to be implemented on devices with low computing power, in particular on microcomputers. Such de-vices will be used both at home and in waste sorting plants.Keywords: waste management, environmental protection, image processing, computer vision
Procedia PDF Downloads 1245374 Nanowire by Ac Electrodeposition Into Nanoporous Alumina Fabrication of High Aspect Ratio Metalic
Authors: M. Beyzaiea, S. Mohammadia
Abstract:
High aspect ratio metallic (silver, cobalt) nanowire arrays were fabricated using ac electrodeposition techniques into the nanoporous alumina template. The template with long pore dept fabricated by hard anodization (HA) and thinned for ac electrodeposition. Template preparation was done in short time by using HA technique and high speed thing process. The TEM and XRD investigation confirm the three dimensional nucleation growth mechanism of metallic nanowire inside the nanoporous alumina that fabricated by HA process.Keywords: metallic, nanowire, nanoporous alumina, ac electrodeposition
Procedia PDF Downloads 2775373 Application of Low-order Modeling Techniques and Neural-Network Based Models for System Identification
Authors: Venkatesh Pulletikurthi, Karthik B. Ariyur, Luciano Castillo
Abstract:
The system identification from the turbulence wakes will lead to the tactical advantage to prepare and also, to predict the trajectory of the opponents’ movements. A low-order modeling technique, POD, is used to predict the object based on the wake pattern and compared with pre-trained image recognition neural network (NN) to classify the wake patterns into objects. It is demonstrated that low-order modeling, POD, is able to predict the objects better compared to pretrained NN by ~30%.Keywords: the bluff body wakes, low-order modeling, neural network, system identification
Procedia PDF Downloads 1875372 Cyclostationary Analysis of Polytime Coded Signals for LPI Radars
Authors: Metuku Shyamsunder, Kakarla Subbarao, P. Prasanna
Abstract:
In radars, an electromagnetic waveform is sent, and an echo of the same signal is received by the receiver. From this received signal, by extracting various parameters such as round trip delay, Doppler frequency it is possible to find distance, speed, altitude, etc. However, nowadays as the technology increases, intruders are intercepting transmitted signal as it reaches them, and they will be extracting the characteristics and trying to modify them. So there is a need to develop a system whose signal cannot be identified by no cooperative intercept receivers. That is why LPI radars came into existence. In this paper, a brief discussion on LPI radar and its modulation (polytime code (PT1)), detection (cyclostationary (DFSM & FAM) techniques such as DFSM, FAM are presented and compared with respect to computational complexity.Keywords: LPI radar, polytime codes, cyclostationary DFSM, FAM
Procedia PDF Downloads 4795371 The ‘Quartered Head Technique’: A Simple, Reliable Way of Maintaining Leg Length and Offset during Total Hip Arthroplasty
Authors: M. Haruna, O. O. Onafowokan, G. Holt, K. Anderson, R. G. Middleton
Abstract:
Background: Requirements for satisfactory outcomes following total hip arthroplasty (THA) include restoration of femoral offset, version, and leg length. Various techniques have been described for restoring these biomechanical parameters, with leg length restoration being the most predominantly described. We describe a “quartered head technique” (QHT) which uses a stepwise series of femoral head osteotomies to identify and preserve the centre of rotation of the femoral head during THA in order to ensure reconstruction of leg length, offset and stem version, such that hip biomechanics are restored as near to normal as possible. This study aims to identify whether using the QHT during hip arthroplasty effectively restores leg length and femoral offset to within acceptable parameters. Methods: A retrospective review of 206 hips was carried out, leaving 124 hips in the final analysis. Power analysis indicated a minimum of 37 patients required. All operations were performed using an anterolateral approach by a single surgeon. All femoral implants were cemented, collarless, polished double taper CPT® stems (Zimmer, Swindon, UK). Both cemented, and uncemented acetabular components were used (Zimmer, Swindon, UK). Leg length, version, and offset were assessed intra-operatively and reproduced using the QHT. Post-operative leg length and femoral offset were determined and compared with the contralateral native hip, and the difference was then calculated. For the determination of leg length discrepancy (LLD), we used the method described by Williamson & Reckling, which has been shown to be reproducible with a measurement error of ±1mm. As a reference, the inferior margin of the acetabular teardrop and the most prominent point of the lesser trochanter were used. A discrepancy of less than 6mm LLD was chosen as acceptable. All peri-operative radiographs were assessed by two independent observers. Results: The mean absolute post-operative difference in leg length from the contralateral leg was +3.58mm. 84% of patients (104/124) had LLD within ±6mm of the contralateral limb. The mean absolute post-operative difference in offset from contralateral leg was +3.88mm (range -15 to +9mm, median 3mm). 90% of patients (112/124) were within ±6mm offset of the contralateral limb. There was no statistical difference noted between observer measurements. Conclusion: The QHT provides a simple, inexpensive yet effective method of maintaining femoral leg length and offset during total hip arthroplasty. Combining this technique with pre-operative templating or other techniques described may enable surgeons to reduce even further the discrepancies between pre-operative state and post-operative outcome.Keywords: leg length discrepancy, technical tip, total hip arthroplasty, operative technique
Procedia PDF Downloads 885370 Evaluation of the Impact of Functional Communication Training on Behaviors of Concern for Students at a Non-Maintained Special School
Authors: Kate Duggan
Abstract:
Introduction: Functional Communication Training (FCT) is an approach which aims to reduce behaviours of concern by teaching more effective ways to communicate. It requires identification of the function of the behaviour of concern, through gathering information from key stakeholders and completing observations of the individual’s behaviour including antecedents to, and consequences of the behaviour. Appropriate communicative alternatives are then identified and taught to the individual using systematic instruction techniques. Behaviours of concern demonstrated by individuals with autism spectrum conditions (ASC) frequently have a communication function. When contributing to positive behavior support plans, speech and language therapists and other professionals working with individuals with ASC need to identify alternative communicative behaviours which are equally reinforcing as the existing behaviours of concern. Successful implementation of FCT is dependent on an effective ‘response match’. The new way of communicating must be equally as effective as the behaviour previously used and require the same amount or less effort from the individual. It must also be understood by the communication partners the individual encounters and be appropriate to their communicative contexts. Method: Four case studies within a non-maintained special school environment were described and analysed. A response match framework was used to identify the effectiveness of functional communication training delivered by the student’s speech and language therapist, teacher and learning support assistants. The success of systematic instruction techniques used to develop new communicative behaviours was evaluated using the CODES framework. Findings: Functional communication training can be used as part of a positive behaviour support approach for students within this setting. All case studies reviewed demonstrated ‘response success’, in that the desired response was gained from the new communicative behaviour. Barriers to the successful embedding of new communicative behaviours were encountered. In some instances, the new communicative behaviour could not be consistently understood across all communication partners which reduced ‘response recognisability’. There was also evidence of increased physical or cognitive difficulty in employing the new communicative behaviour which reduced the ‘response effectivity’. Successful use of ‘thinning schedules of reinforcement’, taught students to tolerate a delay to reinforcement once the new communication behaviour was learned.Keywords: augmentative and alternative communication, autism spectrum conditions, behaviours of concern, functional communication training
Procedia PDF Downloads 1205369 Increasing the Competitiveness of Batik Products as a Ready-To-Wear Cash Material Through Patterned Batik Innovation with Quilting Technique, at Klampar Batik Tourism Village
Authors: Urip Wahyuningsih, Indarti, Yuhri Inang Prihatina
Abstract:
The current development of batik art has given rise to various batik industries. The emergence of the batik industry is in order to meet the needs of the increasing share of the batik fashion market. This gives rise to competitiveness between the batik industry to compete for a share of the existing batik clothing market. Conditions like this also occur in Klampar Pamekasan Maduira Village, as one of the Batik Tourism Villages in Indonesia, it must continue to improve by trying to maintain the characteristics of Klampar Pamekasan Madura batik fashion and must also always innovate so that it remains highly competitive so that it remains one of the places popular batik tourist destination. Ready-to-wear or ready-to-wear clothing is clothing that is mass produced and produced in various sizes and colors, which can be purchased directly and worn easily. Patterned batik cloth is basically batik cloth that has the pattern lines of the clothing parts arranged efficiently, so there is no need to bother designing the pattern layout of the clothing parts on the batik cloth to be cut. Quilting can be defined as the art of combining fabric materials of certain sizes and cuts to form unique motifs. Based on several things above, breakthrough production innovation is needed without abandoning the characteristic of Klampar Pamekasan Madura Batik as one of the Batik Tourism Villages in Indonesia. One innovation that can be done is creating ready-to-wear patterned batik clothing products using a quilting technique. The method used in this research is the Double Diamond Design Process method. This method is divided into 4 phases namely, discover (namely the stage of designing the theme of the ready-to-wear patterned batik fashion innovation concept using quilting techniques in the Batik Village, Klampar Village, Pamekasasan, Madura), define (determine the design summary and present challenges to the design), develop ( presents prototypes developed, tested, reviewed and refined) and deliver (selected designs are produced, pass final tests and are ready to be commercialized). The research produces patterned batik products that are ready to wear with quilting techniques that are validated by experts and accepted by the public.Keywords: competitiveness, ready to wear, innovation, quilting, klampar batik vllage
Procedia PDF Downloads 525368 Synthesis and Characterisation of Bi-Substituted Magnetite Nanoparticles by Mechanochemical Processing (MCP)
Authors: Morteza Mohri Esfahani, Amir S. H. Rozatian, Morteza Mozaffari
Abstract:
Single phase magnetite nanoparticles and Bi-substituted ones were prepared by mechanochemical processing (MCP). The effects of Bi-substitution on the structural and magnetic properties of the nanoparticles were studied by X-ray Diffraction (XRD) and magnetometry techniques, respectively. The XRD results showed that all samples have spinel phase and by increasing Bi content, the main diffraction peaks were shifted to higher angles, which means the lattice parameter decreases from 0.843 to 0.838 nm and then increases to 0.841 nm. Also, the results revealed that increasing Bi content lead to a decrease in saturation magnetization (Ms) from 74.9 to 48.8 emu/g and an increase in coercivity (Hc) from 96.8 to 137.1 Oe.Keywords: bi-substituted magnetite nanoparticles, mechanochemical processing, X-ray diffraction, magnetism
Procedia PDF Downloads 5395367 Landsat Data from Pre Crop Season to Estimate the Area to Be Planted with Summer Crops
Authors: Valdir Moura, Raniele dos Anjos de Souza, Fernando Gomes de Souza, Jose Vagner da Silva, Jerry Adriani Johann
Abstract:
The estimate of the Area of Land to be planted with annual crops and its stratification by the municipality are important variables in crop forecast. Nowadays in Brazil, these information’s are obtained by the Brazilian Institute of Geography and Statistics (IBGE) and published under the report Assessment of the Agricultural Production. Due to the high cloud cover in the main crop growing season (October to March) it is difficult to acquire good orbital images. Thus, one alternative is to work with remote sensing data from dates before the crop growing season. This work presents the use of multitemporal Landsat data gathered on July and September (before the summer growing season) in order to estimate the area of land to be planted with summer crops in an area of São Paulo State, Brazil. Geographic Information Systems (GIS) and digital image processing techniques were applied for the treatment of the available data. Supervised and non-supervised classifications were used for data in digital number and reflectance formats and the multitemporal Normalized Difference Vegetation Index (NDVI) images. The objective was to discriminate the tracts with higher probability to become planted with summer crops. Classification accuracies were evaluated using a sampling system developed basically for this study region. The estimated areas were corrected using the error matrix derived from these evaluations. The classification techniques presented an excellent level according to the kappa index. The proportion of crops stratified by municipalities was derived by a field work during the crop growing season. These proportion coefficients were applied onto the area of land to be planted with summer crops (derived from Landsat data). Thus, it was possible to derive the area of each summer crop by the municipality. The discrepancies between official statistics and our results were attributed to the sampling and the stratification procedures. Nevertheless, this methodology can be improved in order to provide good crop area estimates using remote sensing data, despite the cloud cover during the growing season.Keywords: area intended for summer culture, estimated area planted, agriculture, Landsat, planting schedule
Procedia PDF Downloads 1535366 Data Mining in Medicine Domain Using Decision Trees and Vector Support Machine
Authors: Djamila Benhaddouche, Abdelkader Benyettou
Abstract:
In this paper, we used data mining to extract biomedical knowledge. In general, complex biomedical data collected in studies of populations are treated by statistical methods, although they are robust, they are not sufficient in themselves to harness the potential wealth of data. For that you used in step two learning algorithms: the Decision Trees and Support Vector Machine (SVM). These supervised classification methods are used to make the diagnosis of thyroid disease. In this context, we propose to promote the study and use of symbolic data mining techniques.Keywords: biomedical data, learning, classifier, algorithms decision tree, knowledge extraction
Procedia PDF Downloads 5645365 Analysis of Different Classification Techniques Using WEKA for Diabetic Disease
Authors: Usama Ahmed
Abstract:
Data mining is the process of analyze data which are used to predict helpful information. It is the field of research which solve various type of problem. In data mining, classification is an important technique to classify different kind of data. Diabetes is most common disease. This paper implements different classification technique using Waikato Environment for Knowledge Analysis (WEKA) on diabetes dataset and find which algorithm is suitable for working. The best classification algorithm based on diabetic data is Naïve Bayes. The accuracy of Naïve Bayes is 76.31% and take 0.06 seconds to build the model.Keywords: data mining, classification, diabetes, WEKA
Procedia PDF Downloads 151