Search results for: preposition error detection
2411 Tool Development for Assessing Antineoplastic Drugs Surface Contamination in Healthcare Services and Other Workplaces
Authors: Benoit Atge, Alice Dhersin, Oscar Da Silva Cacao, Beatrice Martinez, Dominique Ducint, Catherine Verdun-Esquer, Isabelle Baldi, Mathieu Molimard, Antoine Villa, Mireille Canal-Raffin
Abstract:
Introduction: Healthcare workers' exposure to antineoplastic drugs (AD) is a burning issue for occupational medicine practitioners. Biological monitoring of occupational exposure (BMOE) is an essential tool for assessing AD contamination of healthcare workers. In addition to BMOE, surface sampling is a useful tool in order to understand how workers get contaminated, to identify sources of environmental contamination, to verify the effectiveness of surface decontamination way and to ensure monitoring of these surfaces. The objective of this work was to develop a complete tool including a kit for surface sampling and a quantification analytical method for AD traces detection. The development was realized with the three following criteria: the kit capacity to sample in every professional environment (healthcare services, veterinaries, etc.), the detection of very low AD traces with a validated analytical method and the easiness of the sampling kit use regardless of the person in charge of sampling. Material and method: AD mostly used in term of quantity and frequency have been identified by an analysis of the literature and consumptions of different hospitals, veterinary services, and home care settings. The kind of adsorbent device, surface moistening solution and mix of solvents for the extraction of AD from the adsorbent device have been tested for a maximal yield. The AD quantification was achieved by an ultra high-performance liquid chromatography method coupled with tandem mass spectrometry (UHPLC-MS/MS). Results: With their high frequencies of use and their good reflect of the diverse activities through healthcare, 15 AD (cyclophosphamide, ifosfamide, doxorubicin, daunorubicin, epirubicin, 5-FU, dacarbazin, etoposide, pemetrexed, vincristine, cytarabine, methothrexate, paclitaxel, gemcitabine, mitomycin C) were selected. The analytical method was optimized and adapted to obtain high sensitivity with very low limits of quantification (25 to 5000ng/mL), equivalent or lowest that those previously published (for 13/15 AD). The sampling kit is easy to use, provided with a didactic support (online video and protocol paper). It showed its effectiveness without inter-individual variation (n=5/person; n= 5 persons; p=0,85; ANOVA) regardless of the person in charge of sampling. Conclusion: This validated tool (sampling kit + analytical method) is very sensitive, easy to use and very didactic in order to control the chemical risk brought by AD. Moreover, BMOE permits a focal prevention. Used in routine, this tool is available for every intervention of occupational health.Keywords: surface contamination, sampling kit, analytical method, sensitivity
Procedia PDF Downloads 1382410 Automatic Segmentation of the Clean Speech Signal
Authors: M. A. Ben Messaoud, A. Bouzid, N. Ellouze
Abstract:
Speech Segmentation is the measure of the change point detection for partitioning an input speech signal into regions each of which accords to only one speaker. In this paper, we apply two features based on multi-scale product (MP) of the clean speech, namely the spectral centroid of MP, and the zero crossings rate of MP. We focus on multi-scale product analysis as an important tool for segmentation extraction. The multi-scale product is based on making the product of the speech wavelet transform coefficients at three successive dyadic scales. We have evaluated our method on the Keele database. Experimental results show the effectiveness of our method presenting a good performance. It shows that the two simple features can find word boundaries, and extracted the segments of the clean speech.Keywords: multiscale product, spectral centroid, speech segmentation, zero crossings rate
Procedia PDF Downloads 5012409 Evaluation of Three Digital Graphical Methods of Baseflow Separation Techniques in the Tekeze Water Basin in Ethiopia
Authors: Alebachew Halefom, Navsal Kumar, Arunava Poddar
Abstract:
The purpose of this work is to specify the parameter values, the base flow index (BFI), and to rank the methods that should be used for base flow separation. Three different digital graphical approaches are chosen and used in this study for the purpose of comparison. The daily time series discharge data were collected from the site for a period of 30 years (1986 up to 2015) and were used to evaluate the algorithms. In order to separate the base flow and the surface runoff, daily recorded streamflow (m³/s) data were used to calibrate procedures and get parameter values for the basin. Additionally, the performance of the model was assessed by the use of the standard error (SE), the coefficient of determination (R²), and the flow duration curve (FDC) and baseflow indexes. The findings indicate that, in general, each strategy can be used worldwide to differentiate base flow; however, the Sliding Interval Method (SIM) performs significantly better than the other two techniques in this basin. The average base flow index was calculated to be 0.72 using the local minimum method, 0.76 using the fixed interval method, and 0.78 using the sliding interval method, respectively.Keywords: baseflow index, digital graphical methods, streamflow, Emba Madre Watershed
Procedia PDF Downloads 872408 Impact of Different Modulation Techniques on the Performance of Free-Space Optics
Authors: Naman Singla, Ajay Pal Singh Chauhan
Abstract:
As the demand for providing high bit rate and high bandwidth is increasing at a rapid rate so there is a need to see in this problem and finds a technology that provides high bit rate and also high bandwidth. One possible solution is by use of optical fiber. Optical fiber technology provides high bandwidth in THz. But the disadvantage of optical fiber is of high cost and not used everywhere because it is not possible to reach all the locations on the earth. Also high maintenance required for usage of optical fiber. It puts a lot of cost. Another technology which is almost similar to optical fiber is Free Space Optics (FSO) technology. FSO is the line of sight technology where modulated optical beam whether infrared or visible is used to transfer information from one point to another through the atmosphere which works as a channel. This paper concentrates on analyzing the performance of FSO in terms of bit error rate (BER) and quality factor (Q) using different modulation techniques like non return to zero on off keying (NRZ-OOK), differential phase shift keying (DPSK) and differential quadrature phase shift keying (DQPSK) using OptiSystem software. The findings of this paper show that FSO system based on DQPSK modulation technique performs better.Keywords: attenuation, bit rate, free space optics, link length
Procedia PDF Downloads 3502407 Flexible Capacitive Sensors Based on Paper Sheets
Authors: Mojtaba Farzaneh, Majid Baghaei Nejad
Abstract:
This article proposes a new Flexible Capacitive Tactile Sensors based on paper sheets. This method combines the parameters of sensor's material and dielectric, and forms a new model of flexible capacitive sensors. The present article tries to present a practical explanation of this method's application and advantages. With the use of this new method, it is possible to make a more flexibility and accurate sensor in comparison with the current models. To assess the performance of this model, the common capacitive sensor is simulated and the proposed model of this article and one of the existing models are assessed. The results of this article indicate that the proposed model of this article can enhance the speed and accuracy of tactile sensor and has less error in comparison with the current models. Based on the results of this study, it can be claimed that in comparison with the current models, the proposed model of this article is capable of representing more flexibility and more accurate output parameters for touching the sensor, especially in abnormal situations and uneven surfaces, and increases accuracy and practicality.Keywords: capacitive sensor, paper sheets, flexible, tactile, uneven
Procedia PDF Downloads 3582406 First Digit Lucas, Fibonacci and Benford Number in Financial Statement
Authors: Teguh Sugiarto, Amir Mohamadian Amiri
Abstract:
Background: This study aims to explore if there is fraud in the company's financial report distribution using the number first digit Lucas, Fibonacci and Benford. Research methods: In this study, the author uses a number model contained in the first digit of the model Lucas, Fibonacci and Benford, to make a distinction between implementation by using the scale above and below 5%, the rate of occurrence of a difference against the digit number contained on Lucas, Fibonacci and Benford. If there is a significant difference above and below 5%, then the process of follow-up and detection of occurrence of fraud against the financial statements can be made. Findings: From research that has been done can be concluded that the number of frequency levels contained in the financial statements of PT Bank BRI Tbk in a year in the same conscientious results for model Lucas, Fibonacci and Benford.Keywords: Lucas, Fibonacci, Benford, first digit
Procedia PDF Downloads 2772405 Performance Evaluation and Dear Based Optimization on Machining Leather Specimens to Reduce Carbonization
Authors: Khaja Moiduddin, Tamer Khalaf, Muthuramalingam Thangaraj
Abstract:
Due to the variety of benefits over traditional cutting techniques, the usage of laser cutting technology has risen substantially in recent years. Hot wire machining can cut the leather in the required shape by controlling the wire by generating thermal energy. In the present study, an attempt has been made to investigate the effects of performance measures in the hot wire machining process on cutting leather specimens. Carbonization and material removal rates were considered as quality indicators. Burning leather during machining might cause carbon particles, reducing product quality. Minimizing the effect of carbon particles is crucial for assuring operator and environmental safety, health, and product quality. Hot wire machining can efficiently cut the specimens by controlling the current through it. Taguchi- DEAR-based optimization was also performed in the process, which resulted in a required Carbonization and material removal rate. Using the DEAR approach, the optimal parameters of the present study were found with 3.7% prediction error accuracy.Keywords: cabronization, leather, MRR, current
Procedia PDF Downloads 672404 Software Defect Analysis- Eclipse Dataset
Authors: Amrane Meriem, Oukid Salyha
Abstract:
The presence of defects or bugs in software can lead to costly setbacks, operational inefficiencies, and compromised user experiences. The integration of Machine Learning(ML) techniques has emerged to predict and preemptively address software defects. ML represents a proactive strategy aimed at identifying potential anomalies, errors, or vulnerabilities within code before they manifest as operational issues. By analyzing historical data, such as code changes, feature im- plementations, and defect occurrences. This en- ables development teams to anticipate and mitigate these issues, thus enhancing software quality, reducing maintenance costs, and ensuring smoother user interactions. In this work, we used a recommendation system to improve the performance of ML models in terms of predicting the code severity and effort estimation.Keywords: software engineering, machine learning, bugs detection, effort estimation
Procedia PDF Downloads 912403 Using Artificial Intelligence Method to Explore the Important Factors in the Reuse of Telecare by the Elderly
Authors: Jui-Chen Huang
Abstract:
This research used artificial intelligence method to explore elderly’s opinions on the reuse of telecare, its effect on their service quality, satisfaction and the relationship between customer perceived value and intention to reuse. This study conducted a questionnaire survey on the elderly. A total of 124 valid copies of a questionnaire were obtained. It adopted Backpropagation Network (BPN) to propose an effective and feasible analysis method, which is different from the traditional method. Two third of the total samples (82 samples) were taken as the training data, and the one third of the samples (42 samples) were taken as the testing data. The training and testing data RMSE (root mean square error) are 0.022 and 0.009 in the BPN, respectively. As shown, the errors are acceptable. On the other hand, the training and testing data RMSE are 0.100 and 0.099 in the regression model, respectively. In addition, the results showed the service quality has the greatest effects on the intention to reuse, followed by the satisfaction, and perceived value. This result of the Backpropagation Network method is better than the regression analysis. This result can be used as a reference for future research.Keywords: artificial intelligence, backpropagation network (BPN), elderly, reuse, telecare
Procedia PDF Downloads 2162402 The Use of Fractional Brownian Motion in the Generation of Bed Topography for Bodies of Water Coupled with the Lattice Boltzmann Method
Authors: Elysia Barker, Jian Guo Zhou, Ling Qian, Steve Decent
Abstract:
A method of modelling topography used in the simulation of riverbeds is proposed in this paper, which removes the need for datapoints and measurements of physical terrain. While complex scans of the contours of a surface can be achieved with other methods, this requires specialised tools, which the proposed method overcomes by using fractional Brownian motion (FBM) as a basis to estimate the real surface within a 15% margin of error while attempting to optimise algorithmic efficiency. This removes the need for complex, expensive equipment and reduces resources spent modelling bed topography. This method also accounts for the change in topography over time due to erosion, sediment transport, and other external factors which could affect the topography of the ground by updating its parameters and generating a new bed. The lattice Boltzmann method (LBM) is used to simulate both stationary and steady flow cases in a side-by-side comparison over the generated bed topography using the proposed method and a test case taken from an external source. The method, if successful, will be incorporated into the current LBM program used in the testing phase, which will allow an automatic generation of topography for the given situation in future research, removing the need for bed data to be specified.Keywords: bed topography, FBM, LBM, shallow water, simulations
Procedia PDF Downloads 1032401 Multiloop Fractional Order PID Controller Tuned Using Cuckoo Algorithm for Two Interacting Conical Tank Process
Authors: U. Sabura Banu, S. K. Lakshmanaprabu
Abstract:
The improvement of meta-heuristic algorithm encourages control engineer to design an optimal controller for industrial process. Most real-world industrial processes are non-linear multivariable process with high interaction. Even in sub-process unit, thousands of loops are available mostly interacting in nature. Optimal controller design for such process are still challenging task. Closed loop controller design by multiloop PID involves a tedious procedure by performing interaction study and then PID auto-tuning the loop with higher interaction. Finally, detuning the controller to accommodate the effects of the other process variables. Fractional order PID controllers are replacing integer order PID controllers recently. Design of Multiloop Fractional Order (MFO) PID controller is still more complicated. Cuckoo algorithm, a swarm intelligence technique is used to optimally tune the MFO PID controller with easiness minimizing Integral Time Absolute Error. The closed loop performance is tested under servo, regulatory and servo-regulatory conditions.Keywords: Cuckoo algorithm, mutliloop fractional order PID controller, two Interacting conical tank process
Procedia PDF Downloads 5022400 A Comparative Study of Generalized Autoregressive Conditional Heteroskedasticity (GARCH) and Extreme Value Theory (EVT) Model in Modeling Value-at-Risk (VaR)
Authors: Longqing Li
Abstract:
The paper addresses the inefficiency of the classical model in measuring the Value-at-Risk (VaR) using a normal distribution or a Student’s t distribution. Specifically, the paper focuses on the one day ahead Value-at-Risk (VaR) of major stock market’s daily returns in US, UK, China and Hong Kong in the most recent ten years under 95% confidence level. To improve the predictable power and search for the best performing model, the paper proposes using two leading alternatives, Extreme Value Theory (EVT) and a family of GARCH models, and compares the relative performance. The main contribution could be summarized in two aspects. First, the paper extends the GARCH family model by incorporating EGARCH and TGARCH to shed light on the difference between each in estimating one day ahead Value-at-Risk (VaR). Second, to account for the non-normality in the distribution of financial markets, the paper applies Generalized Error Distribution (GED), instead of the normal distribution, to govern the innovation term. A dynamic back-testing procedure is employed to assess the performance of each model, a family of GARCH and the conditional EVT. The conclusion is that Exponential GARCH yields the best estimate in out-of-sample one day ahead Value-at-Risk (VaR) forecasting. Moreover, the discrepancy of performance between the GARCH and the conditional EVT is indistinguishable.Keywords: Value-at-Risk, Extreme Value Theory, conditional EVT, backtesting
Procedia PDF Downloads 3262399 Multiple Fault Detection and Classification in a Coupled Motor with Rotor Using Artificial Neural Network
Authors: Mehrdad Nouri Khajavi, Gollamhassan Payganeh, Mohsen Fallah Tafti
Abstract:
Fault diagnosis is an important aspect of maintaining rotating machinery health and increasing productivity. Many researches has been done in this regards. Many faults such as unbalance, misalignment, looseness, bearing faults, etc. have been considered and diagnosed with different techniques. Most of the researches in fault diagnosis of rotating machinery deal with single fault. Where as in reality faults usually occur simultaneously and it is, therefore, necessary to recognize them at the same time. In this research, two of the most common faults namely unbalance and misalignment have been considered simultaneously with different intensity and then identified and classified with the use of Multi-Layer Perception Neural Network (MLPNN). Processed Vibration signals are used as the input to the MLPNN, and the class of mixed unbalancy, and misalignment is the output of the NN.Keywords: unbalance, parallel misalignment, combined faults, vibration signals
Procedia PDF Downloads 3592398 OFDM Radar for High Accuracy Target Tracking
Authors: Mahbube Eghtesad
Abstract:
For a number of years, the problem of simultaneous detection and tracking of a target has been one of the most relevant and challenging issues in a wide variety of military and civilian systems. We develop methods for detecting and tracking a target using an orthogonal frequency division multiplexing (OFDM) based radar. As a preliminary step we introduce the target trajectory and Gaussian noise model in discrete time form. Then resorting to match filter and Kalman filter we derive a detector and target tracker. After that we propose an OFDM radar in order to achieve further improvement in tracking performance. The motivation for employing multiple frequencies is that the different scattering centers of a target resonate differently at each frequency. Numerical examples illustrate our analytical results, demonstrating the achieved performance improvement due to the OFDM signaling method.Keywords: matched filter, target trashing, OFDM radar, Kalman filter
Procedia PDF Downloads 4022397 IoT and Edge Computing for Smog Management and Policy Making
Authors: Farhan Siddiqui
Abstract:
The increasingly complex challenges related to smog require the latest real-time monitoring and mitigation solutions. This paper describes a distributed IoT-edge architecture to improve smog detection, analysis, and policymaking. IoT sensors collect information related to critical air quality indicators while edge nodes perform local analytics with low latency for swift intervention. The system uses predictive algorithms to generate actionable insights to inform adaptive urban management strategies. Field implementations show dramatic improvements, including a 45 percent reduction in processing latency and improved predictive accuracy (R² = 0.92). These results show the potential of the framework to transform urban environmental management and policy making.Keywords: Internet of Things, edge computing, Smog management, air quality, policy making
Procedia PDF Downloads 162396 Comparison Of Data Mining Models To Predict Future Bridge Conditions
Authors: Pablo Martinez, Emad Mohamed, Osama Mohsen, Yasser Mohamed
Abstract:
Highway and bridge agencies, such as the Ministry of Transportation in Ontario, use the Bridge Condition Index (BCI) which is defined as the weighted condition of all bridge elements to determine the rehabilitation priorities for its bridges. Therefore, accurate forecasting of BCI is essential for bridge rehabilitation budgeting planning. The large amount of data available in regard to bridge conditions for several years dictate utilizing traditional mathematical models as infeasible analysis methods. This research study focuses on investigating different classification models that are developed to predict the bridge condition index in the province of Ontario, Canada based on the publicly available data for 2800 bridges over a period of more than 10 years. The data preparation is a key factor to develop acceptable classification models even with the simplest one, the k-NN model. All the models were tested, compared and statistically validated via cross validation and t-test. A simple k-NN model showed reasonable results (within 0.5% relative error) when predicting the bridge condition in an incoming year.Keywords: asset management, bridge condition index, data mining, forecasting, infrastructure, knowledge discovery in databases, maintenance, predictive models
Procedia PDF Downloads 1952395 Study of Nanocrystalline Scintillator for Alpha Particles Detection
Authors: Azadeh Farzaneh, Mohammad Reza Abdi, A. Quaranta, Matteo Dalla Palma, Seyedshahram Mortazavi
Abstract:
We report on the synthesis of cesium-iodide nanoparticles using sol-gel technique. The structural properties of CsI nanoparticles were characterized by X-ray diffraction and Scanning Electron Microscope (SEM) Also, optical properties were followed by optical absorption and UV–vis fluorescence. Intense photoluminescence is also observed, with some spectral tuning possible with ripening time getting a range of emission photon wavelength approximately from 366 to 350 nm. The size effect on CsI luminescence leads to an increase in scintillation light yield, a redshift of the emission bands of the on_center and off_center self_trapped excitons (STEs) and an increase in the contribution of the off_center STEs to the net intrinsic emission yield. The energy transfer from the matrix to CsI nanoparticles is a key characteristic for scintillation detectors. So the scintillation spectra to alpha particles of sample were monitored.Keywords: nanoparticles, luminescence, sol gel, scintillator
Procedia PDF Downloads 6052394 The Impact of the Number of Neurons in the Hidden Layer on the Performance of MLP Neural Network: Application to the Fast Identification of Toxics Gases
Authors: Slimane Ouhmad, Abdellah Halimi
Abstract:
In this work, we have applied neural networks method MLP type to a database from an array of six sensors for the detection of three toxic gases. As the choice of the number of hidden layers and the weight values has a great influence on the convergence of the learning algorithm, we proposed, in this article, a mathematical formulation to determine the optimal number of hidden layers and good weight values based on the method of back propagation of errors. The results of this modeling have improved discrimination of these gases on the one hand, and optimize the computation time on the other hand, the comparison to other results achieved in this case.Keywords: MLP Neural Network, back-propagation, number of neurons in the hidden layer, identification, computing time
Procedia PDF Downloads 3532393 An Approach for Coagulant Dosage Optimization Using Soft Jar Test: A Case Study of Bangkhen Water Treatment Plant
Authors: Ninlawat Phuangchoke, Waraporn Viyanon, Setta Sasananan
Abstract:
The most important process of the water treatment plant process is the coagulation using alum and poly aluminum chloride (PACL), and the value of usage per day is a hundred thousand baht. Therefore, determining the dosage of alum and PACL are the most important factors to be prescribed. Water production is economical and valuable. This research applies an artificial neural network (ANN), which uses the Levenberg–Marquardt algorithm to create a mathematical model (Soft Jar Test) for prediction chemical dose used to coagulation such as alum and PACL, which input data consists of turbidity, pH, alkalinity, conductivity, and, oxygen consumption (OC) of Bangkhen water treatment plant (BKWTP) Metropolitan Waterworks Authority. The data collected from 1 January 2019 to 31 December 2019 cover changing seasons of Thailand. The input data of ANN is divided into three groups training set, test set, and validation set, which the best model performance with a coefficient of determination and mean absolute error of alum are 0.73, 3.18, and PACL is 0.59, 3.21 respectively.Keywords: soft jar test, jar test, water treatment plant process, artificial neural network
Procedia PDF Downloads 1702392 Capability Prediction of Machining Processes Based on Uncertainty Analysis
Authors: Hamed Afrasiab, Saeed Khodaygan
Abstract:
Prediction of machining process capability in the design stage plays a key role to reach the precision design and manufacturing of mechanical products. Inaccuracies in machining process lead to errors in position and orientation of machined features on the part, and strongly affect the process capability in the final quality of the product. In this paper, an efficient systematic approach is given to investigate the machining errors to predict the manufacturing errors of the parts and capability prediction of corresponding machining processes. A mathematical formulation of fixture locators modeling is presented to establish the relationship between the part errors and the related sources. Based on this method, the final machining errors of the part can be accurately estimated by relating them to the combined dimensional and geometric tolerances of the workpiece – fixture system. This method is developed for uncertainty analysis based on the Worst Case and statistical approaches. The application of the presented method is illustrated through presenting an example and the computational results are compared with the Monte Carlo simulation results.Keywords: process capability, machining error, dimensional and geometrical tolerances, uncertainty analysis
Procedia PDF Downloads 3112391 The Effect of Technology on Skin Development and Progress
Authors: Haidy Weliam Megaly Gouda
Abstract:
Dermatology is often a neglected specialty in low-resource settings despite the high morbidity associated with skin disease. This becomes even more significant when associated with HIV infection, as dermatological conditions are more common and aggressive in HIV-positive patients. African countries have the highest HIV infection rates, and skin conditions are frequently misdiagnosed and mismanaged because of a lack of dermatological training and educational material. The frequent lack of diagnostic tests in the African setting renders basic clinical skills all the more vital. This project aimed to improve the diagnosis and treatment of skin disease in the HIV population in a district hospital in Malawi. A basic dermatological clinical tool was developed and produced in collaboration with local staff and based on available literature and data collected from clinics. The aim was to improve diagnostic accuracy and provide guidance for the treatment of skin disease in HIV-positive patients. A literature search within Embassy, Medline and Google Scholar was performed and supplemented through data obtained from attending 5 Antiretroviral clinics. From the literature, conditions were selected for inclusion in the resource if they were described as specific, more prevalent, or extensive in the HIV population or have more adverse outcomes if they develop in HIV patients. Resource-appropriate treatment options were decided using Malawian Ministry of Health guidelines and textbooks specific to African dermatology. After the collection of data and discussion with local clinical and pharmacy staff, a list of 15 skin conditions was included, and a booklet was created using the simple layout of a picture, a diagnostic description of the disease and treatment options. Clinical photographs were collected from local clinics (with full consent of the patient) or from the book ‘Common Skin Diseases in Africa’ (permission granted if fully acknowledged and used in a not-for-profit capacity). This tool was evaluated by the local staff alongside an educational teaching session on skin disease. This project aimed to reduce uncertainty in diagnosis and provide guidance for appropriate treatment in HIV patients by gathering information into one practical and manageable resource. To further this project, we hope to review the effectiveness of the tool in practice.Keywords: prevalence and pattern of skin diseases, impact on quality of life, rural Nepal, interventions, quality switched ruby laser, skin color river blindness, clinical signs, circularity index, grey level run length matrix, grey level co-occurrence matrix, local binary pattern, object detection, ring detection, shape identification
Procedia PDF Downloads 682390 Design of a Universal Wireless Battery Charger
Authors: Ahmad B. Musamih, Ahmad A. Albloushi, Ahmed H. Alshemeili, Abdulaziz Y. Alfili, Ala A. Hussien
Abstract:
This paper proposes a universal wireless battery charger design for portable electronic devices. As the number of portable electronics devices increases, the demand for more flexible and reliable charging techniques is becoming more urgent. A wireless battery charger differs from a traditional charger in the way the power transferred to the battery. In the latter, the power is transferred through electrical wires that connect the charger terminals to the battery terminals, while in the former; the power is transferred by induction without electrical connections. With a detection algorithm that detects the battery size and chemistry, the proposed charger will be able to accommodate a wide range of applications, and will allow a more flexible and reliable option to most of today’s portable electronics.Keywords: efficiency, magnetically-coupled resonators, resonance frequency, wireless power transfer
Procedia PDF Downloads 4572389 Assessing Sydney Tar Ponds Remediation and Natural Sediment Recovery in Nova Scotia, Canada
Authors: Tony R. Walker, N. Devin MacAskill, Andrew Thalhiemer
Abstract:
Sydney Harbour, Nova Scotia has long been subject to effluent and atmospheric inputs of metals, polycyclic aromatic hydrocarbons (PAHs), and polychlorinated biphenyls (PCBs) from a large coking operation and steel plant that operated in Sydney for nearly a century until closure in 1988. Contaminated effluents from the industrial site resulted in the creation of the Sydney Tar Ponds, one of Canada’s largest contaminated sites. Since its closure, there have been several attempts to remediate this former industrial site and finally, in 2004, the governments of Canada and Nova Scotia committed to remediate the site to reduce potential ecological and human health risks to the environment. The Sydney Tar Ponds and Coke Ovens cleanup project has become the most prominent remediation project in Canada today. As an integral part of remediation of the site (i.e., which consisted of solidification/stabilization and associated capping of the Tar Ponds), an extensive multiple media environmental effects program was implemented to assess what effects remediation had on the surrounding environment, and, in particular, harbour sediments. Additionally, longer-term natural sediment recovery rates of select contaminants predicted for the harbour sediments were compared to current conditions. During remediation, potential contributions to sediment quality, in addition to remedial efforts, were evaluated which included a significant harbour dredging project, propeller wash from harbour traffic, storm events, adjacent loading/unloading of coal and municipal wastewater treatment discharges. Two sediment sampling methodologies, sediment grab and gravity corer, were also compared to evaluate the detection of subtle changes in sediment quality. Results indicated that overall spatial distribution pattern of historical contaminants remains unchanged, although at much lower concentrations than previously reported, due to natural recovery. Measurements of sediment indicator parameter concentrations confirmed that natural recovery rates of Sydney Harbour sediments were in broad agreement with predicted concentrations, in spite of ongoing remediation activities. Overall, most measured parameters in sediments showed little temporal variability even when using different sampling methodologies, during three years of remediation compared to baseline, except for the detection of significant increases in total PAH concentrations noted during one year of remediation monitoring. The data confirmed the effectiveness of mitigation measures implemented during construction relative to harbour sediment quality, despite other anthropogenic activities and the dynamic nature of the harbour.Keywords: contaminated sediment, monitoring, recovery, remediation
Procedia PDF Downloads 2382388 A Guide to the Implementation of Ambisonics Super Stereo
Authors: Alessio Mastrorillo, Giuseppe Silvi, Francesco Scagliola
Abstract:
In this work, we introduce an Ambisonics decoder with an implementation of the C-format, also called Super Stereo. This format is an alternative to conventional stereo and binaural decoding. Unlike those, this format conveys audio information from the horizontal plane and works with stereo speakers and headphones. The two C-format channels can also return a reconstructed planar B-format. This work provides an open-source implementation for this format. We implement an all-pass filter for signal quadrature, as required by the decoding equations. This filter works with six Biquads in a cascade configuration, with values for control frequency and quality factor discovered experimentally. The phase response of the filter delivers a small error in the 20-14.000Hz range. The decoder has been tested with audio sources up to 192kHz sample rate, returning pristine sound quality and detailed stereo image. It has been included in the Envelop for Live suite and is available as an open-source repository. This decoder has applications in Virtual Reality and 360° audio productions, music composition, and online streaming.Keywords: ambisonics, UHJ, quadrature filter, virtual reality, Gerzon, decoder, stereo, binaural, biquad
Procedia PDF Downloads 942387 An Analytical Method for Maintenance Cost Estimating Relationships of Helicopters Using Linear Programming
Authors: Meesun Sun, Yongmin Kim
Abstract:
Estimating maintenance cost is crucial in defense management because it affects military budgets and availability of equipment. When it comes to estimating maintenance cost of the deployed equipment, time series forecasting can be applied with the actual historical cost data. It is more difficult issue to estimate maintenance cost of new equipment for which the actual costs are not provided. In this underlying context, this study proposes an analytical method for maintenance cost estimating relationships (CERs) development of helicopters using linear programming. The CERs can be applied to a new helicopter because they use non-cost independent variables such as the number of engines, the empty weight and so on. In the Republic of Korea, the maintenance cost of new equipment has been usually estimated by reflecting maintenance cost to unit price ratio of the legacy equipment. This study confirms that the CERs perform well for the 10 types of airmobile helicopters in terms of mean absolute percentage error by applying leave-one-out cross-validation. The suggested method is very useful to estimate the maintenance cost of new equipment and can help in the affordability assessment of acquisition program portfolios for total life cycle systems management.Keywords: affordability analysis, cost estimating relationship, helicopter, linear programming, maintenance cost
Procedia PDF Downloads 1412386 Review of Hydrologic Applications of Conceptual Models for Precipitation-Runoff Process
Authors: Oluwatosin Olofintoye, Josiah Adeyemo, Gbemileke Shomade
Abstract:
The relationship between rainfall and runoff is an important issue in surface water hydrology therefore the understanding and development of accurate rainfall-runoff models and their applications in water resources planning, management and operation are of paramount importance in hydrological studies. This paper reviews some of the previous works on the rainfall-runoff process modeling. The hydrologic applications of conceptual models and artificial neural networks (ANNs) for the precipitation-runoff process modeling were studied. Gradient training methods such as error back-propagation (BP) and evolutionary algorithms (EAs) are discussed in relation to the training of artificial neural networks and it is shown that application of EAs to artificial neural networks training could be an alternative to other training methods. Therefore, further research interest to exploit the abundant expert knowledge in the area of artificial intelligence for the solution of hydrologic and water resources planning and management problems is needed.Keywords: artificial intelligence, artificial neural networks, evolutionary algorithms, gradient training method, rainfall-runoff model
Procedia PDF Downloads 4602385 The Role of Maladaptive Personality Traits in Obesity Treatment – Quantitative Study
Authors: Judita Konečná, Dagmar Halo, Martin Matoulek
Abstract:
Background: Personality pathology does not have to be a contraindication nor an obstacle in obesity treatment, or eventually, surgical treatment. Detection of specific maladaptive personality traits can help us understand the manner of behavior leading to obesity as well as to address the treatment better. Objective: Using The Personality Inventory for DSM-5 (PID-5) in combination with clinical interviews with the goal of gaining a psychological evaluation to set the treatment procedure. Data was collected from more than 400 patients to detect differences in constellations of maladaptive personality traits based on BMI, DM2 and gender. Conclusions: Besides the fact that a psychological evaluation can help address the treatment better, analyses showed that it is also useful to detect specific groups of patients. Implications for clinical practice are discussed, as well as recommendations for group education programs based on quantitative research.Keywords: bariatric surgery, obesity, personality traits, PID-5, treatment
Procedia PDF Downloads 2342384 Multimodal Employee Attendance Management System
Authors: Khaled Mohammed
Abstract:
This paper presents novel face recognition and identification approaches for the real-time attendance management problem in large companies/factories and government institutions. The proposed uses the Minimum Ratio (MR) approach for employee identification. Capturing the authentic face variability from a sequence of video frames has been considered for the recognition of faces and resulted in system robustness against the variability of facial features. Experimental results indicated an improvement in the performance of the proposed system compared to the Previous approaches at a rate between 2% to 5%. In addition, it decreased the time two times if compared with the Previous techniques, such as Extreme Learning Machine (ELM) & Multi-Scale Structural Similarity index (MS-SSIM). Finally, it achieved an accuracy of 99%.Keywords: attendance management system, face detection and recognition, live face recognition, minimum ratio
Procedia PDF Downloads 1592383 Robust Data Image Watermarking for Data Security
Authors: Harsh Vikram Singh, Ankur Rai, Anand Mohan
Abstract:
In this paper, we propose secure and robust data hiding algorithm based on DCT by Arnold transform and chaotic sequence. The watermark image is scrambled by Arnold cat map to increases its security and then the chaotic map is used for watermark signal spread in middle band of DCT coefficients of the cover image The chaotic map can be used as pseudo-random generator for digital data hiding, to increase security and robustness .Performance evaluation for robustness and imperceptibility of proposed algorithm has been made using bit error rate (BER), normalized correlation (NC), and peak signal to noise ratio (PSNR) value for different watermark and cover images such as Lena, Girl, Tank images and gain factor .We use a binary logo image and text image as watermark. The experimental results demonstrate that the proposed algorithm achieves higher security and robustness against JPEG compression as well as other attacks such as addition of noise, low pass filtering and cropping attacks compared to other existing algorithm using DCT coefficients. Moreover, to recover watermarks in proposed algorithm, there is no need to original cover image.Keywords: data hiding, watermarking, DCT, chaotic sequence, arnold transforms
Procedia PDF Downloads 5172382 Cointegration Dynamics in Asian Stock Markets: Implications for Long-Term Portfolio Management
Authors: Xinyi Xu
Abstract:
This study conducts a detailed examination of Asian stock markets over the period from 2008 to 2023, with a focus on the dynamics of cointegration and their relevance for long-term investment strategies. Specifically, we assess the co-movement and potential for pairs trading—a strategy where investors take opposing positions on two stocks, indices, or financial instruments that historically move together. For example, we explore the relationship between the Nikkei 225 (N225), Japan’s benchmark stock index, and the Straits Times Index (STI) of Singapore, as well as the relationship between the Korea Composite Stock Price Index (KS11) and the STI. The methodology includes tests for normality, stationarity, cointegration, and the application of Vector Error Correction Modeling (VECM). Our findings reveal significant long-term relationships between these pairs, indicating opportunities for pairs trading strategies. Furthermore, the research underscores the challenges posed by model instability and the influence of major global incidents, which are identified as structural breaks. These findings pave the way for further exploration into the intricacies of financial market dynamics.Keywords: normality tests, stationarity, cointegration, VECM, pairs trading
Procedia PDF Downloads 59