Search results for: optimization algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4709

Search results for: optimization algorithms

1169 Effects of Various Wavelet Transforms in Dynamic Analysis of Structures

Authors: Seyed Sadegh Naseralavi, Sadegh Balaghi, Ehsan Khojastehfar

Abstract:

Time history dynamic analysis of structures is considered as an exact method while being computationally intensive. Filtration of earthquake strong ground motions applying wavelet transform is an approach towards reduction of computational efforts, particularly in optimization of structures against seismic effects. Wavelet transforms are categorized into continuum and discrete transforms. Since earthquake strong ground motion is a discrete function, the discrete wavelet transform is applied in the present paper. Wavelet transform reduces analysis time by filtration of non-effective frequencies of strong ground motion. Filtration process may be repeated several times while the approximation induces more errors. In this paper, strong ground motion of earthquake has been filtered once applying each wavelet. Strong ground motion of Northridge earthquake is filtered applying various wavelets and dynamic analysis of sampled shear and moment frames is implemented. The error, regarding application of each wavelet, is computed based on comparison of dynamic response of sampled structures with exact responses. Exact responses are computed by dynamic analysis of structures applying non-filtered strong ground motion.

Keywords: wavelet transform, computational error, computational duration, strong ground motion data

Procedia PDF Downloads 361
1168 Optimization of Poly-β-Hydroxybutyrate Recovery from Bacillus Subtilis Using Solvent Extraction Process by Response Surface Methodology

Authors: Jayprakash Yadav, Nivedita Patra

Abstract:

Polyhydroxybutyrate (PHB) is an interesting material in the field of medical science, pharmaceutical industries, and tissue engineering because of its properties such as biodegradability, biocompatibility, hydrophobicity, and elasticity. PHB is naturally accumulated by several microbes in their cytoplasm during the metabolic process as energy reserve material. PHB can be extracted from cell biomass using halogenated hydrocarbons, chemicals, and enzymes. In this study, a cheaper and non-toxic solvent, acetone, was used for the extraction process. The different parameters like acetone percentage, and solvent pH, process temperature, and incubation periods were optimized using the Response Surface Methodology (RSM). RSM was performed and the determination coefficient (R2) value was found to be 0.8833 from the quadratic regression model with no significant lack of fit. The designed RSM model results indicated that the fitness of the response variable was significant (P-value < 0.0006) and satisfactory to denote the relationship between the responses in terms of PHB recovery and purity with respect to the values of independent variables. Optimum conditions for the maximum PHB recovery and purity were found to be solvent pH 7, extraction temperature - 43 °C, incubation time - 70 minutes, and percentage acetone – 30 % from this study. The maximum predicted PHB recovery was found to be 0.845 g/g biomass dry cell weight and the purity was found to be 97.23 % using the optimized conditions.

Keywords: acetone, PHB, RSM, halogenated hydrocarbons, extraction, bacillus subtilis.

Procedia PDF Downloads 416
1167 Harmonic Assessment and Mitigation in Medical Diagonesis Equipment

Authors: S. S. Adamu, H. S. Muhammad, D. S. Shuaibu

Abstract:

Poor power quality in electrical power systems can lead to medical equipment at healthcare centres to malfunction and present wrong medical diagnosis. Equipment such as X-rays, computerized axial tomography, etc. can pollute the system due to their high level of harmonics production, which may cause a number of undesirable effects like heating, equipment damages and electromagnetic interferences. The conventional approach of mitigation uses passive inductor/capacitor (LC) filters, which has some drawbacks such as, large sizes, resonance problems and fixed compensation behaviours. The current trends of solutions generally employ active power filters using suitable control algorithms. This work focuses on assessing the level of Total Harmonic Distortion (THD) on medical facilities and various ways of mitigation, using radiology unit of an existing hospital as a case study. The measurement of the harmonics is conducted with a power quality analyzer at the point of common coupling (PCC). The levels of measured THD are found to be higher than the IEEE 519-1992 standard limits. The system is then modelled as a harmonic current source using MATLAB/SIMULINK. To mitigate the unwanted harmonic currents a shunt active filter is developed using synchronous detection algorithm to extract the fundamental component of the source currents. Fuzzy logic controller is then developed to control the filter. The THD without the active power filter are validated using the measured values. The THD with the developed filter show that the harmonics are now within the recommended limits.

Keywords: power quality, total harmonics distortion, shunt active filters, fuzzy logic

Procedia PDF Downloads 455
1166 A Generic Middleware to Instantly Sync Intensive Writes of Heterogeneous Massive Data via Internet

Authors: Haitao Yang, Zhenjiang Ruan, Fei Xu, Lanting Xia

Abstract:

Industry data centers often need to sync data changes reliably and instantly from a large-scale of heterogeneous autonomous relational databases accessed via the not-so-reliable Internet, for which a practical universal sync middle of low maintenance and operation costs is most wanted, but developing such a product and adapting it for various scenarios are a very sophisticated and continuous practice. The authors have been devising, applying, and optimizing a generic sync middleware system, named GSMS since 2006, holding the principles or advantages that the middleware must be SyncML-compliant and transparent to data application layer logic, need not refer to implementation details of databases synced, does not rely on host computer operating systems deployed, and its construction is light weighted and hence, of low cost. A series of ultimate experiments with GSMS sync performance were conducted for a persuasive example of a source relational database that underwent a broad range of write loads, say, from one thousand to one million intensive writes within a few minutes. The tests proved that GSMS has achieved an instant sync level of well below a fraction of millisecond per record sync, and GSMS’ smooth performances under ultimate write loads also showed it is feasible and competent.

Keywords: heterogeneous massive data, instantly sync intensive writes, Internet generic middleware design, optimization

Procedia PDF Downloads 101
1165 Performance of the New Laboratory-Based Algorithm for HIV Diagnosis in Southwestern China

Authors: Yanhua Zhao, Chenli Rao, Dongdong Li, Chuanmin Tao

Abstract:

The Chinese Centers for Disease Control and Prevention (CCDC) issued a new laboratory-based algorithm for HIV diagnosis on April 2016, which initially screens with a combination HIV-1/HIV-2 antigen/antibody fourth-generation immunoassay (IA) followed, when reactive, an HIV-1/HIV-2 undifferentiated antibody IA in duplicate. Reactive specimens with concordant results undergo supplemental tests with western blots, or HIV-1 nucleic acid tests (NATs) and non-reactive specimens with discordant results receive HIV-1 NATs or p24 antigen tests or 2-4 weeks follow-up tests. However, little data evaluating the application of the new algorithm have been reported to date. The study was to evaluate the performance of new laboratory-based HIV diagnostic algorithm in an inpatient population of Southwest China over the initial 6 months by compared with the old algorithm. Plasma specimens collected from inpatients from May 1, 2016, to October 31, 2016, are submitted to the laboratory for screening HIV infection performed by both the new HIV testing algorithm and the old version. The sensitivity and specificity of the algorithms and the difference of the categorized numbers of plasmas were calculated. Under the new algorithm for HIV diagnosis, 170 of the total 52 749 plasma specimens were confirmed as positively HIV-infected (0.32%). The sensitivity and specificity of the new algorithm were 100% (170/170) and 100% (52 579/52 579), respectively; while 167 HIV-1 positive specimens were identified by the old algorithm with sensitivity 98.24% (167/170) and 100% (52 579/52 579), respectively. Three acute HIV-1 infections (AHIs) and two early HIV-1 infections (EHIs) were identified by the new algorithm; the former was missed by old procedure. Compared with the old version, the new algorithm produced fewer WB-indeterminate results (2 vs. 16, p = 0.001), which led to fewer follow-up tests. Therefore, the new HIV testing algorithm is more sensitive for detecting acute HIV-1 infections with maintaining the ability to verify the established HIV-1 infections and can dramatically decrease the greater number of WB-indeterminate specimens.

Keywords: algorithm, diagnosis, HIV, laboratory

Procedia PDF Downloads 380
1164 Data-Driven Simulations Tools for Der and Battery Rich Power Grids

Authors: Ali Moradiamani, Samaneh Sadat Sajjadi, Mahdi Jalili

Abstract:

Power system analysis has been a major research topic in the generation and distribution sections, in both industry and academia, for a long time. Several load flow and fault analysis scenarios have been normally performed to study the performance of different parts of the grid in the context of, for example, voltage and frequency control. Software tools, such as PSCAD, PSSE, and PowerFactory DIgSILENT, have been developed to perform these analyses accurately. Distribution grid had been the passive part of the grid and had been known as the grid of consumers. However, a significant paradigm shift has happened with the emergence of Distributed Energy Resources (DERs) in the distribution level. It means that the concept of power system analysis needs to be extended to the distribution grid, especially considering self sufficient technologies such as microgrids. Compared to the generation and transmission levels, the distribution level includes significantly more generation/consumption nodes thanks to PV rooftop solar generation and battery energy storage systems. In addition, different consumption profile is expected from household residents resulting in a diverse set of scenarios. Emergence of electric vehicles will absolutely make the environment more complicated considering their charging (and possibly discharging) requirements. These complexities, as well as the large size of distribution grids, create challenges for the available power system analysis software. In this paper, we study the requirements of simulation tools in the distribution grid and how data-driven algorithms are required to increase the accuracy of the simulation results.

Keywords: smart grids, distributed energy resources, electric vehicles, battery storage systsms, simulation tools

Procedia PDF Downloads 82
1163 Formex Algebra Adaptation into Parametric Design Tools: Dome Structures

Authors: Réka Sárközi, Péter Iványi, Attila B. Széll

Abstract:

The aim of this paper is to present the adaptation of the dome construction tool for formex algebra to the parametric design software Grasshopper. Formex algebra is a mathematical system, primarily used for planning structural systems such like truss-grid domes and vaults, together with the programming language Formian. The goal of the research is to allow architects to plan truss-grid structures easily with parametric design tools based on the versatile formex algebra mathematical system. To produce regular structures, coordinate system transformations are used and the dome structures are defined in spherical coordinate system. Owing to the abilities of the parametric design software, it is possible to apply further modifications on the structures and gain special forms. The paper covers the basic dome types, and also additional dome-based structures using special coordinate-system solutions based on spherical coordinate systems. It also contains additional structural possibilities like making double layer grids in all geometry forms. The adaptation of formex algebra and the parametric workflow of Grasshopper together give the possibility of quick and easy design and optimization of special truss-grid domes.

Keywords: parametric design, structural morphology, space structures, spherical coordinate system

Procedia PDF Downloads 229
1162 Qualitative Detection of HCV and GBV-C Co-infection in Cirrhotic Patients Using a SYBR Green Multiplex Real Time RT-PCR Technique

Authors: Shahzamani Kiana, Esmaeil Lashgarian Hamed, Merat Shahin

Abstract:

HCV and GBV-C belong to the Flaviviridae family of viruses and GBV-C is the closest virus to HCV genetically. Accumulative research is in progress all over the world to clarify clinical aspects of GBV-C. Possibility of interaction between HCV and GBV-C and also its consequence with other liver diseases are the most important clinical aspects which encourage researchers to develop a technique for simultaneous detection of these viruses. In this study a SYBR Green multiplex real time RT-PCR technique as a new economical and sensitive method was optimized for simultaneous detection of HCV/GBV-C in HCV positive plasma samples. After designing and selection of two pairs of specific primers for HCV and GBV-C, SYBR Green Real time RT-PCR technique optimization was performed separately for each virus. Establishment of multiplex PCR was the next step. Finally our technique was performed on positive and negative plasma samples. 89 cirrhotic HCV positive plasma samples (29 of genotype 3 a and 27 of genotype 1a) were collected from patients before receiving treatment. 14% of genotype 3a and 17.1% of genotype 1a showed HCV/GBV-C co-infection. As a result, 13.48% of 89 samples had HCV/GBV-C co-infection that was compatible with other results from all over the world. Data showed no apparent influence of HGV co-infection on the either clinical or virological aspect of HCV infection. Furthermore, with application of multiplex Real time RT-PCR technique, more time and cost could be saved in clinical-research settings.

Keywords: HCV, GBV-C, cirrhotic patients, multiplex real time RT- PCR

Procedia PDF Downloads 277
1161 Recommended Practice for Experimental Evaluation of the Seepage Sensitivity Damage of Coalbed Methane Reservoirs

Authors: Hao Liu, Lihui Zheng, Chinedu J. Okere, Chao Wang, Xiangchun Wang, Peng Zhang

Abstract:

The coalbed methane (CBM) extraction industry (an unconventional energy source) is yet to promulgated an established standard code of practice for the experimental evaluation of sensitivity damage of coal samples. The existing experimental process of previous researches mainly followed the industry standard for conventional oil and gas reservoirs (CIS). However, the existing evaluation method ignores certain critical differences between CBM reservoirs and conventional reservoirs, which could inevitably result in an inaccurate evaluation of sensitivity damage and, eventually, poor decisions regarding the formulation of formation damage prevention measures. In this study, we propose improved experimental guidelines for evaluating seepage sensitivity damage of CBM reservoirs by leveraging on the shortcomings of the existing methods. The proposed method was established via a theoretical analysis of the main drawbacks of the existing methods and validated through comparative experiments. The results show that the proposed evaluation technique provided reliable experimental results that can better reflect actual reservoir conditions and correctly guide future development of CBM reservoirs. This study is pioneering the research on the optimization of experimental parameters for efficient exploration and development of CBM reservoirs.

Keywords: coalbed methane, formation damage, permeability, unconventional energy source

Procedia PDF Downloads 109
1160 Optimization of Wire EDM Parameters for Fabrication of Micro Channels

Authors: Gurinder Singh Brar, Sarbjeet Singh, Harry Garg

Abstract:

Wire Electric Discharge Machining (WEDM) is thermal machining process capable of machining very hard electrically conductive material irrespective of their hardness. WEDM is being widely used to machine micro-scale parts with the high dimensional accuracy and surface finish. The objective of this paper is to optimize the process parameters of wire EDM to fabricate the microchannels and to calculate the surface finish and material removal rate of microchannels fabricated using wire EDM. The material used is aluminum 6061 alloy. The experiments were performed using CNC wire cut electric discharge machine. The effect of various parameters of WEDM like pulse on time (TON) with the levels (100, 150, 200), pulse off time (TOFF) with the levels (25, 35, 45) and current (IP) with the levels (105, 110, 115) were investigated to study the effect on output parameter i.e. Surface Roughness and Material Removal Rate (MRR). Each experiment was conducted under different conditions of a pulse on time, pulse off time and peak current. For material removal rate, TON and Ip were the most significant process parameter. MRR increases with the increase in TON and Ip and decreases with the increase in TOFF. For surface roughness, TON and Ip have the maximum effect and TOFF was found out to be less effective.

Keywords: microchannels, Wire Electric Discharge Machining (WEDM), Metal Removal Rate (MRR), surface finish

Procedia PDF Downloads 475
1159 Cognitive Approach at the Epicenter of Creative Accounting in Cameroonian Companies: The Relevance of the Psycho-Sociological Approach and the Theory of Cognitive Dissonance

Authors: Romuald Temomo Wamba, Robert Wanda

Abstract:

The issue of creative accounting in the psychological and sociological framework has been a mixed subject for over 60 years. The objective of this article is to ensure the existence of creative accounting in Cameroonian entities on the one hand and to understand the strategies used by audit agents to detect errors, omissions, irregularities, or inadequacies in the financial state; optimization techniques used by account preparers to strategically bypass texts on the other hand. To achieve this, we conducted an exploratory study using a cognitive approach, and the data analysis was performed by the software 'decision explorer'. The results obtained challenge the authors' cognition (manifest latent and deceptive behavior). The tax inspectors stress that the entities in Cameroon do not derogate from the rules of piloting in the financial statements. Likewise, they claim a change in current income and net income through depreciation, provisions, inventories, and the spreading of charges over long periods. This suggests the suspicion or intention of manipulating the financial statements. As for the techniques, the account preparers manage the accruals at the end of the year as the basis of the practice of creative accounting. Likewise, management accounts are more favorable to results management.

Keywords: creative accounting, sociocognitive approach, psychological and sociological approach, cognitive dissonance theory, cognitive mapping

Procedia PDF Downloads 171
1158 Fight against Money Laundering with Optical Character Recognition

Authors: Saikiran Subbagari, Avinash Malladhi

Abstract:

Anti Money Laundering (AML) regulations are designed to prevent money laundering and terrorist financing activities worldwide. Financial institutions around the world are legally obligated to identify, assess and mitigate the risks associated with money laundering and report any suspicious transactions to governing authorities. With increasing volumes of data to analyze, financial institutions seek to automate their AML processes. In the rise of financial crimes, optical character recognition (OCR), in combination with machine learning (ML) algorithms, serves as a crucial tool for automating AML processes by extracting the data from documents and identifying suspicious transactions. In this paper, we examine the utilization of OCR for AML and delve into various OCR techniques employed in AML processes. These techniques encompass template-based, feature-based, neural network-based, natural language processing (NLP), hidden markov models (HMMs), conditional random fields (CRFs), binarizations, pattern matching and stroke width transform (SWT). We evaluate each technique, discussing their strengths and constraints. Also, we emphasize on how OCR can improve the accuracy of customer identity verification by comparing the extracted text with the office of foreign assets control (OFAC) watchlist. We will also discuss how OCR helps to overcome language barriers in AML compliance. We also address the implementation challenges that OCR-based AML systems may face and offer recommendations for financial institutions based on the data from previous research studies, which illustrate the effectiveness of OCR-based AML.

Keywords: anti-money laundering, compliance, financial crimes, fraud detection, machine learning, optical character recognition

Procedia PDF Downloads 120
1157 Numerical Simulations of Acoustic Imaging in Hydrodynamic Tunnel with Model Adaptation and Boundary Layer Noise Reduction

Authors: Sylvain Amailland, Jean-Hugh Thomas, Charles Pézerat, Romuald Boucheron, Jean-Claude Pascal

Abstract:

The noise requirements for naval and research vessels have seen an increasing demand for quieter ships in order to fulfil current regulations and to reduce the effects on marine life. Hence, new methods dedicated to the characterization of propeller noise, which is the main source of noise in the far-field, are needed. The study of cavitating propellers in closed-section is interesting for analyzing hydrodynamic performance but could involve significant difficulties for hydroacoustic study, especially due to reverberation and boundary layer noise in the tunnel. The aim of this paper is to present a numerical methodology for the identification of hydroacoustic sources on marine propellers using hydrophone arrays in a large hydrodynamic tunnel. The main difficulties are linked to the reverberation of the tunnel and the boundary layer noise that strongly reduce the signal-to-noise ratio. In this paper it is proposed to estimate the reflection coefficients using an inverse method and some reference transfer functions measured in the tunnel. This approach allows to reduce the uncertainties of the propagation model used in the inverse problem. In order to reduce the boundary layer noise, a cleaning algorithm taking advantage of the low rank and sparse structure of the cross-spectrum matrices of the acoustic and the boundary layer noise is presented. This approach allows to recover the acoustic signal even well under the boundary layer noise. The improvement brought by this method is visible on acoustic maps resulting from beamforming and DAMAS algorithms.

Keywords: acoustic imaging, boundary layer noise denoising, inverse problems, model adaptation

Procedia PDF Downloads 307
1156 Optimal Design of Submersible Permanent Magnet Linear Synchronous Motor Based Design of Experiment and Genetic Algorithm

Authors: Xiao Zhang, Wensheng Xiao, Junguo Cui, Hongmin Wang

Abstract:

Submersible permanent magnet linear synchronous motors (SPMLSMs) are electromagnetic devices, which can directly drive plunger pump to obtain the crude oil. Those motors have been gradually applied in oil fields due to high thrust force density and high efficiency. Since the force performance closely depends on the concrete structural parameters, the seven different structural parameters are investigated in detail. This paper presents an optimum design of an SPMLSM to minimize the detent force and maximize the thrust by using design of experiment (DOE) and genetic algorithm (GA). The three significant structural parameters (air-gap length, slot width, pole-arc coefficient) are separately screened using 27 1/16 fractional factorial design (FFD) to investigate the significant effect of seven parameters used in this research on the force performance. Response surface methodology (RSM) is well adapted to make analytical model of thrust and detent force with constraints of corresponding significant parameters and enable objective function to be easily created, respectively. GA is performed as a searching tool to search for the Pareto-optimal solutions. By finite element analysis, the proposed PMLSM shows merits in improving thrust and reducing the detent force dramatically.

Keywords: optimization, force performance, design of experiment (DOE), genetic algorithm (GA)

Procedia PDF Downloads 271
1155 Chitosan Modified Halloysite Nanomaterials for Efficient and Effective Vaccine Delivery in Farmed Fish

Authors: Saji George, Eng Khuan Seng, Christof Luda

Abstract:

Nanotechnology has been recognized as an important tool for modern agriculture and has the potential to overcome some of the pressing challenges faced by aquaculture industry. A strategy for optimizing nanotechnology-based therapeutic delivery platform for immunizing farmed fish was developed. Accordingly, a compositional library of nanomaterials of natural chemistry (Halloysite (clay), Chitosan, Hydroxyapatite, Mesoporous Silica and a composite material of clay-chitosan) was screened for their toxicity and efficiency in delivering models antigens in cellular and zebrafish embryo models using high throughput screening platforms. Through multi-parametric optimization, chitosan modified halloysite (clay) nanomaterial was identified as an optimal vaccine delivery platform. Further, studies conducted in juvenile seabass showed the potential of clay-chitosan in delivering outer membrane protein of Tenacibaculum maritimum- TIMA (pathogenic bacteria) to and its efficiency in eliciting immune responses in fish. In short, as exemplified by this work, the strategy of using compositional nanomaterial libraries and their biological profiling using high-throughput screening platform could fasten the discovery process of nanomaterials with potential applications in food and agriculture.

Keywords: nanotechnology, fish-vaccine, drug-delivery, halloysite-chitosan

Procedia PDF Downloads 254
1154 Development of a Sequential Multimodal Biometric System for Web-Based Physical Access Control into a Security Safe

Authors: Babatunde Olumide Olawale, Oyebode Olumide Oyediran

Abstract:

The security safe is a place or building where classified document and precious items are kept. To prevent unauthorised persons from gaining access to this safe a lot of technologies had been used. But frequent reports of an unauthorised person gaining access into security safes with the aim of removing document and items from the safes are pointers to the fact that there is still security gap in the recent technologies used as access control for the security safe. In this paper we try to solve this problem by developing a multimodal biometric system for physical access control into a security safe using face and voice recognition. The safe is accessed by the combination of face and speech pattern recognition and also in that sequential order. User authentication is achieved through the use of camera/sensor unit and a microphone unit both attached to the door of the safe. The user face was captured by the camera/sensor while the speech was captured by the use of the microphone unit. The Scale Invariance Feature Transform (SIFT) algorithm was used to train images to form templates for the face recognition system while the Mel-Frequency Cepitral Coefficients (MFCC) algorithm was used to train the speech recognition system to recognise authorise user’s speech. Both algorithms were hosted in two separate web based servers and for automatic analysis of our work; our developed system was simulated in a MATLAB environment. The results obtained shows that the developed system was able to give access to authorise users while declining unauthorised person access to the security safe.

Keywords: access control, multimodal biometrics, pattern recognition, security safe

Procedia PDF Downloads 308
1153 StockTwits Sentiment Analysis on Stock Price Prediction

Authors: Min Chen, Rubi Gupta

Abstract:

Understanding and predicting stock market movements is a challenging problem. It is believed stock markets are partially driven by public sentiments, which leads to numerous research efforts to predict stock market trend using public sentiments expressed on social media such as Twitter but with limited success. Recently a microblogging website StockTwits is becoming increasingly popular for users to share their discussions and sentiments about stocks and financial market. In this project, we analyze the text content of StockTwits tweets and extract financial sentiment using text featurization and machine learning algorithms. StockTwits tweets are first pre-processed using techniques including stopword removal, special character removal, and case normalization to remove noise. Features are extracted from these preprocessed tweets through text featurization process using bags of words, N-gram models, TF-IDF (term frequency-inverse document frequency), and latent semantic analysis. Machine learning models are then trained to classify the tweets' sentiment as positive (bullish) or negative (bearish). The correlation between the aggregated daily sentiment and daily stock price movement is then investigated using Pearson’s correlation coefficient. Finally, the sentiment information is applied together with time series stock data to predict stock price movement. The experiments on five companies (Apple, Amazon, General Electric, Microsoft, and Target) in a duration of nine months demonstrate the effectiveness of our study in improving the prediction accuracy.

Keywords: machine learning, sentiment analysis, stock price prediction, tweet processing

Procedia PDF Downloads 131
1152 City-Wide Simulation on the Effects of Optimal Appliance Scheduling in a Time-of-Use Residential Environment

Authors: Rudolph Carl Barrientos, Juwaln Diego Descallar, Rainer James Palmiano

Abstract:

Household Appliance Scheduling Systems (HASS) coupled with a Time-of-Use (TOU) pricing scheme, a form of Demand Side Management (DSM), is not widely utilized in the Philippines’ residential electricity sector. This paper’s goal is to encourage distribution utilities (DUs) to adopt HASS and TOU by analyzing the effect of household schedulers on the electricity price and load profile in a residential environment. To establish this, a city based on an implemented survey is generated using Monte Carlo Analysis (MCA). Then, a Binary Particle Swarm Optimization (BPSO) algorithm-based HASS is developed considering user satisfaction, electricity budget, appliance prioritization, energy storage systems, solar power, and electric vehicles. The simulations were assessed under varying levels of user compliance. Results showed that the average electricity cost, peak demand, and peak-to-average ratio (PAR) of the city load profile were all reduced. Therefore, the deployment of the HASS and TOU pricing scheme is beneficial for both stakeholders.

Keywords: appliance scheduling, DSM, TOU, BPSO, city-wide simulation, electric vehicle, appliance prioritization, energy storage system, solar power

Procedia PDF Downloads 80
1151 The Magnification of Early Detect Nutrition Case through Local Potential Utilization in Urban Region, Indonesia

Authors: Oktia Woro Kasmini Handayani, Sri Ratna Rahayu, Efa Nugroho, Bertakalswa Hermawati

Abstract:

The double burden of nutrition problem must be faced by Indonesia as developing country. The implemented program did not improve the nutritional status, therefore need to consider to utilize local potential. The objective of this research was to find out the effectivity of magnification model of early detect through local potential utilization in urban region, Semarang, Central Java, Indonesia. The research used an experimental design with the quantitative-qualitative approach. The population was all toddlers under five within the research region, sample determination by purposive sampling, as many as 216 toddlers. Quantitative data analysis used effectively criteria by Sugiono. Qualitative data was analyzed using NVivo. The optimization of local potential in the effort of nutrition status improvement shows number of nutrition case found was increased 225% (very effective), number of cases treated was increased 175% (very effective), number of cases counselled was increased 200% (effective), and number of cases that have improvement increase 75% (effective). The local potential need to be utilized in the effort of nutrition program improvement one of it is through the community empowerment, particularly health care and health high education institution as partner.

Keywords: early detection, nutrition status, local potential, health cadre

Procedia PDF Downloads 251
1150 Measuring the Quality of Business Education: Employment Readiness Assessment

Authors: Gulbakhyt Sultanova

Abstract:

Business education institutions assess the progress of their students by giving them grades for courses completed and calculating a Grade Point Average (GPA). Whether the participation in these courses has led to the development of competences enabling graduates to successfully compete in the labor market should be measured using a new index: Employment Readiness Assessment (ERA). The higher the ERA, the higher the quality of education at a business school. This is applied, empirical research conducted by using a method of linear optimization. The aim of research is to identify factors which lead to the minimization of the deviation of GPA from ERA as well as to the maximization of ERA. ERA is composed of three components resulting from testing proficiency in Business English, testing work and personal skills, and job interview simulation. The quality of education is improving if GPA approximates ERA and ERA increases. Factors which have had a positive effect on quality enhancement are academic mobility of students and staff, practical-oriented courses taught by staff with work experience, and research-based courses taught by staff with research experience. ERA is a better index to measure the quality of business education than traditional indexes such as GPA due to its greater accuracy in assessing the level of graduates’ competences demanded in the labor market. Optimizing the educational process in pursuit of quality enhancement, ERA has to be used in parallel with GPA to find out which changes worked and resulted in improvement.

Keywords: assessment and evaluation, competence evaluation, education quality, employment readiness

Procedia PDF Downloads 429
1149 Rapid Detection and Differentiation of Camel Pox, Contagious Ecthyma and Papilloma Viruses in Clinical Samples of Camels Using a Multiplex PCR

Authors: A. I. Khalafalla, K. A. Al-Busada, I. M. El-Sabagh

Abstract:

Pox and pox-like diseases of camels are a group of exanthematous skin conditions that have become increasingly important economically. They may be caused by three distinct viruses: camelpox virus (CMPV), camel contagious ecthyma virus (CCEV) and camel papillomavirus (CAPV). These diseases are difficult to differentiate based on clinical presentation in disease outbreaks. Molecular methods such as PCR targeting species-specific genes have been developed and used to identify CMPV and CCEV, but not simultaneously in a single tube. Recently, multiplex PCR has gained reputation as a convenient diagnostic method with cost- and time–saving benefits. In the present communication, we describe the development, optimization and validation a multiplex PCR assays able to detect simultaneously the genome of the three viruses in one single test allowing for rapid and efficient molecular diagnosis. The assay was developed based on the evaluation and combination of published and new primer sets, and was applied to the detection of 110 tissue samples. The method showed high sensitivity, and the specificity was confirmed by PCR-product sequencing. In conclusion, this rapid, sensitive and specific assay is considered a useful method for identifying three important viruses in specimens from camels and as part of a molecular diagnostic regime.

Keywords: multiplex PCR, diagnosis, pox and pox-like diseases, camels

Procedia PDF Downloads 444
1148 Surface-Enhanced Raman Spectroscopy on Gold Nanoparticles in the Kidney Disease

Authors: Leonardo C. Pacheco-Londoño, Nataly J Galan-Freyle, Lisandro Pacheco-Lugo, Antonio Acosta-Hoyos, Elkin Navarro, Gustavo Aroca-Martinez, Karin Rondón-Payares, Alberto C. Espinosa-Garavito, Samuel P. Hernández-Rivera

Abstract:

At the Life Science Research Center at Simon Bolivar University, a primary focus is the diagnosis of various diseases, and the use of gold nanoparticles (Au-NPs) in diverse biomedical applications is continually expanding. In the present study, Au-NPs were employed as substrates for Surface-Enhanced Raman Spectroscopy (SERS) aimed at diagnosing kidney diseases arising from Lupus Nephritis (LN), preeclampsia (PC), and Hypertension (H). Discrimination models were developed for distinguishing patients with and without kidney diseases based on the SERS signals from urine samples by partial least squares-discriminant analysis (PLS-DA). A comparative study of the Raman signals across the three conditions was conducted, leading to the identification of potential metabolite signals. Model performance was assessed through cross-validation and external validation, determining parameters like sensitivity and specificity. Additionally, a secondary analysis was performed using machine learning (ML) models, wherein different ML algorithms were evaluated for their efficiency. Models’ validation was carried out using cross-validation and external validation, and other parameters were determined, such as sensitivity and specificity; the models showed average values of 0.9 for both parameters. Additionally, it is not possible to highlight this collaborative effort involved two university research centers and two healthcare institutions, ensuring ethical treatment and informed consent of patient samples.

Keywords: SERS, Raman, PLS-DA, kidney diseases

Procedia PDF Downloads 26
1147 Optimizing Design Works in Construction Consultant Company: A Knowledge-Based Application

Authors: Phan Nghiem Vu, Le Tuan Vu, Ta Quang Tai

Abstract:

The optimal construction design used during the execution of a construction project is a key factor in determining high productivity and customer satisfaction, however, this management process sometimes is carried out without care and the systematic method that it deserves, bringing negative consequences. This study proposes a knowledge management (KM) approach that will enable the intelligent use of experienced and acknowledged engineers to improve the management of construction design works for a project. Then a knowledge-based application to support this decision-making process is proposed and described. To define and design the system for the application, semi-structured interviews were conducted within five construction consulting organizations with the purpose of studying the way that the method’ optimizing process is implemented in practice and the knowledge supported with it. A system of an optimizing construction design works (OCDW) based on knowledge was developed then validated with construction experts. The OCDW was liked as a valuable tool for construction design works’ optimization, by supporting organizations to generate a corporate memory on this issue, reducing the reliance on individual knowledge and also the subjectivity of the decision-making process. The benefits are described as provided by the performance support system, reducing costs and time, improving product design quality, satisfying customer requirements, expanding the brand organization.

Keywords: optimizing construction design work, construction consultant organization, knowledge management, knowledge-based application

Procedia PDF Downloads 104
1146 Creation of an Integrated Development Environment to Assist and Optimize the Learning the Languages C and C++

Authors: Francimar Alves, Marcos Castro, Marllus Lustosa

Abstract:

In the context of the teaching of computer programming, the choice of tool to use is very important in the initiation and continuity of learning a programming language. The literature tools do not always provide usability and pedagogical dynamism clearly and accurately for effective learning. This hypothesis implies fall in productivity and difficulty of learning a particular programming language by students. The integrated development environments (IDEs) Dev-C ++ and Code :: Blocks are widely used in introductory courses for undergraduate courses in Computer Science for learning C and C ++ languages. However, after several years of discontinuity maintaining the source code of Dev-C ++ tool, the continued use of the same in the teaching and learning process of the students of these institutions has led to difficulties, mainly due to the lack of update by the official developers, which resulted in a sequence of problems in using it on educational settings. Much of the users, dissatisfied with the IDE Dev-C ++, migrated to Code :: Blocks platform targeting the more dynamic range in the learning process of the C and C ++ languages. Nevertheless, there is still the need to create a tool that can provide the resources of most IDE's software development literature, however, more interactive, simple, accurate and efficient. This motivation led to the creation of Falcon C ++ tool, IDE that brings with features that turn it into an educational platform, which focuses primarily on increasing student learning index in the early disciplines of programming and algorithms that use the languages ​​C and C ++ . As a working methodology, a field research to prove the truth of the proposed tool was used. The test results and interviews with entry-level students and intermediate in a postsecondary institution gave basis for the composition of this work, demonstrating a positive impact on the use of the tool in teaching programming, showing that the use of Falcon C ++ software is beneficial in the teaching process of the C and C ++ programming languages.

Keywords: ide, education, learning, development, language

Procedia PDF Downloads 423
1145 Optimal Sortation Strategy for a Distribution Network in an E-Commerce Supply Chain

Authors: Pankhuri Dagaonkar, Charumani Singh, Poornima Krothapalli, Krishna Karthik

Abstract:

The backbone of any retail e-commerce success story is a unique design of supply chain network, providing the business an unparalleled speed and scalability. Primary goal of the supply chain strategy is to meet customer expectation by offering fastest deliveries while keeping the cost minimal. Meeting this objective at the large market that India provides is the problem statement that we have targeted here. There are many models and optimization techniques focused on network design to identify the ideal facility location and size, optimizing cost and speed. In this paper we are presenting a tactical approach to optimize cost of an existing network for a predefined speed. We have considered both forward and reverse logistics of a retail e-commerce supply chain consisting of multiple fulfillment (warehouse) and delivery centers, which are connected via sortation nodes. The mathematical model presented here determines if the shipment from a node should get sorted directly for the last mile delivery center or it should travel as consolidated package to another node for further sortation (resort). The objective function minimizes the total cost by varying the resort percentages between nodes and provides the optimal resource allocation and number of sorts at each node.

Keywords: distribution strategy, mathematical model, network design, supply chain management

Procedia PDF Downloads 277
1144 Photocatalytic Degradation of Phenolic Compounds in Wastewater Using Magnetically Recoverable Catalyst

Authors: Ahmed K. Sharaby, Ahmed S. El-Gendy

Abstract:

Phenolic compounds (PCs) exist in the wastewater effluents of some industries such as oil refinery, pharmaceutical and cosmetics. Phenolic compounds are extremely hazardous pollutants that can cause severe problems to the aquatic life and human beings if disposed of without treatment. One of the most efficient treatment methods of PCs is photocatalytic degradation. The current work studies the performance of composite nanomaterial of titanium dioxide with magnetite as a photo-catalyst in the degradation of PCs. The current work aims at optimizing the synthesized photocatalyst dosage and contact time as part of the operational parameters at different initial concentrations of PCs and pH values in the wastewater. The study was performed in a lab-scale batch reactor under fixed conditions of light intensity and aeration rate. The initial concentrations of PCs and the pH values were in the range of (10-200 mg/l) and (3-9), respectively. Results of the study indicate that the dosage of the catalyst and contact time for total mineralization is proportional to the initial concentrations of PCs, while the optimum pH conditions for highly efficient degradation is at pH 3. Exceeding the concentration levels of the catalyst beyond certain limits leads to the decrease in the degradation efficiency due to the dissipation of light. The performance of the catalyst for degradation was also investigated in comparison to the pure TiO2 Degussa (P-25). The dosage required for the synthesized catalyst for photocatalytic degradation was approximately 1.5 times that needed from the pure titania.

Keywords: industrial, optimization, phenolic compounds, photocatalysis, wastewater

Procedia PDF Downloads 297
1143 Conceptual Model of a Residential Waste Collection System Using ARENA Software

Authors: Bruce G. Wilson

Abstract:

The collection of municipal solid waste at the curbside is a complex operation that is repeated daily under varying circumstances around the world. There have been several attempts to develop Monte Carlo simulation models of the waste collection process dating back almost 50 years. Despite this long history, the use of simulation modeling as a planning or optimization tool for waste collection is still extremely limited in practice. Historically, simulation modeling of waste collection systems has been hampered by the limitations of computer hardware and software and by the availability of representative input data. This paper outlines the development of a Monte Carlo simulation model that overcomes many of the limitations contained in previous models. The model uses a general purpose simulation software program that is easily capable of modeling an entire waste collection network. The model treats the stops on a waste collection route as a queue of work to be processed by a collection vehicle (or server). Input data can be collected from a variety of sources including municipal geographic information systems, global positioning system recorders on collection vehicles, and weigh scales at transfer stations or treatment facilities. The result is a flexible model that is sufficiently robust that it can model the collection activities in a large municipality, while providing the flexibility to adapt to changing conditions on the collection route.

Keywords: modeling, queues, residential waste collection, Monte Carlo simulation

Procedia PDF Downloads 383
1142 Taguchi-Based Optimization of Surface Roughness and Dimensional Accuracy in Wire EDM Process with S7 Heat Treated Steel

Authors: Joseph C. Chen, Joshua Cox

Abstract:

This research focuses on the use of the Taguchi method to reduce the surface roughness and improve dimensional accuracy of parts machined by Wire Electrical Discharge Machining (EDM) with S7 heat treated steel material. Due to its high impact toughness, the material is a candidate for a wide variety of tooling applications which require high precision in dimension and desired surface roughness. This paper demonstrates that Taguchi Parameter Design methodology is able to optimize both dimensioning and surface roughness successfully by investigating seven wire-EDM controllable parameters: pulse on time (ON), pulse off time (OFF), servo voltage (SV), voltage (V), servo feed (SF), wire tension (WT), and wire speed (WS). The temperature of the water in the Wire EDM process is investigated as the noise factor in this research. Experimental design and analysis based on L18 Taguchi orthogonal arrays are conducted. This paper demonstrates that the Taguchi-based system enables the wire EDM process to produce (1) high precision parts with an average of 0.6601 inches dimension, while the desired dimension is 0.6600 inches; and (2) surface roughness of 1.7322 microns which is significantly improved from 2.8160 microns.

Keywords: Taguchi Parameter Design, surface roughness, Wire EDM, dimensional accuracy

Procedia PDF Downloads 349
1141 An Intelligent Prediction Method for Annular Pressure Driven by Mechanism and Data

Authors: Zhaopeng Zhu, Xianzhi Song, Gensheng Li, Shuo Zhu, Shiming Duan, Xuezhe Yao

Abstract:

Accurate calculation of wellbore pressure is of great significance to prevent wellbore risk during drilling. The traditional mechanism model needs a lot of iterative solving procedures in the calculation process, which reduces the calculation efficiency and is difficult to meet the demand of dynamic control of wellbore pressure. In recent years, many scholars have introduced artificial intelligence algorithms into wellbore pressure calculation, which significantly improves the calculation efficiency and accuracy of wellbore pressure. However, due to the ‘black box’ property of intelligent algorithm, the existing intelligent calculation model of wellbore pressure is difficult to play a role outside the scope of training data and overreacts to data noise, often resulting in abnormal calculation results. In this study, the multi-phase flow mechanism is embedded into the objective function of the neural network model as a constraint condition, and an intelligent prediction model of wellbore pressure under the constraint condition is established based on more than 400,000 sets of pressure measurement while drilling (MPD) data. The constraint of the multi-phase flow mechanism makes the prediction results of the neural network model more consistent with the distribution law of wellbore pressure, which overcomes the black-box attribute of the neural network model to some extent. The main performance is that the accuracy of the independent test data set is further improved, and the abnormal calculation values basically disappear. This method is a prediction method driven by MPD data and multi-phase flow mechanism, and it is the main way to predict wellbore pressure accurately and efficiently in the future.

Keywords: multiphase flow mechanism, pressure while drilling data, wellbore pressure, mechanism constraints, combined drive

Procedia PDF Downloads 158
1140 Design and Identification of Mycobacterium tuberculosis Glutamate Racemase (MurI) Inhibitors

Authors: Prasanthi Malapati, R. Reshma, Vijay Soni, Perumal Yogeeswari, Dharmarajan Sriram

Abstract:

In the present study, we attempted to develop Mycobacterium tuberculosis (Mtb) inhibitors by exploring the pharmaceutically underexploited enzyme targets which are majorly involved in cell wall biosynthesis of mycobacteria. For this purpose, glutamate racemase (coded by MurI gene) was selected. This enzyme racemize L-glutamate to D-glutamate required for the construction of peptidoglycan in the bacterial cell wall synthesis process. Furthermore this enzyme is neither expressed nor its product, D-glutamate is normally found in mammals, and hence designing inhibitors against this enzyme will not affect the host system as well act as potential antitubercular drugs. A library of BITS in house compounds were screened against Mtb MurI enzyme. Based on docking score, interactions and synthetic feasibility one hit lead was identified. Further optimization of lead was attempted and its derivatives were synthesized. Forty eight derivatives of 2-phenylbenzo[d]oxazole and 2-phenylbenzo[d]thiazole were synthesized and evaluated for Mtb MurI inhibition study, in vitro activities against Mtb, cytotoxicity against RAW 264.7 cell line. Chemical derivatization of the lead resulted in compounds NR-1213 AND NR-1124 as the potent M. tuberculosis glutamate racemase inhibitors with IC50 of 4-5µM which are remarkable and were found to be non-cytotoxic. Molecular dynamics, dormant models and cardiotoxicity studies of the most active molecules are in process.

Keywords: cell wall biosynthesis, dormancy, glutamate racemase, tuberculosis

Procedia PDF Downloads 249