Search results for: optimization algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4721

Search results for: optimization algorithms

1091 Development of Lectin-Based Biosensor for Glycoprofiling of Clinical Samples: Focus on Prostate Cancer

Authors: Dominika Pihikova, Stefan Belicky, Tomas Bertok, Roman Sokol, Petra Kubanikova, Jan Tkac

Abstract:

Since aberrant glycosylation is frequently accompanied by both physiological and pathological processes in a human body (cancer, AIDS, inflammatory diseases, etc.), the analysis of tumor-associated glycan patterns have a great potential for the development of novel diagnostic approaches. Moreover, altered glycoforms may assist as a suitable tool for the specificity and sensitivity enhancement in early-stage prostate cancer diagnosis. In this paper we discuss the construction and optimization of ultrasensitive sandwich biosensor platform employing lectin as glycan-binding protein. We focus on the immunoassay development, reduction of non-specific interactions and final glycoprofiling of human serum samples including both prostate cancer (PCa) patients and healthy controls. The fabricated biosensor was measured by label-free electrochemical impedance spectroscopy (EIS) with further lectin microarray verification. Furthermore, we analyzed different biosensor interfaces with atomic force microscopy (AFM) in nanomechanical mapping mode showing a significant differences in the altitude. These preliminary results revealing an elevated content of α-2,3 linked sialic acid in PCa patients comparing with healthy controls. All these experiments are important step towards development of point-of-care devices and discovery of novel glyco-biomarkers applicable in cancer diagnosis.

Keywords: biosensor, glycan, lectin, prostate cancer

Procedia PDF Downloads 354
1090 Optimal Dynamic Regime for CO Oxidation Reaction Discovered by Policy-Gradient Reinforcement Learning Algorithm

Authors: Lifar M. S., Tereshchenko A. A., Bulgakov A. N., Guda S. A., Guda A. A., Soldatov A. V.

Abstract:

Metal nanoparticles are widely used as heterogeneous catalysts to activate adsorbed molecules and reduce the energy barrier of the reaction. Reaction product yield depends on the interplay between elementary processes - adsorption, activation, reaction, and desorption. These processes, in turn, depend on the inlet feed concentrations, temperature, and pressure. At stationary conditions, the active surface sites may be poisoned by reaction byproducts or blocked by thermodynamically adsorbed gaseous reagents. Thus, the yield of reaction products can significantly drop. On the contrary, the dynamic control accounts for the changes in the surface properties and adjusts reaction parameters accordingly. Therefore dynamic control may be more efficient than stationary control. In this work, a reinforcement learning algorithm has been applied to control the simulation of CO oxidation on a catalyst. The policy gradient algorithm is learned to maximize the CO₂ production rate based on the CO and O₂ flows at a given time step. Nonstationary solutions were found for the regime with surface deactivation. The maximal product yield was achieved for periodic variations of the gas flows, ensuring a balance between available adsorption sites and the concentration of activated intermediates. This methodology opens a perspective for the optimization of catalytic reactions under nonstationary conditions.

Keywords: artificial intelligence, catalyst, co oxidation, reinforcement learning, dynamic control

Procedia PDF Downloads 104
1089 MLProxy: SLA-Aware Reverse Proxy for Machine Learning Inference Serving on Serverless Computing Platforms

Authors: Nima Mahmoudi, Hamzeh Khazaei

Abstract:

Serving machine learning inference workloads on the cloud is still a challenging task at the production level. The optimal configuration of the inference workload to meet SLA requirements while optimizing the infrastructure costs is highly complicated due to the complex interaction between batch configuration, resource configurations, and variable arrival process. Serverless computing has emerged in recent years to automate most infrastructure management tasks. Workload batching has revealed the potential to improve the response time and cost-effectiveness of machine learning serving workloads. However, it has not yet been supported out of the box by serverless computing platforms. Our experiments have shown that for various machine learning workloads, batching can hugely improve the system’s efficiency by reducing the processing overhead per request. In this work, we present MLProxy, an adaptive reverse proxy to support efficient machine learning serving workloads on serverless computing systems. MLProxy supports adaptive batching to ensure SLA compliance while optimizing serverless costs. We performed rigorous experiments on Knative to demonstrate the effectiveness of MLProxy. We showed that MLProxy could reduce the cost of serverless deployment by up to 92% while reducing SLA violations by up to 99% that can be generalized across state-of-the-art model serving frameworks.

Keywords: serverless computing, machine learning, inference serving, Knative, google cloud run, optimization

Procedia PDF Downloads 159
1088 Quantitative Structure-Property Relationship Study of Base Dissociation Constants of Some Benzimidazoles

Authors: Sanja O. Podunavac-Kuzmanović, Lidija R. Jevrić, Strahinja Z. Kovačević

Abstract:

Benzimidazoles are a group of compounds with significant antibacterial, antifungal and anticancer activity. The studied compounds consist of the main benzimidazole structure with different combinations of substituens. This study is based on the two-dimensional and three-dimensional molecular modeling and calculation of molecular descriptors (physicochemical and lipophilicity descriptors) of structurally diverse benzimidazoles. Molecular modeling was carried out by using ChemBio3D Ultra version 14.0 software. The obtained 3D models were subjected to energy minimization using molecular mechanics force field method (MM2). The cutoff for structure optimization was set at a gradient of 0.1 kcal/Åmol. The obtained set of molecular descriptors was used in principal component analysis (PCA) of possible similarities and dissimilarities among the studied derivatives. After the molecular modeling, the quantitative structure-property relationship (QSPR) analysis was applied in order to get the mathematical models which can be used in prediction of pKb values of structurally similar benzimidazoles. The obtained models are based on statistically valid multiple linear regression (MLR) equations. The calculated cross-validation parameters indicate the high prediction ability of the established QSPR models. This study is financially supported by COST action CM1306 and the project No. 114-451-347/2015-02, financially supported by the Provincial Secretariat for Science and Technological Development of Vojvodina.

Keywords: benzimidazoles, chemometrics, molecular modeling, molecular descriptors, QSPR

Procedia PDF Downloads 268
1087 Development, Optimization, and Validation of a Synchronous Fluorescence Spectroscopic Method with Multivariate Calibration for the Determination of Amlodipine and Olmesartan Implementing: Experimental Design

Authors: Noha Ibrahim, Eman S. Elzanfaly, Said A. Hassan, Ahmed E. El Gendy

Abstract:

Objectives: The purpose of the study is to develop a sensitive synchronous spectrofluorimetric method with multivariate calibration after studying and optimizing the different variables affecting the native fluorescence intensity of amlodipine and olmesartan implementing an experimental design approach. Method: In the first step, the fractional factorial design used to screen independent factors affecting the intensity of both drugs. The objective of the second step was to optimize the method performance using a Central Composite Face-centred (CCF) design. The optimal experimental conditions obtained from this study were; a temperature of (15°C ± 0.5), the solvent of 0.05N HCl and methanol with a ratio of (90:10, v/v respectively), Δλ of 42 and the addition of 1.48 % surfactant providing a sensitive measurement of amlodipine and olmesartan. The resolution of the binary mixture with a multivariate calibration method has been accomplished mainly by using partial least squares (PLS) model. Results: The recovery percentage for amlodipine besylate and atorvastatin calcium in tablets dosage form were found to be (102 ± 0.24, 99.56 ± 0.10, for amlodipine and Olmesartan, respectively). Conclusion: Method is valid according to some International Conference on Harmonization (ICH) guidelines, providing to be linear over a range of 200-300, 500-1500 ng mL⁻¹ for amlodipine and Olmesartan. The methods were successful to estimate amlodipine besylate and olmesartan in bulk powder and pharmaceutical preparation.

Keywords: amlodipine, central composite face-centred design, experimental design, fractional factorial design, multivariate calibration, olmesartan

Procedia PDF Downloads 132
1086 Powering Connections: Synergizing Sales and Marketing for Electronics Engineering with Web Development.

Authors: Muhammad Awais Kiani, Abdul Basit Kiani, Maryam Kiani

Abstract:

Synergizing Sales and Marketing for Electronics Engineering with Web Development, explores the dynamic relationship between sales, marketing, and web development within the electronics engineering industry. This study is important for the power of digital platforms to connect with customers. Which increases brand visibility and drives sales. It highlights the need for collaboration between sales and marketing teams, as well as the integration of web development strategies to create seamless user experiences and effective lead generation. Furthermore, It also emphasizes the role of data analytics and customer insights in optimizing sales and marketing efforts in the ever-evolving landscape of electronics engineering. Sales and marketing play a crucial role in driving business growth, and in today's digital landscape, web development has become an integral part of these strategies. Web development enables businesses to create visually appealing and user-friendly websites that effectively showcase their products or services. It allows for the integration of e-commerce functionalities, enabling seamless online transactions. Furthermore, web development helps businesses optimize their online presence through search engine optimization (SEO) techniques, social media integration, and content management systems. This abstract highlights the symbiotic relationship between sales marketing in the electronics industry and web development, emphasizing the importance of a strong online presence in achieving business success.

Keywords: electronics industry, web development, sales, marketing

Procedia PDF Downloads 96
1085 Low-Cost Image Processing System for Evaluating Pavement Surface Distress

Authors: Keerti Kembhavi, M. R. Archana, V. Anjaneyappa

Abstract:

Most asphalt pavement condition evaluation use rating frameworks in which asphalt pavement distress is estimated by type, extent, and severity. Rating is carried out by the pavement condition rating (PCR), which is tedious and expensive. This paper presents the development of a low-cost technique for image pavement distress analysis that permits the identification of pothole and cracks. The paper explores the application of image processing tools for the detection of potholes and cracks. Longitudinal cracking and pothole are detected using Fuzzy-C- Means (FCM) and proceeded with the Spectral Theory algorithm. The framework comprises three phases, including image acquisition, processing, and extraction of features. A digital camera (Gopro) with the holder is used to capture pavement distress images on a moving vehicle. FCM classifier and Spectral Theory algorithms are used to compute features and classify the longitudinal cracking and pothole. The Matlab2016Ra Image preparing tool kit utilizes performance analysis to identify the viability of pavement distress on selected urban stretches of Bengaluru city, India. The outcomes of image evaluation with the utilization semi-computerized image handling framework represented the features of longitudinal crack and pothole with an accuracy of about 80%. Further, the detected images are validated with the actual dimensions, and it is seen that dimension variability is about 0.46. The linear regression model y=1.171x-0.155 is obtained using the existing and experimental / image processing area. The R2 correlation square obtained from the best fit line is 0.807, which is considered in the linear regression model to be ‘large positive linear association’.

Keywords: crack detection, pothole detection, spectral clustering, fuzzy-c-means

Procedia PDF Downloads 164
1084 Aerodynamic Design Optimization of High-Speed Hatchback Cars for Lucrative Commercial Applications

Authors: A. Aravind, M. Vetrivel, P. Abhimanyu, C. A. Akaash Emmanuel Raj, K. Sundararaj, V. R. S. Kumar

Abstract:

The choice of high-speed, low budget hatchback car with diversified options is increasing for meeting the new generation buyers trend. This paper is aimed to augment the current speed of the hatchback cars through the aerodynamic drag reduction technique. The inverted airfoils are facilitated at the bottom of the car for generating the downward force for negating the lift while increasing the current speed range for achieving a better road performance. The numerical simulations have been carried out using a 2D steady pressure-based    k-ɛ realizable model with enhanced wall treatment. In our numerical studies, Reynolds-averaged Navier-Stokes model and its code of solution are used. The code is calibrated and validated using the exact solution of the 2D boundary layer displacement thickness at the Sanal flow choking condition for adiabatic flows. We observed through the parametric analytical studies that the inverted airfoil integrated with the bottom surface at various predesigned locations of Hatchback cars can improve its overall aerodynamic efficiency through drag reduction, which obviously decreases the fuel consumption significantly and ensure an optimum road performance lucratively with maximum permissible speed within the framework of the manufactures constraints.

Keywords: aerodynamics of commercial cars, downward force, hatchback car, inverted airfoil

Procedia PDF Downloads 257
1083 Analyzing the Influence of Hydrometeorlogical Extremes, Geological Setting, and Social Demographic on Public Health

Authors: Irfan Ahmad Afip

Abstract:

This main research objective is to accurately identify the possibility for a Leptospirosis outbreak severity of a certain area based on its input features into a multivariate regression model. The research question is the possibility of an outbreak in a specific area being influenced by this feature, such as social demographics and hydrometeorological extremes. If the occurrence of an outbreak is being subjected to these features, then the epidemic severity for an area will be different depending on its environmental setting because the features will influence the possibility and severity of an outbreak. Specifically, this research objective was three-fold, namely: (a) to identify the relevant multivariate features and visualize the patterns data, (b) to develop a multivariate regression model based from the selected features and determine the possibility for Leptospirosis outbreak in an area, and (c) to compare the predictive ability of multivariate regression model and machine learning algorithms. Several secondary data features were collected locations in the state of Negeri Sembilan, Malaysia, based on the possibility it would be relevant to determine the outbreak severity in the area. The relevant features then will become an input in a multivariate regression model; a linear regression model is a simple and quick solution for creating prognostic capabilities. A multivariate regression model has proven more precise prognostic capabilities than univariate models. The expected outcome from this research is to establish a correlation between the features of social demographic and hydrometeorological with Leptospirosis bacteria; it will also become a contributor for understanding the underlying relationship between the pathogen and the ecosystem. The relationship established can be beneficial for the health department or urban planner to inspect and prepare for future outcomes in event detection and system health monitoring.

Keywords: geographical information system, hydrometeorological, leptospirosis, multivariate regression

Procedia PDF Downloads 97
1082 Optimal Energy Management and Environmental Index Optimization of a Microgrid Operating by Renewable and Sustainable Generation Systems

Authors: Nabil Mezhoud

Abstract:

The economic operation of electric energy generating systems is one of the predominant problems in energy systems. Due to the need for better reliability, high energy quality, lower losses, lower cost and a clean environment, the application of renewable and sustainable energy sources, such as wind energy, solar energy, etc., in recent years has become more widespread. In this work, one of a bio-inspired meta-heuristic algorithm inspired by the flashing behavior of fireflies at night called the Firefly Algorithm (FFA) is applied to solve the Optimal Energy Management (OEM) and the environmental index (EI) problems of a micro-grid (MG) operating by Renewable and Sustainable Generation Systems (RSGS). Our main goal is to minimize the nonlinear objective function of an electrical microgrid, taking into account equality and inequality constraints. The FFA approach was examined and tested on a standard MG system composed of different types of RSGS, such as wind turbines (WT), photovoltaic systems (PV), and non-renewable energy, such as fuel cells (FC), micro turbine (MT), diesel generator (DEG) and loads with energy storage systems (ESS). The results are promising and show the effectiveness and robustness of the proposed approach to solve the OEM and the EI problems. The results of the proposed method have been compared and validated with those known references published recently.

Keywords: renewable energy sources, energy management, distributed generator, micro-grids, firefly algorithm

Procedia PDF Downloads 47
1081 Bone Mineral Density and Quality, Body Composition of Women in the Postmenopausal Period

Authors: Vladyslav Povoroznyuk, Oksana Ivanyk, Nataliia Dzerovych

Abstract:

In the diagnostics of osteoporosis, the gold standard is considered to be bone mineral density; however, X-ray densitometry is not an accurate indicator of osteoporotic fracture risk under all circumstances. In this regard, the search for new methods that could determine the indicators not only of the mineral density, but of the bone tissue quality, is a logical step for diagnostic optimization. One of these methods is the evaluation of trabecular bone quality. The aim of this study was to examine the quality and mineral density of spine bone tissue, femoral neck, and body composition of women depending on the duration of the postmenopausal period, to determine the correlation of body fat with indicators of bone mineral density and quality. The study examined 179 women in premenopausal and postmenopausal periods. The patients were divided into the following groups: Women in the premenopausal period and women in the postmenopausal period at various stages (early, middle, late postmenopause). A general examination and study of the above parameters were conducted with General Electric X-ray densitometer. The results show that bone quality and mineral density probably deteriorate with advancing of postmenopausal period. Total fat and lean mass ratio is not likely to change with age. In the middle and late postmenopausal periods, the bone tissue mineral density of the spine and femoral neck increases along with total fat mass.

Keywords: osteoporosis, bone tissue mineral density, bone quality, fat mass, lean mass, postmenopausal osteoporosis

Procedia PDF Downloads 320
1080 Computational Fluid Dynamics Analysis of Cyclone Separator Performance Using Discrete Phase Model

Authors: Sandeep Mohan Ahuja, Gulshan Kumar Jawa

Abstract:

Cyclone separators are crucial components in various industries tasked with efficiently separating particulate matter from gas streams. Achieving optimal performance hinges on a deep understanding of flow dynamics and particle behaviour within these separators. In this investigation, Computational Fluid Dynamics (CFD) simulations are conducted utilizing the Discrete Phase Model (DPM) to dissect the intricate flow patterns, particle trajectories, and separation efficiency within cyclone separators. The study delves into the influence of pivotal parameters like inlet velocity, particle size distribution, and cyclone geometry on separation efficiency. Through numerical simulations, a comprehensive comprehension of fluid-particle interaction phenomena within cyclone separators is attained, allowing for the assessment of solid collection efficiency across diverse operational conditions and geometrical setups. The insights gleaned from this study promise to advance our understanding of the complex interplay between fluid and particle within cyclone separators, thereby enabling optimization across a wide array of industrial applications. By harnessing the power of CFD simulations and the DPM, this research endeavours to furnish valuable insights for designing, operating, and evaluating the performance of cyclone separators, ultimately fostering greater efficiency and environmental sustainability within industrial processes.

Keywords: cyclone separator, computational fluid dynamics, enhancing efficiency, discrete phase model

Procedia PDF Downloads 24
1079 Destination Port Detection For Vessels: An Analytic Tool For Optimizing Port Authorities Resources

Authors: Lubna Eljabu, Mohammad Etemad, Stan Matwin

Abstract:

Port authorities have many challenges in congested ports to allocate their resources to provide a safe and secure loading/ unloading procedure for cargo vessels. Selecting a destination port is the decision of a vessel master based on many factors such as weather, wavelength and changes of priorities. Having access to a tool which leverages AIS messages to monitor vessel’s movements and accurately predict their next destination port promotes an effective resource allocation process for port authorities. In this research, we propose a method, namely, Reference Route of Trajectory (RRoT) to assist port authorities in predicting inflow and outflow traffic in their local environment by monitoring Automatic Identification System (AIS) messages. Our RRoT method creates a reference route based on historical AIS messages. It utilizes some of the best trajectory similarity measure to identify the destination of a vessel using their recent movement. We evaluated five different similarity measures such as Discrete Fr´echet Distance (DFD), Dynamic Time Warping (DTW), Partial Curve Mapping (PCM), Area between two curves (Area) and Curve length (CL). Our experiments show that our method identifies the destination port with an accuracy of 98.97% and an fmeasure of 99.08% using Dynamic Time Warping (DTW) similarity measure.

Keywords: spatial temporal data mining, trajectory mining, trajectory similarity, resource optimization

Procedia PDF Downloads 99
1078 Understanding the Complexities of Consumer Financial Spinning

Authors: Olivier Mesly

Abstract:

This research presents a conceptual framework termed “Consumer Financial Spinning” (CFS) to analyze consumer behavior in the financial/economic markets. This phenomenon occurs when consumers of high-stakes financial products accumulate unsustainable debt, leading them to detach from their initial financial hierarchy of needs, wealth-related goals, and preferences regarding their household portfolio of assets. The daring actions of these consumers, forming a dark financial triangle, are characterized by three behaviors: overconfidence, the use of rationed rationality, and deceitfulness. We show that we can incorporate CFS into the traditional CAPM and Markovitz’ portfolio optimization models to create a framework that explains such market phenomena as the global financial crisis, highlighting the antecedents and consequences of ill-conceived speculation. Because this is a conceptual paper, there is no methodology with respect to ground studies. However, we apply modeling principles derived from the data percolation methodology, which contains tenets explicating how to structure concepts. A simulation test of the proposed framework is conducted; it demonstrates the conditions under which the relationship between expected returns and risk may deviate from linearity. The analysis and conceptual findings are particularly relevant both theoretically and pragmatically as they shed light on the psychological conditions that drive intense speculation, which can lead to market turmoil. Armed with such understanding, regulators are better equipped to propose solutions before the economic problems become out of control.

Keywords: consumer financial spinning, rationality, deceitfulness, overconfidence, CAPM

Procedia PDF Downloads 26
1077 Six Sigma-Based Optimization of Shrinkage Accuracy in Injection Molding Processes

Authors: Sky Chou, Joseph C. Chen

Abstract:

This paper focuses on using six sigma methodologies to reach the desired shrinkage of a manufactured high-density polyurethane (HDPE) part produced by the injection molding machine. It presents a case study where the correct shrinkage is required to reduce or eliminate defects and to improve the process capability index Cp and Cpk for an injection molding process. To improve this process and keep the product within specifications, the six sigma methodology, design, measure, analyze, improve, and control (DMAIC) approach, was implemented in this study. The six sigma approach was paired with the Taguchi methodology to identify the optimized processing parameters that keep the shrinkage rate within the specifications by our customer. An L9 orthogonal array was applied in the Taguchi experimental design, with four controllable factors and one non-controllable/noise factor. The four controllable factors identified consist of the cooling time, melt temperature, holding time, and metering stroke. The noise factor is the difference between material brand 1 and material brand 2. After the confirmation run was completed, measurements verify that the new parameter settings are optimal. With the new settings, the process capability index has improved dramatically. The purpose of this study is to show that the six sigma and Taguchi methodology can be efficiently used to determine important factors that will improve the process capability index of the injection molding process.

Keywords: injection molding, shrinkage, six sigma, Taguchi parameter design

Procedia PDF Downloads 157
1076 Optimization of Microencapsulation of β-Carotene by Complex Coacervation Technique Using Casein and Gum Tragacanth

Authors: Gargi Ghoshal, Ashay Jain

Abstract:

Microencapsulation of β-carotene was optimized by complex coacervation technique using casein/gum tragacanth (CAS/GT) coating as a function of pH, initial protein to polysaccharide mixing ratio (Pr:Ps), total biopolymer concentration, core material load, zeta potential, and ionic strength. This study was aimed to understand the influence of experimental parameters on the coacervation kinetics, the coacervate yield, and entrapment efficiency. At a Pr:Ps = 2:1, an optimum pH of complex coacervation was found 4.35, at which the intensity of electrostatic interaction was maximum. At these ratios of coating, the phase separation occurred the fastest and the final coacervate yield and entrapment efficiency was the highest. Varying the Pr: Ps shifted the value of optimum pH. This incident was due to the level of charge compensation of the CAS/GT complexes. Finally, electrostatic interaction and formation of coacervates between CAS and GT were confirmed by Fourier transform infra-red (FTIR) spectra. The size and surface properties of coacervates were studied using scanning electron microscopy (SEM). The resultant formulation (β-carotene loaded microcapsules) was evaluated for in vitro release study and antioxidant activity. Stability of encapsulated β-carotene was also evaluated under three levels of temperature (5, 25 and 40 °C) for 3 months. Encapsulation strongly increased the stability of micronutrients. Our results advocate potential of microcapsules as a novel carrier for the safeguard and sustained release of micronutrient.

Keywords: β-carotene, casein, complex coacervation, controlled release, gum tragacanth, microcapsules

Procedia PDF Downloads 249
1075 Benchmarking Machine Learning Approaches for Forecasting Hotel Revenue

Authors: Rachel Y. Zhang, Christopher K. Anderson

Abstract:

A critical aspect of revenue management is a firm’s ability to predict demand as a function of price. Historically hotels have used simple time series models (regression and/or pick-up based models) owing to the complexities of trying to build casual models of demands. Machine learning approaches are slowly attracting attention owing to their flexibility in modeling relationships. This study provides an overview of approaches to forecasting hospitality demand – focusing on the opportunities created by machine learning approaches, including K-Nearest-Neighbors, Support vector machine, Regression Tree, and Artificial Neural Network algorithms. The out-of-sample performances of above approaches to forecasting hotel demand are illustrated by using a proprietary sample of the market level (24 properties) transactional data for Las Vegas NV. Causal predictive models can be built and evaluated owing to the availability of market level (versus firm level) data. This research also compares and contrast model accuracy of firm-level models (i.e. predictive models for hotel A only using hotel A’s data) to models using market level data (prices, review scores, location, chain scale, etc… for all hotels within the market). The prospected models will be valuable for hotel revenue prediction given the basic characters of a hotel property or can be applied in performance evaluation for an existed hotel. The findings will unveil the features that play key roles in a hotel’s revenue performance, which would have considerable potential usefulness in both revenue prediction and evaluation.

Keywords: hotel revenue, k-nearest-neighbors, machine learning, neural network, prediction model, regression tree, support vector machine

Procedia PDF Downloads 115
1074 The Employees' Classification Method in the Space of Their Job Satisfaction, Loyalty and Involvement

Authors: Svetlana Ignatjeva, Jelena Slesareva

Abstract:

The aim of the study is development and adaptation of the method to analyze and quantify the indicators characterizing the relationship between a company and its employees. Diagnostics of such indicators is one of the most complex and actual issues in psychology of labour. The offered method is based on the questionnaire; its indicators reflect cognitive, affective and connotative components of socio-psychological attitude of employees to be as efficient as possible in their professional activities. This approach allows measure not only the selected factors but also such parameters as cognitive and behavioural dissonances. Adaptation of the questionnaire includes factor structure analysis and suitability analysis of phenomena indicators measured in terms of internal consistency of individual factors. Structural validity of the questionnaire was tested by exploratory factor analysis. Extraction Method: Principal Component Analysis. Rotation Method: Varimax with Kaiser Normalization. Factor analysis allows reduce dimension of the phenomena moving from the indicators to aggregative indexes and latent variables. Aggregative indexes are obtained as the sum of relevant indicators followed by standardization. The coefficient Cronbach's Alpha was used to assess the reliability-consistency of the questionnaire items. The two-step cluster analysis in the space of allocated factors allows classify employees according to their attitude to work in the company. The results of psychometric testing indicate possibility of using the developed technique for the analysis of employees’ attitude towards their work in companies and development of recommendations on their optimization.

Keywords: involved in the organization, loyalty, organizations, method

Procedia PDF Downloads 338
1073 Enhancing Industrial Wastewater Treatment through Fe3o4 Nanoparticles-loaded Activated Charcoal: Design and Optimization for Sustainable Development

Authors: Komal Verma, V. S. Moholkar

Abstract:

This paper reports investigations in the mineralization of industrial wastewater (COD = 3246 mg/L, TOC = 2500 mg/L) using a ternary (ultrasound + Fenton + adsorption) hybrid advanced oxidation process. Fe3O4 decorated activated charcoal (Fe3O4@AC) nanocomposites (surface area = 538.88 m2/g; adsorption capacity = 294.31 mg/g) were synthesized using co-precipitation. The wastewater treatment process was optimized using central composite statistical design. At optimum conditions, viz. pH = 4.2, H2O2 loading = 0.71 M, adsorbent dose = 0.34 g/L, reduction in COD and TOC of wastewater were 94.75% and 89%, respectively. This result is essentially a consequence of synergistic interactions among the adsorption of pollutants onto activated charcoal and surface Fenton reactions induced due to the leaching of Fe2+/Fe3+ ions from the Fe3O4 nanoparticles. Microconvection generated due to sonication assisted faster mass transport (adsorption/desorption) of pollutants between Fe₃O₄@AC nanocomposite and the solution. The net result of this synergism was high interactions and reactions among and radicals and pollutants that resulted in the effective mineralization of wastewater The Fe₃O₄@AC showed excellent recovery (> 90 wt%) and reusability (> 90% COD removal) in 5 successive cycles of treatment. LC-MS analysis revealed effective (> 50%) degradation of more than 25 significant contaminants (in the form of herbicides and pesticides) after the treatment with ternary hybrid AOP. Similarly, the toxicity analysis test using the seed germination technique revealed ~ 60% reduction in the toxicity of the wastewater after treatment.

Keywords: Fe₃O₄@AC nanocomposite, RSM, COD;, LC-MS, Toxicity

Procedia PDF Downloads 96
1072 Response Surface Methodology for the Optimization of Sugar Extraction from Phoenix dactylifera L.

Authors: Lila Boulekbache-Makhlouf, Kahina Djaoud, Myriam Tazarourte, Samir Hadjal, Khodir Madani

Abstract:

In Algeria, important quantities of secondary date variety (Phoenix dactylifera L.) are generated in each campaign; their chemical composition is similar to that of commercial dates. The present work aims to valorize this common date variety (Degla-Beida) which is often poorly exploited. In this context, we tried to prepare syrup from the secondary date variety and to evaluate the effect of conventional extraction (CE) or water bath extraction (WBE) and alternative extraction (microwaves assisted extraction (MAE), and ultrasounds assisted extraction (UAE)) on its total sugar content (TSC), using response surface methodology (RSM). Then, the analysis of individual sugars was performed by high-performance liquid chromatography (HPLC). Maximum predicted TSC recoveries under the optimized conditions for MAE, UAE and CE were 233.248 ± 3.594 g/l, 202.889 ± 5.797 g/l, and 233.535 ± 5.412 g/l, respectively, which were close to the experimental values: 233.796 ± 1.898 g/l; 202.037 ± 3.401 g/l and 234.380 ± 2.425 g/l. HPLC analysis revealed high similarity in the sugar composition of date juices obtained by MAE (60.11% sucrose, 16.64% glucose and 23.25% fructose) and CE (50.78% sucrose, 20.67% glucose and 28.55% fructose), although a large difference was detected for that obtained by UAE (0.00% sucrose, 46.94% glucose and 53.06% fructose). Microwave-assisted extraction was the best method for the preparation of date syrup with an optimal recovery of total sugar content. However, ultrasound-assisted extraction was the best one for the preparation of date syrup with high content of reducing sugars.

Keywords: dates, extraction, RSM, sugars, syrup

Procedia PDF Downloads 138
1071 A Comprehensive Theory of Communication with Biological and Non-Biological Intelligence for a 21st Century Curriculum

Authors: Thomas Schalow

Abstract:

It is commonly recognized that our present curriculum is not preparing students to function in the 21st century. This is particularly true in regard to communication needs across cultures - both human and non-human. In this paper, a comprehensive theory of communication-based on communication with non-human cultures and intelligences is presented to meet the following three imminent contingencies: communicating with sentient biological intelligences, communicating with extraterrestrial intelligences, and communicating with artificial super-intelligences. The paper begins with the argument that we need to become much more serious about communicating with the non-human, intelligent life forms that already exists around us here on Earth. We need to broaden our definition of communication and reach out to other sentient life forms in order to provide humanity with a better perspective of its place within our ecosystem. The paper next examines the science and philosophy behind CETI (communication with extraterrestrial intelligences) and how it could prove useful even in the absence of contact with alien life. However, CETI’s assumptions and methodology need to be revised in accordance with the communication theory being proposed in this paper if we are truly serious about finding and communicating with life beyond Earth. The final theme explored in this paper is communication with non-biological super-intelligences. Humanity has never been truly compelled to converse with other species, and our failure to seriously consider such intercourse has left us largely unprepared to deal with communication in a future that will be mediated and controlled by computer algorithms. Fortunately, our experience dealing with other cultures can provide us with a framework for this communication. The basic concepts behind intercultural communication can be applied to the three types of communication envisioned in this paper if we are willing to recognize that we are in fact dealing with other cultures when we interact with other species, alien life, and artificial super-intelligence. The ideas considered in this paper will require a new mindset for humanity, but a new disposition will yield substantial gains. A curriculum that is truly ready for the 21st century needs to be aligned with this new theory of communication.

Keywords: artificial intelligence, CETI, communication, language

Procedia PDF Downloads 341
1070 Screening and Optimization of Pretreatments for Rice Straw and Their Utilization for Bioethanol Production Using Developed Yeast Strain

Authors: Ganesh Dattatraya Saratale, Min Kyu Oh

Abstract:

Rice straw is one of the most abundant lignocellulosic waste materials and its annual production is about 731 Mt in the world. This study treats the subject of effective utilization of this waste biomass for biofuels production. We have showed a comparative assessment of numerous pretreatment strategies for rice straw, comprising of major physical, chemical and physicochemical methods. Among the different methods employed for pretreatment alkaline pretreatment in combination with sodium chlorite/acetic acid delignification found efficient pretreatment with significant improvement in the enzymatic digestibility of rice straw. A cellulase dose of 20 filter paper units (FPU) released a maximum 63.21 g/L of reducing sugar with 94.45% hydrolysis yield and 64.64% glucose yield from rice straw, respectively. The effects of different pretreatment methods on biomass structure and complexity were investigated by FTIR, XRD and SEM analytical techniques. Finally the enzymatic hydrolysate of rice straw was used for ethanol production using developed Saccharomyces cerevisiae SR8. The developed yeast strain enabled efficient fermentation of xylose and glucose and produced higher ethanol production. Thus development of bioethanol production from lignocellulosic waste biomass is generic, applicable methodology and have great implication for using ‘green raw materials’ and producing ‘green products’ much needed today.

Keywords: rice straw, pretreatment, enzymatic hydrolysis, FPU, Saccharomyces cerevisiae SR8, ethanol fermentation

Procedia PDF Downloads 517
1069 Optimization of Highly Oriented Pyrolytic Graphite Crystals for Neutron Optics

Authors: Hao Qu, Xiang Liu, Michael Crosby, Brian Kozak, Andreas K. Freund

Abstract:

The outstanding performance of highly oriented pyrolytic graphite (HOPG) as an optical element for neutron beam conditioning is unequaled by any other crystalline material in the applications of monochromator, analyzer, and filter. This superiority stems from the favorable nuclear properties of carbon (small absorption and incoherent scattering cross-sections, big coherent scattering length) and the specific crystalline structure (small thermal diffuse scattering cross-section, layered crystal structure). The real crystal defect structure revealed by imaging techniques is correlated with the parameters used in the mosaic model (mosaic spread, mosaic block size, uniformity). The diffraction properties (rocking curve width as determined by both the intrinsic mosaic spread and the diffraction process, peak and integrated reflectivity, filter transmission) as a function of neutron wavelength or energy can be predicted with high accuracy and reliability by diffraction theory using empirical primary extinction coefficients extracted from a great amount of existing experimental data. The results of these calculations are given as graphs and tables permitting to optimize HOPG characteristics (mosaic spread, thickness, curvature) for any given experimental situation.

Keywords: neutron optics, pyrolytic graphite, mosaic spread, neutron scattering, monochromator, analyzer

Procedia PDF Downloads 125
1068 Multi-Impairment Compensation Based Deep Neural Networks for 16-QAM Coherent Optical Orthogonal Frequency Division Multiplexing System

Authors: Ying Han, Yuanxiang Chen, Yongtao Huang, Jia Fu, Kaile Li, Shangjing Lin, Jianguo Yu

Abstract:

In long-haul and high-speed optical transmission system, the orthogonal frequency division multiplexing (OFDM) signal suffers various linear and non-linear impairments. In recent years, researchers have proposed compensation schemes for specific impairment, and the effects are remarkable. However, different impairment compensation algorithms have caused an increase in transmission delay. With the widespread application of deep neural networks (DNN) in communication, multi-impairment compensation based on DNN will be a promising scheme. In this paper, we propose and apply DNN to compensate multi-impairment of 16-QAM coherent optical OFDM signal, thereby improving the performance of the transmission system. The trained DNN models are applied in the offline digital signal processing (DSP) module of the transmission system. The models can optimize the constellation mapping signals at the transmitter and compensate multi-impairment of the OFDM decoded signal at the receiver. Furthermore, the models reduce the peak to average power ratio (PAPR) of the transmitted OFDM signal and the bit error rate (BER) of the received signal. We verify the effectiveness of the proposed scheme for 16-QAM Coherent Optical OFDM signal and demonstrate and analyze transmission performance in different transmission scenarios. The experimental results show that the PAPR and BER of the transmission system are significantly reduced after using the trained DNN. It shows that the DNN with specific loss function and network structure can optimize the transmitted signal and learn the channel feature and compensate for multi-impairment in fiber transmission effectively.

Keywords: coherent optical OFDM, deep neural network, multi-impairment compensation, optical transmission

Procedia PDF Downloads 127
1067 Spatio-Temporal Pest Risk Analysis with ‘BioClass’

Authors: Vladimir A. Todiras

Abstract:

Spatio-temporal models provide new possibilities for real-time action in pest risk analysis. It should be noted that estimation of the possibility and probability of introduction of a pest and of its economic consequences involves many uncertainties. We present a new mapping technique that assesses pest invasion risk using online BioClass software. BioClass is a GIS tool designed to solve multiple-criteria classification and optimization problems based on fuzzy logic and level set methods. This research describes a method for predicting the potential establishment and spread of a plant pest into new areas using a case study: corn rootworm (Diabrotica spp.), tomato leaf miner (Tuta absoluta) and plum fruit moth (Grapholita funebrana). Our study demonstrated that in BioClass we can combine fuzzy logic and geographic information systems with knowledge of pest biology and environmental data to derive new information for decision making. Pests are sensitive to a warming climate, as temperature greatly affects their survival and reproductive rate and capacity. Changes have been observed in the distribution, frequency and severity of outbreaks of Helicoverpa armigera on tomato. BioClass has demonstrated to be a powerful tool for applying dynamic models and map the potential future distribution of a species, enable resource to make decisions about dangerous and invasive species management and control.

Keywords: classification, model, pest, risk

Procedia PDF Downloads 268
1066 Numerical Investigation and Optimization of the Effect of Number of Blade and Blade Type on the Suction Pressure and Outlet Mass Flow Rate of a Centrifugal Fan

Authors: Ogan Karabas, Suleyman Yigit

Abstract:

Number of blade and blade type of centrifugal fans are the most decisive factor on the field of application, noise level, suction pressure and outlet mass flow rate. Nowadays, in order to determine these effects on centrifugal fans, numerical studies are carried out in addition to experimental studies. In this study, it is aimed to numerically investigate the changes of suction pressure and outlet mass flow rate values of a centrifugal fan according to the number of blade and blade type. Centrifugal fans of the same size with forward, backward and straight blade type were analyzed by using a simulation program and compared with each other. This analysis was carried out under steady state condition by selecting k-Ɛ turbulence model and air is assumed incompressible. Then, 16, 32 and 48 blade centrifugal fans were again analyzed by using same simulation program, and the optimum number of blades was determined for the suction pressure and the outlet mass flow rate. According to the results of the analysis, it was obtained that the suction pressure in the 32 blade fan was twice the value obtained in the 16 blade fan. In addition, the outlet mass flow rate increased by 45% with the increase in the number of blade from 16 to 32. There is no significant change observed on the suction pressure and outlet mass flow rate when the number of blades increased from 32 to 48. In the light of the analysis results, the optimum blade number was determined as 32.

Keywords: blade type, centrifugal fan, cfd, outlet mass flow rate, suction pressure

Procedia PDF Downloads 386
1065 Quoting Jobshops Due Dates Subject to Exogenous Factors in Developing Nations

Authors: Idris M. Olatunde, Kareem B.

Abstract:

In manufacturing systems, especially job shops, service performance is a key factor that determines customer satisfaction. Service performance depends not only on the quality of the output but on the delivery lead times as well. Besides product quality enhancement, delivery lead time must be minimized for optimal patronage. Quoting accurate due dates is sine quo non for job shop operational survival in a global competitive environment. Quoting accurate due dates in job shops has been a herculean task that nearly defiled solutions from many methods employed due to complex jobs routing nature of the system. This class of NP-hard problems possessed no rigid algorithms that can give an optimal solution. Jobshop operational problem is more complex in developing nations due to some peculiar factors. Operational complexity in job shops emanated from political instability, poor economy, technological know-how, and the non-promising socio-political environment. The mentioned exogenous factors were hardly considered in the previous studies on scheduling problem related to due date determination in job shops. This study has filled the gap created in the past studies by developing a dynamic model that incorporated the exogenous factors for accurate determination of due dates for varying jobs complexity. Real data from six job shops selected from the different part of Nigeria, were used to test the efficacy of the model, and the outcomes were analyzed statistically. The results of the analyzes showed that the model is more promising in determining accurate due dates than the traditional models deployed by many job shops in terms of patronage and lead times minimization.

Keywords: due dates prediction, improved performance, customer satisfaction, dynamic model, exogenous factors, job shops

Procedia PDF Downloads 397
1064 Transmission Line Congestion Management Using Hybrid Fish-Bee Algorithm with Unified Power Flow Controller

Authors: P. Valsalal, S. Thangalakshmi

Abstract:

There is a widespread changeover in the electrical power industry universally from old-style monopolistic outline towards a horizontally distributed competitive structure to come across the demand of rising consumption. When the transmission lines of derestricted system are incapable to oblige the entire service needs, the lines are overloaded or congested. The governor between customer and power producer is nominated as Independent System Operator (ISO) to lessen the congestion without obstructing transmission line restrictions. Among the existing approaches for congestion management, the frequently used approaches are reorganizing the generation and load curbing. There is a boundary for reorganizing the generators, and further loads may not be supplemented with the prevailing resources unless more private power producers are added in the system by considerably raising the cost. Hence, congestion is relaxed by appropriate Flexible AC Transmission Systems (FACTS) devices which boost the existing transfer capacity of transmission lines. The FACTs device, namely, Unified Power Flow Controller (UPFC) is preferred, and the correct placement of UPFC is more vital and should be positioned in the highly congested line. Hence, the weak line is identified by using power flow performance index with the new objective function with proposed hybrid Fish – Bee algorithm. Further, the location of UPFC at appropriate line reduces the branch loading and minimizes the voltage deviation. The power transfer capacity of lines is determined with and without UPFC in the identified congested line of IEEE 30 bus structure and the simulated results are compared with prevailing algorithms. It is observed that the transfer capacity of existing line is increased with the presented algorithm and thus alleviating the congestion.

Keywords: available line transfer capability, congestion management, FACTS device, Hybrid Fish-Bee Algorithm, ISO, UPFC

Procedia PDF Downloads 364
1063 Optimization of Palm Oil Plantation Revitalization in North Sumatera

Authors: Juliza Hidayati, Sukardi, Ani Suryani, Sugiharto, Anas M. Fauzi

Abstract:

The idea of making North Sumatera as a barometer of national oil palm industry requires efforts commodities and agro-industry development of oil palm. One effort that can be done is by successful execution plantation revitalization. The plantation Revitalization is an effort to accelerate the development of smallholder plantations, through expansion and replanting by help of palm Estate Company as business partner and bank financed plantation revitalization fund. Business partner agreement obliged and bound to make at least the same smallholder plantation productivity with business partners, so that the refund rate to banks become larger and prosperous people as a plantation owner. Generally low productivity of smallholder plantations under normal potential caused a lot of old and damaged plants with plant material at random. The purpose of revitalizing oil palm plantations is which are to increase their competitiveness through increased farm productivity. The research aims to identify potential criteria in influencing plantation productivity improvement priorities to be observed and followed up in order to improve the competitiveness of destinations and make North Sumatera barometer of national palm oil can be achieved. Research conducted with Analytical Network Process (ANP), to find the effect of dependency relationships between factors or criteria with the knowledge of the experts in order to produce an objective opinion and relevant depict the actual situation.

Keywords: palm barometer, acceleration of plantation development, productivity, revitalization

Procedia PDF Downloads 661
1062 Emotion Detection in Twitter Messages Using Combination of Long Short-Term Memory and Convolutional Deep Neural Networks

Authors: Bahareh Golchin, Nooshin Riahi

Abstract:

One of the most significant issues as attended a lot in recent years is that of recognizing the sentiments and emotions in social media texts. The analysis of sentiments and emotions is intended to recognize the conceptual information such as the opinions, feelings, attitudes and emotions of people towards the products, services, organizations, people, topics, events and features in the written text. These indicate the greatness of the problem space. In the real world, businesses and organizations are always looking for tools to gather ideas, emotions, and directions of people about their products, services, or events related to their own. This article uses the Twitter social network, one of the most popular social networks with about 420 million active users, to extract data. Using this social network, users can share their information and opinions about personal issues, policies, products, events, etc. It can be used with appropriate classification of emotional states due to the availability of its data. In this study, supervised learning and deep neural network algorithms are used to classify the emotional states of Twitter users. The use of deep learning methods to increase the learning capacity of the model is an advantage due to the large amount of available data. Tweets collected on various topics are classified into four classes using a combination of two Bidirectional Long Short Term Memory network and a Convolutional network. The results obtained from this study with an average accuracy of 93%, show good results extracted from the proposed framework and improved accuracy compared to previous work.

Keywords: emotion classification, sentiment analysis, social networks, deep neural networks

Procedia PDF Downloads 122