Search results for: optimization algorithms
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4760

Search results for: optimization algorithms

3110 An Optimization of Machine Parameters for Modified Horizontal Boring Tool Using Taguchi Method

Authors: Thirasak Panyaphirawat, Pairoj Sapsmarnwong, Teeratas Pornyungyuen

Abstract:

This paper presents the findings of an experimental investigation of important machining parameters for the horizontal boring tool modified to mouth with a horizontal lathe machine to bore an overlength workpiece. In order to verify a usability of a modified tool, design of experiment based on Taguchi method is performed. The parameters investigated are spindle speed, feed rate, depth of cut and length of workpiece. Taguchi L9 orthogonal array is selected for four factors three level parameters in order to minimize surface roughness (Ra and Rz) of S45C steel tubes. Signal to noise ratio analysis and analysis of variance (ANOVA) is performed to study an effect of said parameters and to optimize the machine setting for best surface finish. The controlled factors with most effect are depth of cut, spindle speed, length of workpiece, and feed rate in order. The confirmation test is performed to test the optimal setting obtained from Taguchi method and the result is satisfactory.

Keywords: design of experiment, Taguchi design, optimization, analysis of variance, machining parameters, horizontal boring tool

Procedia PDF Downloads 432
3109 Unveiling Comorbidities in Irritable Bowel Syndrome: A UK BioBank Study utilizing Supervised Machine Learning

Authors: Uswah Ahmad Khan, Muhammad Moazam Fraz, Humayoon Shafique Satti, Qasim Aziz

Abstract:

Approximately 10-14% of the global population experiences a functional disorder known as irritable bowel syndrome (IBS). The disorder is defined by persistent abdominal pain and an irregular bowel pattern. IBS significantly impairs work productivity and disrupts patients' daily lives and activities. Although IBS is widespread, there is still an incomplete understanding of its underlying pathophysiology. This study aims to help characterize the phenotype of IBS patients by differentiating the comorbidities found in IBS patients from those in non-IBS patients using machine learning algorithms. In this study, we extracted samples coding for IBS from the UK BioBank cohort and randomly selected patients without a code for IBS to create a total sample size of 18,000. We selected the codes for comorbidities of these cases from 2 years before and after their IBS diagnosis and compared them to the comorbidities in the non-IBS cohort. Machine learning models, including Decision Trees, Gradient Boosting, Support Vector Machine (SVM), AdaBoost, Logistic Regression, and XGBoost, were employed to assess their accuracy in predicting IBS. The most accurate model was then chosen to identify the features associated with IBS. In our case, we used XGBoost feature importance as a feature selection method. We applied different models to the top 10% of features, which numbered 50. Gradient Boosting, Logistic Regression and XGBoost algorithms yielded a diagnosis of IBS with an optimal accuracy of 71.08%, 71.427%, and 71.53%, respectively. Among the comorbidities most closely associated with IBS included gut diseases (Haemorrhoids, diverticular diseases), atopic conditions(asthma), and psychiatric comorbidities (depressive episodes or disorder, anxiety). This finding emphasizes the need for a comprehensive approach when evaluating the phenotype of IBS, suggesting the possibility of identifying new subsets of IBS rather than relying solely on the conventional classification based on stool type. Additionally, our study demonstrates the potential of machine learning algorithms in predicting the development of IBS based on comorbidities, which may enhance diagnosis and facilitate better management of modifiable risk factors for IBS. Further research is necessary to confirm our findings and establish cause and effect. Alternative feature selection methods and even larger and more diverse datasets may lead to more accurate classification models. Despite these limitations, our findings highlight the effectiveness of Logistic Regression and XGBoost in predicting IBS diagnosis.

Keywords: comorbidities, disease association, irritable bowel syndrome (IBS), predictive analytics

Procedia PDF Downloads 102
3108 Optimization of Hemp Fiber Reinforced Concrete for Various Environmental Conditions

Authors: Zoe Chang, Max Williams, Gautham Das

Abstract:

The purpose of this study is to evaluate the incorporation of hemp fibers (HF) in concrete. Hemp fiber reinforced concrete (HFRC) is becoming more popular as an alternative for regular mix designs. This study was done to evaluate the compressive strength of HFRC regarding mix procedure. Hemp fibers were obtained from the manufacturer and hand-processed to ensure uniformity in width and length. The fibers were added to the concrete as both wet and dry mixes to investigate and optimize the mix design process. Results indicated that the dry mix had a compressive strength of 1157 psi compared to the wet mix of 985 psi. This dry mix compressive strength was within range of the standard mix compressive strength of 1533 psi. The statistical analysis revealed that the mix design process needs further optimization and uniformity concerning the addition of HF. Regression analysis revealed the standard mix design had a coefficient of 0.9 as compared to the dry mix of 0.375, indicating a variation in the mixing process. While completing the dry mix, the addition of plain hemp fibers caused them to intertwine, creating lumps and inconsistency. However, during the wet mixing process, combining water and hemp fibers before incorporation allows the fibers to uniformly disperse within the mix; hence the regression analysis indicated a better coefficient of 0.55. This study concludes that HRFC is a viable alternative to regular mixes; however, more research surrounding its characteristics needs to be conducted.

Keywords: hemp fibers, hemp reinforced concrete, wet & dry, freeze thaw testing, compressive strength

Procedia PDF Downloads 186
3107 Methods and Algorithms of Ensuring Data Privacy in AI-Based Healthcare Systems and Technologies

Authors: Omar Farshad Jeelani, Makaire Njie, Viktoriia M. Korzhuk

Abstract:

Recently, the application of AI-powered algorithms in healthcare continues to flourish. Particularly, access to healthcare information, including patient health history, diagnostic data, and PII (Personally Identifiable Information) is paramount in the delivery of efficient patient outcomes. However, as the exchange of healthcare information between patients and healthcare providers through AI-powered solutions increases, protecting a person’s information and their privacy has become even more important. Arguably, the increased adoption of healthcare AI has resulted in a significant concentration on the security risks and protection measures to the security and privacy of healthcare data, leading to escalated analyses and enforcement. Since these challenges are brought by the use of AI-based healthcare solutions to manage healthcare data, AI-based data protection measures are used to resolve the underlying problems. Consequently, this project proposes AI-powered safeguards and policies/laws to protect the privacy of healthcare data. The project presents the best-in-school techniques used to preserve the data privacy of AI-powered healthcare applications. Popular privacy-protecting methods like Federated learning, cryptographic techniques, differential privacy methods, and hybrid methods are discussed together with potential cyber threats, data security concerns, and prospects. Also, the project discusses some of the relevant data security acts/laws that govern the collection, storage, and processing of healthcare data to guarantee owners’ privacy is preserved. This inquiry discusses various gaps and uncertainties associated with healthcare AI data collection procedures and identifies potential correction/mitigation measures.

Keywords: data privacy, artificial intelligence (AI), healthcare AI, data sharing, healthcare organizations (HCOs)

Procedia PDF Downloads 72
3106 Dynamic Route Optimization in Vehicle Adhoc Networks: A Heuristics Routing Protocol

Authors: Rafi Ullah, Shah Muhammad Emaduddin, Taha Jilani

Abstract:

Vehicle Adhoc Networks (VANET) belongs to a special class of Mobile Adhoc Network (MANET) with high mobility. Network is created by road side vehicles equipped with communication devices like GPS and Wifi etc. Since the environment is highly dynamic due to difference in speed and high mobility of vehicles and weak stability of the network connection, it is a challenging task to design an efficient routing protocol for such an unstable environment. Our proposed algorithm uses heuristic for the calculation of optimal path for routing the packet efficiently in collaboration with several other parameters like geographical location, speed, priority, the distance among the vehicles, communication range, and networks congestion. We have incorporated probabilistic, heuristic and machine learning based approach inconsistency with the relay function of the memory buffer to keep the packet moving towards the destination. These parameters when used in collaboration provide us a very strong and admissible heuristics. We have mathematically proved that the proposed technique is efficient for the routing of packets, especially in a medical emergency situation. These networks can be used for medical emergency, security, entertainment and routing purposes.

Keywords: heuristics routing, intelligent routing, VANET, route optimization

Procedia PDF Downloads 165
3105 Integrated Intensity and Spatial Enhancement Technique for Color Images

Authors: Evan W. Krieger, Vijayan K. Asari, Saibabu Arigela

Abstract:

Video imagery captured for real-time security and surveillance applications is typically captured in complex lighting conditions. These less than ideal conditions can result in imagery that can have underexposed or overexposed regions. It is also typical that the video is too low in resolution for certain applications. The purpose of security and surveillance video is that we should be able to make accurate conclusions based on the images seen in the video. Therefore, if poor lighting and low resolution conditions occur in the captured video, the ability to make accurate conclusions based on the received information will be reduced. We propose a solution to this problem by using image preprocessing to improve these images before use in a particular application. The proposed algorithm will integrate an intensity enhancement algorithm with a super resolution technique. The intensity enhancement portion consists of a nonlinear inverse sign transformation and an adaptive contrast enhancement. The super resolution section is a single image super resolution technique is a Fourier phase feature based method that uses a machine learning approach with kernel regression. The proposed technique intelligently integrates these algorithms to be able to produce a high quality output while also being more efficient than the sequential use of these algorithms. This integration is accomplished by performing the proposed algorithm on the intensity image produced from the original color image. After enhancement and super resolution, a color restoration technique is employed to obtain an improved visibility color image.

Keywords: dynamic range compression, multi-level Fourier features, nonlinear enhancement, super resolution

Procedia PDF Downloads 543
3104 Study of Launch Recovery Control Dynamics of Retro Propulsive Reusable Rockets

Authors: Pratyush Agnihotri

Abstract:

The space missions are very costly because the transportation to the space is highly expensive and therefore there is the need to achieve complete re-usability in our launch vehicles to make the missions highly economic by cost cutting of the material recovered. Launcher reusability is the most efficient approach to decreasing admittance to space access economy, however stays an incredible specialized hurdle for the aerospace industry. Major concern of the difficulties lies in guidance and control procedure and calculations, specifically for those of the controlled landing stage, which should empower an exact landing with low fuel edges. Although cutting edge ways for navigation and control are present viz hybrid navigation and robust control. But for powered descent and landing of first stage of launch vehicle the guidance control is need to enable on board optimization. At first the CAD model of the launch vehicle I.e. space x falcon 9 rocket is presented for better understanding of the architecture that needs to be identified for the guidance and control solution for the recovery of the launcher. The focus is on providing the landing phase guidance scheme for recovery and re usability of first stage using retro propulsion. After reviewing various GNC solutions, to achieve accuracy in pre requisite landing online convex and successive optimization are explored as the guidance schemes.

Keywords: guidance, navigation, control, retro propulsion, reusable rockets

Procedia PDF Downloads 80
3103 Building User Behavioral Models by Processing Web Logs and Clustering Mechanisms

Authors: Madhuka G. P. D. Udantha, Gihan V. Dias, Surangika Ranathunga

Abstract:

Today Websites contain very interesting applications. But there are only few methodologies to analyze User navigations through the Websites and formulating if the Website is put to correct use. The web logs are only used if some major attack or malfunctioning occurs. Web Logs contain lot interesting dealings on users in the system. Analyzing web logs has become a challenge due to the huge log volume. Finding interesting patterns is not as easy as it is due to size, distribution and importance of minor details of each log. Web logs contain very important data of user and site which are not been put to good use. Retrieving interesting information from logs gives an idea of what the users need, group users according to their various needs and improve site to build an effective and efficient site. The model we built is able to detect attacks or malfunctioning of the system and anomaly detection. Logs will be more complex as volume of traffic and the size and complexity of web site grows. Unsupervised techniques are used in this solution which is fully automated. Expert knowledge is only used in validation. In our approach first clean and purify the logs to bring them to a common platform with a standard format and structure. After cleaning module web session builder is executed. It outputs two files, Web Sessions file and Indexed URLs file. The Indexed URLs file contains the list of URLs accessed and their indices. Web Sessions file lists down the indices of each web session. Then DBSCAN and EM Algorithms are used iteratively and recursively to get the best clustering results of the web sessions. Using homogeneity, completeness, V-measure, intra and inter cluster distance and silhouette coefficient as parameters these algorithms self-evaluate themselves to input better parametric values to run the algorithms. If a cluster is found to be too large then micro-clustering is used. Using Cluster Signature Module the clusters are annotated with a unique signature called finger-print. In this module each cluster is fed to Associative Rule Learning Module. If it outputs confidence and support as value 1 for an access sequence it would be a potential signature for the cluster. Then the access sequence occurrences are checked in other clusters. If it is found to be unique for the cluster considered then the cluster is annotated with the signature. These signatures are used in anomaly detection, prevent cyber attacks, real-time dashboards that visualize users, accessing web pages, predict actions of users and various other applications in Finance, University Websites, News and Media Websites etc.

Keywords: anomaly detection, clustering, pattern recognition, web sessions

Procedia PDF Downloads 278
3102 Prediction Modeling of Alzheimer’s Disease and Its Prodromal Stages from Multimodal Data with Missing Values

Authors: M. Aghili, S. Tabarestani, C. Freytes, M. Shojaie, M. Cabrerizo, A. Barreto, N. Rishe, R. E. Curiel, D. Loewenstein, R. Duara, M. Adjouadi

Abstract:

A major challenge in medical studies, especially those that are longitudinal, is the problem of missing measurements which hinders the effective application of many machine learning algorithms. Furthermore, recent Alzheimer's Disease studies have focused on the delineation of Early Mild Cognitive Impairment (EMCI) and Late Mild Cognitive Impairment (LMCI) from cognitively normal controls (CN) which is essential for developing effective and early treatment methods. To address the aforementioned challenges, this paper explores the potential of using the eXtreme Gradient Boosting (XGBoost) algorithm in handling missing values in multiclass classification. We seek a generalized classification scheme where all prodromal stages of the disease are considered simultaneously in the classification and decision-making processes. Given the large number of subjects (1631) included in this study and in the presence of almost 28% missing values, we investigated the performance of XGBoost on the classification of the four classes of AD, NC, EMCI, and LMCI. Using 10-fold cross validation technique, XGBoost is shown to outperform other state-of-the-art classification algorithms by 3% in terms of accuracy and F-score. Our model achieved an accuracy of 80.52%, a precision of 80.62% and recall of 80.51%, supporting the more natural and promising multiclass classification.

Keywords: eXtreme gradient boosting, missing data, Alzheimer disease, early mild cognitive impairment, late mild cognitive impair, multiclass classification, ADNI, support vector machine, random forest

Procedia PDF Downloads 176
3101 Hybrid Model: An Integration of Machine Learning with Traditional Scorecards

Authors: Golnush Masghati-Amoli, Paul Chin

Abstract:

Over the past recent years, with the rapid increases in data availability and computing power, Machine Learning (ML) techniques have been called on in a range of different industries for their strong predictive capability. However, the use of Machine Learning in commercial banking has been limited due to a special challenge imposed by numerous regulations that require lenders to be able to explain their analytic models, not only to regulators but often to consumers. In other words, although Machine Leaning techniques enable better prediction with a higher level of accuracy, in comparison with other industries, they are adopted less frequently in commercial banking especially for scoring purposes. This is due to the fact that Machine Learning techniques are often considered as a black box and fail to provide information on why a certain risk score is given to a customer. In order to bridge this gap between the explain-ability and performance of Machine Learning techniques, a Hybrid Model is developed at Dun and Bradstreet that is focused on blending Machine Learning algorithms with traditional approaches such as scorecards. The Hybrid Model maximizes efficiency of traditional scorecards by merging its practical benefits, such as explain-ability and the ability to input domain knowledge, with the deep insights of Machine Learning techniques which can uncover patterns scorecard approaches cannot. First, through development of Machine Learning models, engineered features and latent variables and feature interactions that demonstrate high information value in the prediction of customer risk are identified. Then, these features are employed to introduce observed non-linear relationships between the explanatory and dependent variables into traditional scorecards. Moreover, instead of directly computing the Weight of Evidence (WoE) from good and bad data points, the Hybrid Model tries to match the score distribution generated by a Machine Learning algorithm, which ends up providing an estimate of the WoE for each bin. This capability helps to build powerful scorecards with sparse cases that cannot be achieved with traditional approaches. The proposed Hybrid Model is tested on different portfolios where a significant gap is observed between the performance of traditional scorecards and Machine Learning models. The result of analysis shows that Hybrid Model can improve the performance of traditional scorecards by introducing non-linear relationships between explanatory and target variables from Machine Learning models into traditional scorecards. Also, it is observed that in some scenarios the Hybrid Model can be almost as predictive as the Machine Learning techniques while being as transparent as traditional scorecards. Therefore, it is concluded that, with the use of Hybrid Model, Machine Learning algorithms can be used in the commercial banking industry without being concerned with difficulties in explaining the models for regulatory purposes.

Keywords: machine learning algorithms, scorecard, commercial banking, consumer risk, feature engineering

Procedia PDF Downloads 123
3100 Design and Implementation of Low-code Model-building Methods

Authors: Zhilin Wang, Zhihao Zheng, Linxin Liu

Abstract:

This study proposes a low-code model-building approach that aims to simplify the development and deployment of artificial intelligence (AI) models. With an intuitive way to drag and drop and connect components, users can easily build complex models and integrate multiple algorithms for training. After the training is completed, the system automatically generates a callable model service API. This method not only lowers the technical threshold of AI development and improves development efficiency but also enhances the flexibility of algorithm integration and simplifies the deployment process of models. The core strength of this method lies in its ease of use and efficiency. Users do not need to have a deep programming background and can complete the design and implementation of complex models with a simple drag-and-drop operation. This feature greatly expands the scope of AI technology, allowing more non-technical people to participate in the development of AI models. At the same time, the method performs well in algorithm integration, supporting many different types of algorithms to work together, which further improves the performance and applicability of the model. In the experimental part, we performed several performance tests on the method. The results show that compared with traditional model construction methods, this method can make more efficient use, save computing resources, and greatly shorten the model training time. In addition, the system-generated model service interface has been optimized for high availability and scalability, which can adapt to the needs of different application scenarios.

Keywords: low-code, model building, artificial intelligence, algorithm integration, model deployment

Procedia PDF Downloads 8
3099 Facile Synthesis and Structure Characterization of Europium (III) Tungstate Nanoparticles

Authors: Mehdi Rahimi-Nasrabadi, Seied Mahdi Pourmortazavi

Abstract:

Taguchi robust design as a statistical method was applied for optimization of the process parameters in order to tunable, simple and fast synthesis of europium (III) tungstate nanoparticles. Europium (III) tungstate nanoparticles were synthesized by a chemical precipitation reaction involving direct addition of europium ion aqueous solution to the tungstate reagent solved in aqueous media. Effects of some synthesis procedure variables i.e., europium and tungstate concentrations, flow rate of cation reagent addition, and temperature of reaction reactor on the particle size of europium (III) tungstate nanoparticles were studied experimentally in order to tune particle size of europium (III) tungstate. Analysis of variance shows the importance of controlling tungstate concentration, cation feeding flow rate and temperature for preparation of europium (III) tungstate nanoparticles by the proposed chemical precipitation reaction. Finally, europium (III) tungstate nanoparticles were synthesized at the optimum conditions of the proposed method and the morphology and chemical composition of the prepared nano-material were characterized by means of X-Ray diffraction, scanning electron microscopy, transmission electron microscopy, FT-IR spectroscopy, and fluorescence.

Keywords: europium (III) tungstate, nano-material, particle size control, procedure optimization

Procedia PDF Downloads 384
3098 Applications for Additive Manufacturing Technology for Reducing the Weight of Body Parts of Gas Turbine Engines

Authors: Liubov Magerramova, Mikhail Petrov, Vladimir Isakov, Liana Shcherbinina, Suren Gukasyan, Daniil Povalyukhin, Olga Klimova-Korsmik, Darya Volosevich

Abstract:

Aircraft engines are developing along the path of increasing resource, strength, reliability, and safety. The building of gas turbine engine body parts is a complex design and technological task. Particularly complex in the design and manufacturing are the casings of the input stages of helicopter gearboxes and central drives of aircraft engines. Traditional technologies, such as precision casting or isothermal forging, are characterized by significant limitations in parts production. For parts like housing, additive technologies guarantee spatial freedom and limitless or flexible design. This article presents the results of computational and experimental studies. These investigations justify the applicability of additive technologies (AT) to reduce the weight of aircraft housing gearbox parts by up to 32%. This is possible due to geometrical optimization compared to the classical, less flexible manufacturing methods and as-casted aircraft parts with over-insured values of safety factors. Using an example of the body of the input stage of an aircraft gearbox, visualization of the layer-by-layer manufacturing of a part based on thermal deformation was demonstrated.

Keywords: additive technologies, gas turbine engines, topological optimization, synthesis process

Procedia PDF Downloads 95
3097 On-Chip Sensor Ellipse Distribution Method and Equivalent Mapping Technique for Real-Time Hardware Trojan Detection and Location

Authors: Longfei Wang, Selçuk Köse

Abstract:

Hardware Trojan becomes great concern as integrated circuit (IC) technology advances and not all manufacturing steps of an IC are accomplished within one company. Real-time hardware Trojan detection is proven to be a feasible way to detect randomly activated Trojans that cannot be detected at testing stage. On-chip sensors serve as a great candidate to implement real-time hardware Trojan detection, however, the optimization of on-chip sensors has not been thoroughly investigated and the location of Trojan has not been carefully explored. On-chip sensor ellipse distribution method and equivalent mapping technique are proposed based on the characteristics of on-chip power delivery network in this paper to address the optimization and distribution of on-chip sensors for real-time hardware Trojan detection as well as to estimate the location and current consumption of hardware Trojan. Simulation results verify that hardware Trojan activation can be effectively detected and the location of a hardware Trojan can be efficiently estimated with less than 5% error for a realistic power grid using our proposed methods. The proposed techniques therefore lay a solid foundation for isolation and even deactivation of hardware Trojans through accurate location of Trojans.

Keywords: hardware trojan, on-chip sensor, power distribution network, power/ground noise

Procedia PDF Downloads 381
3096 Tree-Based Inference for Regionalization: A Comparative Study of Global Topological Perturbation Methods

Authors: Orhun Aydin, Mark V. Janikas, Rodrigo Alves, Renato Assuncao

Abstract:

In this paper, a tree-based perturbation methodology for regionalization inference is presented. Regionalization is a constrained optimization problem that aims to create groups with similar attributes while satisfying spatial contiguity constraints. Similar to any constrained optimization problem, the spatial constraint may hinder convergence to some global minima, resulting in spatially contiguous members of a group with dissimilar attributes. This paper presents a general methodology for rigorously perturbing spatial constraints through the use of random spanning trees. The general framework presented can be used to quantify the effect of the spatial constraints in the overall regionalization result. We compare several types of stochastic spanning trees used in inference problems such as fuzzy regionalization and determining the number of regions. Performance of stochastic spanning trees is juxtaposed against the traditional permutation-based hypothesis testing frequently used in spatial statistics. Inference results for fuzzy regionalization and determining the number of regions is presented on the Local Area Personal Incomes for Texas Counties provided by the Bureau of Economic Analysis.

Keywords: regionalization, constrained clustering, probabilistic inference, fuzzy clustering

Procedia PDF Downloads 215
3095 Field Oriented Control of Electrical Motor for Efficiency Improvement of Aerial Vehicle

Authors: Francois Defay

Abstract:

Uses of Unmanned aerial vehicle (UAV) are increasing for many applicative cases. Long endurance UAVs are required for inspection or transportation in some deserted places. The global optimization of the efficiency is the aim of the works in ISAE-SUPAERO. From the propulsive part until the motor control, the global optimization can increase significantly the global efficiency. This paper deals with the global improvement of the efficiency of the electrical propulsion for the aerial vehicle. The application case of study is a small airplane of 2kg. A global modelization is presented in order to validate the electrical engine in a complete simulation from aerodynamics to battery. The classical control of the synchronous permanent drive is compared to the field-oriented control which is not yet applied for UAVs. The experimental results presented show an increase of more than 10 percent of the efficiency. A complete modelization and simulation based on Matlab/ Simulink are presented in this paper and compared to the experimental study. Finally this paper presents solutions to increase the endurance of the electrical aerial vehicle and provide models to optimize the global consumption for a specific mission. The next step is to use this model and the control to work with distributed propulsion which is the future for small distance plane.

Keywords: electrical propulsion, endurance, field-oriented control, UAV

Procedia PDF Downloads 227
3094 Reliability and Cost Focused Optimization Approach for a Communication Satellite Payload Redundancy Allocation Problem

Authors: Mehmet Nefes, Selman Demirel, Hasan H. Ertok, Cenk Sen

Abstract:

A typical reliability engineering problem regarding communication satellites has been considered to determine redundancy allocation scheme of power amplifiers within payload transponder module, whose dominant function is to amplify power levels of the received signals from the Earth, through maximizing reliability against mass, power, and other technical limitations. Adding each redundant power amplifier component increases not only reliability but also hardware, testing, and launch cost of a satellite. This study investigates a multi-objective approach used in order to solve Redundancy Allocation Problem (RAP) for a communication satellite payload transponder, focusing on design cost due to redundancy and reliability factors. The main purpose is to find the optimum power amplifier redundancy configuration satisfying reliability and capacity thresholds simultaneously instead of analyzing respectively or independently. A mathematical model and calculation approach are instituted including objective function definitions, and then, the problem is solved analytically with different input parameters in MATLAB environment. Example results showed that payload capacity and failure rate of power amplifiers have remarkable effects on the solution and also processing time.

Keywords: communication satellite payload, multi-objective optimization, redundancy allocation problem, reliability, transponder

Procedia PDF Downloads 257
3093 Optimization of Wind Off-Grid System for Remote Area: Egyptian Application

Authors: Marwa M. Ibrahim

Abstract:

The objective of this research is to study the technical and economic performance of wind/diesel/battery (W/D/B) off-grid system supplying a small remote gathering of four families using the HOMER software package. The second objective is to study the effect of wind energy system on the cost of generated electricity considering the cost of reducing CO₂ emissions as external benefit of wind turbines, no pollutant emission through the operational phase. The system consists of a small wind turbine, battery storage, and diesel generator. The electrical energy is to cater to the basic needs for which the daily load pattern is estimated at 8 kW peak. Net Present Cost (NPC) and Cost of Energy (COE) are used as economic criteria, while the measure of performance is % of power shortage. Technical and economic parameters are defined to estimate the feasibility of the system under study. Optimum system configurations are estimated for the selected site in Egypt. Using HOMER software, the simulation results shows that W/D/B systems are economical for the assumed community site as the price of generated electricity is about 0.285 $/kWh, without taking external benefits into considerations and 0.221 if CO₂ emissions taken into consideration W/D/B systems are more economical than alone diesel system as the COE is 0.432 $/kWh for diesel alone.

Keywords: renewable energy, hybrid energy system, on-off grid system, simulation, optimization and environmental impacts

Procedia PDF Downloads 97
3092 Real-Time Data Stream Partitioning over a Sliding Window in Real-Time Spatial Big Data

Authors: Sana Hamdi, Emna Bouazizi, Sami Faiz

Abstract:

In recent years, real-time spatial applications, like location-aware services and traffic monitoring, have become more and more important. Such applications result dynamic environments where data as well as queries are continuously moving. As a result, there is a tremendous amount of real-time spatial data generated every day. The growth of the data volume seems to outspeed the advance of our computing infrastructure. For instance, in real-time spatial Big Data, users expect to receive the results of each query within a short time period without holding in account the load of the system. But with a huge amount of real-time spatial data generated, the system performance degrades rapidly especially in overload situations. To solve this problem, we propose the use of data partitioning as an optimization technique. Traditional horizontal and vertical partitioning can increase the performance of the system and simplify data management. But they remain insufficient for real-time spatial Big data; they can’t deal with real-time and stream queries efficiently. Thus, in this paper, we propose a novel data partitioning approach for real-time spatial Big data named VPA-RTSBD (Vertical Partitioning Approach for Real-Time Spatial Big data). This contribution is an implementation of the Matching algorithm for traditional vertical partitioning. We find, firstly, the optimal attribute sequence by the use of Matching algorithm. Then, we propose a new cost model used for database partitioning, for keeping the data amount of each partition more balanced limit and for providing a parallel execution guarantees for the most frequent queries. VPA-RTSBD aims to obtain a real-time partitioning scheme and deals with stream data. It improves the performance of query execution by maximizing the degree of parallel execution. This affects QoS (Quality Of Service) improvement in real-time spatial Big Data especially with a huge volume of stream data. The performance of our contribution is evaluated via simulation experiments. The results show that the proposed algorithm is both efficient and scalable, and that it outperforms comparable algorithms.

Keywords: real-time spatial big data, quality of service, vertical partitioning, horizontal partitioning, matching algorithm, hamming distance, stream query

Procedia PDF Downloads 150
3091 Effect of Ultrasound on Carotenoids Extraction from Pepper and Process Optimization Using Response Surface Methodology (RSM)

Authors: Elham Mahdian, Reza Karazhian, Rahele Dehghan Tanha

Abstract:

Pepper (Capsicum annum L.) which belong to the family Solananceae, are known for their versatility as a vegetable crop and are consumed both as fresh vegetables or dehydrated for spices. Pepper is considered an excellent source of bioactive nutrients. Ascorbic acid, carotenoids and phenolic compounds are its main antioxidant constituents. Ultrasound assisted extraction is an inexpensive, simple and efficient alternative to conventional extraction techniques. The mechanism of action for ultrasound-assisted extraction are attributed to cavitations, mechanical forces and thermal impact, which result in disruption of cells walls, reduce particle size, and enhance mass transfer across cell membranes. In this study, response surface methodology was used to optimize experimental conditions for ultrasonic assisted extraction of carotenoid compounds from Chili peppers. Variables were included extraction temperatures at 3 levels (30, 40 and 50 °C), extraction times at 3 levels (10, 25 and 40 minutes) and power at 3 levels (30, 60 and 90 %). It was observed that ultrasound waves applied at temperature of 49°C, time of 10 minutes and power 89 % resulted to the highest carotenoids contents (lycopene and β-carotene), while the lowest value was recorded in the control. Thus, results showed that ultrasound waves have strong impact on extraction of carotenoids from pepper.

Keywords: carotenoids, optimization, pepper, response surface methodology

Procedia PDF Downloads 460
3090 Improving Fault Tolerance and Load Balancing in Heterogeneous Grid Computing Using Fractal Transform

Authors: Saad M. Darwish, Adel A. El-Zoghabi, Moustafa F. Ashry

Abstract:

The popularity of the Internet and the availability of powerful computers and high-speed networks as low-cost commodity components are changing the way we use computers today. These technical opportunities have led to the possibility of using geographically distributed and multi-owner resources to solve large-scale problems in science, engineering, and commerce. Recent research on these topics has led to the emergence of a new paradigm known as Grid computing. To achieve the promising potentials of tremendous distributed resources, effective and efficient load balancing algorithms are fundamentally important. Unfortunately, load balancing algorithms in traditional parallel and distributed systems, which usually run on homogeneous and dedicated resources, cannot work well in the new circumstances. In this paper, the concept of a fast fractal transform in heterogeneous grid computing based on R-tree and the domain-range entropy is proposed to improve fault tolerance and load balancing algorithm by improve connectivity, communication delay, network bandwidth, resource availability, and resource unpredictability. A novel two-dimension figure of merit is suggested to describe the network effects on load balance and fault tolerance estimation. Fault tolerance is enhanced by adaptively decrease replication time and message cost while load balance is enhanced by adaptively decrease mean job response time. Experimental results show that the proposed method yields superior performance over other methods.

Keywords: Grid computing, load balancing, fault tolerance, R-tree, heterogeneous systems

Procedia PDF Downloads 477
3089 Proposition of an Intelligent System Based on the Augmented Reality for Warehouse Logistics

Authors: Safa Gharbi, Hayfa Zgaya, Nesrine Zoghlami, Slim Hammadi, Cyril De Barbarin, Laurent Vinatier, Christiane Coupier

Abstract:

Increasing productivity and quality of service, improving the working comfort and ensuring the efficiency of all processes are important challenges for every warehouse. The order picking is recognized to be the most important and costly activity of all the process in warehouses. This paper presents a new approach using Augmented Reality (AR) in the field of logistics. It aims to create a Head-Up Display (HUD) interface with a Warehouse Management System (WMS), using AR glasses. Integrating AR technology allows the optimization of order picking by reducing time of picking process, increasing the efficiency and delivering quickly. The picker will be able to access immediately to all the information needed for his tasks. All the information is displayed when needed in the field of vision (FOV) of the operator, without any action requested from him. These research works are part of the industrial project RASL (Réalité Augmentée au Service de la Logistique) which gathers two major partners: the LAGIS (Laboratory of Automatics, Computer Engineering and Signal Processing in Lille-France) and Genrix Group, European leader in warehouses logistics, who provided his software for implementation, and his logistics expertise.

Keywords: Augmented Reality (AR), logistics and optimization, Warehouse Management System (WMS), Head-Up Display (HUD)

Procedia PDF Downloads 469
3088 Design and Optimization for a Compliant Gripper with Force Regulation Mechanism

Authors: Nhat Linh Ho, Thanh-Phong Dao, Shyh-Chour Huang, Hieu Giang Le

Abstract:

This paper presents a design and optimization for a compliant gripper. The gripper is constructed based on the concept of compliant mechanism with flexure hinge. A passive force regulation mechanism is presented to control the grasping force a micro-sized object instead of using a sensor force. The force regulation mechanism is designed using the planar springs. The gripper is expected to obtain a large range of displacement to handle various sized objects. First of all, the statics and dynamics of the gripper are investigated by using the finite element analysis in ANSYS software. And then, the design parameters of the gripper are optimized via Taguchi method. An orthogonal array L9 is used to establish an experimental matrix. Subsequently, the signal to noise ratio is analyzed to find the optimal solution. Finally, the response surface methodology is employed to model the relationship between the design parameters and the output displacement of the gripper. The design of experiment method is then used to analyze the sensitivity so as to determine the effect of each parameter on the displacement. The results showed that the compliant gripper can move with a large displacement of 213.51 mm and the force regulation mechanism is expected to be used for high precision positioning systems.

Keywords: flexure hinge, compliant mechanism, compliant gripper, force regulation mechanism, Taguchi method, response surface methodology, design of experiment

Procedia PDF Downloads 315
3087 Implementation of Inference Fuzzy System as a Valuation Subsidiary is Based Particle Swarm Optimization for Solves the Issue of Decision Making in Middle Size Soccer Robot League

Authors: Zahra Abdolkarimi, Naser Zouri

Abstract:

Nowadays, there is unbelievable growing of Robots created a collection of complex and motivate subject in robotic and intellectual ornate, also it made a mechatronics style base of theoretical and technical way in Robocop. Additionally, robotics system recommended RoboCup factor as a provider of some standardization and testing method in case of computer discussion widely. The actual purpose of RoboCup is creating independent team of robots in 2050 based of FiFa roles to bring the victory in compare of world star team. In addition, decision making of robots depends to environment reaction, self-player and rival player with using inductive Fuzzy system valuation subsidiary to solve issue of robots in land game. The measure of selection in compare with other methods depends to amount of victories percentage in the same team that plays accidently. Consequences, shows method of our discussion is the best way for Particle Swarm Optimization and Fuzzy system compare to other decision of robotics algorithmic.

Keywords: PSO algorithm, inference fuzzy system, chaos theory, soccer robot league

Procedia PDF Downloads 391
3086 Assignment of Airlines Technical Members under Disruption

Authors: Walid Moudani

Abstract:

The Crew Reserve Assignment Problem (CRAP) considers the assignment of the crew members to a set of reserve activities covering all the scheduled flights in order to ensure a continuous plan so that operations costs are minimized while its solution must meet hard constraints resulting from the safety regulations of Civil Aviation as well as from the airlines internal agreements. The problem considered in this study is of highest interest for airlines and may have important consequences on the service quality and on the economic return of the operations. In this communication, a new mathematical formulation for the CRAP is proposed which takes into account the regulations and the internal agreements. While current solutions make use of Artificial Intelligence techniques run on main frame computers, a low cost approach is proposed to provide on-line efficient solutions to face perturbed operating conditions. The proposed solution method uses a dynamic programming approach for the duties scheduling problem and when applied to the case of a medium airline while providing efficient solutions, shows good potential acceptability by the operations staff. This optimization scheme can then be considered as the core of an on-line Decision Support System for crew reserve assignment operations management.

Keywords: airlines operations management, combinatorial optimization, dynamic programming, crew scheduling

Procedia PDF Downloads 349
3085 A Comparative Analysis of Heuristics Applied to Collecting Used Lubricant Oils Generated in the City of Pereira, Colombia

Authors: Diana Fajardo, Sebastián Ortiz, Oscar Herrera, Angélica Santis

Abstract:

Currently, in Colombia is arising a problem related to collecting used lubricant oils which are generated by the increment of the vehicle fleet. This situation does not allow a proper disposal of this type of waste, which in turn results in a negative impact on the environment. Therefore, through the comparative analysis of various heuristics, the best solution to the VRP (Vehicle Routing Problem) was selected by comparing costs and times for the collection of used lubricant oils in the city of Pereira, Colombia; since there is no presence of management companies engaged in the direct administration of the collection of this pollutant. To achieve this aim, six proposals of through methods of solution of two phases were discussed. First, the assignment of the group of generator points of the residue was made (previously identified). Proposals one and four of through methods are based on the closeness of points. The proposals two and five are using the scanning method and the proposals three and six are considering the restriction of the capacity of collection vehicle. Subsequently, the routes were developed - in the first three proposals by the Clarke and Wright's savings algorithm and in the following proposals by the Traveling Salesman optimization mathematical model. After applying techniques, a comparative analysis of the results was performed and it was determined which of the proposals presented the most optimal values in terms of the distance, cost and travel time.

Keywords: Heuristics, optimization Model, savings algorithm, used vehicular oil, V.R.P.

Procedia PDF Downloads 410
3084 Optimization the Multiplicity of Infection for Large Produce of Lytic Bacteriophage pAh6-C

Authors: Sang Guen Kim, Sib Sankar Giri, Jin Woo Jun, Saekil Yun, Hyoun Joong Kim, Sang Wha Kim, Jung Woo Kang, Se Jin Han, Se Chang Park

Abstract:

Emerging of the super bacteria, bacteriophages are considered to be as an alternative to antibiotics. As the demand of phage increased, economical and large production of phage is becoming one of the critical points. For the therapeutic use, what is important is to eradicate the pathogenic bacteria as fast as possible, so higher concentration of phages is generally needed for effective therapeutic function. On the contrary, for the maximum production, bacteria work as a phage producing factory. As a microbial cell factory, bacteria is needed to last longer producing the phages without eradication. Consequently, killing the bacteria fast has a negative effect on large production. In this study, Multiplicity of Infection (MOI) was manipulated based on initial bacterial inoculation and used phage pAh-6C which has therapeutic effect against Aeromonas hydrophila. 1, 5 and 10 percent of overnight bacterial culture was inoculated and each bacterial culture was co-cultured with the phage of which MOI of 0.01, 0.0001, and 0.000001 respectively. Simply changing the initial MOI as well as bacterial inoculation concentration has regulated the production quantity of the phage without any other changes to culture conditions. It is anticipated that this result can be used as a foundational data for mass production of lytic bacteriophages which can be used as the therapeutic bio-control agent.

Keywords: bacteriophage, multiplicity of infection, optimization, Aeromonas hydrophila

Procedia PDF Downloads 292
3083 Optimization of Fermentation Parameters for Bioethanol Production from Waste Glycerol by Microwave Induced Mutant Escherichia coli EC-MW (ATCC 11105)

Authors: Refal Hussain, Saifuddin M. Nomanbhay

Abstract:

Glycerol is a valuable raw material for the production of industrially useful metabolites. Among many promising applications for the use of glycerol is its bioconversion to high value-added compounds, such as bioethanol through microbial fermentation. Bioethanol is an important industrial chemical with emerging potential as a biofuel to replace vanishing fossil fuels. The yield of liquid fuel in this process was greatly influenced by various parameters viz, temperature, pH, glycerol concentration, organic concentration, and agitation speed were considered. The present study was undertaken to investigate optimum parameters for bioethanol production from raw glycerol by immobilized mutant Escherichia coli (E.coli) (ATCC11505) strain on chitosan cross linked glutaraldehyde optimized by Taguchi statistical method in shake flasks. The initial parameters were set each at four levels and the orthogonal array layout of L16 (45) conducted. The important controlling parameters for optimized the operational fermentation was temperature 38 °C, medium pH 6.5, initial glycerol concentration (250 g/l), and organic source concentration (5 g/l). Fermentation with optimized parameters was carried out in a custom fabricated shake flask. The predicted value of bioethanol production under optimized conditions was (118.13 g/l). Immobilized cells are mainly used for economic benefits of continuous production or repeated use in continuous as well as in batch mode.

Keywords: bioethanol, Escherichia coli, immobilization, optimization

Procedia PDF Downloads 641
3082 Clubhouse: A Minor Rebellion against the Algorithmic Tyranny of the Majority

Authors: Vahid Asadzadeh, Amin Ataee

Abstract:

Since the advent of social media, there has been a wave of optimism among researchers and civic activists about the influence of virtual networks on the democratization process, which has gradually waned. One of the lesser-known concerns is how to increase the possibility of hearing the voices of different minorities. According to the theory of media logic, the media, using their technological capabilities, act as a structure through which events and ideas are interpreted. Social media, through the use of the learning machine and the use of algorithms, has formed a kind of structure in which the voices of minorities and less popular topics are lost among the commotion of the trends. In fact, the recommended systems and algorithms used in social media are designed to help promote trends and make popular content more popular, and content that belongs to minorities is constantly marginalized. As social networks gradually play a more active role in politics, the possibility of freely participating in the reproduction and reinterpretation of structures in general and political structures in particular (as Laclau‎ and Mouffe had in mind‎) can be considered as criteria to democracy in action. The point is that the media logic of virtual networks is shaped by the rule and even the tyranny of the majority, and this logic does not make it possible to design a self-foundation and self-revolutionary model of democracy. In other words, today's social networks, though seemingly full of variety But they are governed by the logic of homogeneity, and they do not have the possibility of multiplicity as is the case in immanent radical democracies (influenced by Gilles Deleuze). However, with the emergence and increasing popularity of Clubhouse as a new social media, there seems to be a shift in the social media space, and that is the diminishing role of algorithms and systems reconditioners as content delivery interfaces. This has led to the fact that in the Clubhouse, the voices of minorities are better heard, and the diversity of political tendencies manifests itself better. The purpose of this article is to show, first, how social networks serve the elimination of minorities in general, and second, to argue that the media logic of social networks must adapt to new interpretations of democracy that give more space to minorities and human rights. Finally, this article will show how the Clubhouse serves the new interpretations of democracy at least in a minimal way. To achieve the mentioned goals, in this article by a descriptive-analytical method, first, the relation between media logic and postmodern democracy will be inquired. The political economy popularity in social media and its conflict with democracy will be discussed. Finally, it will be explored how the Clubhouse provides a new horizon for the concepts embodied in radical democracy, a horizon that more effectively serves the rights of minorities and human rights in general.

Keywords: algorithmic tyranny, Clubhouse, minority rights, radical democracy, social media

Procedia PDF Downloads 138
3081 Optimization of Pregelatinized Taro Boloso-I Starch as a Direct Compression Tablet Excipient

Authors: Tamrat Balcha Balla

Abstract:

Background: Tablets are still the most preferred means of drug delivery. The search for new and improved direct compression tablet excipients is an area of research focus. Taro Boloso-I is a variety of Colocasia esculenta (L. Schott) yielding 67% more than the other varieties (Godare) in Ethiopia. This study aimed to enhance the flowability while keeping the compressibility and compactibility of the pregelatinized Taro Boloso-I starch. Methods: Central composite design was used for the optimization of two factors which were the temperature and duration of pregelatinization against 5 responses. The responses were angle of repose, Hausner ratio, Kawakita compressibility index, mean yield pressure and tablet breaking force. Results and Discussions: An increase in both temperature and time resulted in decrease in the angle of repose. The increase in temperature was shown to decrease the Hausner ratio and to decrease the Kawakita compressibility index. The mean yield pressure was observed to increase with increasing levels of both temperature and time. The pregelatinized (optimized) Taro Boloso-I starch could show desired flow property and compressibility. Conclusions: Pregelatinized Taro Boloso - I starch could be regarded as a potential direct compression excipient in terms of flowability, compressibility and compactibility.

Keywords: starch, compression, pregelatinization, Taro Boloso-I

Procedia PDF Downloads 100