Search results for: Conic optimization
427 A Quality Optimization Approach: An Application on Next Generation Networks
Authors: Gülfem I. Alptekin, S. Emre Alptekin
Abstract:
The next generation wireless systems, especially the cognitive radio networks aim at utilizing network resources more efficiently. They share a wide range of available spectrum in an opportunistic manner. In this paper, we propose a quality management model for short-term sub-lease of unutilized spectrum bands to different service providers. We built our model on competitive secondary market architecture. To establish the necessary conditions for convergent behavior, we utilize techniques from game theory. Our proposed model is based on potential game approach that is suitable for systems with dynamic decision making. The Nash equilibrium point tells the spectrum holders the ideal price values where profit is maximized at the highest level of customer satisfaction. Our numerical results show that the price decisions of the network providers depend on the price and QoS of their own bands as well as the prices and QoS levels of their opponents- bands.Keywords: cognitive radio networks, game theory, nextgeneration wireless networks, spectrum management.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1512426 Determining Optimal Production Plan by Revised Surrogate Worth Trade-off Method
Authors: Tunjo Peric, Zoran Babic
Abstract:
The authors of this work indicate by means of a concrete example that it is possible to apply efficaciously the method of multiple criteria programming in dealing with the problem of determining the optimal production plan for a certain period of time. The work presents: (1) the selection of optimization criteria, (2) the setting of the problem of determining an optimal production plan, (3) the setting of the model of multiple criteria programming in finding a solution to a given problem, (4) the revised surrogate trade-off method, (5) generalized multicriteria model for solving production planning problem and problem of choosing technological variants in the metal manufacturing industry. In the final part of this work the authors reflect on the application of the method of multiple criteria programming while determining the optimal production plan in manufacturing enterprises.
Keywords: multi-criteria programming, production planning, technological variant, Surrogate Worth Trade-off Method.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1853425 Optimization of Multi-Zone Unconventional (Shale) Gas Reservoir Using Hydraulic Fracturing Technique
Authors: F.C. Amadi, G. C. Enyi, G. G. Nasr
Abstract:
Hydraulic fracturing is one of the most important stimulation techniques available to the petroleum engineer to extract hydrocarbons in tight gas sandstones. It allows more oil and gas production in tight reservoirs as compared to conventional means. The main aim of the study is to optimize the hydraulic fracturing as technique and for this purpose three multi-zones layer formation is considered and fractured contemporaneously. The three zones are named as Zone1 (upper zone), Zone2 (middle zone) and Zone3 (lower zone) respectively and they all occur in shale rock. Simulation was performed with Mfrac integrated software which gives a variety of 3D fracture options. This simulation process yielded an average fracture efficiency of 93.8%for the three respective zones and an increase of the average permeability of the rock system. An average fracture length of 909 ft with net height (propped height) of 210 ft (average) was achieved. Optimum fracturing results was also achieved with maximum fracture width of 0.379 inches at an injection rate of 13.01 bpm with 17995 Mscf of gas production.Keywords: Hydraulic fracturing, Mfrac, Optimisation, Tight reservoir.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1959424 Quantum Computing: A New Era of Computing
Authors: Jyoti Chaturvedi Gursaran
Abstract:
Nature conducts its action in a very private manner. To reveal these actions classical science has done a great effort. But classical science can experiment only with the things that can be seen with eyes. Beyond the scope of classical science quantum science works very well. It is based on some postulates like qubit, superposition of two states, entanglement, measurement and evolution of states that are briefly described in the present paper. One of the applications of quantum computing i.e. implementation of a novel quantum evolutionary algorithm(QEA) to automate the time tabling problem of Dayalbagh Educational Institute (Deemed University) is also presented in this paper. Making a good timetable is a scheduling problem. It is NP-hard, multi-constrained, complex and a combinatorial optimization problem. The solution of this problem cannot be obtained in polynomial time. The QEA uses genetic operators on the Q-bit as well as updating operator of quantum gate which is introduced as a variation operator to converge toward better solutions.
Keywords: Quantum computing, qubit, superposition, entanglement, measurement of states, evolution of states, Scheduling problem, hard and soft constraints, evolutionary algorithm, quantum evolutionary algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2662423 Sparse-View CT Reconstruction Based on Nonconvex L1 − L2 Regularizations
Authors: Ali Pour Yazdanpanah, Farideh Foroozandeh Shahraki, Emma Regentova
Abstract:
The reconstruction from sparse-view projections is one of important problems in computed tomography (CT) limited by the availability or feasibility of obtaining of a large number of projections. Traditionally, convex regularizers have been exploited to improve the reconstruction quality in sparse-view CT, and the convex constraint in those problems leads to an easy optimization process. However, convex regularizers often result in a biased approximation and inaccurate reconstruction in CT problems. Here, we present a nonconvex, Lipschitz continuous and non-smooth regularization model. The CT reconstruction is formulated as a nonconvex constrained L1 − L2 minimization problem and solved through a difference of convex algorithm and alternating direction of multiplier method which generates a better result than L0 or L1 regularizers in the CT reconstruction. We compare our method with previously reported high performance methods which use convex regularizers such as TV, wavelet, curvelet, and curvelet+TV (CTV) on the test phantom images. The results show that there are benefits in using the nonconvex regularizer in the sparse-view CT reconstruction.Keywords: Computed tomography, sparse-view reconstruction, L1 −L2 minimization, non-convex, difference of convex functions.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2031422 Mechanical Structure Design Optimization by Blind Number Theory: Time-dependent Reliability
Authors: Zakari Yaou, Lirong Cui
Abstract:
In a product development process, understanding the functional behavior of the system, the role of components in achieving functions and failure modes if components/subsystem fails its required function will help develop appropriate design validation and verification program for reliability assessment. The integration of these three issues will help design and reliability engineers in identifying weak spots in design and planning future actions and testing program. This case study demonstrate the advantage of unascertained theory described in the subjective cognition uncertainty, and then applies blind number (BN) theory in describing the uncertainty of the mechanical system failure process and the same time used the same theory in bringing out another mechanical reliability system model. The practical calculations shows the BN Model embodied the characters of simply, small account of calculation but betterforecasting capability, which had the value of macroscopic discussion to some extent.
Keywords: Mechanical structure Design, time-dependent stochastic process, unascertained information, blind number theory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1471421 Optimization of Growth Conditions for Acidic Protease Production from Rhizopus oligosporus through Solid State Fermentation of Sunflower Meal
Authors: Abdul Rauf, Muhammad Irfan, Muhammad Nadeem, Ishtiaq Ahmed, Hafiz Muhammad Nasir Iqbal
Abstract:
Rhizopus oligosporus was used in the present study for the production of protease enzyme under SSF. Sunflower meal was used as by-product of oil industry incorporated with organic salts was employed for the production of protease enzyme. The main purpose of the present was to study different parameters of protease productivity, its yields and to optimize basal fermentation conditions. The optimal conditions found for protease production using sunflower meal as a substrate in the present study were inoculum size (1%), substrate (Sunflower meal), substrate concentration (20 g), pH (3), cultivation period (72 h), incubation temperature (35oC), substrate to diluent-s ratio (1:2) and tween 81 (1 mL). The maximum production of protease in the presence of cheaper substrate at low concentration and stability at acidic pH, these characteristics make the strain and its enzymes useful in different industry.Keywords: Acidic protease, Rhizopus oligosporus, Mediaoptimization, Solid state Fermentation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3002420 Inferential Reasoning for Heterogeneous Multi-Agent Mission
Authors: Sagir M. Yusuf, Chris Baber
Abstract:
We describe issues bedeviling the coordination of heterogeneous (different sensors carrying agents) multi-agent missions such as belief conflict, situation reasoning, etc. We applied Bayesian and agents' presumptions inferential reasoning to solve the outlined issues with the heterogeneous multi-agent belief variation and situational-base reasoning. Bayesian Belief Network (BBN) was used in modeling the agents' belief conflict due to sensor variations. Simulation experiments were designed, and cases from agents’ missions were used in training the BBN using gradient descent and expectation-maximization algorithms. The output network is a well-trained BBN for making inferences for both agents and human experts. We claim that the Bayesian learning algorithm prediction capacity improves by the number of training data and argue that it enhances multi-agents robustness and solve agents’ sensor conflicts.Keywords: Distributed constraint optimization problem, multi-agent system, multi-robot coordination, autonomous system, swarm intelligence.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 640419 A New Controlling Parameter in Design of Above Knee Prosthesis
Abstract:
In this paper after reviewing some previous studies, in order to optimize the above knee prosthesis, beside the inertial properties a new controlling parameter is informed. This controlling parameter makes the prosthesis able to act as a multi behavior system when the amputee is opposing to different environments. This active prosthesis with the new controlling parameter can simplify the control of prosthesis and reduce the rate of energy consumption in comparison to recently presented similar prosthesis “Agonistantagonist active knee prosthesis". In this paper three models are generated, a passive, an active, and an optimized active prosthesis. Second order Taylor series is the numerical method in solution of the models equations and the optimization procedure is genetic algorithm. Modeling the prosthesis which comprises this new controlling parameter (SEP) during the swing phase represents acceptable results in comparison to natural behavior of shank. Reported results in this paper represent 3.3 degrees as the maximum deviation of models shank angle from the natural pattern. The natural gait pattern belongs to walking at the speed of 81 m/min.Keywords: Above knee prosthesis, active controlling parameter, ballistic motion, swing phase.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1872418 Enhancement Throughput of Unplanned Wireless Mesh Networks Deployment Using Partitioning Hierarchical Cluster (PHC)
Authors: Ahmed K. Hasan, A. A. Zaidan, Anas Majeed, B. B. Zaidan, Rosli Salleh, Omar Zakaria, Ali Zuheir
Abstract:
Wireless mesh networks based on IEEE 802.11 technology are a scalable and efficient solution for next generation wireless networking to provide wide-area wideband internet access to a significant number of users. The deployment of these wireless mesh networks may be within different authorities and without any planning, they are potentially overlapped partially or completely in the same service area. The aim of the proposed model is design a new model to Enhancement Throughput of Unplanned Wireless Mesh Networks Deployment Using Partitioning Hierarchical Cluster (PHC), the unplanned deployment of WMNs are determinates there performance. We use throughput optimization approach to model the unplanned WMNs deployment problem based on partitioning hierarchical cluster (PHC) based architecture, in this paper the researcher used bridge node by allowing interworking traffic between these WMNs as solution for performance degradation.Keywords: Wireless Mesh Networks, 802.11s Internetworking, partitioning Hierarchical Cluste.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1533417 Simulation of the Reactive Rotational Molding Using Smoothed Particle Hydrodynamics
Authors: A. Hamidi, S. Khelladi, L. Illoul, A. Tcharkhtchi
Abstract:
Reactive rotational molding (RRM) is a process to manufacture hollow plastic parts with reactive material has several advantages compared to conventional roto molding of thermoplastic powders: process cycle time is shorter; raw material is less expensive because polymerization occurs during processing and high-performance polymers may be used such as thermosets, thermoplastics or blends. However, several phenomena occur during this process which makes the optimization of the process quite complex. In this study, we have used a mixture of isocyanate and polyol as a reactive system. The chemical transformation of this system to polyurethane has been studied by thermal analysis and rheology tests. Thanks to these results of the curing process and rheological measurements, the kinetic and rheokinetik of polyurethane was identified. Smoothed Particle Hydrodynamics, a Lagrangian meshless method, was chosen to simulate reactive fluid flow in 2 and 3D configurations of the polyurethane during the process taking into account the chemical, and chemiorehological results obtained experimentally in this study.Keywords: Reactive rotational molding, free surface flows, simulation, smoothed particle hydrodynamics, surface tension.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1077416 Deep Reinforcement Learning for Optimal Decision-making in Supply Chains
Authors: Nitin Singh, Meng Ling, Talha Ahmed, Tianxia Zhao, Reinier van de Pol
Abstract:
We propose the use of Reinforcement Learning (RL) as a viable alternative for optimizing supply chain management, particularly in scenarios with stochasticity in product demands. RL’s adaptability to changing conditions and its demonstrated success in diverse fields of sequential decision-making make it a promising candidate for addressing supply chain problems. We investigate the impact of demand fluctuations in a multi-product supply chain system and develop RL agents with learned generalizable policies. We provide experimentation details for training RL agents and a statistical analysis of the results. We study generalization ability of RL agents for different demand uncertainty scenarios and observe superior performance compared to the agents trained with fixed demand curves. The proposed methodology has the potential to lead to cost reduction and increased profit for companies dealing with frequent inventory movement between supply and demand nodes.
Keywords: Inventory Management, Reinforcement Learning, Supply Chain Optimization, Uncertainty.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 383415 Real Time Compensation of Machining Errors for Machine Tools NC based on Systematic Dispersion
Authors: M. Rahou, A. Cheikh, F. Sebaa
Abstract:
Manufacturing tolerancing is intended to determine the intermediate geometrical and dimensional states of the part during its manufacturing process. These manufacturing dimensions also serve to satisfy not only the functional requirements given in the definition drawing, but also the manufacturing constraints, for example geometrical defects of the machine, vibration and the wear of the cutting tool. In this paper, an experimental study on the influence of the wear of the cutting tool (systematic dispersions) is explored. This study was carried out on three stages .The first stage allows machining without elimination of dispersions (random, systematic) so the tolerances of manufacture according to total dispersions. In the second stage, the results of the first stage are filtered in such way to obtain the tolerances according to random dispersions. Finally, from the two previous stages, the systematic dispersions are generated. The objective of this study is to model by the least squares method the error of manufacture based on systematic dispersion. Finally, an approach of optimization of the manufacturing tolerances was developed for machining on a CNC machine toolKeywords: Dispersions, Compensation, modeling, manufacturing Tolerance, machine tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2336414 MIMO System Order Reduction Using Real-Coded Genetic Algorithm
Authors: Swadhin Ku. Mishra, Sidhartha Panda, Simanchala Padhy, C. Ardil
Abstract:
In this paper, real-coded genetic algorithm (RCGA) optimization technique has been applied for large-scale linear dynamic multi-input-multi-output (MIMO) system. The method is based on error minimization technique where the integral square error between the transient responses of original and reduced order models has been minimized by RCGA. The reduction procedure is simple computer oriented and the approach is comparable in quality with the other well-known reduction techniques. Also, the proposed method guarantees stability of the reduced model if the original high-order MIMO system is stable. The proposed approach of MIMO system order reduction is illustrated with the help of an example and the results are compared with the recently published other well-known reduction techniques to show its superiority.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2262413 Optimization of Carbon Nanotube Content of Asphalt Nanocomposites with Regard to Resistance to Permanent Deformation
Authors: João V. Staub de Melo, Glicério Trichês, Liseane P. Thives
Abstract:
This paper presents the results of the development of asphalt nanocomposites containing carbon nanotubes (CNTs) with high resistance to permanent deformation, aiming to increase the performance of asphalt surfaces in relation to the rutting problem. Asphalt nanocomposites were prepared with the addition of different proportions of CNTs (1%, 2% and 3%) in relation to the weight of asphalt binder. The base binder used was a conventional binder (50-70 penetration) classified as PG 58-22. The optimum percentage of CNT addition in the asphalt binder (base) was determined through the evaluation of the rheological and empirical characteristics of the nanocomposites produced. In order to evaluate the contribution and the effects of the nanocomposite (optimized) in relation to the rutting, the conventional and nanomodified asphalt mixtures were tested in a French traffic simulator (Orniéreur). The results obtained demonstrate the efficient contribution of the asphalt nanocomposite containing CNTs to the resistance to permanent deformation of the asphalt mixture.
Keywords: Asphalt nanocomposites, asphalt mixtures, carbon nanotubes, nanotechnology, permanent deformation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1489412 Evolution of Fuzzy Neural Networks Using an Evolution Strategy with Fuzzy Genotype Values
Authors: Hidehiko Okada
Abstract:
Evolution strategy (ES) is a well-known instance of evolutionary algorithms, and there have been many studies on ES. In this paper, the author proposes an extended ES for solving fuzzy-valued optimization problems. In the proposed ES, genotype values are not real numbers but fuzzy numbers. Evolutionary processes in the ES are extended so that it can handle genotype instances with fuzzy numbers. In this study, the proposed method is experimentally applied to the evolution of neural networks with fuzzy weights and biases. Results reveal that fuzzy neural networks evolved using the proposed ES with fuzzy genotype values can model hidden target fuzzy functions even though no training data are explicitly provided. Next, the proposed method is evaluated in terms of variations in specifying fuzzy numbers as genotype values. One of the mostly adopted fuzzy numbers is a symmetric triangular one that can be specified by its lower and upper bounds (LU) or its center and width (CW). Experimental results revealed that the LU model contributed better to the fuzzy ES than the CW model, which indicates that the LU model should be adopted in future applications of the proposed method.
Keywords: Evolutionary algorithm, evolution strategy, fuzzy number, feedforward neural network, neuroevolution.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1545411 Improving TNT Curing Process by Using Infrared Camera
Authors: O. Srihakulung, Y. Soongsumal
Abstract:
Among the chemicals used for ammunition production, TNT (Trinitrotoluene) play a significant role since World War I and II. Various types of military weapon utilize TNT in casting process. However, the TNT casting process for warhead is difficult to control the cooling rate of the liquid TNT. This problem occurs because the casting process lacks the equipment to detect the temperature during the casting procedure This study presents the temperature detected by infrared camera to illustrate the cooling rate and cooling zone of curing, and demonstrates the optimization of TNT condition to reduce the risk of air gap occurred in the warhead which can result in the destruction afterward. Premature initiation of explosive-filled projectiles in response to set-back forces during gunfiring cause by casting defects. Finally the study can help improving the process of the TNT casting. The operators can control the curing of TNT inside the case by rising up the heating rod at the proper time. Consequently this can reduce tremendous time of rework if the air gaps occur and increase strength to lower elastic modulus. Therefore, it can be clearly concluded that the use of Infrared Cameras in this process is another method to improve the casting procedure.
Keywords: Infrared camera, TNT casting, warhead, curing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2265410 Properties of Composite Nanofiber Produced by Single and Coaxial Nozzle Method used for Electrospinning Technique
Authors: Onur Ayaz, Nuray Ucar, Elif Bahar, Oguzhan Ucar, Mustafa Oksuz, Aysen Onen, Mehmet Ucar, Ezgi İşmar, Ali Demir
Abstract:
In this study, single nozzle method used for electrospinning technique which composite polymer solution with cellulose nanowiskers (CNW) was treated by ultrasonic sonificator have been compared with coaxial (double) nozzle method, in terms of mechanical, thermal and morphological properties of composite nanofiber. The effect of water content in composite polymer solution on properties of nanofiber has also been examined. It has been seen that single nozzle method which polymer solution does not contain water has better results than that of coaxial method, in terms of mechanical, thermal and morphological properties of nanofiber. However, it is necessary to make an optimization study on setting condition of ultrasonic treatment to get better dispersion of CNW in composite nanofiber and to get better mechanical and thermal propertiesKeywords: cellulose nanowhiskers, coaxial nozzle, composite nanofiber, electrospinning
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2088409 Analyses and Optimization of Physical and Mechanical Properties of Direct Recycled Aluminium Alloy (AA6061) Wastes by ANOVA Approach
Authors: Mohammed H. Rady, Mohd Sukri Mustapa, S Shamsudin, M. A. Lajis, A. Wagiman
Abstract:
The present study is aimed at investigating microhardness and density of aluminium alloy chips when subjected to various settings of preheating temperature and preheating time. Three values of preheating temperature were taken as 450 °C, 500 °C, and 550 °C. On the other hand, three values of preheating time were chosen (1, 2, 3) hours. The influences of the process parameters (preheating temperature and time) were analyzed using Design of Experiments (DOE) approach whereby full factorial design with center point analysis was adopted. The total runs were 11 and they comprise of two factors of full factorial design with 3 center points. The responses were microhardness and density. The results showed that the density and microhardness increased with decreasing the preheating temperature. The results also found that the preheating temperature is more important to be controlled rather than the preheating time in microhardness analysis while both the preheating temperature and preheating time are important in density analysis. It can be concluded that setting temperature at 450 °C for 1 hour resulted in the optimum responses.
Keywords: AA6061, density, DOE, hot extrusion, microhardness.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 703408 Numerical Optimization of Pin-Fin Heat Sink with Forced Cooling
Authors: Y. T. Yang, H. S. Peng, H. T. Hsu
Abstract:
This study presents the numerical simulation of optimum pin-fin heat sink with air impinging cooling by using Taguchi method. 9 L ( 4 3 ) orthogonal array is selected as a plan for the four design-parameters with three levels. The governing equations are discretized by using the control-volume-based-finite-difference method with a power-law scheme on the non-uniform staggered grid. We solved the coupling of the velocity and the pressure terms of momentum equations using SIMPLEC algorithm. We employ the k −ε two-equations turbulence model to describe the turbulent behavior. The parameters studied include fin height H (35mm-45mm), inter-fin spacing a , b , and c (2 mm-6.4 mm), and Reynolds number ( Re = 10000- 25000). The objective of this study is to examine the effects of the fin spacings and fin height on the thermal resistance and to find the optimum group by using the Taguchi method. We found that the fin spacings from the center to the edge of the heat sink gradually extended, and the longer the fin’s height the better the results. The optimum group is 3 1 2 3 H a b c . In addition, the effects of parameters are ranked by importance as a , H , c , and b .
Keywords: Heat sink, Optimum, Electronics cooling, CFD.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3705407 CPU Architecture Based on Static Hardware Scheduler Engine and Multiple Pipeline Registers
Authors: Ionel Zagan, Vasile Gheorghita Gaitan
Abstract:
The development of CPUs and of real-time systems based on them made it possible to use time at increasingly low resolutions. Together with the scheduling methods and algorithms, time organizing has been improved so as to respond positively to the need for optimization and to the way in which the CPU is used. This presentation contains both a detailed theoretical description and the results obtained from research on improving the performances of the nMPRA (Multi Pipeline Register Architecture) processor by implementing specific functions in hardware. The proposed CPU architecture has been developed, simulated and validated by using the FPGA Virtex-7 circuit, via a SoC project. Although the nMPRA processor hardware structure with five pipeline stages is very complex, the present paper presents and analyzes the tests dedicated to the implementation of the CPU and of the memory on-chip for instructions and data. In order to practically implement and test the entire SoC project, various tests have been performed. These tests have been performed in order to verify the drivers for peripherals and the boot module named Bootloader.
Keywords: Hardware scheduler, nMPRA processor, real-time systems, scheduling methods.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1096406 Reinforced Concrete Slab under Static and Dynamic Loadings
Authors: Aaron Aboshio, Jianqioa Ye
Abstract:
In this study, static and dynamic responses of a typical reinforced concrete solid slab, designed to British Standard (BS 8110: 1997) and under self and live loadings for dance halls are reported. Linear perturbation analysis using finite element method was employed for modal, impulse loading and frequency response analyses of the slab under the aforementioned loading condition. Results from the static and dynamic analyses, comprising of the slab fundamental frequencies and mode shapes, dynamic amplification factor, maximum deflection, stress distributions among other valuable outcomes are presented and discussed. These were gauged with the limiting provisions in the design code with a view of justifying valid optimization objective function for the structure that can ensure both adequate strength and economical section for large clear span slabs. This is necessary owing to the continued increase in cost of erecting building structures and the squeeze on public finance globally.Keywords: Economical design, Finite element method, Modal dynamics, Reinforced concrete, Slab.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4617405 Community Detection-based Analysis of the Human Interactome Network
Authors: Razvan Bocu, Sabin Tabirca
Abstract:
The study of proteomics reached unexpected levels of interest, as a direct consequence of its discovered influence over some complex biological phenomena, such as problematic diseases like cancer. This paper presents a new technique that allows for an accurate analysis of the human interactome network. It is basically a two-step analysis process that involves, at first, the detection of each protein-s absolute importance through the betweenness centrality computation. Then, the second step determines the functionallyrelated communities of proteins. For this purpose, we use a community detection technique that is based on the edge betweenness calculation. The new technique was thoroughly tested on real biological data and the results prove some interesting properties of those proteins that are involved in the carcinogenesis process. Apart from its experimental usefulness, the novel technique is also computationally effective in terms of execution times. Based on the analysis- results, some topological features of cancer mutated proteins are presented and a possible optimization solution for cancer drugs design is suggested.Keywords: Betweenness centrality, interactome networks, proteinprotein interactions, protein communities, cancer.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1293404 System Security Impact on the Dynamic Characteristics of Measurement Sensors in Smart Grids
Authors: Yiyang Su, Jörg Neumann, Jan Wetzlich, Florian Thiel
Abstract:
Smart grid is a term used to describe the next generation power grid. New challenges such as integration of renewable and decentralized energy sources, the requirement for continuous grid estimation and optimization, as well as the use of two-way flows of energy have been brought to the power gird. In order to achieve efficient, reliable, sustainable, as well as secure delivery of electric power more and more information and communication technologies are used for the monitoring and the control of power grids. Consequently, the need for cybersecurity is dramatically increased and has converged into several standards which will be presented here. These standards for the smart grid must be designed to satisfy both performance and reliability requirements. An in depth investigation of the effect of retrospectively embedded security in existing grids on it’s dynamic behavior is required. Therefore, a retrofitting plan for existing meters is offered, and it’s performance in a test low voltage microgrid is investigated. As a result of this, integration of security measures into measurement architectures of smart grids at the design phase is strongly recommended.Keywords: Cyber security, performance, protocols, security standards, smart grid.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 884403 Self-Organizing Map Network for Wheeled Robot Movement Optimization
Authors: Boguslaw Schreyer
Abstract:
The paper investigates the application of the Kohonen’s Self-Organizing Map (SOM) to the wheeled robot starting and braking dynamic states. In securing wheeled robot stability as well as minimum starting and braking time, it is important to ensure correct torque distribution as well as proper slope of braking and driving moments. In this paper, a correct movement distribution has been formulated, securing optimum adhesion coefficient and good transversal stability of a wheeled robot. A neural tuner has been proposed to secure the above properties, although most of the attention is attached to the SOM network application. If the delay of the torque application or torque release is not negligible, it is important to change the rising and falling slopes of the torque. The road/surface condition is also paramount in robot dynamic states control. As the road conditions may randomly change in time, application of the SOM network has been suggested in order to classify the actual road conditions.
Keywords: SOM network, torque distribution, torque slope, wheeled robots.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 590402 Feature Vector Fusion for Image Based Human Age Estimation
Authors: D. Karthikeyan, G. Balakrishnan
Abstract:
Human faces, as important visual signals, express a significant amount of nonverbal info for usage in human-to-human communication. Age, specifically, is more significant among these properties. Human age estimation using facial image analysis as an automated method which has numerous potential real‐world applications. In this paper, an automated age estimation framework is presented. Support Vector Regression (SVR) strategy is utilized to investigate age prediction. This paper depicts a feature extraction taking into account Gray Level Co-occurrence Matrix (GLCM), which can be utilized for robust face recognition framework. It applies GLCM operation to remove the face's features images and Active Appearance Models (AAMs) to assess the human age based on image. A fused feature technique and SVR with GA optimization are proposed to lessen the error in age estimation.
Keywords: Support vector regression, feature extraction, gray level co-occurrence matrix, active appearance models.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1314401 A Location-Allocation-Routing Model for a Home Health Care Supply Chain Problem
Authors: Amir Mohammad Fathollahi Fard, Mostafa Hajiaghaei-Keshteli, Mohammad Mahdi Paydar
Abstract:
With increasing life expectancy in developed countries, the role of home care services is highlighted by both academia and industrial contributors in Home Health Care Supply Chain (HHCSC) companies. The main decisions in such supply chain systems are the location of pharmacies, the allocation of patients to these pharmacies and also the routing and scheduling decisions of nurses to visit their patients. In this study, for the first time, an integrated model is proposed to consist of all preliminary and necessary decisions in these companies, namely, location-allocation-routing model. This model is a type of NP-hard one. Therefore, an Imperialist Competitive Algorithm (ICA) is utilized to solve the model, especially in large sizes. Results confirm the efficiency of the developed model for HHCSC companies as well as the performance of employed ICA.
Keywords: Home health care supply chain, location-allocation-routing problem, imperialist competitive algorithm, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1055400 Control Improvement of a C Sugar Cane Crystallization Using an Auto-Tuning PID Controller Based on Linearization of a Neural Network
Authors: S. Beyou, B. Grondin-Perez, M. Benne, C. Damour, J.-P. Chabriat
Abstract:
The industrial process of the sugar cane crystallization produces a residual that still contains a lot of soluble sucrose and the objective of the factory is to improve its extraction. Therefore, there are substantial losses justifying the search for the optimization of the process. Crystallization process studied on the industrial site is based on the “three massecuites process". The third step of this process constitutes the final stage of exhaustion of the sucrose dissolved in the mother liquor. During the process of the third step of crystallization (Ccrystallization), the phase that is studied and whose control is to be improved, is the growing phase (crystal growth phase). The study of this process on the industrial site is a problem in its own. A control scheme is proposed to improve the standard PID control law used in the factory. An auto-tuning PID controller based on instantaneous linearization of a neural network is then proposed.
Keywords: Auto-tuning, PID, Instantaneous linearization, Neural network, Non linear process, C-crystallisation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1467399 A New Hybrid RMN Image Segmentation Algorithm
Authors: Abdelouahab Moussaoui, Nabila Ferahta, Victor Chen
Abstract:
The development of aid's systems for the medical diagnosis is not easy thing because of presence of inhomogeneities in the MRI, the variability of the data from a sequence to the other as well as of other different source distortions that accentuate this difficulty. A new automatic, contextual, adaptive and robust segmentation procedure by MRI brain tissue classification is described in this article. A first phase consists in estimating the density of probability of the data by the Parzen-Rozenblatt method. The classification procedure is completely automatic and doesn't make any assumptions nor on the clusters number nor on the prototypes of these clusters since these last are detected in an automatic manner by an operator of mathematical morphology called skeleton by influence zones detection (SKIZ). The problem of initialization of the prototypes as well as their number is transformed in an optimization problem; in more the procedure is adaptive since it takes in consideration the contextual information presents in every voxel by an adaptive and robust non parametric model by the Markov fields (MF). The number of bad classifications is reduced by the use of the criteria of MPM minimization (Maximum Posterior Marginal).Keywords: Clustering, Automatic Classification, SKIZ, MarkovFields, Image segmentation, Maximum Posterior Marginal (MPM).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1412398 Optimizing the Capacity of a Convolutional Neural Network for Image Segmentation and Pattern Recognition
Authors: Yalong Jiang, Zheru Chi
Abstract:
In this paper, we study the factors which determine the capacity of a Convolutional Neural Network (CNN) model and propose the ways to evaluate and adjust the capacity of a CNN model for best matching to a specific pattern recognition task. Firstly, a scheme is proposed to adjust the number of independent functional units within a CNN model to make it be better fitted to a task. Secondly, the number of independent functional units in the capsule network is adjusted to fit it to the training dataset. Thirdly, a method based on Bayesian GAN is proposed to enrich the variances in the current dataset to increase its complexity. Experimental results on the PASCAL VOC 2010 Person Part dataset and the MNIST dataset show that, in both conventional CNN models and capsule networks, the number of independent functional units is an important factor that determines the capacity of a network model. By adjusting the number of functional units, the capacity of a model can better match the complexity of a dataset.Keywords: CNN, capsule network, capacity optimization, character recognition, data augmentation; semantic segmentation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 701