Search results for: large opening
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2318

Search results for: large opening

2078 Predictive Semi-Empirical NOx Model for Diesel Engine

Authors: Saurabh Sharma, Yong Sun, Bruce Vernham

Abstract:

Accurate prediction of NOx emission is a continuous challenge in the field of diesel engine-out emission modeling. Performing experiments for each conditions and scenario cost significant amount of money and man hours, therefore model-based development strategy has been implemented in order to solve that issue. NOx formation is highly dependent on the burn gas temperature and the O2 concentration inside the cylinder. The current empirical models are developed by calibrating the parameters representing the engine operating conditions with respect to the measured NOx. This makes the prediction of purely empirical models limited to the region where it has been calibrated. An alternative solution to that is presented in this paper, which focus on the utilization of in-cylinder combustion parameters to form a predictive semi-empirical NOx model. The result of this work is shown by developing a fast and predictive NOx model by using the physical parameters and empirical correlation. The model is developed based on the steady state data collected at entire operating region of the engine and the predictive combustion model, which is developed in Gamma Technology (GT)-Power by using Direct Injected (DI)-Pulse combustion object. In this approach, temperature in both burned and unburnt zone is considered during the combustion period i.e. from Intake Valve Closing (IVC) to Exhaust Valve Opening (EVO). Also, the oxygen concentration consumed in burnt zone and trapped fuel mass is also considered while developing the reported model.  Several statistical methods are used to construct the model, including individual machine learning methods and ensemble machine learning methods. A detailed validation of the model on multiple diesel engines is reported in this work. Substantial numbers of cases are tested for different engine configurations over a large span of speed and load points. Different sweeps of operating conditions such as Exhaust Gas Recirculation (EGR), injection timing and Variable Valve Timing (VVT) are also considered for the validation. Model shows a very good predictability and robustness at both sea level and altitude condition with different ambient conditions. The various advantages such as high accuracy and robustness at different operating conditions, low computational time and lower number of data points requires for the calibration establishes the platform where the model-based approach can be used for the engine calibration and development process. Moreover, the focus of this work is towards establishing a framework for the future model development for other various targets such as soot, Combustion Noise Level (CNL), NO2/NOx ratio etc.

Keywords: Diesel engine, machine learning, NOx emission, semi-empirical.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 848
2077 Experimental Investigation of Plane Jets Exiting Five Parallel Channels with Large Aspect Ratio

Authors: Laurentiu Moruz, Jens Kitzhofer, Mircea Dinulescu

Abstract:

The paper aims to extend the knowledge about jet behavior and jet interaction between five plane unventilated jets with large aspect ratio (AR). The distance between the single plane jets is two times the channel height. The experimental investigation applies 2D Particle Image Velocimetry (PIV) and static pressure measurements. Our study focuses on the influence of two different outlet nozzle geometries (triangular shape with 2 x 7.5° and blunt geometry) with respect to variation of Reynolds number from 5500 - 12000. It is shown that the outlet geometry has a major influence on the jet formation in terms of uniformity of velocity profiles downstream of the sudden expansion. Furthermore, we describe characteristic regions like converging region, merging region and combined region. The triangular outlet geometry generates most uniform velocity distributions in comparison to a blunt outlet nozzle geometry. The blunt outlet geometry shows an unstable behavior where the jets tend to attach to one side of the walls (ceiling) generating a large recirculation region on the opposite side. Static pressure measurements confirm the observation and indicate that the recirculation region is connected to larger pressure drop.

Keywords: 2D particle image velocimetry, parallel jet interaction, pressure drop, sudden expansion.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 873
2076 Post-Cracking Behaviour of High Strength Fiber Concrete Prediction and Validation

Authors: Andrejs Krasnikovs, Olga Kononova, Amjad Khabbaz, Edgar Machanovsky, Artur Machanovsky

Abstract:

Fracture process in mechanically loaded steel fiber reinforced high-strength (SFRHSC) concrete is characterized by fibers bridging the crack providing resistance to its opening. Structural SFRHSC fracture model was created; material fracture process was modeled, based on single fiber pull-out laws, which were determined experimentally (for straight fibers, fibers with end hooks (Dramix), and corrugated fibers (Tabix)) as well as obtained numerically ( using FEM simulations). For this purpose experimental program was realized and pull-out force versus pull-out fiber length was obtained (for fibers embedded into concrete at different depth and under different angle). Model predictions were validated by 15x15x60cm prisms 4 point bending tests. Fracture surfaces analysis was realized for broken prisms with the goal to improve elaborated model assumptions. Optimal SFRHSC structures were recognized.

Keywords: crack, fiber concrete, fiber pull-out, strength.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2093
2075 Inductance Characteristic of Annealed Titanium Dioxide on Silicon Substrate

Authors: Chih Chin Yang, Lan Hui Huang, Bo Shum Chen, Jia Liang Ke, Chung Lun Tsai

Abstract:

The control of oxygen flow rate during growth of titanium dioxide by mass flow controller in DC plasma sputtering growth system is studied. The impedance of TiO2 films for inductance effect is influenced by annealing time and oxygen flow rate. As annealing time is increased, the inductance of TiO2 film is the more. The growth condition of optimum and maximum inductance for TiO2 film to serve as sensing device are oxygen flow rate of 15 sccm and large annealing time. The large inductance of TiO2 film will be adopted to fabricate the biosensor to obtain the high sensitivity of sensing in biology.

Keywords: Annealed, Inductance, Silicon substarte, Titanium dioxide

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1959
2074 A Simple Knowledge Management Strategy Model for SMEs in Developing Countries

Authors: Kitimaporn Choochote, Roger Nurse

Abstract:

The area of knowledge management has been in the highlight for enterprises over the past three decades. Many enterprises would like to have knowledge management and work hard to achieve it, however they are often confused about which direction to take to be successful and this point is especially true for Small and Medium Enterprises (SMEs) in developing countries. Many large companies have realized that knowledge is one of the richest resources which an organization possesses and knowledge management is a part of the foundation for a sustainable competitive advantage. Much work has been done in the area of knowledge management, but most of it has served large enterprises. This research provides a Model of knowledge management strategy for SMEs. It is based on analysis, insights and recommendations and it is presented so that SMEs in developing countries can easily understand and implement this model.

Keywords: Competitive advantage, Developing Country, Knowledge Management Strategy, Small and Medium Enterprise

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2097
2073 Real-Time 3D City Generation using Shape Grammars with LOD Variations

Authors: Pearl Goswell, Jun Jo

Abstract:

Creating3D environments, including characters and cities, is a significantly time consuming process due to a large amount of workinvolved in designing and modelling.There have been a number of attempts to automatically generate 3D objects employing shape grammars. However it is still too early to apply the mechanism to real problems such as real-time computer games.The purpose of this research is to introduce a time efficient and cost effective method to automatically generatevarious 3D objects for real-time 3D games. This Shape grammar-based real-time City Generation (RCG) model is a conceptual model for generating 3Denvironments in real-time and can be applied to 3D gamesoranimations. The RCG system can generate even a large cityby applying fundamental principles of shape grammars to building elementsin various levels of detailin real-time.

Keywords: real-time city generation, shape grammars, 3D games, 3D modelling.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2319
2072 Evaluation of Hancornia speciosa Gomes Lyophilization at Different Stages of Maturation

Authors: D. C. Soares, J. T. S. Santos, D. G. Costa, A. K. S. Abud, T. P. Nunes, A. V. D. Figueiredo, A. M. de Oliveira Junior

Abstract:

Mangabeira (Hancornia speciosa Gomes), a native plant in Brazil, is found growing spontaneously in various regions of the country. The high perishability of tropical fruits such as mangaba, causes it to be necessary to use technologies that promote conservation, aiming to increase the shelf life of this fruit and add value. The objective of this study was to compare the mangabas lyophilization curves behaviors with different sizes and maturation stages. The fruits were freeze-dried for a period of approximately 45 hours at lyophilizer Liotop brand, model L -108. It has been considered large the fruits between 38 and 58 mm diameter and small, between 23 and 28 mm diameter and the two states of maturation, intermediate and mature. Large size mangabas drying curves in both states of maturation were linear behavior at all process, while the kinetic drying curves related to small fruits, independent of maturation state, had a typical behavior of drying, with all the well-defined steps. With these results it was noted that the time of lyophilization was suitable for small mangabas, a fact that did not happen with the larger one. This may indicate that the large mangabas require a longer time to freeze until reaches the equilibrium level, as it happens with the small fruits, going to have constant moisture at the end of the process. For both types of fruit were analyzed water activity, acidity, protein, lipid, and vitamin C before and after the process.

Keywords: Freeze dryer, mangaba, conservation, chemical characteristics.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2124
2071 Parallelization of Ensemble Kalman Filter (EnKF) for Oil Reservoirs with Time-lapse Seismic Data

Authors: Md Khairullah, Hai-Xiang Lin, Remus G. Hanea, Arnold W. Heemink

Abstract:

In this paper we describe the design and implementation of a parallel algorithm for data assimilation with ensemble Kalman filter (EnKF) for oil reservoir history matching problem. The use of large number of observations from time-lapse seismic leads to a large turnaround time for the analysis step, in addition to the time consuming simulations of the realizations. For efficient parallelization it is important to consider parallel computation at the analysis step. Our experiments show that parallelization of the analysis step in addition to the forecast step has good scalability, exploiting the same set of resources with some additional efforts.

Keywords: EnKF, Data assimilation, Parallel computing, Parallel efficiency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2277
2070 Application of Model Free Adaptive Control in Main Steam Temperature System of Thermal Power Plant

Authors: Khaing Yadana Swe, Lillie Dewan

Abstract:

At present, the cascade PID control is widely used to control the superheating temperature (main steam temperature). As Main Steam Temperature has the characteristics of large inertia, large time-delay and time varying, etc., conventional PID control strategy cannot achieve good control performance. In order to overcome the bad performance and deficiencies of main steam temperature control system, Model Free Adaptive Control (MFAC) - P cascade control system is proposed in this paper. By substituting MFAC in PID of the main control loop of the main steam temperature control, it can overcome time delays, non-linearity, disturbance and time variation.

Keywords: Model free Adaptive Control, Cascade Control, Adaptive Control, PID.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2793
2069 Production Planning and Scheduling and SME

Authors: M. Heck, H. Vettiger

Abstract:

Small and medium-sized enterprises (SME) are the backbone of central Europe’s economies and have a significant contribution to the gross domestic product. Production planning and scheduling (PPS) is still a crucial element in manufacturing industries of the 21st century even though this area of research is more than a century old. The topic of PPS is well researched especially in the context of large enterprises in the manufacturing industry. However the implementation of PPS methodologies within SME is mostly unobserved. This work analyzes how PPS is implemented in SME with the geographical focus on Switzerland and its vicinity. Based on restricted resources compared to large enterprises, SME have to face different challenges. The real problem areas of selected enterprises in regards of PPS are identified and evaluated. For the identified real-life problem areas of SME clear and detailed recommendations are created, covering concepts and best practices and the efficient usage of PPS. Furthermore the economic and entrepreneurial value for companies is lined out and why the implementation of the introduced recommendations is advised.

Keywords: Central Europe, PPS, Production Planning, SME.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2172
2068 Calibration of the Discrete Element Method Using a Large Shear Box

Authors: Corné J. Coetzee, Etienne Horn

Abstract:

One of the main challenges in using the Discrete Element Method (DEM) is to specify the correct input parameter values. In general, the models are sensitive to the input parameter values and accurate results can only be achieved if the correct values are specified. For the linear contact model, micro-parameters such as the particle density, stiffness, coefficient of friction, as well as the particle size and shape distributions are required. There is a need for a procedure to accurately calibrate these parameters before any attempt can be made to accurately model a complete bulk materials handling system. Since DEM is often used to model applications in the mining and quarrying industries, a calibration procedure was developed for materials that consist of relatively large (up to 40 mm in size) particles. A coarse crushed aggregate was used as the test material. Using a specially designed large shear box with a diameter of 590 mm, the confined Young’s modulus (bulk stiffness) and internal friction angle of the material were measured by means of the confined compression test and the direct shear test respectively. DEM models of the experimental setup were developed and the input parameter values were varied iteratively until a close correlation between the experimental and numerical results was achieved. The calibration process was validated by modelling the pull-out of an anchor from a bed of material. The model results compared well with experimental measurement.

Keywords: Discrete Element Method (DEM), calibration, shear box, anchor pull-out.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2666
2067 An Intelligent Approach of Rough Set in Knowledge Discovery Databases

Authors: Hrudaya Ku. Tripathy, B. K. Tripathy, Pradip K. Das

Abstract:

Knowledge Discovery in Databases (KDD) has evolved into an important and active area of research because of theoretical challenges and practical applications associated with the problem of discovering (or extracting) interesting and previously unknown knowledge from very large real-world databases. Rough Set Theory (RST) is a mathematical formalism for representing uncertainty that can be considered an extension of the classical set theory. It has been used in many different research areas, including those related to inductive machine learning and reduction of knowledge in knowledge-based systems. One important concept related to RST is that of a rough relation. In this paper we presented the current status of research on applying rough set theory to KDD, which will be helpful for handle the characteristics of real-world databases. The main aim is to show how rough set and rough set analysis can be effectively used to extract knowledge from large databases.

Keywords: Data mining, Data tables, Knowledge discovery in database (KDD), Rough sets.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2328
2066 Improving Convergence of Parameter Tuning Process of the Additive Fuzzy System by New Learning Strategy

Authors: Thi Nguyen, Lee Gordon-Brown, Jim Peterson, Peter Wheeler

Abstract:

An additive fuzzy system comprising m rules with n inputs and p outputs in each rule has at least t m(2n + 2 p + 1) parameters needing to be tuned. The system consists of a large number of if-then fuzzy rules and takes a long time to tune its parameters especially in the case of a large amount of training data samples. In this paper, a new learning strategy is investigated to cope with this obstacle. Parameters that tend toward constant values at the learning process are initially fixed and they are not tuned till the end of the learning time. Experiments based on applications of the additive fuzzy system in function approximation demonstrate that the proposed approach reduces the learning time and hence improves convergence speed considerably.

Keywords: Additive fuzzy system, improving convergence, parameter learning process, unsupervised learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1509
2065 Investigating the Impact of the Laundry and Sterilization Process on the Performance of Reusable Surgical Gowns

Authors: N. Khomarloo, F. Mousazadegan, M. Latifi, N. Hemmatinejad

Abstract:

Recently, the utilization of reusable surgical gowns in order to decrease costs, environmental protection and enhance surgeon’s comfort is considered. One of the concerns in applying this kind of medical protective clothing is reduction of their resistance to bacterial penetration especially in wet state, after repeated laundering and sterilizing process. The purpose of this study is to investigate the effect of the laundering and sterilizing process on the reusable surgical gown’s resistance against bacterial wet penetration. To this end, penetration of Staphylococcus aureus bacteria in wet state after 70 washing and sterilizing cycles was evaluated on the two single-layer and three-layer reusable gowns. The outcomes reveal that up to 20 laundering and sterilizing cycles, protective property of samples improves due to fabric shrinkage, after that because of the fabric’s construction opening, the bacterial penetration increase. However, the three-layer gown presents higher protective performance comparing to the single-layer one.

Keywords: Reusable surgical gown, laundry, sterilization, wet bacterial penetration.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1743
2064 Design of the Large Dimension Cold Shield Cooled by G-M Cryocooler

Authors: Gong Jie, Yu Qianxu, Liu Min, Shan Weiwei

Abstract:

The design of methods of the 20 K large dimension cold shield used for infrared radiation demarcating in space environment simulation test were introduced in this paper. The cold shield were cooled by five G-M cryocoolers , and the dimension of the cold shield is the largest in our country.Cold shield installation and distribution and compensator for contraction on cooling were introduced detailedly. The temperature distribution and cool-down time of cold shield surface were also calculated and analysed in this paper. The design of cold shield resolves the difficulty of compensator for contraction on cooling successfully. Test results show that the actual technical performance indicators of cold shield met and exceeded the design requirements.

Keywords: cold shield, G-M cryocooler,infrared radiometer demarcating, satellite, space environment simulation equipments

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1689
2063 Computational Fluid Dynamics Simulation and Comparison of Flow through Mechanical Heart Valve Using Newtonian and Non-Newtonian Fluid

Authors: D. Šedivý, S. Fialová

Abstract:

The main purpose of this study is to show differences between the numerical solution of the flow through the artificial heart valve using Newtonian or non-Newtonian fluid. The simulation was carried out by a commercial computational fluid dynamics (CFD) package based on finite-volume method. An aortic bileaflet heart valve (Sorin Bicarbon) was used as a pattern for model of real heart valve replacement. Computed tomography (CT) was used to gain the accurate parameters of the valve. Data from CT were transferred in the commercial 3D designer, where the model for CFD was made. Carreau rheology model was applied as non-Newtonian fluid. Physiological data of cardiac cycle were used as boundary conditions. Outputs were taken the leaflets excursion from opening to closure and the fluid dynamics through the valve. This study also includes experimental measurement of pressure fields in ambience of valve for verification numerical outputs. Results put in evidence a favorable comparison between the computational solutions of flow through the mechanical heart valve using Newtonian and non-Newtonian fluid.

Keywords: Computational modeling, dynamic mesh, mechanical heart valve, non-Newtonian fluid, SDOF.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1614
2062 Research on the Protection and Reuse Model of Historical Buildings in Chinese Airports

Authors: Jie Ouyang, Chen Nie

Abstract:

China had constructed a large number of military and civilian airports before and after World War II, and then began large-scale repairs, reconstructions or relocation of airports after the baptism of wars after World War I and World War II. The airport's historical area and its historical buildings such as terminals, hangars, and towers have adopted different protection strategies and reuse application strategies. This paper is based on the judgment of the value of airport historical buildings to study different protection and reuse strategies. The protection and reuse models of historical buildings are classified in three dimensions: the airport historical area, the airport historical building complex and its individual buildings, and combined with specific examples to discuss and summarize the technical characteristics, protection strategies and successful experiences of different modes of protection and reuse of historical areas and historical buildings of airports.

Keywords: Airport, airport area, historic airport building, protection, reuse model.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 662
2061 Reducing the Number of Constraints in Non Safe Petri Net

Authors: M. Zareiee, A. Dideban

Abstract:

This paper addresses the problem of forbidden states in non safe Petri Nets. In the system, for preventing it from entering the forbidden states, some linear constraints can be assigned to them. Then these constraints can be enforced on the system using control places. But when the number of constraints in the system is large, a large number of control places must be added to the model of system. This concept complicates the model of system. There are some methods for reducing the number of constraints in safe Petri Nets. But there is no a systematic method for non safe Petri Nets. In this paper we propose a method for reducing the number of constraints in non safe Petri Nets which is based on solving an integer linear programming problem.

Keywords: discrete event system, Supervisory control, Petri Net, Constraint

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1312
2060 Multidimensional Data Mining by Means of Randomly Travelling Hyper-Ellipsoids

Authors: Pavel Y. Tabakov, Kevin Duffy

Abstract:

The present study presents a new approach to automatic data clustering and classification problems in large and complex databases and, at the same time, derives specific types of explicit rules describing each cluster. The method works well in both sparse and dense multidimensional data spaces. The members of the data space can be of the same nature or represent different classes. A number of N-dimensional ellipsoids are used for enclosing the data clouds. Due to the geometry of an ellipsoid and its free rotation in space the detection of clusters becomes very efficient. The method is based on genetic algorithms that are used for the optimization of location, orientation and geometric characteristics of the hyper-ellipsoids. The proposed approach can serve as a basis for the development of general knowledge systems for discovering hidden knowledge and unexpected patterns and rules in various large databases.

Keywords: Classification, clustering, data minig, genetic algorithms.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1768
2059 A Genetic Algorithm for Clustering on Image Data

Authors: Qin Ding, Jim Gasvoda

Abstract:

Clustering is the process of subdividing an input data set into a desired number of subgroups so that members of the same subgroup are similar and members of different subgroups have diverse properties. Many heuristic algorithms have been applied to the clustering problem, which is known to be NP Hard. Genetic algorithms have been used in a wide variety of fields to perform clustering, however, the technique normally has a long running time in terms of input set size. This paper proposes an efficient genetic algorithm for clustering on very large data sets, especially on image data sets. The genetic algorithm uses the most time efficient techniques along with preprocessing of the input data set. We test our algorithm on both artificial and real image data sets, both of which are of large size. The experimental results show that our algorithm outperforms the k-means algorithm in terms of running time as well as the quality of the clustering.

Keywords: Clustering, data mining, genetic algorithm, image data.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2043
2058 Lightweight Mirrors for Space X-Ray Telescopes

Authors: M. Mika, L. Pina, M. Landova, L. Sveda, R. Havlikova, V. Marsikova, R. Hudec, A. Inneman

Abstract:

Future astronomical projects on large space x-ray imaging telescopes require novel substrates and technologies for the construction of their reflecting mirrors. The mirrors must be lightweight and precisely shaped to achieve large collecting area with high angular resolution. The new materials and technologies must be cost-effective. Currently, the most promising materials are glass or silicon foils. We focused on precise shaping these foils by thermal forming process. We studied free and forced slumping in the temperature region of hot plastic deformation and compared the shapes obtained by the different slumping processes. We measured the shapes and the surface quality of the foils. In the experiments, we varied both heat-treatment temperature and time following our experiment design. The obtained data and relations we can use for modeling and optimizing the thermal forming procedure.

Keywords: Glass, silicon, thermal forming, x-ray

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1381
2057 Performance Comparison of Different Regression Methods for a Polymerization Process with Adaptive Sampling

Authors: Florin Leon, Silvia Curteanu

Abstract:

Developing complete mechanistic models for polymerization reactors is not easy, because complex reactions occur simultaneously; there is a large number of kinetic parameters involved and sometimes the chemical and physical phenomena for mixtures involving polymers are poorly understood. To overcome these difficulties, empirical models based on sampled data can be used instead, namely regression methods typical of machine learning field. They have the ability to learn the trends of a process without any knowledge about its particular physical and chemical laws. Therefore, they are useful for modeling complex processes, such as the free radical polymerization of methyl methacrylate achieved in a batch bulk process. The goal is to generate accurate predictions of monomer conversion, numerical average molecular weight and gravimetrical average molecular weight. This process is associated with non-linear gel and glass effects. For this purpose, an adaptive sampling technique is presented, which can select more samples around the regions where the values have a higher variation. Several machine learning methods are used for the modeling and their performance is compared: support vector machines, k-nearest neighbor, k-nearest neighbor and random forest, as well as an original algorithm, large margin nearest neighbor regression. The suggested method provides very good results compared to the other well-known regression algorithms.

Keywords: Adaptive sampling, batch bulk methyl methacrylate polymerization, large margin nearest neighbor regression, machine learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1394
2056 Gravitino Dark Matter in (nearly) SLagy D3/D7 m-Split SUSY

Authors: Mansi Dhuria, Aalok Misra

Abstract:

In the context of large volume Big Divisor (nearly) SLagy D3/D7 μ-Split SUSY [1], after an explicit identification of first generation of SM leptons and quarks with fermionic superpartners of four Wilson line moduli, we discuss the identification of gravitino as a potential dark matter candidate by explicitly calculating the decay life times of gravitino (LSP) to be greater than age of universe and lifetimes of decays of the co-NLSPs (the first generation squark/slepton and a neutralino) to the LSP (the gravitino) to be very small to respect BBN constraints. Interested in non-thermal production mechanism of gravitino, we evaluate the relic abundance of gravitino LSP in terms of that of the co-NLSP-s by evaluating their (co-)annihilation cross sections and hence show that the former satisfies the requirement for a potential Dark Matter candidate. We also show that it is possible to obtain a 125 GeV light Higgs in our setup.

Keywords: Split Supersymmetry, Large Volume Swiss-Cheese Calabi-Yau's, Dark Matter, (N)LSP decays, relic abundance.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1579
2055 MONARC: A Case Study on Simulation Analysis for LHC Activities

Authors: Ciprian Dobre

Abstract:

The scale, complexity and worldwide geographical spread of the LHC computing and data analysis problems are unprecedented in scientific research. The complexity of processing and accessing this data is increased substantially by the size and global span of the major experiments, combined with the limited wide area network bandwidth available. We present the latest generation of the MONARC (MOdels of Networked Analysis at Regional Centers) simulation framework, as a design and modeling tool for large scale distributed systems applied to HEP experiments. We present simulation experiments designed to evaluate the capabilities of the current real-world distributed infrastructure to support existing physics analysis processes and the means by which the experiments bands together to meet the technical challenges posed by the storage, access and computing requirements of LHC data analysis within the CMS experiment.

Keywords: Modeling and simulation, evaluation, large scale distributed systems, LHC experiments, CMS.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1807
2054 Effect of Surface Stress on the Deformation around a Nanosized Elliptical Hole: a Finite Element Study

Authors: Weifeng Wang, Xianwei Zeng, Jianping Ding

Abstract:

When the characteristic length of an elastic solid is down to the nanometer level, its deformation behavior becomes size dependent. Surface energy /surface stress have recently been applied to explain such dependency. In this paper, the effect of strain-independent surface stress on the deformation of an isotropic elastic solid containing a nanosized elliptical hole is studied by the finite element method. Two loading cases are considered, in the first case, hoop stress along the rim of the elliptical hole induced by pure surface stress is studied, in the second case, hoop stress around the elliptical opening under combined remote tension and surface stress is investigated. It has been shown that positive surface stress induces compressive hoop stress along the hole, and negative surface stress has opposite effect, maximum hoop stress occurs near the major semi-axes of the ellipse. Under combined loading of remote tension and surface stress, stress concentration around the hole can be either intensified or weakened depending on the sign of the surface stress.

Keywords: Surface stress, finite element method, stress concentration, nanosized elliptical hole

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2071
2053 A Large Ion Collider Experiment (ALICE) Diffractive Detector Control System for RUN-II at the Large Hadron Collider

Authors: J. C. Cabanillas-Noris, M. I. Martínez-Hernández, I. León-Monzón

Abstract:

The selection of diffractive events in the ALICE experiment during the first data taking period (RUN-I) of the Large Hadron Collider (LHC) was limited by the range over which rapidity gaps occur. It would be possible to achieve better measurements by expanding the range in which the production of particles can be detected. For this purpose, the ALICE Diffractive (AD0) detector has been installed and commissioned for the second phase (RUN-II). Any new detector should be able to take the data synchronously with all other detectors and be operated through the ALICE central systems. One of the key elements that must be developed for the AD0 detector is the Detector Control System (DCS). The DCS must be designed to operate safely and correctly this detector. Furthermore, the DCS must also provide optimum operating conditions for the acquisition and storage of physics data and ensure these are of the highest quality. The operation of AD0 implies the configuration of about 200 parameters, from electronics settings and power supply levels to the archiving of operating conditions data and the generation of safety alerts. It also includes the automation of procedures to get the AD0 detector ready for taking data in the appropriate conditions for the different run types in ALICE. The performance of AD0 detector depends on a certain number of parameters such as the nominal voltages for each photomultiplier tube (PMT), their threshold levels to accept or reject the incoming pulses, the definition of triggers, etc. All these parameters define the efficiency of AD0 and they have to be monitored and controlled through AD0 DCS. Finally, AD0 DCS provides the operator with multiple interfaces to execute these tasks. They are realized as operating panels and scripts running in the background. These features are implemented on a SCADA software platform as a distributed control system which integrates to the global control system of the ALICE experiment.

Keywords: AD0, ALICE, DCS, LHC.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1386
2052 Learning Monte Carlo Data for Circuit Path Length

Authors: Namal A. Senanayake, A. Beg, Withana C. Prasad

Abstract:

This paper analyzes the patterns of the Monte Carlo data for a large number of variables and minterms, in order to characterize the circuit path length behavior. We propose models that are determined by training process of shortest path length derived from a wide range of binary decision diagram (BDD) simulations. The creation of the model was done use of feed forward neural network (NN) modeling methodology. Experimental results for ISCAS benchmark circuits show an RMS error of 0.102 for the shortest path length complexity estimation predicted by the NN model (NNM). Use of such a model can help reduce the time complexity of very large scale integrated (VLSI) circuitries and related computer-aided design (CAD) tools that use BDDs.

Keywords: Monte Carlo data, Binary decision diagrams, Neural network modeling, Shortest path length estimation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1591
2051 Efficient Variants of Square Contour Algorithm for Blind Equalization of QAM Signals

Authors: Ahmad Tariq Sheikh, Shahzad Amin Sheikh

Abstract:

A new distance-adjusted approach is proposed in which static square contours are defined around an estimated symbol in a QAM constellation, which create regions that correspond to fixed step sizes and weighting factors. As a result, the equalizer tap adjustment consists of a linearly weighted sum of adaptation criteria that is scaled by a variable step size. This approach is the basis of two new algorithms: the Variable step size Square Contour Algorithm (VSCA) and the Variable step size Square Contour Decision-Directed Algorithm (VSDA). The proposed schemes are compared with existing blind equalization algorithms in the SCA family in terms of convergence speed, constellation eye opening and residual ISI suppression. Simulation results for 64-QAM signaling over empirically derived microwave radio channels confirm the efficacy of the proposed algorithms. An RTL implementation of the blind adaptive equalizer based on the proposed schemes is presented and the system is configured to operate in VSCA error signal mode, for square QAM signals up to 64-QAM.

Keywords: Adaptive filtering, Blind Equalization, Square Contour Algorithm.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1848
2050 Stochastic Control of Decentralized Singularly Perturbed Systems

Authors: Walid S. Alfuhaid, Saud A. Alghamdi, John M. Watkins, M. Edwin Sawan

Abstract:

Designing a controller for stochastic decentralized interconnected large scale systems usually involves a high degree of complexity and computation ability. Noise, observability, and controllability of all system states, connectivity, and channel bandwidth are other constraints to design procedures for distributed large scale systems. The quasi-steady state model investigated in this paper is a reduced order model of the original system using singular perturbation techniques. This paper results in an optimal control synthesis to design an observer based feedback controller by standard stochastic control theory techniques using Linear Quadratic Gaussian (LQG) approach and Kalman filter design with less complexity and computation requirements. Numerical example is given at the end to demonstrate the efficiency of the proposed method.

Keywords: Decentralized, optimal control, output, singular perturb.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1565
2049 An Approach for Transient Response Calculation of large Nonproportionally Damped Structures using Component Mode Synthesis

Authors: Alexander A. Muravyov

Abstract:

A minimal complexity version of component mode synthesis is presented that requires simplified computer programming, but still provides adequate accuracy for modeling lower eigenproperties of large structures and their transient responses. The novelty is that a structural separation into components is done along a plane/surface that exhibits rigid-like behavior, thus only normal modes of each component is sufficient to use, without computing any constraint, attachment, or residual-attachment modes. The approach requires only such input information as a few (lower) natural frequencies and corresponding undamped normal modes of each component. A novel technique is shown for formulation of equations of motion, where a double transformation to generalized coordinates is employed and formulation of nonproportional damping matrix in generalized coordinates is shown.

Keywords: component mode synthesis, finite element models, transient response, nonproportional damping

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1800