Search results for: Large Volume Swiss-Cheese Calabi-Yau's
32853 Evaluation of Hancornia speciosa Gomes Lyophilization at Different Stages of Maturation
Authors: D. C. Soares, J. T. S. Santos, D. G. Costa, A. K. S. Abud, T. P. Nunes, A. V. D. Figueiredo, A. M. de Oliveira Junior
Abstract:
Mangabeira (Hancornia speciosa Gomes), a native plant in Brazil, is found growing spontaneously in various regions of the country. The high perishability of tropical fruits such as mangaba, causes it to be necessary to use technologies that promote conservation, aiming to increase the shelf life of this fruit and add value. The objective of this study was to compare the mangabas lyophilization curves behaviors with different sizes and maturation stages. The fruits were freeze-dried for a period of approximately 45 hours at lyophilizer Liotop brand, model L -108. It has been considered large the fruits between 38 and 58 mm diameter and small, between 23 and 28 mm diameter and the two states of maturation, intermediate and mature. Large size mangabas drying curves in both states of maturation were linear behavior at all process, while the kinetic drying curves related to small fruits, independent of maturation state, had a typical behavior of drying, with all the well-defined steps. With these results it was noted that the time of lyophilization was suitable for small mangabas, a fact that did not happen with the larger one. This may indicate that the large mangabas require a longer time to freeze until reaches the equilibrium level, as it happens with the small fruits, going to have constant moisture at the end of the process. For both types of fruit were analyzed water activity, acidity, protein, lipid, and vitamin C before and after the process.
Keywords: Freeze dryer, mangaba, conservation, chemical characteristics.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 212832852 Considering Aerosol Processes in Nuclear Transport Package Containment Safety Cases
Authors: Andrew Cummings, Rhianne Boag, Sarah Bryson, Gordon Turner
Abstract:
Packages designed for transport of radioactive material must satisfy rigorous safety regulations specified by the International Atomic Energy Agency (IAEA). Higher Activity Waste (HAW) transport packages have to maintain containment of their contents during normal and accident conditions of transport (NCT and ACT). To ensure containment criteria is satisfied these packages are required to be leak-tight in all transport conditions to meet allowable activity release rates. Package design safety reports are the safety cases that provide the claims, evidence and arguments to demonstrate that packages meet the regulations and once approved by the competent authority (in the UK this is the Office for Nuclear Regulation) a licence to transport radioactive material is issued for the package(s). The standard approach to demonstrating containment in the RWM transport safety case is set out in BS EN ISO 12807. In this document a method for measuring a leak rate from the package is explained by way of a small interspace test volume situated between two O-ring seals on the underside of the package lid. The interspace volume is pressurised and a pressure drop measured. A small interspace test volume makes the method more sensitive enabling the measurement of smaller leak rates. By ascertaining the activity of the contents, identifying a releasable fraction of material and by treating that fraction of material as a gas, allowable leak rates for NCT and ACT are calculated. The adherence to basic safety principles in ISO12807 is very pessimistic and current practice in the demonstration of transport safety, which is accepted by the UK regulator. It is UK government policy that management of HAW will be through geological disposal. It is proposed that the intermediate level waste be transported to the geological disposal facility (GDF) in large cuboid packages. This poses a challenge for containment demonstration because such packages will have long seals and therefore large interspace test volumes. There is also uncertainty on the releasable fraction of material within the package ullage space. This is because the waste may be in many different forms which makes it difficult to define the fraction of material released by the waste package. Additionally because of the large interspace test volume, measuring the calculated leak rates may not be achievable. For this reason a justification for a lower releasable fraction of material is sought. This paper considers the use of aerosol processes to reduce the releasable fraction for both NCT and ACT. It reviews the basic coagulation and removal processes and applies the dynamic aerosol balance equation. The proposed solution includes only the most well understood physical processes namely; Brownian coagulation and gravitational settling. Other processes have been eliminated either on the basis that they would serve to reduce the release to the environment further (pessimistically in keeping with the essence of nuclear transport safety cases) or that they are not credible in the conditions of transport considered.Keywords: Aerosol processes, Brownian coagulation, gravitational settling, transport regulations.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 65232851 Energy-Level Structure of a Confined Electron-Positron Pair in Nanostructure
Authors: Tokuei Sako, Paul-Antoine Hervieux
Abstract:
The energy-level structure of a pair of electron and positron confined in a quasi-one-dimensional nano-scale potential well has been investigated focusing on its trend in the small limit of confinement strength ω, namely, the Wigner molecular regime. An anisotropic Gaussian-type basis functions supplemented by high angular momentum functions as large as l = 19 has been used to obtain reliable full configuration interaction (FCI) wave functions. The resultant energy spectrum shows a band structure characterized by ω for the large ω regime whereas for the small ω regime it shows an energy-level pattern dominated by excitation into the in-phase motion of the two particles. The observed trend has been rationalized on the basis of the nodal patterns of the FCI wave functions.
Keywords: Confined systems, positron, wave function, Wigner molecule, quantum dots.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 185332850 Calibration of the Discrete Element Method Using a Large Shear Box
Authors: Corné J. Coetzee, Etienne Horn
Abstract:
One of the main challenges in using the Discrete Element Method (DEM) is to specify the correct input parameter values. In general, the models are sensitive to the input parameter values and accurate results can only be achieved if the correct values are specified. For the linear contact model, micro-parameters such as the particle density, stiffness, coefficient of friction, as well as the particle size and shape distributions are required. There is a need for a procedure to accurately calibrate these parameters before any attempt can be made to accurately model a complete bulk materials handling system. Since DEM is often used to model applications in the mining and quarrying industries, a calibration procedure was developed for materials that consist of relatively large (up to 40 mm in size) particles. A coarse crushed aggregate was used as the test material. Using a specially designed large shear box with a diameter of 590 mm, the confined Young’s modulus (bulk stiffness) and internal friction angle of the material were measured by means of the confined compression test and the direct shear test respectively. DEM models of the experimental setup were developed and the input parameter values were varied iteratively until a close correlation between the experimental and numerical results was achieved. The calibration process was validated by modelling the pull-out of an anchor from a bed of material. The model results compared well with experimental measurement.Keywords: Discrete Element Method (DEM), calibration, shear box, anchor pull-out.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 267232849 Large-Dimensional Shells under Mining Tremors from Various Mining Regions in Poland
Authors: Joanna M. Dulińska, Maria Fabijańska
Abstract:
In the paper a detailed analysis of the dynamic response of a cooling tower shell to mining tremors originated from two main regions of mining activity in Poland (Upper Silesian Coal Basin and Legnica-Glogow Copper District) was presented. The representative time histories registered in the both regions were used as ground motion data in calculations of the dynamic response of the structure. It was proved that the dynamic response of the shell is strongly dependent not only on the level of vibration amplitudes but on the dominant frequency range of the mining shock typical for the mining region as well. Also a vertical component of vibrations occurred to have considerable influence on the total dynamic response of the shell. Finally, it turned out that non-uniformity of kinematic excitation resulting from spatial variety of ground motion plays a significant role in dynamic analysis of large-dimensional shells under mining shocks.Keywords: Cooling towers, dynamic response, mining tremors, non-uniform kinematic excitation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 142132848 A Tabu Search Heuristic for Scratch-Pad Memory Management
Authors: Maha Idrissi Aouad, Rene Schott, Olivier Zendra
Abstract:
Reducing energy consumption of embedded systems requires careful memory management. It has been shown that Scratch- Pad Memories (SPMs) are low size, low cost, efficient (i.e. energy saving) data structures directly managed at the software level. In this paper, the focus is on heuristic methods for SPMs management. A method is efficient if the number of accesses to SPM is as large as possible and if all available space (i.e. bits) is used. A Tabu Search (TS) approach for memory management is proposed which is, to the best of our knowledge, a new original alternative to the best known existing heuristic (BEH). In fact, experimentations performed on benchmarks show that the Tabu Search method is as efficient as BEH (in terms of energy consumption) but BEH requires a sorting which can be computationally expensive for a large amount of data. TS is easy to implement and since no sorting is necessary, unlike BEH, the corresponding sorting time is saved. In addition to that, in a dynamic perspective where the maximum capacity of the SPM is not known in advance, the TS heuristic will perform better than BEH.
Keywords: Energy consumption, memory allocation management, optimization, tabu search heuristic.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 167932847 Estimating Shortest Circuit Path Length Complexity
Authors: Azam Beg, P. W. Chandana Prasad, S.M.N.A Senenayake
Abstract:
When binary decision diagrams are formed from uniformly distributed Monte Carlo data for a large number of variables, the complexity of the decision diagrams exhibits a predictable relationship to the number of variables and minterms. In the present work, a neural network model has been used to analyze the pattern of shortest path length for larger number of Monte Carlo data points. The neural model shows a strong descriptive power for the ISCAS benchmark data with an RMS error of 0.102 for the shortest path length complexity. Therefore, the model can be considered as a method of predicting path length complexities; this is expected to lead to minimum time complexity of very large-scale integrated circuitries and related computer-aided design tools that use binary decision diagrams.Keywords: Monte Carlo circuit simulation data, binary decision diagrams, neural network modeling, shortest path length estimation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 137832846 Music-Inspired Harmony Search Algorithm for Fixed Outline Non-Slicing VLSI Floorplanning
Authors: K. Sivasubramanian, K. B. Jayanthi
Abstract:
Floorplanning plays a vital role in the physical design process of Very Large Scale Integrated (VLSI) chips. It is an essential design step to estimate the chip area prior to the optimized placement of digital blocks and their interconnections. Since VLSI floorplanning is an NP-hard problem, many optimization techniques were adopted in the literature. In this work, a music-inspired Harmony Search (HS) algorithm is used for the fixed die outline constrained floorplanning, with the aim of reducing the total chip area. HS draws inspiration from the musical improvisation process of searching for a perfect state of harmony. Initially, B*-tree is used to generate the primary floorplan for the given rectangular hard modules and then HS algorithm is applied to obtain an optimal solution for the efficient floorplan. The experimental results of the HS algorithm are obtained for the MCNC benchmark circuits.Keywords: Floor planning, harmony search, non-slicing floorplan, very large scale integrated circuits.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 195632845 Characterization of Catalagzi Fly Ash for Heavy Metal Adsorption
Authors: Nurcan Tugrul, Nil Baran Acarali, Seyma Kolemen, Emek Moroydor Derun, Sabriye Piskin
Abstract:
Fly ash is a significant waste that is released of thermal power plants and defined as very fine particles that are drifted upward with up taken by the flue gases due to the burning of used coal [1]. The fly-ash is capable of removing organic contaminants in consequence of high carbon content, a large surface area per unit volume and contained heavy metals. Therefore, fly ash is used as an effective coagulant and adsorbent by pelletization [2, 3]. In this study, the possibility of use of fly ash taken from Turkey like low-cost adsorbent for adsorption of zinc ions found in waste water was investigated. The fly ash taken from Turkey was pelletized with bentonite and molass to evaluate the adsorption capaticity. For this purpose; analyses such as sieve analysis, XRD, XRF, FTIR and SEM were performed. As a result, it was seen that pellets prepared from fly ash, bentonite and molass would be used for zinc adsorption.Keywords: Fly ash, heavy metal, sieve, adsorbent.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 259332844 CFD Simulation of Hydrodynamic Behaviors and Gas-Liquid Mass Transfer in a Stirred Airlift Bioreactor
Authors: Sérgio S. de Jesus, Edgar Leonardo Martínez, Aulus R.R. Binelli, Aline Santana, Rubens Maciel Filho
Abstract:
The speed profiles, gas holdup (eG) and global oxygen transfer coefficient (kLa) from a stirred airlift bioreactor using water as the fluid model, was investigated by computational fluid dynamics modeling. The parameters predicted by the computer model were validated with the experimental dates. The CFD results were very close to those obtained experimentally. During the simulation it was verified a prevalent impeller effect at low speeds, propelling a large volume of fluid against the walls of the vessel, which without recirculation, results in low values of eG and kLa; however, by increasing air velocity, the impeller effect is smaller with the air flow being greater, in the region of the riser, causing fluid recirculation, which explains the increase in eG and kLa.
Keywords: CFD, Hydrodynamics, Mass transfer, Stirred airlift bioreactor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 375132843 Li-Fi Technology: Data Transmission through Visible Light
Authors: Shahzad Hassan, Kamran Saeed
Abstract:
People are always in search of Wi-Fi hotspots because Internet is a major demand nowadays. But like all other technologies, there is still room for improvement in the Wi-Fi technology with regards to the speed and quality of connectivity. In order to address these aspects, Harald Haas, a professor at the University of Edinburgh, proposed what we know as the Li-Fi (Light Fidelity). Li-Fi is a new technology in the field of wireless communication to provide connectivity within a network environment. It is a two-way mode of wireless communication using light. Basically, the data is transmitted through Light Emitting Diodes which can vary the intensity of light very fast, even faster than the blink of an eye. From the research and experiments conducted so far, it can be said that Li-Fi can increase the speed and reliability of the transfer of data. This paper pays particular attention on the assessment of the performance of this technology. In other words, it is a 5G technology which uses LED as the medium of data transfer. For coverage within the buildings, Wi-Fi is good but Li-Fi can be considered favorable in situations where large amounts of data are to be transferred in areas with electromagnetic interferences. It brings a lot of data related qualities such as efficiency, security as well as large throughputs to the table of wireless communication. All in all, it can be said that Li-Fi is going to be a future phenomenon where the presence of light will mean access to the Internet as well as speedy data transfer.
Keywords: Communication, LED, Li-Fi, Wi-Fi.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 216832842 Pose-Dependency of Machine Tool Structures: Appearance, Consequences, and Challenges for Lightweight Large-Scale Machines
Authors: S. Apprich, F. Wulle, A. Lechler, A. Pott, A. Verl
Abstract:
Large-scale machine tools for the manufacturing of large work pieces, e.g. blades, casings or gears for wind turbines, feature pose-dependent dynamic behavior. Small structural damping coefficients lead to long decay times for structural vibrations that have negative impacts on the production process. Typically, these vibrations are handled by increasing the stiffness of the structure by adding mass. This is counterproductive to the needs of sustainable manufacturing as it leads to higher resource consumption both in material and in energy. Recent research activities have led to higher resource efficiency by radical mass reduction that is based on controlintegrated active vibration avoidance and damping methods. These control methods depend on information describing the dynamic behavior of the controlled machine tools in order to tune the avoidance or reduction method parameters according to the current state of the machine. This paper presents the appearance, consequences and challenges of the pose-dependent dynamic behavior of lightweight large-scale machine tool structures in production. It starts with the theoretical introduction of the challenges of lightweight machine tool structures resulting from reduced stiffness. The statement of the pose-dependent dynamic behavior is corroborated by the results of the experimental modal analysis of a lightweight test structure. Afterwards, the consequences of the pose-dependent dynamic behavior of lightweight machine tool structures for the use of active control and vibration reduction methods are explained. Based on the state of the art of pose-dependent dynamic machine tool models and the modal investigation of an FE-model of the lightweight test structure, the criteria for a pose-dependent model for use in vibration reduction are derived. The description of the approach for a general posedependent model of the dynamic behavior of large lightweight machine tools that provides the necessary input to the aforementioned vibration avoidance and reduction methods to properly tackle machine vibrations is the outlook of the paper.Keywords: Dynamic behavior, lightweight, machine tool, pose-dependency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 284432841 Performance Analysis of List Scheduling in Heterogeneous Computing Systems
Authors: Keqin Li
Abstract:
Given a parallel program to be executed on a heterogeneous computing system, the overall execution time of the program is determined by a schedule. In this paper, we analyze the worst-case performance of the list scheduling algorithm for scheduling tasks of a parallel program in a mixed-machine heterogeneous computing system such that the total execution time of the program is minimized. We prove tight lower and upper bounds for the worst-case performance ratio of the list scheduling algorithm. We also examine the average-case performance of the list scheduling algorithm. Our experimental data reveal that the average-case performance of the list scheduling algorithm is much better than the worst-case performance and is very close to optimal, except for large systems with large heterogeneity. Thus, the list scheduling algorithm is very useful in real applications.Keywords: Average-case performance, list scheduling algorithm, mixed-machine heterogeneous computing system, worst-case performance.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 134932840 Performance Comparison of Different Regression Methods for a Polymerization Process with Adaptive Sampling
Authors: Florin Leon, Silvia Curteanu
Abstract:
Developing complete mechanistic models for polymerization reactors is not easy, because complex reactions occur simultaneously; there is a large number of kinetic parameters involved and sometimes the chemical and physical phenomena for mixtures involving polymers are poorly understood. To overcome these difficulties, empirical models based on sampled data can be used instead, namely regression methods typical of machine learning field. They have the ability to learn the trends of a process without any knowledge about its particular physical and chemical laws. Therefore, they are useful for modeling complex processes, such as the free radical polymerization of methyl methacrylate achieved in a batch bulk process. The goal is to generate accurate predictions of monomer conversion, numerical average molecular weight and gravimetrical average molecular weight. This process is associated with non-linear gel and glass effects. For this purpose, an adaptive sampling technique is presented, which can select more samples around the regions where the values have a higher variation. Several machine learning methods are used for the modeling and their performance is compared: support vector machines, k-nearest neighbor, k-nearest neighbor and random forest, as well as an original algorithm, large margin nearest neighbor regression. The suggested method provides very good results compared to the other well-known regression algorithms.Keywords: Adaptive sampling, batch bulk methyl methacrylate polymerization, large margin nearest neighbor regression, machine learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 140032839 An Agent-based Model for Analyzing Interaction of Two Stable Social Networks
Authors: Masatora Daito, Hisashi Kojima
Abstract:
In this research, the authors analyze network stability using agent-based simulation. Firstly, the authors focus on analyzing large networks (eight agents) by connecting different two stable small social networks (A small stable network is consisted on four agents.). Secondly, the authors analyze the network (eight agents) shape which is added one agent to a stable network (seven agents). Thirdly, the authors analyze interpersonal comparison of utility. The “star-network "was not found on the result of interaction among stable two small networks. On the other hand, “decentralized network" was formed from several combination. In case of added one agent to a stable network (seven agents), if the value of “c"(maintenance cost of per a link) was larger, the number of patterns of stable network was also larger. In this case, the authors identified the characteristics of a large stable network. The authors discovered the cases of decreasing personal utility under condition increasing total utility.Keywords: Social Network, Symmetric Situation, Network Stability, Agent-Based Modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 153332838 Hybrid Modulation Technique for Fingerprinting
Authors: Hae-Yeoun Lee, In-Koo Kang, Heung-Kyu Lee
Abstract:
This paper addresses an efficient technique to embed and detect digital fingerprint code. Orthogonal modulation method is a straightforward and widely used approach for digital fingerprinting but shows several limitations in computational cost and signal efficiency. Coded modulation method can solve these limitations in theory. However it is difficult to perform well in practice if host signals are not available during tracing colluders, other kinds of attacks are applied, and the size of fingerprint code becomes large. In this paper, we propose a hybrid modulation method, in which the merits of or-thogonal modulation and coded modulation method are combined so that we can achieve low computational cost and high signal efficiency. To analyze the performance, we design a new fingerprint code based on GD-PBIBD theory and modulate this code into images by our method using spread-spectrum watermarking on frequency domain. The results show that the proposed method can efficiently handle large fingerprint code and trace colluders against averaging attacks.
Keywords: Fingerprinting, GD-PBIBD theory, Hybrid modulationtechnique.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 138032837 Analysis of Wave Propagation in Two-dimensional Phononic Crystals with Hollow Cylinders
Authors: Zi-Gui Huang, Tsung-Tsong Wu
Abstract:
Large full frequency band gaps of surface and bulk acoustic waves in two-dimensional phononic band structures with hollow cylinders are addressed in this paper. It is well-known that absolute frequency band gaps are difficultly obtained in a band structure consisted of low-acoustic-impedance cylinders in high-acoustic-impedance host materials such as PMMA/Ni band structures. Phononic band structures with hollow cylinders are analyzed and discussed to obtain large full frequency band gaps not only for bulk modes but also for surface modes. The tendency of absolute frequency band gaps of surface and bulk acoustic waves is also addressed by changing the inner radius of hollow cylinders in this paper. The technique and this kind of band structure are useful for tuning the frequency band gaps and the design of acoustic waveguides.Keywords: Phononic crystals, Band gap, SAW, BAW.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 198832836 The Scale of Farms and Development Perspectives in Georgia
Authors: M. Chavleishvili, E. Kharaishvili, G. Erkomaishvili
Abstract:
The article presents the development trends of farms, estimates on the optimal scope of farming, as well as the experience of local and foreign countries in this area. As well, the advantages of small and large farms are discussed; herewith, the scales of farms are compared to the local reality. The study analyzes the results of farm operations and the possibilities of diversification of farms. The indicators of an effective use of land resources and land fragmentation are measured; also, a comparative analysis with other countries is presented, in particular, the measurements of agricultural lands for farming, as well as the indicators of population ensuring. The conducted research shows that most of the farms in Georgia are small and their development is at the initial stage, which outlines that the country has a high resource potential to increase the scale of the farming industry and its full integration into market relations. On the basis of the obtained results, according to the research on the scale of farming in Georgia and the identification of hampering factors of farming development, the conclusions are presented and the relevant recommendations are suggested.
Keywords: Farm cooperatives, farms, farm scale, land fragmentation, small and large farms.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 82632835 Kinematic Behavior of Geogrid Reinforcements during Earthquakes
Authors: Ahmed Hosny Abdel-Rahman, Mohamed Abdel-Moneim
Abstract:
Reinforced earth structures are generally subjected to cyclic loading generated from earthquakes. This paper presents a summary of the results and analyses of a testing program carried out in a large-scale multi-function geosynthetic testing apparatus that accommodates soil samples up to 1.0 m3. This apparatus performs different shear and pullout tests under both static and cyclic loading. The testing program was carried out to investigate the controlling factors affecting soil/geogrid interaction under cyclic loading. The extensibility of the geogrids, the applied normal stresses, the characteristics of the cyclic loading (frequency, and amplitude), and initial static load within the geogrid sheet were considered in the testing program. Based on the findings of the testing program, the effect of these parameters on the pullout resistance of geogrids, as well as the displacement mobility under cyclic loading were evaluated. Conclusions and recommendations for the design of reinforced earth walls under cyclic loading are presented.Keywords: Geogrid, Soil, Interface, Cyclic Loading, Pullout, and Large scale Testing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 185232834 Influence of Inertial Forces of Large Bearings Utilized in Wind Energy Assemblies
Authors: S. Barabas, F. Sarbu, B. Barabas, A. Fota
Abstract:
Main objective of this paper is to establish a link between inertial forces of the bearings used in construction of wind power plant and its behavior. Using bearings with lower inertial forces has the immediate effect of decreasing inertia rotor system, with significant results in increased energy efficiency, due to decreased friction forces between rollers and raceways. The F.E.M. analysis shows the appearance of uniform contact stress at the ends of the rollers, demonstrated the necessity of production of low mass bearings. Favorable results are expected in the economic field, by reducing material consumption and by increasing the durability of bearings. Using low mass bearings with hollow rollers instead of solid rollers has an impact on working temperature, on vibrations and noise which decrease. Implementation of types of hollow rollers of cylindrical tubular type, instead of expensive rollers with logarithmic profile, will bring significant inertial forces decrease with large benefits in behavior of wind power plant.Keywords: Inertial forces, Von Mises stress, hollow rollers.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 226732833 A Large Ion Collider Experiment (ALICE) Diffractive Detector Control System for RUN-II at the Large Hadron Collider
Authors: J. C. Cabanillas-Noris, M. I. Martínez-Hernández, I. León-Monzón
Abstract:
The selection of diffractive events in the ALICE experiment during the first data taking period (RUN-I) of the Large Hadron Collider (LHC) was limited by the range over which rapidity gaps occur. It would be possible to achieve better measurements by expanding the range in which the production of particles can be detected. For this purpose, the ALICE Diffractive (AD0) detector has been installed and commissioned for the second phase (RUN-II). Any new detector should be able to take the data synchronously with all other detectors and be operated through the ALICE central systems. One of the key elements that must be developed for the AD0 detector is the Detector Control System (DCS). The DCS must be designed to operate safely and correctly this detector. Furthermore, the DCS must also provide optimum operating conditions for the acquisition and storage of physics data and ensure these are of the highest quality. The operation of AD0 implies the configuration of about 200 parameters, from electronics settings and power supply levels to the archiving of operating conditions data and the generation of safety alerts. It also includes the automation of procedures to get the AD0 detector ready for taking data in the appropriate conditions for the different run types in ALICE. The performance of AD0 detector depends on a certain number of parameters such as the nominal voltages for each photomultiplier tube (PMT), their threshold levels to accept or reject the incoming pulses, the definition of triggers, etc. All these parameters define the efficiency of AD0 and they have to be monitored and controlled through AD0 DCS. Finally, AD0 DCS provides the operator with multiple interfaces to execute these tasks. They are realized as operating panels and scripts running in the background. These features are implemented on a SCADA software platform as a distributed control system which integrates to the global control system of the ALICE experiment.Keywords: AD0, ALICE, DCS, LHC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 139832832 Using Radial Basis Function Neural Networks to Calibrate Water Quality Model
Authors: Lihui Ma, Kunlun Xin, Suiqing Liu
Abstract:
Modern managements of water distribution system (WDS) need water quality models that are able to accurately predict the dynamics of water quality variations within the distribution system environment. Before water quality models can be applied to solve system problems, they should be calibrated. Although former researchers use GA solver to calibrate relative parameters, it is difficult to apply on the large-scale or medium-scale real system for long computational time. In this paper a new method is designed which combines both macro and detailed model to optimize the water quality parameters. This new combinational algorithm uses radial basis function (RBF) metamodeling as a surrogate to be optimized for the purpose of decreasing the times of time-consuming water quality simulation and can realize rapidly the calibration of pipe wall reaction coefficients of chlorine model of large-scaled WDS. After two cases study this method is testified to be more efficient and promising, and deserve to generalize in the future.Keywords: Metamodeling, model calibration, radial basisfunction, water distribution system, water quality model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 202132831 Density Clustering Based On Radius of Data (DCBRD)
Authors: A.M. Fahim, A. M. Salem, F. A. Torkey, M. A. Ramadan
Abstract:
Clustering algorithms are attractive for the task of class identification in spatial databases. However, the application to large spatial databases rises the following requirements for clustering algorithms: minimal requirements of domain knowledge to determine the input parameters, discovery of clusters with arbitrary shape and good efficiency on large databases. The well-known clustering algorithms offer no solution to the combination of these requirements. In this paper, a density based clustering algorithm (DCBRD) is presented, relying on a knowledge acquired from the data by dividing the data space into overlapped regions. The proposed algorithm discovers arbitrary shaped clusters, requires no input parameters and uses the same definitions of DBSCAN algorithm. We performed an experimental evaluation of the effectiveness and efficiency of it, and compared this results with that of DBSCAN. The results of our experiments demonstrate that the proposed algorithm is significantly efficient in discovering clusters of arbitrary shape and size.
Keywords: Clustering Algorithms, Arbitrary Shape of clusters, cluster Analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 187532830 Formal Verification of Cache System Using a Novel Cache Memory Model
Authors: Guowei Hou, Lixin Yu, Wei Zhuang, Hui Qin, Xue Yang
Abstract:
Formal verification is proposed to ensure the correctness of the design and make functional verification more efficient. As cache plays a vital role in the design of System on Chip (SoC), and cache with Memory Management Unit (MMU) and cache memory unit makes the state space too large for simulation to verify, then a formal verification is presented for such system design. In the paper, a formal model checking verification flow is suggested and a new cache memory model which is called “exhaustive search model” is proposed. Instead of using large size ram to denote the whole cache memory, exhaustive search model employs just two cache blocks. For cache system contains data cache (Dcache) and instruction cache (Icache), Dcache memory model and Icache memory model are established separately using the same mechanism. At last, the novel model is employed to the verification of a cache which is module of a custom-built SoC system that has been applied in practical, and the result shows that the cache system is verified correctly using the exhaustive search model, and it makes the verification much more manageable and flexible.
Keywords: Cache system, formal verification, novel model, System on Chip (SoC).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 229832829 A Serializability Condition for Multi-step Transactions Accessing Ordered Data
Authors: Rafat Alshorman, Walter Hussak
Abstract:
In mobile environments, unspecified numbers of transactions arrive in continuous streams. To prove correctness of their concurrent execution a method of modelling an infinite number of transactions is needed. Standard database techniques model fixed finite schedules of transactions. Lately, techniques based on temporal logic have been proposed as suitable for modelling infinite schedules. The drawback of these techniques is that proving the basic serializability correctness condition is impractical, as encoding (the absence of) conflict cyclicity within large sets of transactions results in prohibitively large temporal logic formulae. In this paper, we show that, under certain common assumptions on the graph structure of data items accessed by the transactions, conflict cyclicity need only be checked within all possible pairs of transactions. This results in formulae of considerably reduced size in any temporal-logic-based approach to proving serializability, and scales to arbitrary numbers of transactions.Keywords: multi-step transactions, serializability, directed graph.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 135832828 The Determinants of Voluntary Disclosure in Croatia
Authors: Zeljana Aljinovic Barac, Marina Granic, Tina Vuko
Abstract:
Study investigates the level and extent of voluntary disclosure practice in Croatia. The research was conducted on the sample of 130 medium and large companies. Findings indicate that two thirds of the companies analyzed disclose below-average number of additional information. The explanatory analyses has shown that firm size, listing status and industrial sector significantly and positively affect the level and extent of voluntary disclosure in the annual report of Croatian companies. On the other hand, profitability and ownership structure were found statistically insignificant. Unlike previous studies, this paper deals with level of voluntary disclosure of medium and large companies, as well as companies whose shares are not listed on the organized capital market, which can be found as our contribution. Also, the research makes contribution by providing the insights into voluntary disclosure practices in Croatia, as a case of macro-oriented accounting system economy, i.e. bank oriented economy with an emerging capital market.
Keywords: Annual report, Croatian companies, Disclosure index, Voluntary disclosure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 252532827 An Efficient 3D Animation Data Reduction Using Frame Removal
Authors: Jinsuk Yang, Choongjae Joo, Kyoungsu Oh
Abstract:
Existing methods in which the animation data of all frames are stored and reproduced as with vertex animation cannot be used in mobile device environments because these methods use large amounts of the memory. So 3D animation data reduction methods aimed at solving this problem have been extensively studied thus far and we propose a new method as follows. First, we find and remove frames in which motion changes are small out of all animation frames and store only the animation data of remaining frames (involving large motion changes). When playing the animation, the removed frame areas are reconstructed using the interpolation of the remaining frames. Our key contribution is to calculate the accelerations of the joints of individual frames and the standard deviations of the accelerations using the information of joint locations in the relevant 3D model in order to find and delete frames in which motion changes are small. Our methods can reduce data sizes by approximately 50% or more while providing quality which is not much lower compared to original animations. Therefore, our method is expected to be usefully used in mobile device environments or other environments in which memory sizes are limited.
Keywords: Data Reduction, Interpolation, Vertex Animation, 3D Animation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 166132826 Organizational Dimensions as Determinant Factors of KM Approaches in SMEs
Authors: Mehdi Shami Zanjani, Shohreh Mehrasa, Mandana Modiri
Abstract:
In the current economy of increasing global competition, many organizations are attempting to use knowledge as one of the means to gain sustainable competitive advantage. Besides large organizations, the success of SMEs can be linked to how well they manage their knowledge. Despite the profusion of research about knowledge management within large organizations, fewer studies tried to analyze KM in SMEs. This research proposes a new framework showing the determinant role of organizational dimensions onto KM approaches. The paper and its propositions are based on a literature review and analysis. In this research, personalization versus codification, individualization versus institutionalization and IT-based versus non IT-based are highlighted as three distinct dimensions of knowledge management approaches. The study contributes to research by providing a more nuanced classification of KM approaches and provides guidance to managers about the types of KM approaches that should be adopted based on the size, geographical dispersion and task nature of SMEs. To the author-s knowledge, the paper is the first of its kind to examine if there are suitable configurations of KM approaches for SMEs with different dimensions. It gives valuable information, which hopefully will help SME sector to accomplish KM.Keywords: Knowledge Management, Knowledge Management Approach, SME, Organizational Dimension.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 160632825 A Smart-Visio Microphone for Audio-Visual Speech Recognition “Vmike“
Abstract:
The practical implementation of audio-video coupled speech recognition systems is mainly limited by the hardware complexity to integrate two radically different information capturing devices with good temporal synchronisation. In this paper, we propose a solution based on a smart CMOS image sensor in order to simplify the hardware integration difficulties. By using on-chip image processing, this smart sensor can calculate in real time the X/Y projections of the captured image. This on-chip projection reduces considerably the volume of the output data. This data-volume reduction permits a transmission of the condensed visual information via the same audio channel by using a stereophonic input available on most of the standard computation devices such as PC, PDA and mobile phones. A prototype called VMIKE (Visio-Microphone) has been designed and realised by using standard 0.35um CMOS technology. A preliminary experiment gives encouraged results. Its efficiency will be further investigated in a large variety of applications such as biometrics, speech recognition in noisy environments, and vocal control for military or disabled persons, etc.
Keywords: Audio-Visual Speech recognition, CMOS Smartsensor, On-Chip image processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 182632824 Induction Motor Efficiency Estimation using Genetic Algorithm
Authors: Khalil Banan, Mohammad B.B. Sharifian, Jafar Mohammadi
Abstract:
Due to the high percentage of induction motors in industrial market, there exist a large opportunity for energy savings. Replacement of working induction motors with more efficient ones can be an important resource for energy savings. A calculation of energy savings and payback periods, as a result of such a replacement, based on nameplate motor efficiency or manufacture-s data can lead to large errors [1]. Efficiency of induction motors (IMs) can be extracted using some procedures that use the no-load test results. In the cases that we must estimate the efficiency on-line, some of these procedures can-t be efficient. In some cases the efficiency estimates using the rating values of the motor, but these procedures can have errors due to the different working condition of the motor. In this paper the efficiency of an IM estimated by using the genetic algorithm. The results are compared with the measured values of the torque and power. The results show smaller errors for this procedure compared with the conventional classical procedures, hence the cost of the equipments is reduced and on-line estimation of the efficiency can be made.
Keywords: Genetic algorithm, induction motor, efficiency.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2602