Search results for: pairing computation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 566

Search results for: pairing computation

476 Abnormal Features of Two Quasiparticle Rotational Bands in Rare Earths

Authors: Kawalpreet Kalra, Alpana Goel

Abstract:

The behaviour of the rotational bands should be smooth but due to large amount of inertia and decreased pairing it is not so. Many experiments have been done in the last few decades, and a large amount of data is available for comprehensive study in this region. Peculiar features like signature dependence, signature inversion, and signature reversal are observed in many two quasiparticle rotational bands of doubly odd and doubly even nuclei. At high rotational frequencies, signature and parity are the only two good quantum numbers available to label a state. Signature quantum number is denoted by α. Even-angular momentum states of a rotational band have α =0, and the odd-angular momentum states have α =1. It has been observed that the odd-spin members lie lower in energy up to a certain spin Ic; the normal signature dependence is restored afterwards. This anomalous feature is termed as signature inversion. The systematic of signature inversion in high-j orbitals for doubly odd rare earth nuclei have been done. Many unusual features like signature dependence, signature inversion and signature reversal are observed in rotational bands of even-even/odd-odd nuclei. Attempts have been made to understand these phenomena using several models. These features have been analyzed within the framework of the Two Quasiparticle Plus Rotor Model (TQPRM).

Keywords: rotational bands, signature dependence, signature quantum number, two quasiparticle

Procedia PDF Downloads 136
475 Investigating the Contemporary Architecture Education Challenges in India

Authors: Vriddhi Prasad

Abstract:

The paper briefly outlines the nature of contemporary Architecture Education in India and its present challenges with theoretically feasible solutions. It explores in detail the arduous position of architecture education owing to, privatization of higher education institutes in India, every changing demand of the technology driven industry and discipline, along with regional and cultural resources that should be explored academically for the enrichment of graduates. With the government's education policy of supporting privatization, a comprehensive role for the regulating body of Architecture Education becomes imperative. The paper provides key insights through empirical research into the nature of these roles and the areas which need attention in light of the problems. With the aid of critically acclaimed education model like Design Build, contextual retrofits for Indian institutes can be stressed for inclusion in the curriculum. The pairing of a private institute and public industry/research body and vice versa can lead to pro-economic and pro-social research environment. These reforms if stressed by an autonomous nationwide regulating body rather than the state will lead to uniformity and flexibility of curriculum which promotes the creation of fresh graduates who are adaptable to the changing needs.

Keywords: architecture education, building information modelling, design build, pedagogy

Procedia PDF Downloads 185
474 Accelerating Quantum Chemistry Calculations: Machine Learning for Efficient Evaluation of Electron-Repulsion Integrals

Authors: Nishant Rodrigues, Nicole Spanedda, Chilukuri K. Mohan, Arindam Chakraborty

Abstract:

A crucial objective in quantum chemistry is the computation of the energy levels of chemical systems. This task requires electron-repulsion integrals as inputs, and the steep computational cost of evaluating these integrals poses a major numerical challenge in efficient implementation of quantum chemical software. This work presents a moment-based machine-learning approach for the efficient evaluation of electron-repulsion integrals. These integrals were approximated using linear combinations of a small number of moments. Machine learning algorithms were applied to estimate the coefficients in the linear combination. A random forest approach was used to identify promising features using a recursive feature elimination approach, which performed best for learning the sign of each coefficient but not the magnitude. A neural network with two hidden layers were then used to learn the coefficient magnitudes along with an iterative feature masking approach to perform input vector compression, identifying a small subset of orbitals whose coefficients are sufficient for the quantum state energy computation. Finally, a small ensemble of neural networks (with a median rule for decision fusion) was shown to improve results when compared to a single network.

Keywords: quantum energy calculations, atomic orbitals, electron-repulsion integrals, ensemble machine learning, random forests, neural networks, feature extraction

Procedia PDF Downloads 61
473 Development and Evaluation of Novel Diagnostic Methods for Infectious Rhinotracheitis of Cattle

Authors: Wenxiao Liu, Kun Zhang, Yongqing Li

Abstract:

Bovine herpesvirus 1, a member of the genus Variellovirus of the subfamily Alphaherpesvirinae, has caused severe economic cost to the bovine industry. In this study, BoHV-1 glycerol protein gD was expressed in insect cells, and the purified gD was immunized in the Balb/C mice to generate monoclonal antibodies. Based on hybridoma cell fusion techniques, 20 monoclonal antibodies against Bovine herpesvirus 1 have been obtained. Further, mAb 3F8 with neutralizing activity and gD were applied to develop a blocking enzyme-linked immunosorbent assay (Elisa) for detecting neutralizing antibodies against BoHV-1, which shows a significant correlation between the blocking Elisa and VNT. The sensitivity and specificity of the test were estimated to be 94.59% and 93.42%, respectively. Furthermore, antibody pairing tests revealed that mAb 1B6 conjugated to fluorescence microspheres was used as the capture antibody, and mAb 3F9 was used as the detectable antibody to establish the immunochromatographic assay (ICS). The ICS was conducted to detect BoHV-1 in bovine samples with high sensitivity, specificity, and good stability. Clinical sample testing revealed that the results of ICS and real-time PCR have a coincidence rate of 95.42%. Our research confirmed that the ICS is a rapid and reliable method for the diagnosis of BoHV-1. In conclusion, our results lay a solid foundation for the prevention and control of BoHV-1 infection.

Keywords: bovine disease, BoHV-1, ELISA, ICS assay

Procedia PDF Downloads 29
472 Heuristic Algorithms for Time Based Weapon-Target Assignment Problem

Authors: Hyun Seop Uhm, Yong Ho Choi, Ji Eun Kim, Young Hoon Lee

Abstract:

Weapon-target assignment (WTA) is a problem that assigns available launchers to appropriate targets in order to defend assets. Various algorithms for WTA have been developed over past years for both in the static and dynamic environment (denoted by SWTA and DWTA respectively). Due to the problem requirement to be solved in a relevant computational time, WTA has suffered from the solution efficiency. As a result, SWTA and DWTA problems have been solved in the limited situation of the battlefield. In this paper, the general situation under continuous time is considered by Time based Weapon Target Assignment (TWTA) problem. TWTA are studied using the mixed integer programming model, and three heuristic algorithms; decomposed opt-opt, decomposed opt-greedy, and greedy algorithms are suggested. Although the TWTA optimization model works inefficiently when it is characterized by a large size, the decomposed opt-opt algorithm based on the linearization and decomposition method extracted efficient solutions in a reasonable computation time. Because the computation time of the scheduling part is too long to solve by the optimization model, several algorithms based on greedy is proposed. The models show lower performance value than that of the decomposed opt-opt algorithm, but very short time is needed to compute. Hence, this paper proposes an improved method by applying decomposition to TWTA, and more practical and effectual methods can be developed for using TWTA on the battlefield.

Keywords: air and missile defense, weapon target assignment, mixed integer programming, piecewise linearization, decomposition algorithm, military operations research

Procedia PDF Downloads 303
471 Protective Effect of hsa-miR-124 against to Bacillus anthracis Toxins on Human Macrophage Cells

Authors: Ali Oztuna, Meral Sarper, Deniz Torun, Fatma Bayrakdar, Selcuk Kilic, Mehmet Baysallar

Abstract:

Bacillus anthracis is one of the biological agents most likely to be used in case of bioterrorist attack as well as being the cause of anthrax. The bacterium's major virulence factors are the anthrax toxins and an antiphagocytic polyglutamic capsule. TEM8 (ANTXR1) and CMG2 (ANTXR2) are ubiquitously expressed type I transmembrane proteins, and ANTXR2 is the major receptor for anthrax toxins. MicroRNAs are 21-24 bp small noncoding RNAs that regulate gene expression by base pairing with the 3' UTR (untranslated regions) of their target mRNAs resulting in mRNA degradation and/or translational repression. MicroRNAs contribute to regulation of most biological processes and influence numerous pathological states like infectious disease. In this study, post-exposure (toxins) protective effect of the hsa-miR-124-3p against Bacillus anthracis was examined. In this context, i) THP-1 and U937 cells were differentiated to MΦ macrophage, ii) miRNA transfection efficiencies were evaluated by flow cytometry and qPCR, iii) protection against Bacillus anthracis toxins were investigated by XTT, cAMP ELISA and MEK2 cleavage assays. Acknowledgements: This work was supported by the Scientific and Technological Research Council of Turkey (TUBITAK) under Grant SBAG-218S467.

Keywords: ANTXR2, hsa-miR-124-3p, MΦ macrophage, THP-1, U937

Procedia PDF Downloads 116
470 Analysis of Replication Protein A (RPA): The Role of Homolog Interaction and Recombination during Meiosis

Authors: Jeong Hwan Joo, Keun Pil Kim

Abstract:

During meiosis, meiotic recombination is initiated by Spo11-mediated DSB formation and exonuclease-mediated DSB resection occurs to expose single stranded DNA formation. RPA is further required to inhibit secondary structure formation of ssDNA that can be formed Watson-Crick pairing. Rad51-Dmc1, RecA homologs in eukaryote and their accessory factors involve in searching homolog templates to mediate strand exchange. In this study, we investigate the recombinational roles of replication protein A (RPA), which is heterotrimeric protein that is composed of RPA1, RPA2, and RPA3. Here, we investigated meiotic recombination using DNA physical analysis at the HIS4LEU2 hot spot. In rfa1-119 (K45E, N316S) cells, crossover (CO) and non-crossover (NCO) products reduced than WT. rfa1-119 delayed in single end invasion-to-double holiday junction (SEI-to-dHJ) transition and exhibits a defect in second-end capture that is also modulated by Rad52. In the further experiment, we observed that in rfa1-119 mutant, RPA could not be released in timely manner. Furthermore, rfa1-119 exhibits failure in the second end capture, implying reduction of COs and NCOs. In this talk, we will discuss more detail how RPA involves in chromatin axis association via formation of axis-bridge and why RPA is required for Rad52-mediated second-end capture progression.

Keywords: homolog interaction, meiotic recombination, replication protein A, RPA1

Procedia PDF Downloads 172
469 Towards Human-Interpretable, Automated Learning of Feedback Control for the Mixing Layer

Authors: Hao Li, Guy Y. Cornejo Maceda, Yiqing Li, Jianguo Tan, Marek Morzynski, Bernd R. Noack

Abstract:

We propose an automated analysis of the flow control behaviour from an ensemble of control laws and associated time-resolved flow snapshots. The input may be the rich database of machine learning control (MLC) optimizing a feedback law for a cost function in the plant. The proposed methodology provides (1) insights into the control landscape, which maps control laws to performance, including extrema and ridge-lines, (2) a catalogue of representative flow states and their contribution to cost function for investigated control laws and (3) visualization of the dynamics. Key enablers are classification and feature extraction methods of machine learning. The analysis is successfully applied to the stabilization of a mixing layer with sensor-based feedback driving an upstream actuator. The fluctuation energy is reduced by 26%. The control replaces unforced Kelvin-Helmholtz vortices with subsequent vortex pairing by higher-frequency Kelvin-Helmholtz structures of lower energy. These efforts target a human interpretable, fully automated analysis of MLC identifying qualitatively different actuation regimes, distilling corresponding coherent structures, and developing a digital twin of the plant.

Keywords: machine learning control, mixing layer, feedback control, model-free control

Procedia PDF Downloads 185
468 Heat Transfer and Diffusion Modelling

Authors: R. Whalley

Abstract:

The heat transfer modelling for a diffusion process will be considered. Difficulties in computing the time-distance dynamics of the representation will be addressed. Incomplete and irrational Laplace function will be identified as the computational issue. Alternative approaches to the response evaluation process will be provided. An illustration application problem will be presented. Graphical results confirming the theoretical procedures employed will be provided.

Keywords: heat, transfer, diffusion, modelling, computation

Procedia PDF Downloads 517
467 Numerical Studies for Standard Bi-Conjugate Gradient Stabilized Method and the Parallel Variants for Solving Linear Equations

Authors: Kuniyoshi Abe

Abstract:

Bi-conjugate gradient (Bi-CG) is a well-known method for solving linear equations Ax = b, for x, where A is a given n-by-n matrix, and b is a given n-vector. Typically, the dimension of the linear equation is high and the matrix is sparse. A number of hybrid Bi-CG methods such as conjugate gradient squared (CGS), Bi-CG stabilized (Bi-CGSTAB), BiCGStab2, and BiCGstab(l) have been developed to improve the convergence of Bi-CG. Bi-CGSTAB has been most often used for efficiently solving the linear equation, but we have seen the convergence behavior with a long stagnation phase. In such cases, it is important to have Bi-CG coefficients that are as accurate as possible, and the stabilization strategy, which stabilizes the computation of the Bi-CG coefficients, has been proposed. It may avoid stagnation and lead to faster computation. Motivated by a large number of processors in present petascale high-performance computing hardware, the scalability of Krylov subspace methods on parallel computers has recently become increasingly prominent. The main bottleneck for efficient parallelization is the inner products which require a global reduction. The resulting global synchronization phases cause communication overhead on parallel computers. The parallel variants of Krylov subspace methods reducing the number of global communication phases and hiding the communication latency have been proposed. However, the numerical stability, specifically, the convergence speed of the parallel variants of Bi-CGSTAB may become worse than that of the standard Bi-CGSTAB. In this paper, therefore, we compare the convergence speed between the standard Bi-CGSTAB and the parallel variants by numerical experiments and show that the convergence speed of the standard Bi-CGSTAB is faster than the parallel variants. Moreover, we propose the stabilization strategy for the parallel variants.

Keywords: bi-conjugate gradient stabilized method, convergence speed, Krylov subspace methods, linear equations, parallel variant

Procedia PDF Downloads 127
466 University Level Spanish Heritage Language Students' Use of Metaphor in Writing: Exploring Auto-Biographical Linguistic Narratives

Authors: Lorraine Ramos

Abstract:

The question of heritage language learners in foreign language classrooms has been widely debated in second language education, especially with Spanish in a U.S. Instructors of Spanish as a foreign language have brought pedagogical focus to Spanish heritage language students in order to retain, develop and maintain their first language. This paper proposes a thorough examination of the use of conceptual metaphors within autobiographical linguistic narratives as a key indicator of the writing development of advanced Spanish-language students. By pairing genre theory from Systemic Functional Linguistics with metaphor theory, this paper will examine the metaphors used by 3rd and 4th year university Spanish students within the narrative genre from a corpus of 16, 091 words. The investigation has found that heritage language students use a variety of bicultural metaphors, transferred from both languages to conceptualize their linguistic development, in addition to using metaphor in specific narrative stages as a literary strategy. Since it has been found that the metaphors used were transcultural, the use of conceptual metaphors in heritage language learners can be further examined to help these students achieve their linguistic and academic goals in the Spanish by transferring from their knowledge in English. In conclusion, by closely examining the function of student discourse through their multicultural metaphoric competence, this study provides important insights on how to enable instructors to best further their students’ writing development in the target language.

Keywords: academic writing development, heritage language learners, language attitudes and ideologies, metaphor

Procedia PDF Downloads 196
465 Numerical Study of Flapping-Wing Flight of Hummingbird Hawkmoth during Hovering: Longitudinal Dynamics

Authors: Yao Jie, Yeo Khoon Seng

Abstract:

In recent decades, flapping wing aerodynamics has attracted great interest. Understanding the physics of biological flyers such as birds and insects can help improve the performance of micro air vehicles. The present research focuses on the aerodynamics of insect-like flapping wing flight with the approach of numerical computation. Insect model of hawkmoth is adopted in the numerical study with rigid wing assumption currently. The numerical model integrates the computational fluid dynamics of the flow and active control of wing kinematics to achieve stable flight. The computation grid is a hybrid consisting of background Cartesian nodes and clouds of mesh-free grids around immersed boundaries. The generalized finite difference method is used in conjunction with single value decomposition (SVD-GFD) in computational fluid dynamics solver to study the dynamics of a free hovering hummingbird hawkmoth. The longitudinal dynamics of the hovering flight is governed by three control parameters, i.e., wing plane angle, mean positional angle and wing beating frequency. In present work, a PID controller works out the appropriate control parameters with the insect motion as input. The controller is adjusted to acquire desired maneuvering of the insect flight. The numerical scheme in present study is proven to be accurate and stable to simulate the flight of the hummingbird hawkmoth, which has relatively high Reynolds number. The PID controller is responsive to provide feedback to the wing kinematics during the hovering flight. The simulated hovering flight agrees well with the real insect flight. The present numerical study offers a promising route to investigate the free flight aerodynamics of insects, which could overcome some of the limitations of experiments.

Keywords: aerodynamics, flight control, computational fluid dynamics (CFD), flapping-wing flight

Procedia PDF Downloads 316
464 Inner Derivations of Low-Dimensional Diassociative Algebras

Authors: M. A. Fiidow, Ahmad M. Alenezi

Abstract:

In this work, we study the inner derivations of diassociative algebras in dimension two and three, an algorithmic approach is adopted for the computation of inner derivation, using some results from the derivation of finite dimensional diassociative algebras. Some basic properties of inner derivation of finite dimensional diassociative algebras are also provided.

Keywords: diassociative algebras, inner derivations, derivations, left and right operator

Procedia PDF Downloads 236
463 Learning from Dendrites: Improving the Point Neuron Model

Authors: Alexander Vandesompele, Joni Dambre

Abstract:

The diversity in dendritic arborization, as first illustrated by Santiago Ramon y Cajal, has always suggested a role for dendrites in the functionality of neurons. In the past decades, thanks to new recording techniques and optical stimulation methods, it has become clear that dendrites are not merely passive electrical components. They are observed to integrate inputs in a non-linear fashion and actively participate in computations. Regardless, in simulations of neural networks dendritic structure and functionality are often overlooked. Especially in a machine learning context, when designing artificial neural networks, point neuron models such as the leaky-integrate-and-fire (LIF) model are dominant. These models mimic the integration of inputs at the neuron soma, and ignore the existence of dendrites. In this work, the LIF point neuron model is extended with a simple form of dendritic computation. This gives the LIF neuron increased capacity to discriminate spatiotemporal input sequences, a dendritic functionality as observed in another study. Simulations of the spiking neurons are performed using the Bindsnet framework. In the common LIF model, incoming synapses are independent. Here, we introduce a dependency between incoming synapses such that the post-synaptic impact of a spike is not only determined by the weight of the synapse, but also by the activity of other synapses. This is a form of short term plasticity where synapses are potentiated or depressed by the preceding activity of neighbouring synapses. This is a straightforward way to prevent inputs from simply summing linearly at the soma. To implement this, each pair of synapses on a neuron is assigned a variable,representing the synaptic relation. This variable determines the magnitude ofthe short term plasticity. These variables can be chosen randomly or, more interestingly, can be learned using a form of Hebbian learning. We use Spike-Time-Dependent-Plasticity (STDP), commonly used to learn synaptic strength magnitudes. If all neurons in a layer receive the same input, they tend to learn the same through STDP. Adding inhibitory connections between the neurons creates a winner-take-all (WTA) network. This causes the different neurons to learn different input sequences. To illustrate the impact of the proposed dendritic mechanism, even without learning, we attach five input neurons to two output neurons. One output neuron isa regular LIF neuron, the other output neuron is a LIF neuron with dendritic relationships. Then, the five input neurons are allowed to fire in a particular order. The membrane potentials are reset and subsequently the five input neurons are fired in the reversed order. As the regular LIF neuron linearly integrates its inputs at the soma, the membrane potential response to both sequences is similar in magnitude. In the other output neuron, due to the dendritic mechanism, the membrane potential response is different for both sequences. Hence, the dendritic mechanism improves the neuron’s capacity for discriminating spa-tiotemporal sequences. Dendritic computations improve LIF neurons even if the relationships between synapses are established randomly. Ideally however, a learning rule is used to improve the dendritic relationships based on input data. It is possible to learn synaptic strength with STDP, to make a neuron more sensitive to its input. Similarly, it is possible to learn dendritic relationships with STDP, to make the neuron more sensitive to spatiotemporal input sequences. Feeding structured data to a WTA network with dendritic computation leads to a significantly higher number of discriminated input patterns. Without the dendritic computation, output neurons are less specific and may, for instance, be activated by a sequence in reverse order.

Keywords: dendritic computation, spiking neural networks, point neuron model

Procedia PDF Downloads 95
462 Continuous Processing Approaches for Tunable Asymmetric Photochemical Synthesis

Authors: Amanda C. Evans

Abstract:

Enabling technologies such as continuous processing (CP) approaches can provide the tools needed to control and manipulate reactivities and transform chemical reactions into micro-controlled in-flow processes. Traditional synthetic approaches can be radically transformed by the application of CP, facilitating the pairing of chemical methodologies with technologies from other disciplines. CP supports sustainable processes that controllably generate reaction specificity utilizing supramolecular interactions. Continuous photochemical processing is an emerging field of investigation. The use of light to drive chemical reactivity is not novel, but the controlled use of specific and tunable wavelengths of light to selectively generate molecular structure under continuous processing conditions is an innovative approach towards chemical synthesis. This investigation focuses on the use of circularly polarized (cp) light as a sustainable catalyst for the CP generation of asymmetric molecules. Chiral photolysis has already been achieved under batch, solid-phase conditions: using synchrotron-sourced cp light, asymmetric photolytic selectivities of up to 4.2% enantiomeric excess (e.e.) have been reported. In order to determine the optimal wavelengths to use for irradiation with cp light for any given molecular building block, CD and anisotropy spectra for each building block of interest have been generated in two different solvents (water, hexafluoroisopropanol) across a range of wavelengths (130-400 nm). These spectra are being used to support a series of CP experiments using cp light to generate enantioselectivity.

Keywords: anisotropy, asymmetry, flow chemistry, active pharmaceutical ingredients

Procedia PDF Downloads 123
461 Performance Comparison and Visualization of COMSOL Multiphysics, Matlab, and Fortran for Predicting the Reservoir Pressure on Oil Production in a Multiple Leases Reservoir with Boundary Element Method

Authors: N. Alias, W. Z. W. Muhammad, M. N. M. Ibrahim, M. Mohamed, H. F. S. Saipol, U. N. Z. Ariffin, N. A. Zakaria, M. S. Z. Suardi

Abstract:

This paper presents the performance comparison of some computation software for solving the boundary element method (BEM). BEM formulation is the numerical technique and high potential for solving the advance mathematical modeling to predict the production of oil well in arbitrarily shaped based on multiple leases reservoir. The limitation of data validation for ensuring that a program meets the accuracy of the mathematical modeling is considered as the research motivation of this paper. Thus, based on this limitation, there are three steps involved to validate the accuracy of the oil production simulation process. In the first step, identify the mathematical modeling based on partial differential equation (PDE) with Poisson-elliptic type to perform the BEM discretization. In the second step, implement the simulation of the 2D BEM discretization using COMSOL Multiphysic and MATLAB programming languages. In the last step, analyze the numerical performance indicators for both programming languages by using the validation of Fortran programming. The performance comparisons of numerical analysis are investigated in terms of percentage error, comparison graph and 2D visualization of pressure on oil production of multiple leases reservoir. According to the performance comparison, the structured programming in Fortran programming is the alternative software for implementing the accurate numerical simulation of BEM. As a conclusion, high-level language for numerical computation and numerical performance evaluation are satisfied to prove that Fortran is well suited for capturing the visualization of the production of oil well in arbitrarily shaped.

Keywords: performance comparison, 2D visualization, COMSOL multiphysic, MATLAB, Fortran, modelling and simulation, boundary element method, reservoir pressure

Procedia PDF Downloads 456
460 Selective Solvent Extraction of Co from Ni and Mn through Outer-Sphere Interactions

Authors: Korban Oosthuizen, Robert C. Luckay

Abstract:

Due to the growing popularity of electric vehicles and the importance of cobalt as part of the cathode material for lithium-ion batteries, demand for this metal is on the rise. Recycling of the cathode materials by means of solvent extraction is an attractive means of recovering cobalt and easing the pressure on limited natural resources. In this study, a series of straight chain and macrocyclic diamine ligands were developed for the selective recovery of cobalt from the solution containing nickel and manganese by means of solvent extraction. This combination of metals is the major cathode material used in electric vehicle batteries. The ligands can be protonated and function as ion-pairing ligands targeting the anionic [CoCl₄]²⁻, a species which is not observed for Ni or Mn. Selectivity for Co was found to be good at very high chloride concentrations and low pH. Longer chains or larger macrocycles were found to enhance selectivity, and linear chains on the amide side groups also resulted in greater selectivity over the branched groups. The cation of the chloride salt used for adjusting chloride concentrations seems to play a major role in extraction through salting-out effects. The ligands developed in this study show good selectivity for Co over Ni and Mn but require very high chloride concentrations to function. This research does, however, open the door for further investigations into using diamines as solvent extraction ligands for the recovery of cobalt from spent lithium-ion batteries.

Keywords: hydrometallurgy, solvent extraction, cobalt, lithium-ion batteries

Procedia PDF Downloads 31
459 A Study of the Effects of Nurse Innovation on Service Quality and Service Experience

Authors: Rhay-Hung Weng, Ching-Yuan Huang, Wan-Ping Chen

Abstract:

Recently, many hospitals have put much emphasis upon the development of nurse innovation. The present study aimed to clarify how nurse innovation is related to medical service quality and medical service experience. This study adopted questionnaire-survey method with nurses and customers of the inpatient wards from three Taiwanese hospitals as the research subjects. After pairing, there were 294 valid questionnaires. Hierarchical regression analysis was utilized to test the possible impact of nurse innovation on medical service quality and experience. In terms of the dimensions of nurse innovation, “innovation behavior” ranked the highest (3.24), followed by knowledge creation and innovation diffusion; in terms of the degree of the medical service quality, 'reliability' ranked the highest (4.35). As for the degree of the medical service experience, 'feel experience' ranked the highest (4.44). All dimensions of nurse innovation have no significant effects on medical service quality and medical service experience. Of these three dimensions of nurse innovation, the level of innovation behavior was perceived by the nurses as the highest. The study found that nurse innovation has no significant effects on medical service quality and medical service experience. Managers shall provide sufficient resources and budget for fostering innovation development and encourage their nurses to develop nursing innovation for patents. The education and training courses on “patient-centered ” shall be enhanced among hospital nurses. Health care managers shall also explore the difficulties about innovation diffusion and find the solutions for nurses.

Keywords: innovation, employee innovative behavior, service quality, service experience

Procedia PDF Downloads 299
458 A Novel Algorithm for Production Scheduling

Authors: Ali Mohammadi Bolban Abad, Fariborz Ahmadi

Abstract:

Optimization in manufacture is a method to use limited resources to obtain the best performance and reduce waste. In this paper a new algorithm based on eurygaster life is introduced to obtain a plane in which task order and completion time of resources are defined. Evaluation results show our approach has less make span when the resources are allocated with some products in comparison to genetic algorithm.

Keywords: evolutionary computation, genetic algorithm, particle swarm optimization, NP-Hard problems, production scheduling

Procedia PDF Downloads 345
457 Genetic Programming: Principles, Applications and Opportunities for Hydrological Modelling

Authors: Oluwaseun K. Oyebode, Josiah A. Adeyemo

Abstract:

Hydrological modelling plays a crucial role in the planning and management of water resources, most especially in water stressed regions where the need to effectively manage the available water resources is of critical importance. However, due to the complex, nonlinear and dynamic behaviour of hydro-climatic interactions, achieving reliable modelling of water resource systems and accurate projection of hydrological parameters are extremely challenging. Although a significant number of modelling techniques (process-based and data-driven) have been developed and adopted in that regard, the field of hydrological modelling is still considered as one that has sluggishly progressed over the past decades. This is majorly as a result of the identification of some degree of uncertainty in the methodologies and results of techniques adopted. In recent times, evolutionary computation (EC) techniques have been developed and introduced in response to the search for efficient and reliable means of providing accurate solutions to hydrological related problems. This paper presents a comprehensive review of the underlying principles, methodological needs and applications of a promising evolutionary computation modelling technique – genetic programming (GP). It examines the specific characteristics of the technique which makes it suitable to solving hydrological modelling problems. It discusses the opportunities inherent in the application of GP in water related-studies such as rainfall estimation, rainfall-runoff modelling, streamflow forecasting, sediment transport modelling, water quality modelling and groundwater modelling among others. Furthermore, the means by which such opportunities could be harnessed in the near future are discussed. In all, a case for total embracement of GP and its variants in hydrological modelling studies is made so as to put in place strategies that would translate into achieving meaningful progress as it relates to modelling of water resource systems, and also positively influence decision-making by relevant stakeholders.

Keywords: computational modelling, evolutionary algorithms, genetic programming, hydrological modelling

Procedia PDF Downloads 259
456 Numerical Simulation of Filtration Gas Combustion: Front Propagation Velocity

Authors: Yuri Laevsky, Tatyana Nosova

Abstract:

The phenomenon of filtration gas combustion (FGC) had been discovered experimentally at the beginning of 80’s of the previous century. It has a number of important applications in such areas as chemical technologies, fire-explosion safety, energy-saving technologies, oil production. From the physical point of view, FGC may be defined as the propagation of region of gaseous exothermic reaction in chemically inert porous medium, as the gaseous reactants seep into the region of chemical transformation. The movement of the combustion front has different modes, and this investigation is focused on the low-velocity regime. The main characteristic of the process is the velocity of the combustion front propagation. Computation of this characteristic encounters substantial difficulties because of the strong heterogeneity of the process. The mathematical model of FGC is formed by the energy conservation laws for the temperature of the porous medium and the temperature of gas and the mass conservation law for the relative concentration of the reacting component of the gas mixture. In this case the homogenization of the model is performed with the use of the two-temperature approach when at each point of the continuous medium we specify the solid and gas phases with a Newtonian heat exchange between them. The construction of a computational scheme is based on the principles of mixed finite element method with the usage of a regular mesh. The approximation in time is performed by an explicit–implicit difference scheme. Special attention was given to determination of the combustion front propagation velocity. Straight computation of the velocity as grid derivative leads to extremely unstable algorithm. It is worth to note that the term ‘front propagation velocity’ makes sense for settled motion when some analytical formulae linking velocity and equilibrium temperature are correct. The numerical implementation of one of such formulae leading to the stable computation of instantaneous front velocity has been proposed. The algorithm obtained has been applied in subsequent numerical investigation of the FGC process. This way the dependence of the main characteristics of the process on various physical parameters has been studied. In particular, the influence of the combustible gas mixture consumption on the front propagation velocity has been investigated. It also has been reaffirmed numerically that there is an interval of critical values of the interfacial heat transfer coefficient at which a sort of a breakdown occurs from a slow combustion front propagation to a rapid one. Approximate boundaries of such an interval have been calculated for some specific parameters. All the results obtained are in full agreement with both experimental and theoretical data, confirming the adequacy of the model and the algorithm constructed. The presence of stable techniques to calculate the instantaneous velocity of the combustion wave allows considering the semi-Lagrangian approach to the solution of the problem.

Keywords: filtration gas combustion, low-velocity regime, mixed finite element method, numerical simulation

Procedia PDF Downloads 272
455 Microgrid Design Under Optimal Control With Batch Reinforcement Learning

Authors: Valentin Père, Mathieu Milhé, Fabien Baillon, Jean-Louis Dirion

Abstract:

Microgrids offer potential solutions to meet the need for local grid stability and increase isolated networks autonomy with the integration of intermittent renewable energy production and storage facilities. In such a context, sizing production and storage for a given network is a complex task, highly depending on input data such as power load profile and renewable resource availability. This work aims at developing an operating cost computation methodology for different microgrid designs based on the use of deep reinforcement learning (RL) algorithms to tackle the optimal operation problem in stochastic environments. RL is a data-based sequential decision control method based on Markov decision processes that enable the consideration of random variables for control at a chosen time scale. Agents trained via RL constitute a promising class of Energy Management Systems (EMS) for the operation of microgrids with energy storage. Microgrid sizing (or design) is generally performed by minimizing investment costs and operational costs arising from the EMS behavior. The latter might include economic aspects (power purchase, facilities aging), social aspects (load curtailment), and ecological aspects (carbon emissions). Sizing variables are related to major constraints on the optimal operation of the network by the EMS. In this work, an islanded mode microgrid is considered. Renewable generation is done with photovoltaic panels; an electrochemical battery ensures short-term electricity storage. The controllable unit is a hydrogen tank that is used as a long-term storage unit. The proposed approach focus on the transfer of agent learning for the near-optimal operating cost approximation with deep RL for each microgrid size. Like most data-based algorithms, the training step in RL leads to important computer time. The objective of this work is thus to study the potential of Batch-Constrained Q-learning (BCQ) for the optimal sizing of microgrids and especially to reduce the computation time of operating cost estimation in several microgrid configurations. BCQ is an off-line RL algorithm that is known to be data efficient and can learn better policies than on-line RL algorithms on the same buffer. The general idea is to use the learned policy of agents trained in similar environments to constitute a buffer. The latter is used to train BCQ, and thus the agent learning can be performed without update during interaction sampling. A comparison between online RL and the presented method is performed based on the score by environment and on the computation time.

Keywords: batch-constrained reinforcement learning, control, design, optimal

Procedia PDF Downloads 85
454 Using Derivative Free Method to Improve the Error Estimation of Numerical Quadrature

Authors: Chin-Yun Chen

Abstract:

Numerical integration is an essential tool for deriving different physical quantities in engineering and science. The effectiveness of a numerical integrator depends on different factors, where the crucial one is the error estimation. This work presents an error estimator that combines a derivative free method to improve the performance of verified numerical quadrature.

Keywords: numerical quadrature, error estimation, derivative free method, interval computation

Procedia PDF Downloads 425
453 Ideology and Lexicogrammar: Discourse Against the Power in Lyrical Texts (XIII, XVII and XX Centuries)

Authors: Ulisses Tadeu Vaz de Oliveira

Abstract:

The development of multifunctional studies in the theoretical-methodological perspective of the Systemic-Functional Grammar (SFG) and the increasing number of critical literary studies have introduced new opportunities for the study of ideologies and societies, but also brought up new challenges across and within many areas. In this regard, the Critical Linguistics researches allow a form of pairing a textual linguistic analysis method (micro level) with a social language theory in political and ideological processes (macro level), presented in the literature. This presentation will report on strategies to criticize power holders in literary productions from three distinct eras, namely: (a) Satirical Galego-Portuguese chants of Gil Pérez Conde (thirteenth century), (b) Poems of Gregorio de Matos Guerra (seventeenth century), and (c) Songs of Chico Buarque de Holanda (twentieth century). The analysis of these productions is based on the SFG proposals, which considers the clause as a social event. Therefore, the structure serves to realize three concurrent meanings (metafunctions): Ideational, Interpersonal and Textual. The presenter aims to shed light on the core issues relevant to the successes of the authors to criticize authorities in repressive times while caring about face-threatening and politeness. The effective and meaningful critical discourse was a way of moving the society`s chains towards new ideologies reflected in the lexicogrammatical choices made and the rhetorical functions of the persuasive structures used by the authors.

Keywords: ideology, literature, persuasion, systemic-functional grammar

Procedia PDF Downloads 381
452 Application of the Standard Deviation in Regulating Design Variation of Urban Solutions Generated through Evolutionary Computation

Authors: Mohammed Makki, Milad Showkatbakhsh, Aiman Tabony

Abstract:

Computational applications of natural evolutionary processes as problem-solving tools have been well established since the mid-20th century. However, their application within architecture and design has only gained ground in recent years, with an increasing number of academics and professionals in the field electing to utilize evolutionary computation to address problems comprised from multiple conflicting objectives with no clear optimal solution. Recent advances in computer science and its consequent constructive influence on the architectural discourse has led to the emergence of multiple algorithmic processes capable of simulating the evolutionary process in nature within an efficient timescale. Many of the developed processes of generating a population of candidate solutions to a design problem through an evolutionary based stochastic search process are often driven through the application of both environmental and architectural parameters. These methods allow for conflicting objectives to be simultaneously, independently, and objectively optimized. This is an essential approach in design problems with a final product that must address the demand of a multitude of individuals with various requirements. However, one of the main challenges encountered through the application of an evolutionary process as a design tool is the ability for the simulation to maintain variation amongst design solutions in the population while simultaneously increasing in fitness. This is most commonly known as the ‘golden rule’ of balancing exploration and exploitation over time; the difficulty of achieving this balance in the simulation is due to the tendency of either variation or optimization being favored as the simulation progresses. In such cases, the generated population of candidate solutions has either optimized very early in the simulation, or has continued to maintain high levels of variation to which an optimal set could not be discerned; thus, providing the user with a solution set that has not evolved efficiently to the objectives outlined in the problem at hand. As such, the experiments presented in this paper seek to achieve the ‘golden rule’ by incorporating a mathematical fitness criterion for the development of an urban tissue comprised from the superblock as its primary architectural element. The mathematical value investigated in the experiments is the standard deviation factor. Traditionally, the standard deviation factor has been used as an analytical value rather than a generative one, conventionally used to measure the distribution of variation within a population by calculating the degree by which the majority of the population deviates from the mean. A higher standard deviation value delineates a higher number of the population is clustered around the mean and thus limited variation within the population, while a lower standard deviation value is due to greater variation within the population and a lack of convergence towards an optimal solution. The results presented will aim to clarify the extent to which the utilization of the standard deviation factor as a fitness criterion can be advantageous to generating fitter individuals in a more efficient timeframe when compared to conventional simulations that only incorporate architectural and environmental parameters.

Keywords: architecture, computation, evolution, standard deviation, urban

Procedia PDF Downloads 103
451 Timetabling for Interconnected LRT Lines: A Package Solution Based on a Real-world Case

Authors: Huazhen Lin, Ruihua Xu, Zhibin Jiang

Abstract:

In this real-world case, timetabling the LRT network as a whole is rather challenging for the operator: they are supposed to create a timetable to avoid various route conflicts manually while satisfying a given interval and the number of rolling stocks, but the outcome is not satisfying. Therefore, the operator adopts a computerised timetabling tool, the Train Plan Maker (TPM), to cope with this problem. However, with various constraints in the dual-line network, it is still difficult to find an adequate pairing of turnback time, interval and rolling stocks’ number, which requires extra manual intervention. Aiming at current problems, a one-off model for timetabling is presented in this paper to simplify the procedure of timetabling. Before the timetabling procedure starts, this paper presents how the dual-line system with a ring and several branches is turned into a simpler structure. Then, a non-linear programming model is presented in two stages. In the first stage, the model sets a series of constraints aiming to calculate a proper timing for coordinating two lines by adjusting the turnback time at termini. Then, based on the result of the first stage, the model introduces a series of inequality constraints to avoid various route conflicts. With this model, an analysis is conducted to reveal the relation between the ratio of trains in different directions and the possible minimum interval, observing that the more imbalance the ratio is, the less possible to provide frequent service under such strict constraints.

Keywords: light rail transit (LRT), non-linear programming, railway timetabling, timetable coordination

Procedia PDF Downloads 32
450 Fault Tolerant and Testable Designs of Reversible Sequential Building Blocks

Authors: Vishal Pareek, Shubham Gupta, Sushil Chandra Jain

Abstract:

With increasing high-speed computation demand the power consumption, heat dissipation and chip size issues are posing challenges for logic design with conventional technologies. Recovery of bit loss and bit errors is other issues that require reversibility and fault tolerance in the computation. The reversible computing is emerging as an alternative to conventional technologies to overcome the above problems and helpful in a diverse area such as low-power design, nanotechnology, quantum computing. Bit loss issue can be solved through unique input-output mapping which require reversibility and bit error issue require the capability of fault tolerance in design. In order to incorporate reversibility a number of combinational reversible logic based circuits have been developed. However, very few sequential reversible circuits have been reported in the literature. To make the circuit fault tolerant, a number of fault model and test approaches have been proposed for reversible logic. In this paper, we have attempted to incorporate fault tolerance in sequential reversible building blocks such as D flip-flop, T flip-flop, JK flip-flop, R-S flip-flop, Master-Slave D flip-flop, and double edge triggered D flip-flop by making them parity preserving. The importance of this proposed work lies in the fact that it provides the design of reversible sequential circuits completely testable for any stuck-at fault and single bit fault. In our opinion our design of reversible building blocks is superior to existing designs in term of quantum cost, hardware complexity, constant input, garbage output, number of gates and design of online testable D flip-flop have been proposed for the first time. We hope our work can be extended for building complex reversible sequential circuits.

Keywords: parity preserving gate, quantum computing, fault tolerance, flip-flop, sequential reversible logic

Procedia PDF Downloads 514
449 Detection of Curvilinear Structure via Recursive Anisotropic Diffusion

Authors: Sardorbek Numonov, Hyohun Kim, Dongwha Shin, Yeonseok Kim, Ji-Su Ahn, Dongeun Choi, Byung-Woo Hong

Abstract:

The detection of curvilinear structures often plays an important role in the analysis of images. In particular, it is considered as a crucial step for the diagnosis of chronic respiratory diseases to localize the fissures in chest CT imagery where the lung is divided into five lobes by the fissures that are characterized by linear features in appearance. However, the characteristic linear features for the fissures are often shown to be subtle due to the high intensity variability, pathological deformation or image noise involved in the imaging procedure, which leads to the uncertainty in the quantification of anatomical or functional properties of the lung. Thus, it is desired to enhance the linear features present in the chest CT images so that the distinctiveness in the delineation of the lobe is improved. We propose a recursive diffusion process that prefers coherent features based on the analysis of structure tensor in an anisotropic manner. The local image features associated with certain scales and directions can be characterized by the eigenanalysis of the structure tensor that is often regularized via isotropic diffusion filters. However, the isotropic diffusion filters involved in the computation of the structure tensor generally blur geometrically significant structure of the features leading to the degradation of the characteristic power in the feature space. Thus, it is required to take into consideration of local structure of the feature in scale and direction when computing the structure tensor. We apply an anisotropic diffusion in consideration of scale and direction of the features in the computation of the structure tensor that subsequently provides the geometrical structure of the features by its eigenanalysis that determines the shape of the anisotropic diffusion kernel. The recursive application of the anisotropic diffusion with the kernel the shape of which is derived from the structure tensor leading to the anisotropic scale-space where the geometrical features are preserved via the eigenanalysis of the structure tensor computed from the diffused image. The recursive interaction between the anisotropic diffusion based on the geometry-driven kernels and the computation of the structure tensor that determines the shape of the diffusion kernels yields a scale-space where geometrical properties of the image structure are effectively characterized. We apply our recursive anisotropic diffusion algorithm to the detection of curvilinear structure in the chest CT imagery where the fissures present curvilinear features and define the boundary of lobes. It is shown that our algorithm yields precise detection of the fissures while overcoming the subtlety in defining the characteristic linear features. The quantitative evaluation demonstrates the robustness and effectiveness of the proposed algorithm for the detection of fissures in the chest CT in terms of the false positive and the true positive measures. The receiver operating characteristic curves indicate the potential of our algorithm as a segmentation tool in the clinical environment. This work was supported by the MISP(Ministry of Science and ICT), Korea, under the National Program for Excellence in SW (20170001000011001) supervised by the IITP(Institute for Information and Communications Technology Promotion).

Keywords: anisotropic diffusion, chest CT imagery, chronic respiratory disease, curvilinear structure, fissure detection, structure tensor

Procedia PDF Downloads 201
448 Fog Computing- Network Based Computing

Authors: Navaneeth Krishnan, Chandan N. Bhagwat, Aparajit P. Utpat

Abstract:

Cloud Computing provides us a means to upload data and use applications over the internet. As the number of devices connecting to the cloud grows, there is undue pressure on the cloud infrastructure. Fog computing or Network Based Computing or Edge Computing allows to move a part of the processing in the cloud to the network devices present along the node to the cloud. Therefore the nodes connected to the cloud have a better response time. This paper proposes a method of moving the computation from the cloud to the network by introducing an android like appstore on the networking devices.

Keywords: cloud computing, fog computing, network devices, appstore

Procedia PDF Downloads 341
447 A Parallel Implementation of k-Means in MATLAB

Authors: Dimitris Varsamis, Christos Talagkozis, Alkiviadis Tsimpiris, Paris Mastorocostas

Abstract:

The aim of this work is the parallel implementation of k-means in MATLAB, in order to reduce the execution time. Specifically, a new function in MATLAB for serial k-means algorithm is developed, which meets all the requirements for the conversion to a function in MATLAB with parallel computations. Additionally, two different variants for the definition of initial values are presented. In the sequel, the parallel approach is presented. Finally, the performance tests for the computation times respect to the numbers of features and classes are illustrated.

Keywords: K-means algorithm, clustering, parallel computations, Matlab

Procedia PDF Downloads 338