Search results for: array redistribution
662 Array Type Miniaturized Ultrasonic Sensors for Detecting Sinkhole in the City
Authors: Won Young Choi, Kwan Kyu Park
Abstract:
Recently, the road depression happening in the urban area is different from the cause of the sink hole and the generation mechanism occurring in the limestone area. The main cause of sinkholes occurring in the city center is the loss of soil due to the damage of old underground buried materials and groundwater discharge due to large underground excavation works. The method of detecting the sinkhole in the urban area is mostly using the Ground Penetration Radar (GPR). However, it is challenging to implement compact system and detecting watery state since it is based on electromagnetic waves. Although many ultrasonic underground detection studies have been conducted, near-ground detection (several tens of cm to several meters) has been developed for bulk systems using geophones as a receiver. The goal of this work is to fabricate a miniaturized sinkhole detecting system based on low-cost ultrasonic transducers of 40 kHz resonant frequency with high transmission pressure and receiving sensitivity. Motived by biomedical ultrasonic imaging methods, we detect air layers below the ground such as asphalt through the pulse-echo method. To improve image quality using multi-channel, linear array system is implemented, and image is acquired by classical synthetic aperture imaging method. We present the successful feasibility test of multi-channel sinkhole detector based on ultrasonic transducer. In this work, we presented and analyzed image results which are imaged by single channel pulse-echo imaging, synthetic aperture imaging.Keywords: road depression, sinkhole, synthetic aperture imaging, ultrasonic transducer
Procedia PDF Downloads 144661 One Step Further: Pull-Process-Push Data Processing
Authors: Romeo Botes, Imelda Smit
Abstract:
In today’s modern age of technology vast amounts of data needs to be processed in real-time to keep users satisfied. This data comes from various sources and in many formats, including electronic and mobile devices such as GPRS modems and GPS devices. They make use of different protocols including TCP, UDP, and HTTP/s for data communication to web servers and eventually to users. The data obtained from these devices may provide valuable information to users, but are mostly in an unreadable format which needs to be processed to provide information and business intelligence. This data is not always current, it is mostly historical data. The data is not subject to implementation of consistency and redundancy measures as most other data usually is. Most important to the users is that the data are to be pre-processed in a readable format when it is entered into the database. To accomplish this, programmers build processing programs and scripts to decode and process the information stored in databases. Programmers make use of various techniques in such programs to accomplish this, but sometimes neglect the effect some of these techniques may have on database performance. One of the techniques generally used,is to pull data from the database server, process it and push it back to the database server in one single step. Since the processing of the data usually takes some time, it keeps the database busy and locked for the period of time that the processing takes place. Because of this, it decreases the overall performance of the database server and therefore the system’s performance. This paper follows on a paper discussing the performance increase that may be achieved by utilizing array lists along with a pull-process-push data processing technique split in three steps. The purpose of this paper is to expand the number of clients when comparing the two techniques to establish the impact it may have on performance of the CPU storage and processing time.Keywords: performance measures, algorithm techniques, data processing, push data, process data, array list
Procedia PDF Downloads 245660 Optimization of MAG Welding Process Parameters Using Taguchi Design Method on Dead Mild Steel
Authors: Tadele Tesfaw, Ajit Pal Singh, Abebaw Mekonnen Gezahegn
Abstract:
Welding is a basic manufacturing process for making components or assemblies. Recent welding economics research has focused on developing the reliable machinery database to ensure optimum production. Research on welding of materials like steel is still critical and ongoing. Welding input parameters play a very significant role in determining the quality of a weld joint. The metal active gas (MAG) welding parameters are the most important factors affecting the quality, productivity and cost of welding in many industrial operations. The aim of this study is to investigate the optimization process parameters for metal active gas welding for 60x60x5mm dead mild steel plate work-piece using Taguchi method to formulate the statistical experimental design using semi-automatic welding machine. An experimental study was conducted at Bishoftu Automotive Industry, Bishoftu, Ethiopia. This study presents the influence of four welding parameters (control factors) like welding voltage (volt), welding current (ampere), wire speed (m/min.), and gas (CO2) flow rate (lit./min.) with three different levels for variability in the welding hardness. The objective functions have been chosen in relation to parameters of MAG welding i.e., welding hardness in final products. Nine experimental runs based on an L9 orthogonal array Taguchi method were performed. An orthogonal array, signal-to-noise (S/N) ratio and analysis of variance (ANOVA) are employed to investigate the welding characteristics of dead mild steel plate and used in order to obtain optimum levels for every input parameter at 95% confidence level. The optimal parameters setting was found is welding voltage at 22 volts, welding current at 125 ampere, wire speed at 2.15 m/min and gas flow rate at 19 l/min by using the Taguchi experimental design method within the constraints of the production process. Finally, six conformations welding have been carried out to compare the existing values; the predicated values with the experimental values confirm its effectiveness in the analysis of welding hardness (quality) in final products. It is found that welding current has a major influence on the quality of welded joints. Experimental result for optimum setting gave a better hardness of welding condition than initial setting. This study is valuable for different material and thickness variation of welding plate for Ethiopian industries.Keywords: Weld quality, metal active gas welding, dead mild steel plate, orthogonal array, analysis of variance, Taguchi method
Procedia PDF Downloads 482659 High-Frequency Modulation of Light-Emitting Diodes for New Ultraviolet Communications
Authors: Meng-Chyi Wu, Bonn Lin, Jyun-Hao Liao, Chein-Ju Chen, Yu-Cheng Jhuang, Mau-Phon Houng, Fang-Hsing Wang, Min-Chu Liu, Cheng-Fu Yang, Cheng-Shong Hong
Abstract:
Since the use of wireless communications has become critical nowadays, the available RF spectrum has become limited. Ultraviolet (UV) communication system can alleviate the spectrum constraint making UV communication system a potential alternative to future communication demands. Also, UV links can provide faster communication rate and can be used in combination with existing RF communication links, providing new communications diversity with higher user capacity. The UV region of electromagnetic spectrum has been of interest to detector, imaging and communication technologies because the stratospheric ozone layer effectively absorbs some solar UV radiation from reaching the earth surface. The wavebands where most of UV radiation is absorbed by the ozone are commonly known as the solar blind region. By operating in UV-C band (200-280 nm) the communication system can minimize the transmission power consumption since it will have less radiation noise. UV communication uses the UV ray as the medium. Electric signal is carried on this band after being modulated and then be transmitted within the atmosphere as channel. Though the background noise of UV-C communication is very low owing to the solar-blind feature, it leads to a large propagation loss. The 370 nm UV provides a much lower propagation loss than that the UV-C does and the recent device technology for UV source on this band is more mature. The fabricated 370 nm AlGaN light-emitting diodes (LEDs) with an aperture size of 45 m exhibit a modulation bandwidth of 165 MHz at 30 mA and a high power of 7 W/cm2 at 230 A/cm2. In order to solve the problem of low power in single UV LED, a UV LED array is presented in.Keywords: ultraviolet (UV) communication, light-emitting diodes (LEDs), modulation bandwidth, LED array, 370 nm
Procedia PDF Downloads 414658 Modeling of the Mechanism of Ion Channel Opening of the Visual Receptor's Rod on the Light and Allosteric Effect of Rhodopsin in the Phosphorylation Process
Authors: N. S. Vassilieva-Vashakmadze, R. A. Gakhokidze, I. M. Khachatryan
Abstract:
In the first part of the paper it is shown that both the depolarization of the cytoplasmic membrane of rods observed in invertebrates and hyperpolarization characteristic of vertebrates on the light may activate the functioning of ion (Na+) channels of cytoplasmic membrane of rods and thus provide the emergence of nerve impulse and its transfer to the neighboring neuron etc. In the second part, using the quantum mechanical program for modeling of the molecular processes, we got a clear picture demonstrating the effect of charged phosphate groups on the protein components of α-helical subunits of the visual rhodopsin receptor. The analysis shows that the phosphorylation of terminal amino acid of seventh α-helical subunits of the visual rhodopsin causes a redistribution of electron density on the atoms, i.e. polarization of subunits, also the changing the configuration of the nuclear subsystem, which corresponds to the deformation process in the molecule. Based on the use of models it can be concluded that this system has an internal relationship between polarization and deformation processes that indicates on the allosteric effect. The allosteric effect is based on quantum-mechanical principle of the self-consistency of the molecules.Keywords: membrane potential, ion channels, visual rhodopsin, allosteric effect
Procedia PDF Downloads 271657 Constant Factor Approximation Algorithm for p-Median Network Design Problem with Multiple Cable Types
Authors: Chaghoub Soraya, Zhang Xiaoyan
Abstract:
This research presents the first constant approximation algorithm to the p-median network design problem with multiple cable types. This problem was addressed with a single cable type and there is a bifactor approximation algorithm for the problem. To the best of our knowledge, the algorithm proposed in this paper is the first constant approximation algorithm for the p-median network design with multiple cable types. The addressed problem is a combination of two well studied problems which are p-median problem and network design problem. The introduced algorithm is a random sampling approximation algorithm of constant factor which is conceived by using some random sampling techniques form the literature. It is based on a redistribution Lemma from the literature and a steiner tree problem as a subproblem. This algorithm is simple, and it relies on the notions of random sampling and probability. The proposed approach gives an approximation solution with one constant ratio without violating any of the constraints, in contrast to the one proposed in the literature. This paper provides a (21 + 2)-approximation algorithm for the p-median network design problem with multiple cable types using random sampling techniques.Keywords: approximation algorithms, buy-at-bulk, combinatorial optimization, network design, p-median
Procedia PDF Downloads 204656 Monoallelic and Biallelic Deletions of 13q14 in a Group of 36 CLL Patients Investigated by CGH Haematological Cancer and SNP Array (8x60K)
Authors: B. Grygalewicz, R. Woroniecka, J. Rygier, K. Borkowska, A. Labak, B. Nowakowska, B. Pienkowska-Grela
Abstract:
Introduction: Chronic lymphocytic leukemia (CLL) is the most common form of adult leukemia in the Western world. Hemizygous and or homozygous loss at 13q14 occur in more than half of cases and constitute the most frequent chromosomal abnormality in CLL. It is believed that deletions 13q14 play a role in CLL pathogenesis. Two microRNA genes miR-15a and miR- 16-1 are targets of 13q14 deletions and plays a tumor suppressor role by targeting antiapoptotic BCL2 gene. Deletion size, as a single change detected in FISH analysis, has haprognostic significance. Patients with small deletions, without RB1 gene involvement, have the best prognosis and the longest overall survival time (OS 133 months). In patients with bigger deletion region, containing RB1 gene, prognosis drops to intermediate, like in patients with normal karyotype and without changes in FISH with overall survival 111 months. Aim: Precise delineation of 13q14 deletions regions in two groups of CLL patients, with mono- and biallelic deletions and qualifications of their prognostic significance. Methods: Detection of 13q14 deletions was performed by FISH analysis with CLL probe panel (D13S319, LAMP1, TP53, ATM, CEP-12). Accurate deletion size detection was performed by CGH Haematological Cancer and SNP array (8x60K). Results: Our investigated group of CLL patients with the 13q14 deletion, detected by FISH analysis, comprised two groups: 18 patients with monoallelic deletions and 18 patients with biallelic deletions. In FISH analysis, in the monoallelic group the range of cells with deletion, was 43% to 97%, while in biallelic group deletion was detected in 11% to 94% of cells. Microarray analysis revealed precise deletion regions. In the monoallelic group, the range of size was 348,12 Kb to 34,82 Mb, with median deletion size 7,93 Mb. In biallelic group discrepancy of total deletions, size was 135,27 Kb to 33,33 Mb, with median deletion size 2,52 Mb. The median size of smaller deletion regions on one copy chromosome 13 was 1,08 Mb while the average region of bigger deletion on the second chromosome 13 was 4,04 Mb. In the monoallelic group, in 8/18 deletion region covered RB1 gene. In the biallelic group, in 4/18 cases, revealed deletion on one copy of biallelic deletion and in 2/18 showed deletion of RB1 gene on both deleted 13q14 regions. All minimal deleted regions included miR-15a and miR-16-1 genes. Genetic results will be correlated with clinical data. Conclusions: Application of CGH microarrays technique in CLL allows accurately delineate the size of 13q14 deletion regions, what have a prognostic value. All deleted regions included miR15a and miR-16-1, what confirms the essential role of these genes in CLL pathogenesis. In our investigated groups of CLL patients with mono- and biallelic 13q14 deletions, patients with biallelic deletion presented smaller deletion sizes (2,52 Mb vs 7,93 Mb), what is connected with better prognosis.Keywords: CLL, deletion 13q14, CGH microarrays, SNP array
Procedia PDF Downloads 256655 SnSₓ, Cu₂ZnSnS₄ Nanostructured Thin Layers for Thin-Film Solar Cells
Authors: Elena A. Outkina, Marina V. Meledina, Aliaksandr A. Khodin
Abstract:
Nanostructured thin films of SnSₓ, Cu₂ZnSnS₄ (CZTS) semiconductors were fabricated by chemical processing to produce thin-film photoactive layers for photocells as a prospective lowest-cost and environment-friendly alternative to Si, Cu(In, Ga)Se₂, and other traditional solar cells materials. To produce SnSₓ layers, the modified successive ionic layer adsorption and reaction (SILAR) technique were investigated, including successive cyclic dipping into Na₂S solution and SnCl₂, NaCl, triethanolamine solution. To fabricate CZTS layers, the cyclic dipping into CuSO₄ with ZnSO₄, SnCl₂, and Na₂S solutions was used with intermediate rinsing in distilled water. The nano-template aluminum/alumina substrate was used to control deposition processes. Micromorphology and optical characteristics of the fabricated layers have been investigated. Analysis of 2D-like layers deposition features using nano-template substrate is presented, including the effect of nanotips in a template on surface charge redistribution and transport.Keywords: kesterite, nanotemplate, SILAR, solar cell, tin sulphide
Procedia PDF Downloads 143654 Shaking Force Balancing of Mechanisms: An Overview
Authors: Vigen Arakelian
Abstract:
The balancing of mechanisms is a well-known problem in the field of mechanical engineering because the variable dynamic loads cause vibrations, as well as noise, wear and fatigue of the machines. A mechanical system with unbalance shaking force and shaking moment transmits substantial vibration to the frame. Therefore, the objective of the balancing is to cancel or reduce the variable dynamic reactions transmitted to the frame. The resolution of this problem consists in the balancing of the shaking force and shaking moment. It can be fully or partially, by internal mass redistribution via adding counterweights or by modification of the mechanism's architecture via adding auxiliary structures. The balancing problems are of continue interest to researchers. Several laboratories around the world are very active in this area and new results are published regularly. However, despite its ancient history, mechanism balancing theory continues to be developed and new approaches and solutions are constantly being reported. Various surveys have been published that disclose particularities of balancing methods. The author believes that this is an appropriate moment to present a state of the art of the shaking force balancing studies completed by new research results. This paper presents an overview of methods devoted to the shaking force balancing of mechanisms, as well as the historical aspects of the origins and the evolution of the balancing theory of mechanisms.Keywords: inertial forces, shaking forces, balancing, dynamics, mechanism design
Procedia PDF Downloads 127653 Flexible and Color Tunable Inorganic Light Emitting Diode Array for High Resolution Optogenetic Devices
Authors: Keundong Lee, Dongha Yoo, Youngbin Tchoe, Gyu-Chul Yi
Abstract:
Light emitting diode (LED) array is an ideal optical stimulation tool for optogenetics, which controls inhibition and excitation of specific neurons with light-sensitive ion channels or pumps. Although a fiber-optic cable with an external light source, either a laser or LED mechanically connected to the end of the fiber-optic cable has widely been used for illumination on neural tissue, a new approach to use micro LEDs (µLEDs) has recently been demonstrated. The LEDs can be placed directly either on the cortical surface or within the deep brain using a penetrating depth probe. Accordingly, this method would not need a permanent opening in the skull if the LEDs are integrated with miniature electrical power source and wireless communication. In addition, multiple color generation from single µLED cell would enable to excite and/or inhibit neurons in localized regions. Here, we demonstrate flexible and color tunable µLEDs for the optogenetic device applications. The flexible and color tunable LEDs was fabricated using multifaceted gallium nitride (GaN) nanorod arrays with GaN nanorods grown on InxGa1−xN/GaN single quantum well structures (SQW) anisotropically formed on the nanorod tips and sidewalls. For various electroluminescence (EL) colors, current injection paths were controlled through a continuous p-GaN layer depending on the applied bias voltage. The electric current was injected through different thickness and composition, thus changing the color of light from red to blue that the LED emits. We believe that the flexible and color tunable µLEDs enable us to control activities of the neuron by emitting various colors from the single µLED cell.Keywords: light emitting diode, optogenetics, graphene, flexible optoelectronics
Procedia PDF Downloads 211652 Estimation of Seismic Ground Motion and Shaking Parameters Based on Microtremor Measurements at Palu City, Central Sulawesi Province, Indonesia
Authors: P. S. Thein, S. Pramumijoyo, K. S. Brotopuspito, J. Kiyono, W. Wilopo, A. Furukawa, A. Setianto
Abstract:
In this study, we estimated the seismic ground motion parameters based on microtremor measurements at Palu City. Several earthquakes have struck along the Palu-Koro Fault during recent years. The USGS epicenter, magnitude Mw 6.3 event that occurred on January 23, 2005 caused several casualties. We conducted a microtremor survey to estimate the strong ground motion distribution during the earthquake. From this survey we produced a map of the peak ground acceleration, velocity, seismic vulnerability index and ground shear strain maps in Palu City. We performed single observations of microtremor at 151 sites in Palu City. We also conducted 8-site microtremors array investigation to gain a representative determination of the soil condition of subsurface structures in Palu City. From the array observations, Palu City corresponds to relatively soil condition with Vs ≤ 300 m/s, the predominant periods due to horizontal vertical ratios (HVSRs) are in the range of 0.4 to 1.8 s and the frequency are in the range of 0.7 to 3.3 Hz. Strong ground motions of the Palu area were predicted based on the empirical stochastic green’s function method. Peak ground acceleration and velocity becomes more than 400 gal and 30 kine in some areas, which causes severe damage for buildings in high probability. Microtremor survey results showed that in hilly areas had low seismic vulnerability index and ground shear strain, whereas in coastal alluvium was composed of material having a high seismic vulnerability and ground shear strain indication.Keywords: Palu-Koro fault, microtremor, peak ground acceleration, peak ground velocity, seismic vulnerability index
Procedia PDF Downloads 405651 Crack Width Analysis of Reinforced Concrete Members under Shrinkage Effect by Pseudo-Discrete Crack Model
Authors: F. J. Ma, A. K. H. Kwan
Abstract:
Crack caused by shrinkage movement of concrete is a serious problem especially when restraint is provided. It may cause severe serviceability and durability problems. The existing prediction methods for crack width of concrete due to shrinkage movement are mainly numerical methods under simplified circumstances, which do not agree with each other. To get a more unified prediction method applicable to more sophisticated circumstances, finite element crack width analysis for shrinkage effect should be developed. However, no existing finite element analysis can be carried out to predict the crack width of concrete due to shrinkage movement because of unsolved reasons of conventional finite element analysis. In this paper, crack width analysis implemented by finite element analysis is presented with pseudo-discrete crack model, which combines traditional smeared crack model and newly proposed crack queuing algorithm. The proposed pseudo-discrete crack model is capable of simulating separate and single crack without adopting discrete crack element. And the improved finite element analysis can successfully simulate the stress redistribution when concrete is cracked, which is crucial for predicting crack width, crack spacing and crack number.Keywords: crack queuing algorithm, crack width analysis, finite element analysis, shrinkage effect
Procedia PDF Downloads 419650 Analytical Solutions of Josephson Junctions Dynamics in a Resonant Cavity for Extended Dicke Model
Authors: S.I.Mukhin, S. Seidov, A. Mukherjee
Abstract:
The Dicke model is a key tool for the description of correlated states of quantum atomic systems, excited by resonant photon absorption and subsequently emitting spontaneous coherent radiation in the superradiant state. The Dicke Hamiltonian (DH) is successfully used for the description of the dynamics of the Josephson Junction (JJ) array in a resonant cavity under applied current. In this work, we have investigated a generalized model, which is described by DH with a frustrating interaction term. This frustrating interaction term is explicitly the infinite coordinated interaction between all the spin half in the system. In this work, we consider an array of N superconducting islands, each divided into two sub-islands by a Josephson Junction, taken in a charged qubit / Cooper Pair Box (CPB) condition. The array is placed inside the resonant cavity. One important aspect of the problem lies in the dynamical nature of the physical observables involved in the system, such as condensed electric field and dipole moment. It is important to understand how these quantities behave with time to define the quantum phase of the system. The Dicke model without frustrating term is solved to find the dynamical solutions of the physical observables in analytic form. We have used Heisenberg’s dynamical equations for the operators and on applying newly developed Rotating Holstein Primakoff (HP) transformation and DH we have arrived at the four coupled nonlinear dynamical differential equations for the momentum and spin component operators. It is possible to solve the system analytically using two-time scales. The analytical solutions are expressed in terms of Jacobi's elliptic functions for the metastable ‘bound luminosity’ dynamic state with the periodic coherent beating of the dipoles that connect the two double degenerate dipolar ordered phases discovered previously. In this work, we have proceeded the analysis with the extended DH with a frustrating interaction term. Inclusion of the frustrating term involves complexity in the system of differential equations and it gets difficult to solve analytically. We have solved semi-classical dynamic equations using the perturbation technique for small values of Josephson energy EJ. Because the Hamiltonian contains parity symmetry, thus phase transition can be found if this symmetry is broken. Introducing spontaneous symmetry breaking term in the DH, we have derived the solutions which show the occurrence of finite condensate, showing quantum phase transition. Our obtained result matches with the existing results in this scientific field.Keywords: Dicke Model, nonlinear dynamics, perturbation theory, superconductivity
Procedia PDF Downloads 135649 Transcriptomine: The Nuclear Receptor Signaling Transcriptome Database
Authors: Scott A. Ochsner, Christopher M. Watkins, Apollo McOwiti, David L. Steffen Lauren B. Becnel, Neil J. McKenna
Abstract:
Understanding signaling by nuclear receptors (NRs) requires an appreciation of their cognate ligand- and tissue-specific transcriptomes. While target gene regulation data are abundant in this field, they reside in hundreds of discrete publications in formats refractory to routine query and analysis and, accordingly, their full value to the NR signaling community has not been realized. One of the mandates of the Nuclear Receptor Signaling Atlas (NURSA) is to facilitate access of the community to existing public datasets. Pursuant to this mandate we are developing a freely-accessible community web resource, Transcriptomine, to bring together the sum total of available expression array and RNA-Seq data points generated by the field in a single location. Transcriptomine currently contains over 25,000,000 gene fold change datapoints from over 1200 contrasts relevant to over 100 NRs, ligands and coregulators in over 200 tissues and cell lines. Transcriptomine is designed to accommodate a spectrum of end users ranging from the bench researcher to those with advanced bioinformatic training. Visualization tools allow users to build custom charts to compare and contrast patterns of gene regulation across different tissues and in response to different ligands. Our resource affords an entirely new paradigm for leveraging gene expression data in the NR signaling field, empowering users to query gene fold changes across diverse regulatory molecules, tissues and cell lines, target genes, biological functions and disease associations, and that would otherwise be prohibitive in terms of time and effort. Transcriptomine will be regularly updated with gene lists from future genome-wide expression array and expression-sequencing datasets in the NR signaling field.Keywords: target gene database, informatics, gene expression, transcriptomics
Procedia PDF Downloads 275648 Comparisons of Co-Seismic Gravity Changes between GRACE Observations and the Predictions from the Finite-Fault Models for the 2012 Mw = 8.6 Indian Ocean Earthquake Off-Sumatra
Authors: Armin Rahimi
Abstract:
The Gravity Recovery and Climate Experiment (GRACE) has been a very successful project in determining math redistribution within the Earth system. Large deformations caused by earthquakes are in the high frequency band. Unfortunately, GRACE is only capable to provide reliable estimate at the low-to-medium frequency band for the gravitational changes. In this study, we computed the gravity changes after the 2012 Mw8.6 Indian Ocean earthquake off-Sumatra using the GRACE Level-2 monthly spherical harmonic (SH) solutions released by the University of Texas Center for Space Research (UTCSR). Moreover, we calculated gravity changes using different fault models derived from teleseismic data. The model predictions showed non-negligible discrepancies in gravity changes. However, after removing high-frequency signals, using Gaussian filtering 350 km commensurable GRACE spatial resolution, the discrepancies vanished, and the spatial patterns of total gravity changes predicted from all slip models became similar at the spatial resolution attainable by GRACE observations, and predicted-gravity changes were consistent with the GRACE-detected gravity changes. Nevertheless, the fault models, in which give different slip amplitudes, proportionally lead to different amplitude in the predicted gravity changes.Keywords: undersea earthquake, GRACE observation, gravity change, dislocation model, slip distribution
Procedia PDF Downloads 356647 Effects of Soil Erosion on Vegetation Development
Authors: Josephine Wanja Nyatia
Abstract:
The relationship between vegetation and soil erosion deserves attention due to its scientific importance and practical applications. A great deal of information is available about the mechanisms and benefits of vegetation in the control of soil erosion, but the effects of soil erosion on vegetation development and succession is poorly documented. Research shows that soil erosion is the most important driving force for the degradation of upland and mountain ecosystems. Soil erosion interferes with the process of plant community development and vegetation succession, commencing with seed formation and impacting throughout the whole growth phase and affecting seed availability, dispersal, germination and establishment, plant community structure and spatial distribution. There have been almost no studies on the effects of soil erosion on seed development and availability, of surface flows on seed movement and redistribution, and their influences on soil seed bank and on vegetation establishment and distribution. However, these effects may be the main cause of low vegetation cover in regions of high soil erosion activity, and these issues need to be investigated. Moreover, soil erosion is not only a negative influence on vegetation succession and restoration but also a driving force of plant adaptation and evolution. Consequently, we need to study the effects of soil erosion on ecological processes and on development and regulation of vegetation succession from the points of view of pedology and vegetation, plant and seed ecology, and to establish an integrated theory and technology for deriving practical solutions to soil erosion problemsKeywords: soil erosion, vegetation, development, seed availability
Procedia PDF Downloads 86646 Hardware Implementation on Field Programmable Gate Array of Two-Stage Algorithm for Rough Set Reduct Generation
Authors: Tomasz Grzes, Maciej Kopczynski, Jaroslaw Stepaniuk
Abstract:
The rough sets theory developed by Prof. Z. Pawlak is one of the tools that can be used in the intelligent systems for data analysis and processing. Banking, medicine, image recognition and security are among the possible fields of utilization. In all these fields, the amount of the collected data is increasing quickly, but with the increase of the data, the computation speed becomes the critical factor. Data reduction is one of the solutions to this problem. Removing the redundancy in the rough sets can be achieved with the reduct. A lot of algorithms of generating the reduct were developed, but most of them are only software implementations, therefore have many limitations. Microprocessor uses the fixed word length, consumes a lot of time for either fetching as well as processing of the instruction and data; consequently, the software based implementations are relatively slow. Hardware systems don’t have these limitations and can process the data faster than a software. Reduct is the subset of the decision attributes that provides the discernibility of the objects. For the given decision table there can be more than one reduct. Core is the set of all indispensable condition attributes. None of its elements can be removed without affecting the classification power of all condition attributes. Moreover, every reduct consists of all the attributes from the core. In this paper, the hardware implementation of the two-stage greedy algorithm to find the one reduct is presented. The decision table is used as an input. Output of the algorithm is the superreduct which is the reduct with some additional removable attributes. First stage of the algorithm is calculating the core using the discernibility matrix. Second stage is generating the superreduct by enriching the core with the most common attributes, i.e., attributes that are more frequent in the decision table. Described above algorithm has two disadvantages: i) generating the superreduct instead of reduct, ii) additional first stage may be unnecessary if the core is empty. But for the systems focused on the fast computation of the reduct the first disadvantage is not the key problem. The core calculation can be achieved with a combinational logic block, and thus add respectively little time to the whole process. Algorithm presented in this paper was implemented in Field Programmable Gate Array (FPGA) as a digital device consisting of blocks that process the data in a single step. Calculating the core is done by the comparators connected to the block called 'singleton detector', which detects if the input word contains only single 'one'. Calculating the number of occurrences of the attribute is performed in the combinational block made up of the cascade of the adders. The superreduct generation process is iterative and thus needs the sequential circuit for controlling the calculations. For the research purpose, the algorithm was also implemented in C language and run on a PC. The times of execution of the reduct calculation in a hardware and software were considered. Results show increase in the speed of data processing.Keywords: data reduction, digital systems design, field programmable gate array (FPGA), reduct, rough set
Procedia PDF Downloads 220645 Apoptosis Pathway Targeted by Thymoquinone in MCF7 Breast Cancer Cell Line
Authors: M. Marjaneh, M. Y. Narazah, H. Shahrul
Abstract:
Array-based gene expression analysis is a powerful tool to profile expression of genes and to generate information on therapeutic effects of new anti-cancer compounds. Anti-apoptotic effect of thymoquinone was studied in MCF7 breast cancer cell line using gene expression profiling with cDNA micro array. The purity and yield of RNA samples were determined using RNeasyPlus Mini kit. The Agilent RNA 6000 Nano LabChip kit evaluated the quantity of the RNA samples. AffinityScript RT oligo-dT promoter primer was used to generate cDNA strands. T7 RNA polymerase was used to convert cDNA to cRNA. The cRNA samples and human universal reference RNA were labelled with Cy-3-CTP and Cy-5-CTP, respectively. Feature Extraction and GeneSpring software analysed the data. The single experiment analysis revealed involvement of 64 pathways with up-regulated genes and 78 pathways with down-regulated genes. The MAPK and p38-MAPK pathways were inhibited due to the up-regulation of PTPRR gene. The inhibition of p38-MAPK suggested up-regulation of TGF-ß pathway. Inhibition of p38 - MAPK caused up-regulation of TP53 and down-regulation of Bcl2 genes indicating involvement of intrinsic apoptotic pathway. Down-regulation of CARD16 gene as an adaptor molecule regulated CASP1 and suggested necrosis-like programmed cell death and involvement of caspase in apoptosis. Furthermore, down-regulation of GPCR, EGF-EGFR signalling pathways suggested reduction of ER. Involvement of AhR pathway which control cytochrome P450 and glucuronidation pathways showed metabolism of Thymoquinone. The findings showed differential expression of several genes in apoptosis pathways with thymoquinone treatment in estrogen receptor-positive breast cancer cells.Keywords: cDNA microarray, thymoquinone, CARD16, PTPRR, CASP10
Procedia PDF Downloads 348644 Examples of RC Design with Eurocode2
Authors: Carla Ferreira, Helena Barros
Abstract:
The paper termed “Design of reinforced concrete with Eurocode 2” presents the theory regarding the design of reinforced concrete sections and the development of the tables and abacuses to verify the concrete section to the ultimate limit and service limit states. This paper is a complement of it, showing how to use the previous tools. Different numerical results are shown, proving the capability of the methodology. When a section of a beam is already chosen, the computer program presents the reinforcing steel in many locations along the structure, and it is the engineer´s task to choose the layout available for the construction, considering the maximum regular kind of reinforcing bars. There are many computer programs available for this task, but the interest of the present kind of tools is the fast and easy way of making the design and choose the optimal solution. Another application of these design tools is in the definition of the section dimensions, in a way that when stresses are evaluated, the final design is acceptable. In the design offices, these are considered by the engineers a very quick and useful way of designing reinforced concrete sections, employing variable strength concrete and higher steel classes. Examples of nonlinear analyses and redistribution of the bending moment will be considered, according to the Eurocode 2 recommendations, for sections under bending moment and axial forces. Examples of the evaluation of the service limit state will be presented.Keywords: design examples, eurocode 2, reinforced concrete, section design
Procedia PDF Downloads 73643 Numerical Analysis of Shear Crack Propagation in a Concrete Beam without Transverse Reinforcement
Authors: G. A. Rombach, A. Faron
Abstract:
Crack formation and growth in reinforced concrete members are, in many cases, the cause of the collapse of technical structures. Such serious failures impair structural behavior and can also damage property and persons. An intensive investigation of the crack propagation is indispensable. Numerical methods are being developed to analyze crack growth in an element and to detect fracture failure at an early stage. For reinforced concrete components, however, further research and action are required in the analysis of shear cracks. This paper presents numerical simulations and continuum mechanical modeling of bending shear crack propagation in a three-dimensional reinforced concrete beam without transverse reinforcement. The analysis will provide a further understanding of crack growth and redistribution of inner forces in concrete members. As a numerical method to map discrete cracks, the extended finite element method (XFEM) is applied. The crack propagation is compared with the smeared crack approach using concrete damage plasticity. For validation, the crack patterns of real experiments are compared with the results of the different finite element models. The evaluation is based on single span beams under bending. With the analysis, it is possible to predict the fracture behavior of concrete members.Keywords: concrete damage plasticity, crack propagation, extended finite element method, fracture mechanics
Procedia PDF Downloads 119642 Investigation of a Natural Convection Heat Sink for LEDs Based on Micro Heat Pipe Array-Rectangular Channel
Authors: Wei Wang, Yaohua Zhao, Yanhua Diao
Abstract:
The exponential growth of the lighting industry has rendered traditional thermal technologies inadequate for addressing the thermal management challenges inherent to high-power light-emitting diode (LED) technology. To enhance the thermal management of LEDs, this study proposes a heat sink configuration that integrates a miniature heat pipe array based on phase change technology with rectangular channels. The thermal performance of the heat sink was evaluated through experimental testing, and the results demonstrated that when the input power was 100W, 150W, and 200W, the temperatures of the LED substrate were 47.64℃, 56.78℃, and 69.06℃, respectively. Additionally, the maximum temperature difference of the MHPA in the vertical direction was observed to be 0.32℃, 0.30℃, and 0.30℃, respectively. The results demonstrate that the heat sink not only effectively dissipates the heat generated by the LEDs, but also exhibits excellent temperature uniformity. In consideration of the experimental measurement outcomes, a corresponding numerical model was developed as part of this study. Following the model validation, the effect of the structural parameters of the heat sink on its heat dissipation efficacy was examined through the use of response surface methodology (RSM) analysis. The rectangular channel width, channel height, channel length, number of channel cross-sections, and channel cross-section spacing were selected as the input parameters, while the LED substrate temperature and the total mass of the heat sink were regarded as the response variables. Subsequently, the response was subjected to an analysis of variance (ANOVA), which yielded a regression model that predicted the response based on the input variables. This offers some direction for the design of the radiator.Keywords: light-emitting diodes, heat transfer, heat pipe, natural convection, response surface methodology
Procedia PDF Downloads 37641 Parametric Study on the Behavior of Reinforced Concrete Continuous Beams Flexurally Strengthened with FRP Plates
Authors: Mohammed A. Sakr, Tarek M. Khalifa, Walid N. Mansour
Abstract:
External bonding of fiber reinforced polymer (FRP) plates to reinforced concrete (RC) beams is an effective technique for flexural strengthening. This paper presents an analytical parametric study on the behavior of RC continuous beams flexurally strengthened with externally bonded FRP plates on the upper and lower fibers, conducted using simple uniaxial nonlinear finite element model (UNFEM). UNFEM is able to estimate the load-carrying capacity, different failure modes and the interfacial stresses of RC continuous beams flexurally strengthened with externally bonded FRP plates on the upper and lower fibers. The study investigated the effect of five key parameters on the behavior and moment redistribution of FRP-reinforced continuous beams. The investigated parameters were the length of the FRP plate, the width and the thickness of the FRP plate, the ratio between the area of the FRP plate to the concrete area, the cohesive shear strength of the adhesive layer, and the concrete compressive strength. The investigation resulted in a number of important conclusions reflecting the effects of the studied parameters on the behavior of RC continuous beams flexurally strengthened with externally bonded FRP plates.Keywords: continuous beams, parametric study, finite element, fiber reinforced polymer
Procedia PDF Downloads 371640 Sensor and Sensor System Design, Selection and Data Fusion Using Non-Deterministic Multi-Attribute Tradespace Exploration
Authors: Matthew Yeager, Christopher Willy, John Bischoff
Abstract:
The conceptualization and design phases of a system lifecycle consume a significant amount of the lifecycle budget in the form of direct tasking and capital, as well as the implicit costs associated with unforeseeable design errors that are only realized during downstream phases. Ad hoc or iterative approaches to generating system requirements oftentimes fail to consider the full array of feasible systems or product designs for a variety of reasons, including, but not limited to: initial conceptualization that oftentimes incorporates a priori or legacy features; the inability to capture, communicate and accommodate stakeholder preferences; inadequate technical designs and/or feasibility studies; and locally-, but not globally-, optimized subsystems and components. These design pitfalls can beget unanticipated developmental or system alterations with added costs, risks and support activities, heightening the risk for suboptimal system performance, premature obsolescence or forgone development. Supported by rapid advances in learning algorithms and hardware technology, sensors and sensor systems have become commonplace in both commercial and industrial products. The evolving array of hardware components (i.e. sensors, CPUs, modular / auxiliary access, etc…) as well as recognition, data fusion and communication protocols have all become increasingly complex and critical for design engineers during both concpetualization and implementation. This work seeks to develop and utilize a non-deterministic approach for sensor system design within the multi-attribute tradespace exploration (MATE) paradigm, a technique that incorporates decision theory into model-based techniques in order to explore complex design environments and discover better system designs. Developed to address the inherent design constraints in complex aerospace systems, MATE techniques enable project engineers to examine all viable system designs, assess attribute utility and system performance, and better align with stakeholder requirements. Whereas such previous work has been focused on aerospace systems and conducted in a deterministic fashion, this study addresses a wider array of system design elements by incorporating both traditional tradespace elements (e.g. hardware components) as well as popular multi-sensor data fusion models and techniques. Furthermore, statistical performance features to this model-based MATE approach will enable non-deterministic techniques for various commercial systems that range in application, complexity and system behavior, demonstrating a significant utility within the realm of formal systems decision-making.Keywords: multi-attribute tradespace exploration, data fusion, sensors, systems engineering, system design
Procedia PDF Downloads 189639 Thermal Hydraulic Analysis of Sub-Channels of Pressurized Water Reactors with Hexagonal Array: A Numerical Approach
Authors: Md. Asif Ullah, M. A. R. Sarkar
Abstract:
This paper illustrates 2-D and 3-D simulations of sub-channels of a Pressurized Water Reactor (PWR) having hexagonal array of fuel rods. At a steady state, the temperature of outer surface of the cladding of fuel rod is kept about 1200°C. The temperature of this isothermal surface is taken as boundary condition for simulation. Water with temperature of 290°C is given as a coolant inlet to the primary water circuit which is pressurized upto 157 bar. Turbulent flow of pressurized water is used for heat removal. In 2-D model, temperature, velocity, pressure and Nusselt number distributions are simulated in a vertical sectional plane through the sub-channels of a hexagonal fuel rod assembly. Temperature, Nusselt number and Y-component of convective heat flux along a line in this plane near the end of fuel rods are plotted for different Reynold’s number. A comparison between X-component and Y-component of convective heat flux in this vertical plane is analyzed. Hexagonal fuel rod assembly has three types of sub-channels according to geometrical shape whose boundary conditions are different too. In 3-D model, temperature, velocity, pressure, Nusselt number, total heat flux magnitude distributions for all the three sub-channels are studied for a suitable Reynold’s number. A horizontal sectional plane is taken from each of the three sub-channels to study temperature, velocity, pressure, Nusselt number and convective heat flux distribution in it. Greater values of temperature, Nusselt number and Y-component of convective heat flux are found for greater Reynold’s number. X-component of convective heat flux is found to be non-zero near the bottom of fuel rod and zero near the end of fuel rod. This indicates that the convective heat transfer occurs totally along the direction of flow near the outlet. As, length to radius ratio of sub-channels is very high, simulation for a short length of the sub-channels are done for graphical interface advantage. For the simulations, Turbulent Flow (K-Є ) module and Heat Transfer in Fluids (ht) module of COMSOL MULTIPHYSICS 5.0 are used.Keywords: sub-channels, Reynold’s number, Nusselt number, convective heat transfer
Procedia PDF Downloads 361638 Potential Antibacterial Applications and Synthesis, Structural, Magnetic, Optical, and Dielectric Characterization of Nickel-Substituted Cobalt Ferrite Nanoparticles
Authors: Tesfay Gebremichael Reda
Abstract:
Nanoparticle technology is fast progressing and is being employed in innumerable medical applications. At this time, the public's health is seriously threatened by the rise of bacterial strains resistant to several medications. Metal nanoparticles are a potential alternate approach for tackling this global concern, and this is the main focus of this study. The citrate precursor sol-gel synthesis method was used to synthesize the, Niₓ Co(₁-ₓ) Fe₂ O₄, (where x = 0.0:0.2:1.0) nanoparticle. XRD identified the development of the cubic crystal structure to have a preferential orientation along (311), and the average particle size was found to be 29-38 nm. The average crystallizes assessed with ImageJ software and origin 22 of the SEM are nearly identical to the XRD results. In the created NCF NPs, the FT-IR spectroscopy reveals structural examinations and the redistribution of cations between octahedral (505-428 cm-1) and tetrahedral (653-603 cm-1) locales. Finally, the decrease of coercive fields HC, 2384 Oe to 241.93 Oe replacement of Co²+ cation with Ni²+. Band gap energy rises as Ni concentration increases, which may be attributed to the fact that the ionic radii of Ni²+ ions are smaller than that of Co²+ ions, which results in a strong electrostatic interaction. On the contrary, except at x = 0.4, the dielectric constant decreases as the nickel concentration increases. According to the findings of this research work, nanoparticles composed of Ni₀.₄ Co₀.₄ Fe₂ O₄ have demonstrated a promising value against S. aureus and E. coli, and it suggests a proposed model for their potential use as a new source of antibacterial agents.Keywords: antimicrobial, band gap, citrate precursor, dielectric, nanoparticle
Procedia PDF Downloads 29637 Liquefaction Potential Assessment Using Screw Driving Testing and Microtremor Data: A Case Study in the Philippines
Authors: Arturo Daag
Abstract:
The Philippine Institute of Volcanology and Seismology (PHIVOLCS) is enhancing its liquefaction hazard map towards a detailed probabilistic approach using SDS and geophysical data. Target sites for liquefaction assessment are public schools in Metro Manila. Since target sites are in highly urbanized-setting, the objective of the project is to conduct both non-destructive geotechnical studies using Screw Driving Testing (SDFS) combined with geophysical data such as refraction microtremor array (ReMi), 3 component microtremor Horizontal to Vertical Spectral Ratio (HVSR), and ground penetrating RADAR (GPR). Initial test data was conducted in liquefaction impacted areas from the Mw 6.1 earthquake in Central Luzon last April 22, 2019 Province of Pampanga. Numerous accounts of liquefaction events were documented areas underlain by quaternary alluvium and mostly covered by recent lahar deposits. SDS estimated values showed a good correlation to actual SPT values obtained from available borehole data. Thus, confirming that SDS can be an alternative tool for liquefaction assessment and more efficient in terms of cost and time compared to SPT and CPT. Conducting borehole may limit its access in highly urbanized areas. In order to extend or extrapolate the SPT borehole data, non-destructive geophysical equipment was used. A 3-component microtremor obtains a subsurface velocity model in 1-D seismic shear wave velocity of the upper 30 meters of the profile (Vs30). For the ReMi, 12 geophone array with 6 to 8-meter spacing surveys were conducted. Microtremor data were computed through the Factor of Safety, which is the quotient of Cyclic Resistance Ratio (CRR) and Cyclic Stress Ratio (CSR). Complementary GPR was used to study the subsurface structure and used to inferred subsurface structures and groundwater conditions.Keywords: screw drive testing, microtremor, ground penetrating RADAR, liquefaction
Procedia PDF Downloads 203636 Development of a Social Assistive Robot for Elderly Care
Authors: Edwin Foo, Woei Wen, Lui, Meijun Zhao, Shigeru Kuchii, Chin Sai Wong, Chung Sern Goh, Yi Hao He
Abstract:
This presentation presents an elderly care and assistive social robot development work. We named this robot JOS and he is restricted to table top operation. JOS is designed to have a maximum volume of 3600 cm3 with its base restricted to 250 mm and his mission is to provide companion, assist and help the elderly. In order for JOS to accomplish his mission, he will be equipped with perception, reaction and cognition capability. His appearance will be not human like but more towards cute and approachable type. JOS will also be designed to be neutral gender. However, the robot will still have eyes, eyelid and a mouth. For his eyes and eyelids, they will be built entirely with Robotis Dynamixel AX18 motor. To realize this complex task, JOS will be also be equipped with micro-phone array, vision camera and Intel i5 NUC computer and a powered by a 12 V lithium battery that will be self-charging. His face is constructed using 1 motor each for the eyelid, 2 motors for the eyeballs, 3 motors for the neck mechanism and 1 motor for the lips movement. The vision senor will be house on JOS forehead and the microphone array will be somewhere below the mouth. For the vision system, Omron latest OKAO vision sensor is used. It is a compact and versatile sensor that is only 60mm by 40mm in size and operates with only 5V supply. In addition, OKAO vision sensor is capable of identifying the user and recognizing the expression of the user. With these functions, JOS is able to track and identify the user. If he cannot recognize the user, JOS will ask the user if he would want him to remember the user. If yes, JOS will store the user information together with the capture face image into a database. This will allow JOS to recognize the user the next time the user is with JOS. In addition, JOS is also able to interpret the mood of the user through the facial expression of the user. This will allow the robot to understand the user mood and behavior and react according. Machine learning will be later incorporated to learn the behavior of the user so as to understand the mood of the user and requirement better. For the speech system, Microsoft speech and grammar engine is used for the speech recognition. In order to use the speech engine, we need to build up a speech grammar database that captures the commonly used words by the elderly. This database is built from research journals and literature on elderly speech and also interviewing elderly what do they want to robot to assist them with. Using the result from the interview and research from journal, we are able to derive a set of common words the elderly frequently used to request for the help. It is from this set that we build up our grammar database. In situation where there is more than one person near JOS, he is able to identify the person who is talking to him through an in-house developed microphone array structure. In order to make the robot more interacting, we have also included the capability for the robot to express his emotion to the user through the facial expressions by changing the position and movement of the eyelids and mouth. All robot emotions will be in response to the user mood and request. Lastly, we are expecting to complete this phase of project and test it with elderly and also delirium patient by Feb 2015.Keywords: social robot, vision, elderly care, machine learning
Procedia PDF Downloads 441635 Implementation of Statistical Parameters to Form an Entropic Mathematical Models
Authors: Gurcharan Singh Buttar
Abstract:
It has been discovered that although these two areas, statistics, and information theory, are independent in their nature, they can be combined to create applications in multidisciplinary mathematics. This is due to the fact that where in the field of statistics, statistical parameters (measures) play an essential role in reference to the population (distribution) under investigation. Information measure is crucial in the study of ambiguity, assortment, and unpredictability present in an array of phenomena. The following communication is a link between the two, and it has been demonstrated that the well-known conventional statistical measures can be used as a measure of information.Keywords: probability distribution, entropy, concavity, symmetry, variance, central tendency
Procedia PDF Downloads 157634 Distributional and Dynamic impact of Energy Subsidy Reform
Authors: Ali Hojati Najafabadi, Mohamad Hosein Rahmati, Seyed Ali Madanizadeh
Abstract:
Governments execute energy subsidy reforms by either increasing energy prices or reducing energy price dispersion. These policies make less use of energy per plant (intensive margin), vary the total number of firms (extensive margin), promote technological progress (technology channel), and make additional resources to redistribute (resource channel). We estimate a structural dynamic firm model with endogenous technology adaptation using data from the manufacturing firms in Iran and a country ranked the second-largest energy subsidy plan by the IMF. The findings show significant dynamics and distributional effects due to an energy reform plan. The price elasticity of energy consumption in the industrial sector is about -2.34, while it is -3.98 for large firms. The dispersion elasticity, defined as the amounts of changes in energy consumption by a one-percent reduction in the standard error of energy price distribution, is about 1.43, suggesting significant room for a distributional policy. We show that the intensive margin is the main driver of energy price elasticity, whereas the other channels mostly offset it. In contrast, the labor response is mainly through the extensive margin. Total factor productivity slightly improves in light of the reduction in energy consumption if, at the same time, the redistribution policy boosts the aggregate demands.Keywords: energy reform, firm dynamics, structural estimation, subsidy policy
Procedia PDF Downloads 96633 Reading and Writing Memories in Artificial and Human Reasoning
Authors: Ian O'Loughlin
Abstract:
Memory networks aim to integrate some of the recent successes in machine learning with a dynamic memory base that can be updated and deployed in artificial reasoning tasks. These models involve training networks to identify, update, and operate over stored elements in a large memory array in order, for example, to ably perform question and answer tasks parsing real-world and simulated discourses. This family of approaches still faces numerous challenges: the performance of these network models in simulated domains remains considerably better than in open, real-world domains, wide-context cues remain elusive in parsing words and sentences, and even moderately complex sentence structures remain problematic. This innovation, employing an array of stored and updatable ‘memory’ elements over which the system operates as it parses text input and develops responses to questions, is a compelling one for at least two reasons: first, it addresses one of the difficulties that standard machine learning techniques face, by providing a way to store a large bank of facts, offering a way forward for the kinds of long-term reasoning that, for example, recurrent neural networks trained on a corpus have difficulty performing. Second, the addition of a stored long-term memory component in artificial reasoning seems psychologically plausible; human reasoning appears replete with invocations of long-term memory, and the stored but dynamic elements in the arrays of memory networks are deeply reminiscent of the way that human memory is readily and often characterized. However, this apparent psychological plausibility is belied by a recent turn in the study of human memory in cognitive science. In recent years, the very notion that there is a stored element which enables remembering, however dynamic or reconstructive it may be, has come under deep suspicion. In the wake of constructive memory studies, amnesia and impairment studies, and studies of implicit memory—as well as following considerations from the cognitive neuroscience of memory and conceptual analyses from the philosophy of mind and cognitive science—researchers are now rejecting storage and retrieval, even in principle, and instead seeking and developing models of human memory wherein plasticity and dynamics are the rule rather than the exception. In these models, storage is entirely avoided by modeling memory using a recurrent neural network designed to fit a preconceived energy function that attains zero values only for desired memory patterns, so that these patterns are the sole stable equilibrium points in the attractor network. So although the array of long-term memory elements in memory networks seem psychologically appropriate for reasoning systems, they may actually be incurring difficulties that are theoretically analogous to those that older, storage-based models of human memory have demonstrated. The kind of emergent stability found in the attractor network models more closely fits our best understanding of human long-term memory than do the memory network arrays, despite appearances to the contrary.Keywords: artificial reasoning, human memory, machine learning, neural networks
Procedia PDF Downloads 272