Search results for: computational domain
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3665

Search results for: computational domain

2015 Understanding the Scope of Architects in Disaster Risk Reduction: The Case of Bhuj

Authors: Sweta Kandari

Abstract:

Predominantly, the conventional role of an architect is to design and construct. However, in a post-disaster scenario, the prevalent role expands and includes many other responsibilities. Agencies collaborating in post-disaster reconstruction face the challenge of building back quickly while requiring them to listen, reflect, develop and deliver as per the needs and requirements of the people. The question of the role of an architect has been extensively discussed in the reconstruction field. Discourses about the role of an architect in post-disaster scenario revolve around the ignorance by the profession, their professional abilities and inabilities. Within this domain, this paper aims at analyzing and recognizing the roles, responsibilities, scope, limitations, skillsets applied and required by an architect while working in a post-disaster situation. Four projects rebuilt after the 2001 Bhuj earthquake in Gujarat, India were examined for this research. Based on the analysis of the case study, areas of intervention of an architect in the various stages of rebuilding were identified. It was reinforced that within the areas of intervention identified, there is a vast gap between the prescribed, the prevalent notion and the performed responsibilities of an architect. This paper brings forth the specific gaps in the rebuilding process while exploring and understanding the relationship between various stakeholders that influence the role of an architect.

Keywords: rebuilding, role of an architect, Bhuj, post-disaster

Procedia PDF Downloads 132
2014 Design and Implementation of Medium Access Control Based Routing on Real Wireless Sensor Networks Testbed

Authors: Smriti Agarwal, Ashish Payal, B. V. R. Reddy

Abstract:

IEEE 802.15.4 is a Low Rate Wireless Personal Area Networks (LR-WPAN) standard combined with ZigBee, which is going to enable new applications in Wireless Sensor Networks (WSNs) and Internet of Things (IoT) domain. In recent years, it has become a popular standard for WSNs. Wireless communication among sensor motes, enabled by IEEE 802.15.4 standard, is extensively replacing the existing wired technology in a wide range of monitoring and control applications. Researchers have proposed a routing framework and mechanism that interacts with the IEEE 802.15.4 standard using software platform. In this paper, we have designed and implemented MAC based routing (MBR) based on IEEE 802.15.4 standard using a hardware platform “SENSEnuts”. The experimental results include data through light and temperature sensors obtained from communication between PAN coordinator and source node through coordinator, MAC address of some modules used in the experimental setup, topology of the network created for simulation and the remaining battery power of the source node. Our experimental effort on a WSN Testbed has helped us in bridging the gap between theoretical and practical aspect of implementing IEEE 802.15.4 for WSNs applications.

Keywords: IEEE 802.15.4, routing, WSN, ZigBee

Procedia PDF Downloads 406
2013 Blockchain for IoT Security and Privacy in Healthcare Sector

Authors: Umair Shafique, Hafiz Usman Zia, Fiaz Majeed, Samina Naz, Javeria Ahmed, Maleeha Zainab

Abstract:

The Internet of Things (IoT) has become a hot topic for the last couple of years. This innovative technology has shown promising progress in various areas, and the world has witnessed exponential growth in multiple application domains. Researchers are working to investigate its aptitudes to get the best from it by harnessing its true potential. But at the same time, IoT networks open up a new aspect of vulnerability and physical threats to data integrity, privacy, and confidentiality. It's is due to centralized control, data silos approach for handling information, and a lack of standardization in the IoT networks. As we know, blockchain is a new technology that involves creating secure distributed ledgers to store and communicate data. Some of the benefits include resiliency, integrity, anonymity, decentralization, and autonomous control. The potential for blockchain technology to provide the key to managing and controlling IoT has created a new wave of excitement around the idea of putting that data back into the hands of the end-users. In this manuscript, we have proposed a model that combines blockchain and IoT networks to address potential security and privacy issues in the healthcare domain. Then we try to describe various application areas, challenges, and future directions in the healthcare sector where blockchain platforms merge with IoT networks.

Keywords: IoT, blockchain, cryptocurrency, healthcare, consensus, data

Procedia PDF Downloads 180
2012 Limitation of Parallel Flow in Three-Dimensional Elongated Porous Domain Subjected to Cross Heat and Mass Flux

Authors: Najwa Mimouni, Omar Rahli, Rachid Bennacer, Salah Chikh

Abstract:

In the present work 2D and 3D numerical simulations of double diffusion natural convection in an elongated enclosure filled with a binary fluid saturating a porous medium are carried out. In the formulation of the problem, the Boussinesq approximation is considered and cross Neumann boundary conditions are specified for heat and mass walls conditions. The numerical method is based on the control volume approach with the third order QUICK scheme. Full approximation storage (FAS) with full multigrid (FMG) method is used to solve the problem. For the explored large range of the controlling parameters, we clearly evidenced that the increase in the depth of the cavity i.e. the lateral aspect ratio has an important effect on the flow patterns. The 2D perfect parallel flows obtained for a small lateral aspect ratio are drastically destabilized by increasing the cavity lateral dimension. This yields a 3D fluid motion with a much more complicated flow pattern and the classically studied 2D parallel flows are impossible.

Keywords: bifurcation, natural convection, heat and mass transfer, parallel flow, porous media

Procedia PDF Downloads 474
2011 Conceptual Model Providing More Information on the Contact Situation between Crime Victim and the Police

Authors: M. Inzunza

Abstract:

In contemporary society, victims of crime has been given more recognition, which have contributed to advancing the knowledge on the effects of crime. There exists a complexity of who gets the status of victim and that the typology of good versus bad can interfere with the contact situation of the victim with the police. The aim of this study is to identify the most central areas affecting the contact situation between crime victims and the police to develop a conceptual model to be useful empirically. By considering previously documented problem areas and different theoretical domains, a conceptual model has been developed. Preliminary findings suggest that an area that should be given attention is to get a better understanding of the victim, not only in terms of demographics but also in terms of risk behavior and social network. This area has been considered to influence the status of the crime victim. Another domain of value is the type of crime and the context of the incident in more detail. The police officer approach style in the contact situation is also a pertinent area that is influenced by how the police based victim services are organized and how individual police officers are suited for the mission. Suitability includes constructs from empathy models adapted to the police context and especially focusing on sub-constructs such as perspective taking. Discussion will focus on how these findings can be operationalized in practice and how they are used in ongoing empirical studies.

Keywords: empathy, perspective taking, police contact, victim of crime

Procedia PDF Downloads 138
2010 Integrating Assurance and Risk Management of Complex Systems

Authors: Odd Ivar Haugen

Abstract:

This paper explores the relationship between assurance, risk, and risk management in the context of complex safety-related systems. It introduces a nuanced understanding of assurance and argues that the foundation for grounds for justified confidence in claims made about a complex system is related to the system behaviour. It emphasises the importance of knowledge as the cornerstone of assurance. The paper addresses the challenges of epistemic and aleatory uncertainties inherent in safety-critical systems. A systems approach is proposed to model emergent properties and complexity using the composition, environment, structure, mechanisms (CESM) metamodel, offering a structured framework for analysing system behaviour. The interplay between assurance and risk management is conceptualised through two models: the domain model and the control model. Assurance and risk management are mutually dependent on each other to reduce uncertainty and control risk levels. This work highlights the dual roles of assurance in risk management, acting as an epistemic actuator on the one side and providing feedback about the strength of the justification on the other. Assurance and risk management have inseparable roles in ensuring safety in complex systems.

Keywords: assurance, CESM metamodel, confidence, emergent properties, knowledge, objectivity, risk, system behaviour, system safety

Procedia PDF Downloads 6
2009 A Study of Cloud Computing Solution for Transportation Big Data Processing

Authors: Ilgin Gökaşar, Saman Ghaffarian

Abstract:

The need for fast processed big data of transportation ridership (eg., smartcard data) and traffic operation (e.g., traffic detectors data) which requires a lot of computational power is incontrovertible in Intelligent Transportation Systems. Nowadays cloud computing is one of the important subjects and popular information technology solution for data processing. It enables users to process enormous measure of data without having their own particular computing power. Thus, it can also be a good selection for transportation big data processing as well. This paper intends to examine how the cloud computing can enhance transportation big data process with contrasting its advantages and disadvantages, and discussing cloud computing features.

Keywords: big data, cloud computing, Intelligent Transportation Systems, ITS, traffic data processing

Procedia PDF Downloads 469
2008 Pulse Method for Investigation of Zr-C Phase Diagram at High Carbon Content Domain under High Temperatures

Authors: Arseniy M. Kondratyev, Sergey V. Onufriev, Alexander I. Savvatimskiy

Abstract:

The microsecond electrical pulse heating technique which provides uniform energy input into an investigated specimen is considered. In the present study we investigated ZrC+C carbide specimens in a form of a thin layer (about 5 microns thick) that were produced using a method of magnetron sputtering on insulating substrates. Specimens contained (at. %): Zr–17.88; C–67.69; N–8.13; O–5.98. Current through the specimen, voltage drop across it and radiation at the wavelength of 856 nm were recorded in the experiments. It enabled us to calculate the input energy, specific heat (from 2300 to 4500 K) and resistivity (referred to the initial dimensions of a specimen). To obtain the true temperature a black body specimen was used. Temperature of the beginning and completion of a phase transition (solid–liquid) was measured.Temperature of the onset of melting was 3150 K at the input energy 2.65 kJ/g; temperature of the completion of melting was 3450 K at the input energy 5.2 kJ/g. The specific heat of the solid phase of investigated carbide calculated using our data on temperature and imparted energy, is close to 0.75 J/gК for temperature range 2100–2800 K. Our results are considered together with the equilibrium Zr-C phase diagram.

Keywords: pulse heating, zirconium carbide, high temperatures, melting

Procedia PDF Downloads 323
2007 Characterization of the Dispersion Phenomenon in an Optical Biosensor

Authors: An-Shik Yang, Chin-Ting Kuo, Yung-Chun Yang, Wen-Hsin Hsieh, Chiang-Ho Cheng

Abstract:

Optical biosensors have become a powerful detection and analysis tool for wide-ranging applications in biomedical research, pharmaceuticals and environmental monitoring. This study carried out the computational fluid dynamics (CFD)-based simulations to explore the dispersion phenomenon in the microchannel of a optical biosensor. The predicted time sequences of concentration contours were utilized to better understand the dispersion development occurred in different geometric shapes of microchannels. The simulation results showed the surface concentrations at the sensing probe (with the best performance of a grating coupler) in respect of time to appraise the dispersion effect and therefore identify the design configurations resulting in minimum dispersion.

Keywords: CFD simulations, dispersion, microfluidic, optical waveguide sensors

Procedia PDF Downloads 545
2006 Element-Independent Implementation for Method of Lagrange Multipliers

Authors: Gil-Eon Jeong, Sung-Kie Youn, K. C. Park

Abstract:

Treatment for the non-matching interface is an important computational issue. To handle this problem, the method of Lagrange multipliers including classical and localized versions are the most popular technique. It essentially imposes the interface compatibility conditions by introducing Lagrange multipliers. However, the numerical system becomes unstable and inefficient due to the Lagrange multipliers. The interface element-independent formulation that does not include the Lagrange multipliers can be obtained by modifying the independent variables mathematically. Through this modification, more efficient and stable system can be achieved while involving equivalent accuracy comparing with the conventional method. A numerical example is conducted to verify the validity of the presented method.

Keywords: element-independent formulation, interface coupling, methods of Lagrange multipliers, non-matching interface

Procedia PDF Downloads 403
2005 A Parallel Cellular Automaton Model of Tumor Growth for Multicore and GPU Programming

Authors: Manuel I. Capel, Antonio Tomeu, Alberto Salguero

Abstract:

Tumor growth from a transformed cancer-cell up to a clinically apparent mass spans through a range of spatial and temporal magnitudes. Through computer simulations, Cellular Automata (CA) can accurately describe the complexity of the development of tumors. Tumor development prognosis can now be made -without making patients undergo through annoying medical examinations or painful invasive procedures- if we develop appropriate CA-based software tools. In silico testing mainly refers to Computational Biology research studies of application to clinical actions in Medicine. To establish sound computer-based models of cellular behavior, certainly reduces costs and saves precious time with respect to carrying out experiments in vitro at labs or in vivo with living cells and organisms. These aim to produce scientifically relevant results compared to traditional in vitro testing, which is slow, expensive, and does not generally have acceptable reproducibility under the same conditions. For speeding up computer simulations of cellular models, specific literature shows recent proposals based on the CA approach that include advanced techniques, such the clever use of supporting efficient data structures when modeling with deterministic stochastic cellular automata. Multiparadigm and multiscale simulation of tumor dynamics is just beginning to be developed by the concerned research community. The use of stochastic cellular automata (SCA), whose parallel programming implementations are open to yield a high computational performance, are of much interest to be explored up to their computational limits. There have been some approaches based on optimizations to advance in multiparadigm models of tumor growth, which mainly pursuit to improve performance of these models through efficient memory accesses guarantee, or considering the dynamic evolution of the memory space (grids, trees,…) that holds crucial data in simulations. In our opinion, the different optimizations mentioned above are not decisive enough to achieve the high performance computing power that cell-behavior simulation programs actually need. The possibility of using multicore and GPU parallelism as a promising multiplatform and framework to develop new programming techniques to speed-up the computation time of simulations is just starting to be explored in the few last years. This paper presents a model that incorporates parallel processing, identifying the synchronization necessary for speeding up tumor growth simulations implemented in Java and C++ programming environments. The speed up improvement that specific parallel syntactic constructs, such as executors (thread pools) in Java, are studied. The new tumor growth parallel model is proved using implementations with Java and C++ languages on two different platforms: chipset Intel core i-X and a HPC cluster of processors at our university. The parallelization of Polesczuk and Enderling model (normally used by researchers in mathematical oncology) proposed here is analyzed with respect to performance gain. We intend to apply the model and overall parallelization technique presented here to solid tumors of specific affiliation such as prostate, breast, or colon. Our final objective is to set up a multiparadigm model capable of modelling angiogenesis, or the growth inhibition induced by chemotaxis, as well as the effect of therapies based on the presence of cytotoxic/cytostatic drugs.

Keywords: cellular automaton, tumor growth model, simulation, multicore and manycore programming, parallel programming, high performance computing, speed up

Procedia PDF Downloads 244
2004 Flow Characterization in Complex Terrain for Aviation Safety

Authors: Adil Rasheed, Mandar Tabib

Abstract:

The paper describes the ability of a high-resolution Computational Fluid Dynamics model to predict terrain-induced turbulence and wind shear close to the ground. Various sensitivity studies to choose the optimal simulation setup for modeling the flow characteristics in a complex terrain are presented. The capabilities of the model are demonstrated by applying it to the Sandnessjøen Airport, Stokka in Norway, an airport that is located in a mountainous area. The model is able to forecast turbulence in real time and trigger an alert when atmospheric conditions might result in high wind shear and turbulence.

Keywords: aviation safety, terrain-induced turbulence, atmospheric flow, alert system

Procedia PDF Downloads 416
2003 Alteration of Bone Strength in Osteoporosis of Mouse Femora: Computational Study Based on Micro CT Images

Authors: Changsoo Chon, Sangkuy Han, Donghyun Seo, Jihyung Park, Bokku Kang, Hansung Kim, Keyoungjin Chun, Cheolwoong Ko

Abstract:

The purpose of the study is to develop a finite element model based on 3D bone structural images of Micro-CT and to analyze the stress distribution for the osteoporosis mouse femora. In this study, results of finite element analysis show that the early osteoporosis of mouse model decreased a bone density in trabecular region; however, the bone density in cortical region increased.

Keywords: micro-CT, finite element analysis, osteoporosis, bone strength

Procedia PDF Downloads 363
2002 Hierarchical Checkpoint Protocol in Data Grids

Authors: Rahma Souli-Jbali, Minyar Sassi Hidri, Rahma Ben Ayed

Abstract:

Grid of computing nodes has emerged as a representative means of connecting distributed computers or resources scattered all over the world for the purpose of computing and distributed storage. Since fault tolerance becomes complex due to the availability of resources in decentralized grid environment, it can be used in connection with replication in data grids. The objective of our work is to present fault tolerance in data grids with data replication-driven model based on clustering. The performance of the protocol is evaluated with Omnet++ simulator. The computational results show the efficiency of our protocol in terms of recovery time and the number of process in rollbacks.

Keywords: data grids, fault tolerance, clustering, chandy-lamport

Procedia PDF Downloads 341
2001 Solving the Pseudo-Geometric Traveling Salesman Problem with the “Union Husk” Algorithm

Authors: Boris Melnikov, Ye Zhang, Dmitrii Chaikovskii

Abstract:

This study explores the pseudo-geometric version of the extensively researched Traveling Salesman Problem (TSP), proposing a novel generalization of existing algorithms which are traditionally confined to the geometric version. By adapting the "onion husk" method and introducing auxiliary algorithms, this research fills a notable gap in the existing literature. Through computational experiments using randomly generated data, several metrics were analyzed to validate the proposed approach's efficacy. Preliminary results align with expected outcomes, indicating a promising advancement in TSP solutions.

Keywords: optimization problems, traveling salesman problem, heuristic algorithms, “onion husk” algorithm, pseudo-geometric version

Procedia PDF Downloads 207
2000 Meditation-Based Interventions in the Workplace

Authors: Louise Fitzgerald, John Allman

Abstract:

Introduction: Having previously engaged in a meditation-based programme (MBP) for staff in general practice, we explore the evidence and extent to which MBPs are employed in the workplace. Aim of the study: We aim to understand the current workplace MBP intervention literature, which will help inform the suitability of these interventions within the workplace domain. Objectives: Uptake of MBPs in the workplace has grown as organizations look to support employee health, wellbeing, and performance. We will discuss the current MBP literature, including the large variability across MBPs and the associated difficulties in evaluating their efficacy. Learning points: 1) MBPs have a positive impact on cognitive function including concentration and memory and as such job performance. MBPs appear to have a positive impact on objective and subjective job satisfaction, productivity, motivation and work engagement. Meditation in the workplace may have positive impacts on mental health issues - including stress reduction and depression. 2) From our review MBPs appear to be implementable in a wide range of professions and work contexts - regardless of individual factors. Given many companies are focusing on health and wellbeing of employees, this could be included in employee wellbeing programmes. 3) Despite the benefits of mindfulness and meditation interventions in psychosocial workplace health and work performance the long-term efficacy has yet to be fully determined.

Keywords: meditation-based programmes, mindfulness, meditation, well-being

Procedia PDF Downloads 140
1999 A Research Using Remote Monitoring Technology for Pump Output Monitoring in Distributed Fuel Stations in Nigeria

Authors: Ofoegbu Ositadinma Edward

Abstract:

This research paper discusses a web based monitoring system that enables effective monitoring of fuel pump output and sales volume from distributed fuel stations under the domain of a single company/organization. The traditional method of operation by these organizations in Nigeria is non-automated and accounting for dispensed product is usually approximated and manual as there is little or no technology implemented to presently provide information relating to the state of affairs in the station both to on-ground staff and to supervisory staff that are not physically present in the station. This results in unaccountable losses in product and revenue as well as slow decision making. Remote monitoring technology as a vast research field with numerous application areas incorporating various data collation techniques and sensor networks can be applied to provide information relating to fuel pump status in distributed fuel stations reliably. Thus, the proposed system relies upon a microcontroller, keypad and pump to demonstrate the traditional fuel dispenser. A web-enabled PC with an accompanying graphic user interface (GUI) was designed using virtual basic which is connected to the microcontroller via the serial port which is to provide the web implementation.

Keywords: fuel pump, microcontroller, GUI, web

Procedia PDF Downloads 434
1998 Measuring Audit Quality Using Text Analysis: An Empirical Study of Indian Companies

Authors: Leesa Mohanty, Ashok Banerjee

Abstract:

Better audit quality signifies the financial statements of the auditee firm reflect true and fair view of their actual state of affairs, which reduces information asymmetry between management and shareholders, as a result, helps protect interests of shareholders. This study examines the impact of joint audit on audit quality. It is motivated by the ongoing debate where The Institute of Chartered Accountants of India (ICAI), the regulatory body governing auditors, has advocated the finance ministry and the Reserve Bank of India (RBI) for the mandatory use of joint audit in private banks to enhance the quality of audit. Earlier, the Government of India had rejected the plea by ICAI for mandatory joint audits in large companies stating it is not a viable option for promoting domestic firms. We introduce a new measure of audit quality. Drawing from the domain of text analytics, we use relevant phrases in audit reports to gauge audit quality and demonstrate that joint audit improves audit quality. We also, for robustness, use prevalent proxy for audit quality (Big N Auditor, ratio of audit fees to total fees) and find negative effect of joint audit on audit quality. We, therefore highlight that different proxy for audit quality show opposite effect of joint audit.

Keywords: audit fees, audit quality, Big N. Auditor, joint audit

Procedia PDF Downloads 358
1997 One-Dimensional Performance Improvement of a Single-Stage Transonic Compressor

Authors: A. Shahsavari, M. Nili-Ahmadabadi

Abstract:

This paper presents an innovative one-dimensional optimization of a transonic compressor based on the radial equilibrium theory by means of increasing blade loading. Firstly, the rotor blade of the transonic compressor is redesigned based on the constant span-wise deHaller number and diffusion. The code is applied to extract compressor meridional plane and blade to blade geometry containing rotor and stator in order to design blade three-dimensional view. A structured grid is generated for the numerical domain of fluid. Finer grids are used for regions near walls to capture boundary layer effects and behavior. RANS equations are solved by finite volume method for rotating zones (rotor) and stationary zones (stator). The experimental data, available for the performance map of NASA Rotor67, is used to validate the results of simulations. Then, the capability of the design method is validated by CFD that is capable of predicting the performance map. The numerical results of new geometry show about 19% increase in pressure ratio and 11% improvement in overall efficiency of the transonic stage; however, the design point mass flow rate of the new compressor is 5.7% less than that of the original compressor.

Keywords: deHaller number, one dimensional design, radial equilibrium, transonic compressor

Procedia PDF Downloads 341
1996 Design Guidelines for an Enhanced Interaction Experience in the Domain of Smartphone-Based Applications for Sport and Fitness

Authors: Paolo Pilloni, Fabrizio Mulas, Salvatore Carta

Abstract:

Nowadays, several research studies point up that an active lifestyle is essential for physical and mental health benefits. Mobile phones have greatly influenced people’s habits and attitudes also in the way they exercise. Our research work is mainly focused on investigating how to exploit mobile technologies to favour people’s exertion experience. To this end, we developed an exertion framework users can exploit through a real world mobile application, called BLINDED, designed to act as a virtual personal trainer to support runners during their trainings. In this work, inspired by both previous findings in the field of interaction design for people with visual impairments, feedback gathered from real users of our framework, and positive results obtained from two experimentations, we present some new interaction facilities we designed to enhance the interaction experience during a training. The positive obtained results helped us to derive some interaction design recommendations we believe will be a valid support for designers of future mobile systems conceived to be used in circumstances where there are limited possibilities of interaction.

Keywords: human computer interaction, interaction design guidelines, persuasive mobile technologies for sport and health

Procedia PDF Downloads 532
1995 Empirical Green’s Function Technique for Accelerogram Synthesis: The Problem of the Use for Marine Seismic Hazard Assessment

Authors: Artem A. Krylov

Abstract:

Instrumental seismological researches in water areas are complicated and expensive, that leads to the lack of strong motion records in most offshore regions. In the same time the number of offshore industrial infrastructure objects, such as oil rigs, subsea pipelines, is constantly increasing. The empirical Green’s function technique proved to be very effective for accelerograms synthesis under the conditions of poorly described seismic wave propagation medium. But the selection of suitable small earthquake record in offshore regions as an empirical Green’s function is a problem because of short seafloor instrumental seismological investigation results usually with weak micro-earthquakes recordings. An approach based on moving average smoothing in the frequency domain is presented for preliminary processing of weak micro-earthquake records before using it as empirical Green’s function. The method results in significant waveform correction for modeled event. The case study for 2009 L’Aquila earthquake was used to demonstrate the suitability of the method. This work was supported by the Russian Foundation of Basic Research (project № 18-35-00474 mol_a).

Keywords: accelerogram synthesis, empirical Green's function, marine seismology, microearthquakes

Procedia PDF Downloads 324
1994 Application of Wavelet Based Approximation for the Solution of Partial Integro-Differential Equation Arising from Viscoelasticity

Authors: Somveer Singh, Vineet Kumar Singh

Abstract:

This work contributes a numerical method based on Legendre wavelet approximation for the treatment of partial integro-differential equation (PIDE). Operational matrices of Legendre wavelets reduce the solution of PIDE into the system of algebraic equations. Some useful results concerning the computational order of convergence and error estimates associated to the suggested scheme are presented. Illustrative examples are provided to show the effectiveness and accuracy of proposed numerical method.

Keywords: legendre wavelets, operational matrices, partial integro-differential equation, viscoelasticity

Procedia PDF Downloads 448
1993 A Hebbian Neural Network Model of the Stroop Effect

Authors: Vadim Kulikov

Abstract:

The classical Stroop effect is the phenomenon that it takes more time to name the ink color of a printed word if the word denotes a conflicting color than if it denotes the same color. Over the last 80 years, there have been many variations of the experiment revealing various mechanisms behind semantic, attentional, behavioral and perceptual processing. The Stroop task is known to exhibit asymmetry. Reading the words out loud is hardly dependent on the ink color, but naming the ink color is significantly influenced by the incongruent words. This asymmetry is reversed, if instead of naming the color, one has to point at a corresponding color patch. Another debated aspects are the notions of automaticity and how much of the effect is due to semantic and how much due to response stage interference. Is automaticity a continuous or an all-or-none phenomenon? There are many models and theories in the literature tackling these questions which will be discussed in the presentation. None of them, however, seems to capture all the findings at once. A computational model is proposed which is based on the philosophical idea developed by the author that the mind operates as a collection of different information processing modalities such as different sensory and descriptive modalities, which produce emergent phenomena through mutual interaction and coherence. This is the framework theory where ‘framework’ attempts to generalize the concepts of modality, perspective and ‘point of view’. The architecture of this computational model consists of blocks of neurons, each block corresponding to one framework. In the simplest case there are four: visual color processing, text reading, speech production and attention selection modalities. In experiments where button pressing or pointing is required, a corresponding block is added. In the beginning, the weights of the neural connections are mostly set to zero. The network is trained using Hebbian learning to establish connections (corresponding to ‘coherence’ in framework theory) between these different modalities. The amount of data fed into the network is supposed to mimic the amount of practice a human encounters, in particular it is assumed that converting written text into spoken words is a more practiced skill than converting visually perceived colors to spoken color-names. After the training, the network performs the Stroop task. The RT’s are measured in a canonical way, as these are continuous time recurrent neural networks (CTRNN). The above-described aspects of the Stroop phenomenon along with many others are replicated. The model is similar to some existing connectionist models but as will be discussed in the presentation, has many advantages: it predicts more data, the architecture is simpler and biologically more plausible.

Keywords: connectionism, Hebbian learning, artificial neural networks, philosophy of mind, Stroop

Procedia PDF Downloads 267
1992 Cuckoo Search Optimization for Black Scholes Option Pricing

Authors: Manas Shah

Abstract:

Black Scholes option pricing model is one of the most important concepts in modern world of computational finance. However, its practical use can be challenging as one of the input parameters must be estimated; implied volatility of the underlying security. The more precisely these values are estimated, the more accurate their corresponding estimates of theoretical option prices would be. Here, we present a novel model based on Cuckoo Search Optimization (CS) which finds more precise estimates of implied volatility than Particle Swarm Optimization (PSO) and Genetic Algorithm (GA).

Keywords: black scholes model, cuckoo search optimization, particle swarm optimization, genetic algorithm

Procedia PDF Downloads 453
1991 Effect of the Applied Bias on Miniband Structures in Dimer Fibonacci Inas/Ga1-Xinxas Superlattices

Authors: Z. Aziz, S. Terkhi, Y. Sefir, R. Djelti, S. Bentata

Abstract:

The effect of a uniform electric field across multibarrier systems (InAs/InxGa1-xAs) is exhaustively explored by a computational model using exact airy function formalism and the transfer-matrix technique. In the case of biased DFHBSL structure a strong reduction in transmission properties was observed and the width of the miniband structure linearly decreases with the increase of the applied bias. This is due to the confinement of the states in the miniband structure, which becomes increasingly important (Wannier-Stark Effect).

Keywords: dimer fibonacci height barrier superlattices, singular extended state, exact airy function, transfer matrix formalism

Procedia PDF Downloads 307
1990 Urban Traffic: Understanding the Traffic Flow Factor Through Fluid Dynamics

Authors: Sathish Kumar Jayaraj

Abstract:

The study of urban traffic dynamics, underpinned by the principles of fluid dynamics, offers a distinct perspective to comprehend and enhance the efficiency of traffic flow within bustling cityscapes. Leveraging the concept of the Traffic Flow Factor (TFF) as an analog to the Reynolds number, this research delves into the intricate interplay between traffic density, velocity, and road category, drawing compelling parallels to fluid dynamics phenomena. By introducing the notion of Vehicle Shearing Resistance (VSR) as an analogy to dynamic viscosity, the study sheds light on the multifaceted influence of traffic regulations, lane management, and road infrastructure on the smoothness and resilience of traffic flow. The TFF equation serves as a comprehensive metric for quantifying traffic dynamics, enabling the identification of congestion hotspots, the optimization of traffic signal timings, and the formulation of data-driven traffic management strategies. The study underscores the critical significance of integrating fluid dynamics principles into the domain of urban traffic management, fostering sustainable transportation practices, and paving the way for a more seamless and resilient urban mobility ecosystem.

Keywords: traffic flow factor (TFF), urban traffic dynamics, fluid dynamics principles, vehicle shearing resistance (VSR), traffic congestion management, sustainable urban mobility

Procedia PDF Downloads 62
1989 Artificial Intelligent Tax Simulator to Minimize Tax Liability for Multinational Corporations

Authors: Sean Goltz, Michael Mayo

Abstract:

The purpose of this research is to use Global-Regulation.com database of the world laws, focusing on tax treaties between countries, in order to create an AI-driven tax simulator that will run an AI agent through potential tax scenarios across countries. The AI agent goal is to identify the scenario that will result in minimum tax liability based on tax treaties between countries. The results will be visualized by a three dimensional matrix. This will be an online web application. Multinational corporations are running their business through multiple countries. These countries, in turn, have a tax treaty with many other countries to regulate the payment of taxes on income that is transferred between these countries. As a result, planning the best tax scenario across multiple countries and numerous tax treaties is almost impossible. This research propose to use Global-Regulation.com database of word laws in English (machine translated by Google and Microsoft API’s) in order to create a simulator that will include the information in the tax treaties. Once ready, an AI agent will be sent through the simulator to identify the scenario that will result in minimum tax liability. Identifying the best tax scenario across countries may save multinational corporations, like Google, billions of dollars annually. Given the nature of the raw data and the domain of taxes (i.e., numbers), this is a promising ground to employ artificial intelligence towards a practical and beneficial purpose.

Keywords: taxation, law, multinational, corporation

Procedia PDF Downloads 199
1988 Adaptation of Hough Transform Algorithm for Text Document Skew Angle Detection

Authors: Kayode A. Olaniyi, Olabanji F. Omotoye, Adeola A. Ogunleye

Abstract:

The skew detection and correction form an important part of digital document analysis. This is because uncompensated skew can deteriorate document features and can complicate further document image processing steps. Efficient text document analysis and digitization can rarely be achieved when a document is skewed even at a small angle. Once the documents have been digitized through the scanning system and binarization also achieved, document skew correction is required before further image analysis. Research efforts have been put in this area with algorithms developed to eliminate document skew. Skew angle correction algorithms can be compared based on performance criteria. Most important performance criteria are accuracy of skew angle detection, range of skew angle for detection, speed of processing the image, computational complexity and consequently memory space used. The standard Hough Transform has successfully been implemented for text documentation skew angle estimation application. However, the standard Hough Transform algorithm level of accuracy depends largely on how much fine the step size for the angle used. This consequently consumes more time and memory space for increase accuracy and, especially where number of pixels is considerable large. Whenever the Hough transform is used, there is always a tradeoff between accuracy and speed. So a more efficient solution is needed that optimizes space as well as time. In this paper, an improved Hough transform (HT) technique that optimizes space as well as time to robustly detect document skew is presented. The modified algorithm of Hough Transform presents solution to the contradiction between the memory space, running time and accuracy. Our algorithm starts with the first step of angle estimation accurate up to zero decimal place using the standard Hough Transform algorithm achieving minimal running time and space but lacks relative accuracy. Then to increase accuracy, suppose estimated angle found using the basic Hough algorithm is x degree, we then run again basic algorithm from range between ±x degrees with accuracy of one decimal place. Same process is iterated till level of desired accuracy is achieved. The procedure of our skew estimation and correction algorithm of text images is implemented using MATLAB. The memory space estimation and process time are also tabulated with skew angle assumption of within 00 and 450. The simulation results which is demonstrated in Matlab show the high performance of our algorithms with less computational time and memory space used in detecting document skew for a variety of documents with different levels of complexity.

Keywords: hough-transform, skew-detection, skew-angle, skew-correction, text-document

Procedia PDF Downloads 159
1987 Optimizing E-commerce Retention: A Detailed Study of Machine Learning Techniques for Churn Prediction

Authors: Saurabh Kumar

Abstract:

In the fiercely competitive landscape of e-commerce, understanding and mitigating customer churn has become paramount for sustainable business growth. This paper presents a thorough investigation into the application of machine learning techniques for churn prediction in e-commerce, aiming to provide actionable insights for businesses seeking to enhance customer retention strategies. We conduct a comparative study of various machine learning algorithms, including traditional statistical methods and ensemble techniques, leveraging a rich dataset sourced from Kaggle. Through rigorous evaluation, we assess the predictive performance, interpretability, and scalability of each method, elucidating their respective strengths and limitations in capturing the intricate dynamics of customer churn. We identified the XGBoost classifier to be the best performing. Our findings not only offer practical guidelines for selecting suitable modeling approaches but also contribute to the broader understanding of customer behavior in the e-commerce domain. Ultimately, this research equips businesses with the knowledge and tools necessary to proactively identify and address churn, thereby fostering long-term customer relationships and sustaining competitive advantage.

Keywords: customer churn, e-commerce, machine learning techniques, predictive performance, sustainable business growth

Procedia PDF Downloads 29
1986 Computational and Experimental Determination of Acoustic Impedance of Internal Combustion Engine Exhaust

Authors: A. O. Glazkov, A. S. Krylova, G. G. Nadareishvili, A. S. Terenchenko, S. I. Yudin

Abstract:

The topic of the presented materials concerns the design of the exhaust system for a certain internal combustion engine. The exhaust system can be divided into two parts. The first is the engine exhaust manifold, turbocharger, and catalytic converters, which are called “hot part.” The second part is the gas exhaust system, which contains elements exclusively for reducing exhaust noise (mufflers, resonators), the accepted designation of which is the "cold part." The design of the exhaust system from the point of view of acoustics, that is, reducing the exhaust noise to a predetermined level, consists of working on the second part. Modern computer technology and software make it possible to design "cold part" with high accuracy in a given frequency range but with the condition of accurately specifying the input parameters, namely, the amplitude spectrum of the input noise and the acoustic impedance of the noise source in the form of an engine with a "hot part". Getting this data is a difficult problem: high temperatures, high exhaust gas velocities (turbulent flows), and high sound pressure levels (non-linearity mode) do not allow the calculated results to be applied with sufficient accuracy. The aim of this work is to obtain the most reliable acoustic output parameters of an engine with a "hot part" based on a complex of computational and experimental studies. The presented methodology includes several parts. The first part is a finite element simulation of the "cold part" of the exhaust system (taking into account the acoustic impedance of radiation of outlet pipe into open space) with the result in the form of the input impedance of "cold part". The second part is a finite element simulation of the "hot part" of the exhaust system (taking into account acoustic characteristics of catalytic units and geometry of turbocharger) with the result in the form of the input impedance of the "hot part". The next third part of the technique consists of the mathematical processing of the results according to the proposed formula for the convergence of the mathematical series of summation of multiple reflections of the acoustic signal "cold part" - "hot part". This is followed by conducting a set of tests on an engine stand with two high-temperature pressure sensors measuring pulsations in the nozzle between "hot part" and "cold part" of the exhaust system and subsequent processing of test results according to a well-known technique in order to separate the "incident" and "reflected" waves. The final stage consists of the mathematical processing of all calculated and experimental data to obtain a result in the form of a spectrum of the amplitude of the engine noise and its acoustic impedance.

Keywords: acoustic impedance, engine exhaust system, FEM model, test stand

Procedia PDF Downloads 59