Search results for: algorithm symbol recognition
1040 Pilot Induced Oscillations Adaptive Suppression in Fly-By-Wire Systems
Authors: Herlandson C. Moura, Jorge H. Bidinotto, Eduardo M. Belo
Abstract:
The present work proposes the development of an adaptive control system which enables the suppression of Pilot Induced Oscillations (PIO) in Digital Fly-By-Wire (DFBW) aircrafts. The proposed system consists of a Modified Model Reference Adaptive Control (M-MRAC) integrated with the Gain Scheduling technique. The PIO oscillations are detected using a Real Time Oscillation Verifier (ROVER) algorithm, which then enables the system to switch between two reference models; one in PIO condition, with low proneness to the phenomenon and another one in normal condition, with high (or medium) proneness. The reference models are defined in a closed loop condition using the Linear Quadratic Regulator (LQR) control methodology for Multiple-Input-Multiple-Output (MIMO) systems. The implemented algorithms are simulated in software implementations with state space models and commercial flight simulators as the controlled elements and with pilot dynamics models. A sequence of pitch angles is considered as the reference signal, named as Synthetic Task (Syntask), which must be tracked by the pilot models. The initial outcomes show that the proposed system can detect and suppress (or mitigate) the PIO oscillations in real time before it reaches high amplitudes.Keywords: adaptive control, digital Fly-By-Wire, oscillations suppression, PIO
Procedia PDF Downloads 1341039 Robust Medical Image Watermarking based on Contourlet and Extraction Using ICA
Authors: S. Saju, G. Thirugnanam
Abstract:
In this paper, a medical image watermarking algorithm based on contourlet is proposed. Medical image watermarking is a special subcategory of image watermarking in the sense that images have special requirements. Watermarked medical images should not differ perceptually from their original counterparts because clinical reading of images must not be affected. Watermarking techniques based on wavelet transform are reported in many literatures but robustness and security using contourlet are better when compared to wavelet transform. The main challenge in exploring geometry in images comes from the discrete nature of the data. In this paper, original image is decomposed to two level using contourlet and the watermark is embedded in the resultant sub-bands. Sub-band selection is based on the value of Peak Signal to Noise Ratio (PSNR) that is calculated between watermarked and original image. To extract the watermark, Kernel ICA is used and it has a novel characteristic is that it does not require the transformation process to extract the watermark. Simulation results show that proposed scheme is robust against attacks such as Salt and Pepper noise, Median filtering and rotation. The performance measures like PSNR and Similarity measure are evaluated and compared with Discrete Wavelet Transform (DWT) to prove the robustness of the scheme. Simulations are carried out using Matlab Software.Keywords: digital watermarking, independent component analysis, wavelet transform, contourlet
Procedia PDF Downloads 5291038 Deploying a Transformative Learning Model in Technological University Dublin to Assess Transversal Skills
Authors: Sandra Thompson, Paul Dervan
Abstract:
Ireland’s first Technological University (TU Dublin) was established on 1st January 2019, and its creation is an exciting new milestone in Irish Higher Education. TU Dublin is now Ireland’s biggest University supporting 29,000 students across three campuses with 3,500 staff. The University aspires to create work-ready graduates who are socially responsible, open-minded global thinkers who are ambitious to change the world for the better. As graduates, they will be enterprising and daring in all their endeavors, ready to play their part in transforming the future. Feedback from Irish employers and students coupled with evidence from other authoritative sources such as the World Economic Forum points to a need for greater focus on the development of students’ employability skills as they prepare for today’s work environment. Moreover, with an increased focus on Universal Design for Learning (UDL) and inclusiveness, there is recognition that students are more than a numeric grade value. Robust grading systems have been developed to track a student’s performance around discipline knowledge but there is little or no global consensus on a definition of transversal skills nor on a unified framework to assess transversal skills. Education and industry sectors are often assessing one or two skills, and some are developing their own frameworks to capture the learner’s achievement in this area. Technological University Dublin (TU Dublin) have discovered and implemented a framework to allow students to develop, assess and record their transversal skills using transformative learning theory. The model implemented is an adaptation of Student Transformative Learning Record - STLR which originated in the University of Central Oklahoma (UCO). The purpose of this paper therefore, is to examine the views of students, staff and employers in the context of deploying a Transformative Learning model within the University to assess transversal skills. It will examine the initial impact the transformative learning model is having socially, personally and on the University as an organization. Crucially also, to identify lessons learned from the deployment in order to assist other Universities and Higher Education Institutes who may be considering a focused adoption of Transformative Learning to meet the challenge of preparing students for today’s work environment.Keywords: assessing transversal skills, higher education, transformative learning, students
Procedia PDF Downloads 1301037 Blood Oxygen Saturation Measurement System Using Broad-Band Light Source with LabVIEW Program
Authors: Myoung Ah Kim, Dong Ho Sin, Chul Gyu Song
Abstract:
Blood oxygen saturation system is a well-established, noninvasive photoplethysmographic method to monitor vital signs. Conventional blood oxygen saturation measurements for the two LED light source is the ambiguity of the oxygen saturation measurement principle and the measurement results greatly influenced and heat and motion artifact. A high accuracy in order to solve these problems blood oxygen saturation measuring method has been proposed using a broadband light source that can be easily understood by the algorithm. The measurement of blood oxygen saturation based on broad-band light source has advantage of simple testing facility and easy understanding. Broadband light source based on blood oxygen saturation measuring program proposed in this paper is a combination of LabVIEW and MATLAB. Using the wavelength range of 450 nm-750 nm using a floating light absorption of oxyhemoglobin and deoxyhemoglobin to measure the blood oxygen saturation. Hand movement is to fix the probe to the motor stage in order to prevent oxygen saturation measurement that affect the sample and probe kept constant interval. Experimental results show that the proposed method noticeably increases the accuracy and saves time compared with the conventional methods.Keywords: oxygen saturation, broad-band light source, CCD, light reflectance theory
Procedia PDF Downloads 4611036 Research of Control System for Space Intelligent Robot Based on Vision Servo
Authors: Changchun Liang, Xiaodong Zhang, Xin Liu, Pengfei Sun
Abstract:
Space intelligent robotic systems are expected to play an increasingly important role in the future. The robotic on-orbital service, whose key is the tracking and capturing technology, becomes research hot in recent years. In this paper, the authors propose a vision servo control system for target capturing. Robotic manipulator will be an intelligent robotic system with large-scale movement, functional agility, and autonomous ability, and it can be operated by astronauts in the space station or be controlled by the ground operator in the remote operation mode. To realize the autonomous movement and capture mission of SRM, a kind of autonomous programming strategy based on multi-camera vision fusion is designed and the selection principle of object visual position and orientation measurement information is defined for the better precision. Distributed control system hierarchy is designed and reliability is considering to guarantee the abilities of control system. At last, a ground experiment system is set up based on the concept of robotic control system. With that, the autonomous target capturing experiments are conducted. The experiment results validate the proposed algorithm, and demonstrates that the control system can fulfill the needs of function, real-time and reliability.Keywords: control system, on-orbital service, space robot, vision servo
Procedia PDF Downloads 4191035 “Everything, Everywhere, All at Once” Hollywoodization and Lack of Authenticity in Today’s Mainstream Cinema
Authors: Haniyeh Parhizkar
Abstract:
When Sarris came up with the "auteur theory" in 1962, he emphasized that the utmost premise of auteur theory is the inner meanings and concepts of a film and that a film is purely an art form. Today's mainstream movies are conceptually closer to what the Frankfurt School scholars regarded as "reproduced" and "mass culture" years ago. Hollywood goes on to be a huge movie-making machine that leads the dominant paradigms of films throughout the world and cinema is far from art. Although there are still movies, directors, and audiences who favor art cinema over Hollywood and mainstream movies, it's an almost undeniable fact that, for the most part, people's perception of movies is widely influenced by their American depiction and Hollywood's legacy of mass culture. With the uprising of Hollywood studios as the forerunners of the movie industry and cinema being largely dependent on economics rather than artistic values, this distinctive role of cinema has diminished and is replaced with a global standard. The Blockbuster 2022 film, 'Everything, Everywhere, All at Once' is now the most-awarded movie of all time, winning seven Oscars at the 95th Academy Awards. Despite its main cast being Asian, the movie is produced by American incorporation and is heavily influenced by Hollywood's dominant themes of superheroes, fantasy, action, and adventure. The New Yorker film critic, Richard Brody, called the movie "a pitch for a Marvel" and critiqued the film for being "universalized" and "empty of history and culture". Other critics of Variety pinpointed the movie's similarities to Marvel, particularly in their storylines of multi-universe which manifest traces of American legacy. As argued by these critics, 'Everything, Everywhere, All at Once' might appear as a unique and authentic film at first glance, but it can be argued that it is yet another version of a Marvel movie. While the movie's universal acclaim was regarded as recognition and an acknowledgment of its Asian cast, the issue that arises here is when the Hollywood influences and American themes are so robust in the film, is the movie industry honoring another culture or is it yet another celebration of Hollywood's dominant paradigm. This essay will employ a critical approach to Hollywood's dominance and mass-produced culture, which has deprived authenticity of non-American movies and is constantly reproducing the same formula of success.Keywords: hollywoodization, universalization, blockbuster, dominant paradigm, marvel, authenticity, diversity
Procedia PDF Downloads 901034 High Performance of Direct Torque and Flux Control of a Double Stator Induction Motor Drive with a Fuzzy Stator Resistance Estimator
Authors: K. Kouzi
Abstract:
In order to have stable and high performance of direct torque and flux control (DTFC) of double star induction motor drive (DSIM), proper on-line adaptation of the stator resistance is very important. This is inevitably due to the variation of the stator resistance during operating conditions, which introduces error in estimated flux position and the magnitude of the stator flux. Error in the estimated stator flux deteriorates the performance of the DTFC drive. Also, the effect of error in estimation is very important especially at low speed. Due to this, our aim is to overcome the sensitivity of the DTFC to the stator resistance variation by proposing on-line fuzzy estimation stator resistance. The fuzzy estimation method is based on an on-line stator resistance correction through the variations of the stator current estimation error and its variations. The fuzzy logic controller gives the future stator resistance increment at the output. The main advantage of the suggested algorithm control is to avoid the drive instability that may occur in certain situations and ensure the tracking of the actual stator resistance. The validity of the technique and the improvement of the whole system performance are proved by the results.Keywords: direct torque control, dual stator induction motor, Fuzzy Logic estimation, stator resistance adaptation
Procedia PDF Downloads 3281033 A Semantic and Concise Structure to Represent Human Actions
Authors: Tobias Strübing, Fatemeh Ziaeetabar
Abstract:
Humans usually manipulate objects with their hands. To represent these actions in a simple and understandable way, we need to use a semantic framework. For this purpose, the Semantic Event Chain (SEC) method has already been presented which is done by consideration of touching and non-touching relations between manipulated objects in a scene. This method was improved by a computational model, the so-called enriched Semantic Event Chain (eSEC), which incorporates the information of static (e.g. top, bottom) and dynamic spatial relations (e.g. moving apart, getting closer) between objects in an action scene. This leads to a better action prediction as well as the ability to distinguish between more actions. Each eSEC manipulation descriptor is a huge matrix with thirty rows and a massive set of the spatial relations between each pair of manipulated objects. The current eSEC framework has so far only been used in the category of manipulation actions, which eventually involve two hands. Here, we would like to extend this approach to a whole body action descriptor and make a conjoint activity representation structure. For this purpose, we need to do a statistical analysis to modify the current eSEC by summarizing while preserving its features, and introduce a new version called Enhanced eSEC or (e2SEC). This summarization can be done from two points of the view: 1) reducing the number of rows in an eSEC matrix, 2) shrinking the set of possible semantic spatial relations. To achieve these, we computed the importance of each matrix row in an statistical way, to see if it is possible to remove a particular one while all manipulations are still distinguishable from each other. On the other hand, we examined which semantic spatial relations can be merged without compromising the unity of the predefined manipulation actions. Therefore by performing the above analyses, we made the new e2SEC framework which has 20% fewer rows, 16.7% less static spatial and 11.1% less dynamic spatial relations. This simplification, while preserving the salient features of a semantic structure in representing actions, has a tremendous impact on the recognition and prediction of complex actions, as well as the interactions between humans and robots. It also creates a comprehensive platform to integrate with the body limbs descriptors and dramatically increases system performance, especially in complex real time applications such as human-robot interaction prediction.Keywords: enriched semantic event chain, semantic action representation, spatial relations, statistical analysis
Procedia PDF Downloads 1261032 A Numerical Study on Electrophoresis of a Soft Particle with Charged Core Coated with Polyelectrolyte Layer
Authors: Partha Sarathi Majee, S. Bhattacharyya
Abstract:
Migration of a core-shell soft particle under the influence of an external electric field in an electrolyte solution is studied numerically. The soft particle is coated with a positively charged polyelectrolyte layer (PEL) and the rigid core is having a uniform surface charge density. The Darcy-Brinkman extended Navier-Stokes equations are solved for the motion of the ionized fluid, the non-linear Nernst-Planck equations for the ion transport and the Poisson equation for the electric potential. A pressure correction based iterative algorithm is adopted for numerical computations. The effects of convection on double layer polarization (DLP) and diffusion dominated counter ions penetration are investigated for a wide range of Debye layer thickness, PEL fixed surface charge density, and permeability of the PEL. Our results show that when the Debye layer is in order of the particle size, the DLP effect is significant and produces a reduction in electrophoretic mobility. However, the double layer polarization effect is negligible for a thin Debye layer or low permeable cases. The point of zero mobility and the existence of mobility reversal depending on the electrolyte concentration are also presented.Keywords: debye length, double layer polarization, electrophoresis, mobility reversal, soft particle
Procedia PDF Downloads 3461031 DNA Nano Wires: A Charge Transfer Approach
Authors: S. Behnia, S. Fathizadeh, A. Akhshani
Abstract:
In the recent decades, DNA has increasingly interested in the potential technological applications that not directly related to the coding for functional proteins that is the expressed in form of genetic information. One of the most interesting applications of DNA is related to the construction of nanostructures of high complexity, design of functional nanostructures in nanoelectronical devices, nanosensors and nanocercuits. In this field, DNA is of fundamental interest to the development of DNA-based molecular technologies, as it possesses ideal structural and molecular recognition properties for use in self-assembling nanodevices with a definite molecular architecture. Also, the robust, one-dimensional flexible structure of DNA can be used to design electronic devices, serving as a wire, transistor switch, or rectifier depending on its electronic properties. In order to understand the mechanism of the charge transport along DNA sequences, numerous studies have been carried out. In this regard, conductivity properties of DNA molecule could be investigated in a simple, but chemically specific approach that is intimately related to the Su-Schrieffer-Heeger (SSH) model. In SSH model, the non-diagonal matrix element dependence on intersite displacements is considered. In this approach, the coupling between the charge and lattice deformation is along the helix. This model is a tight-binding linear nanoscale chain established to describe conductivity phenomena in doped polyethylene. It is based on the assumption of a classical harmonic interaction between sites, which is linearly coupled to a tight-binding Hamiltonian. In this work, the Hamiltonian and corresponding motion equations are nonlinear and have high sensitivity to initial conditions. Then, we have tried to move toward the nonlinear dynamics and phase space analysis. Nonlinear dynamics and chaos theory, regardless of any approximation, could open new horizons to understand the conductivity mechanism in DNA. For a detailed study, we have tried to study the current flowing in DNA and investigated the characteristic I-V diagram. As a result, It is shown that there are the (quasi-) ohmic areas in I-V diagram. On the other hand, the regions with a negative differential resistance (NDR) are detectable in diagram.Keywords: DNA conductivity, Landauer resistance, negative dierential resistance, Chaos theory, mean Lyapunov exponent
Procedia PDF Downloads 4261030 Bank Liquidity Creation in a Dual Banking System: An Empirical Investigation
Authors: Lianne M. Q. Lee, Mohammed Sharaf Shaiban
Abstract:
The importance of bank liquidity management took center stage as policy makers promoted a more resilient global banking system after the market turmoil of 2007. The growing recognition of Islamic banks’ function of intermediating funds in the economy warrants the need to investigate its balance sheet structure which is distinct from its conventional counterparts. Given that asymmetric risk, transformation is inevitable; Islamic banks need to identify the liquidity risk within their distinctive balance sheet structure. Thus, there is a strong need to quantify and assess the liquidity position to ensure proper functioning of a financial institution. It is vital to measure bank liquidity because liquid banks face less liquidity risk. We examine this issue by using two alternative quantitative measures of liquidity creation “cat fat” and “cat nonfat” constructed by Berger and Bouwman (2009). “Cat fat” measures all on balance sheet items including off balance sheet, whilst the latter measures only on balance sheet items. Liquidity creation is measured over the period 2007-2014 in 14 countries where Islamic and conventional commercial banks coexist. Also, separately by bank size class as empirical studies have shown that liquidity creation varies by bank size. An interesting and important finding shows that all size class of Islamic banks, on average have increased creation of aggregate liquidity in real dollar terms over the years for both liquidity creation measures especially for large banks indicating that Islamic banks actually generates more liquidity to the economy compared to its conventional counterparts, including from off-balance sheet items. The liquidity creation for off-balance sheets by conventional banks may have been affected by the global financial crisis when derivatives markets were severely hit. The results also suggest that Islamic banks have the higher volume of assets and deposits and that borrowing/issues of bonds are less in Islamic banks compared to conventional banks because most products are interest-based. As Islamic banks appear to create more liquidity than conventional banks under both measures, it translates that the development of Islamic banking is significant over the decades since its inception. This finding is encouraging as, despite Islamic banking’s overall size, it represents growth opportunities for these countries.Keywords: financial institution, liquidity creation, liquidity risk, policy and regulation
Procedia PDF Downloads 3531029 Turn-Taking and Leading Roles in Early Cognition: Interaction of Social Cognition and Language in Development
Authors: Zsuzsanna Schnell, Francesca Ervas
Abstract:
Background: Our study aims to clarify how language fosters further cognitive development and how we eventually arrive at the complex human specific skill of pragmatic competence and reveal what levels of mentalization and theory of mind are in place before language. Method: Our experimental pragmatic investigation maps the interaction of mentalization and pragmatic competence. We map the different levels of mentalization that empower different levels of pragmatic meaning construction and evaluate the results with statistical analysis (MannWhitney and ANOVA). Analyzing the comprehension of literal and non-compositional (figurative) utterances, we apply linguistic trials, among them metaphor-, irony-, irony with surface cue-, humor- and the recognition of maxim infringements trial in neurotypical (NT) preschoolers with a coherent and comparative methodology. Results: The findings reveal the relationship and direction of interaction between Language and theory of mind. On the one hand social-cognitive skills enhance, facilitate and provide a basis for language acquisition, and in return linguistic structures (DeVilliers 2000, 2007) provide a framework for further development of mentalizing skills. Conclusions: Findings confirm that this scaffolding becomes a mutually supportive system where language and social cognition develops in interaction. Certain stages in ToM development serve as a precursor of understanding grammatically complex sentences, like embedded phrases which mirror embedded mental states; which, in turn, facilitates the development of pragmatic competence, thus, the social use of language, integrating social, cognitive, linguistic and psychological factors in discourse. Future implications: Our investigation functions as a differential-diagnostic measure, with typically developing results thus serve as a baseline in further empirical research for atypical cases. This enables the study of populations where language and ToM development is disturbed, reveals how language and ToM are acquired and interact, and gives an insight into what this has to do with clinical symptoms. This in turn can reveal the causal link to the syndrome at hand, which can set directions for therapeutic development and training.Keywords: theory of mind, language development, mentalization, language philosophy, experimental pragmatics
Procedia PDF Downloads 321028 Blockchain-Resilient Framework for Cloud-Based Network Devices within the Architecture of Self-Driving Cars
Authors: Mirza Mujtaba Baig
Abstract:
Artificial Intelligence (AI) is evolving rapidly, and one of the areas in which this field has influenced is automation. The automobile, healthcare, education, and robotic industries deploy AI technologies constantly, and the automation of tasks is beneficial to allow time for knowledge-based tasks and also introduce convenience to everyday human endeavors. The paper reviews the challenges faced with the current implementations of autonomous self-driving cars by exploring the machine learning, robotics, and artificial intelligence techniques employed for the development of this innovation. The controversy surrounding the development and deployment of autonomous machines, e.g., vehicles, begs the need for the exploration of the configuration of the programming modules. This paper seeks to add to the body of knowledge of research assisting researchers in decreasing the inconsistencies in current programming modules. Blockchain is a technology of which applications are mostly found within the domains of financial, pharmaceutical, manufacturing, and artificial intelligence. The registering of events in a secured manner as well as applying external algorithms required for the data analytics are especially helpful for integrating, adapting, maintaining, and extending to new domains, especially predictive analytics applications.Keywords: artificial intelligence, automation, big data, self-driving cars, machine learning, neural networking algorithm, blockchain, business intelligence
Procedia PDF Downloads 1201027 Predicting Destination Station Based on Public Transit Passenger Profiling
Authors: Xuyang Song, Jun Yin
Abstract:
The smart card has been an extremely universal tool in public transit. It collects a large amount of data on buses, urban railway transit, and ferries and provides possibilities for passenger profiling. This paper combines offline analysis of passenger profiling and real-time prediction to propose a method that can accurately predict the destination station in real-time when passengers tag on. Firstly, this article constructs a static database of user travel characteristics after identifying passenger travel patterns based on the Density-Based Spatial Clustering of Applications with Noise (DBSCAN). The dual travel passenger habits are identified: OD travel habits and D station travel habits. Then a rapid real-time prediction algorithm based on Transit Passenger Profiling is proposed, which can predict the destination of in-board passengers. This article combines offline learning with online prediction, providing a technical foundation for real-time passenger flow prediction, monitoring and simulation, and short-term passenger behavior and demand prediction. This technology facilitates the efficient and real-time acquisition of passengers' travel destinations and demand. The last, an actual case was simulated and demonstrated feasibility and efficiency.Keywords: travel behavior, destination prediction, public transit, passenger profiling
Procedia PDF Downloads 211026 Linguistic Competencies of Students with Hearing Impairment
Authors: Munawar Malik, Muntaha Ahmad, Khalil Ullah Khan
Abstract:
Linguistic abilities in students with hearing impairment yet remain a concern for educationists. The emerging technological support and provisions in recent era vows to have addressed the situation and claims significant contribution in terms of linguistic repertoire. Being a descriptive and quantitative paradigm of study, the purpose of this research set forth was to assess linguistic competencies of students with hearing impairment in English language. The goals were further broken down to identify level of reading abilities in the subject population. The population involved students with HI studying at higher secondary level in Lahore. Simple random sampling technique was used to choose a sample of fifty students. A purposive curriculum-based assessment was designed in line with accelerated learning program by Punjab Government, to assess Linguistic competence among the sample. Further to it, an Informal Reading Inventory (IRI) corresponding to reading levels was also developed by researchers duly validated and piloted before the final use. Descriptive and inferential statistics were utilized to reach to the findings. Spearman’s correlation was used to find out relationship between degree of hearing loss, grade level, gender and type of amplification device. Independent sample t-test was used to compare means among groups. Major findings of the study revealed that students with hearing impairment exhibit significant deviation from the mean scores when compared in terms of grades, severity and amplification device. The study divulged that respective students with HI have yet failed to qualify an independent level of reading according to their grades as majority falls at frustration level of word recognition and passage comprehension. The poorer performance can be attributed to lower linguistic competence as it shows in the frustration levels of reading, writing and comprehension. The correlation analysis did reflect an improved performance grade wise, however scores could only correspond to frustration level and independent levels was never achieved. Reported achievements at instructional level of subject population may further to linguistic skills if practiced purposively.Keywords: linguistic competence, hearing impairment, reading levels, educationist
Procedia PDF Downloads 691025 Field-Programmable Gate Arrays Based High-Efficiency Oriented Fast and Rotated Binary Robust Independent Elementary Feature Extraction Method Using Feature Zone Strategy
Authors: Huang Bai-Cheng
Abstract:
When deploying the Oriented Fast and Rotated Binary Robust Independent Elementary Feature (BRIEF) (ORB) extraction algorithm on field-programmable gate arrays (FPGA), the access of global storage for 31×31 pixel patches of the features has become the bottleneck of the system efficiency. Therefore, a feature zone strategy has been proposed. Zones are searched as features are detected. Pixels around the feature zones are extracted from global memory and distributed into patches corresponding to feature coordinates. The proposed FPGA structure is targeted on a Xilinx FPGA development board of Zynq UltraScale+ series, and multiple datasets are tested. Compared with the streaming pixel patch extraction method, the proposed architecture obtains at least two times acceleration consuming extra 3.82% Flip-Flops (FFs) and 7.78% Look-Up Tables (LUTs). Compared with the non-streaming one, the proposed architecture saves 22.3% LUT and 1.82% FF, causing a latency of only 0.2ms and a drop in frame rate for 1. Compared with the related works, the proposed strategy and hardware architecture have the superiority of keeping a balance between FPGA resources and performance.Keywords: feature extraction, real-time, ORB, FPGA implementation
Procedia PDF Downloads 1221024 Hybrid Genetic Approach for Solving Economic Dispatch Problems with Valve-Point Effect
Authors: Mohamed I. Mahrous, Mohamed G. Ashmawy
Abstract:
Hybrid genetic algorithm (HGA) is proposed in this paper to determine the economic scheduling of electric power generation over a fixed time period under various system and operational constraints. The proposed technique can outperform conventional genetic algorithms (CGAs) in the sense that HGA make it possible to improve both the quality of the solution and reduce the computing expenses. In contrast, any carefully designed GA is only able to balance the exploration and the exploitation of the search effort, which means that an increase in the accuracy of a solution can only occure at the sacrifice of convergent speed, and vice visa. It is unlikely that both of them can be improved simultaneously. The proposed hybrid scheme is developed in such a way that a simple GA is acting as a base level search, which makes a quick decision to direct the search towards the optimal region, and a local search method (pattern search technique) is next employed to do the fine tuning. The aim of the strategy is to achieve the cost reduction within a reasonable computing time. The effectiveness of the proposed hybrid technique is verified on two real public electricity supply systems with 13 and 40 generator units respectively. The simulation results obtained with the HGA for the two real systems are very encouraging with regard to the computational expenses and the cost reduction of power generation.Keywords: genetic algorithms, economic dispatch, pattern search
Procedia PDF Downloads 4461023 Gestalt in Music and Brain: A Non-Linear Chaos Based Study with Detrended/Adaptive Fractal Analysis
Authors: Shankha Sanyal, Archi Banerjee, Sayan Biswas, Sourya Sengupta, Sayan Nag, Ranjan Sengupta, Dipak Ghosh
Abstract:
The term ‘gestalt’ has been widely used in the field of psychology which defined the perception of human mind to group any object not in part but as a 'unified' whole. Music, in general, is polyphonic - i.e. a combination of a number of pure tones (frequencies) mixed together in a manner that sounds harmonious. The study of human brain response due to different frequency groups of the acoustic signal can give us an excellent insight regarding the neural and functional architecture of brain functions. Hence, the study of music cognition using neuro-biosensors is becoming a rapidly emerging field of research. In this work, we have tried to analyze the effect of different frequency bands of music on the various frequency rhythms of human brain obtained from EEG data. Four widely popular Rabindrasangeet clips were subjected to Wavelet Transform method for extracting five resonant frequency bands from the original music signal. These frequency bands were initially analyzed with Detrended/Adaptive Fractal analysis (DFA/AFA) methods. A listening test was conducted on a pool of 100 respondents to assess the frequency band in which the music becomes non-recognizable. Next, these resonant frequency bands were presented to 20 subjects as auditory stimulus and EEG signals recorded simultaneously in 19 different locations of the brain. The recorded EEG signals were noise cleaned and subjected again to DFA/AFA technique on the alpha, theta and gamma frequency range. Thus, we obtained the scaling exponents from the two methods in alpha, theta and gamma EEG rhythms corresponding to different frequency bands of music. From the analysis of music signal, it is seen that loss of recognition is proportional to the loss of long range correlation in the signal. From the EEG signal analysis, we obtain frequency specific arousal based response in different lobes of brain as well as in specific EEG bands corresponding to musical stimuli. In this way, we look to identify a specific frequency band beyond which the music becomes non-recognizable and below which in spite of the absence of other bands the music is perceivable to the audience. This revelation can be of immense importance when it comes to the field of cognitive music therapy and researchers of creativity.Keywords: AFA, DFA, EEG, gestalt in music, Hurst exponent
Procedia PDF Downloads 3321022 Experimental and Numerical Investigation of Heat Transfer in THTL Test Loop Shell and Tube Heat Exchanger
Authors: M. Moody, R. Mahmoodi, A. R. Zolfaghari, A. Aminottojari
Abstract:
In this study, flow inside the shell side of a shell-and-tube heat exchanger is simulated numerically for laminar and turbulent flows in both steady state and transient mode. Governing equations of fluid flow are discrete using finite volume method and central difference scheme and solved with simple algorithm which is staggered grid by using MATLAB programming language. The heat transfer coefficient is obtained using velocity field from equation Dittus-Bolter. In comparison with, heat exchanger is simulated with ANSYS CFX software and experimental data measured in the THTL test loop. Numerical results obtained from the study show good agreement with experimental data and ANSYS CFX results. In addition, by deliberation the effect of the baffle spacing and the baffle cut on the heat transfer rate for turbulent flow, it is illustrated that the heat transfer rate depends on the baffle spacing and the baffle cut directly. In other word in spied of large turbulence, if these two parameters are not selected properly in the heat exchanger, the heat transfer rate can reduce.Keywords: shell-and-tube heat exchanger, flow and heat transfer, laminar and turbulence flow, turbulence model, baffle spacing, baffle cut
Procedia PDF Downloads 5381021 Comparative Study on Manet Using Soft Computing Techniques
Authors: Amarjit Singh, Tripatdeep Singh Dua, Vikas Attri
Abstract:
Mobile Ad-hoc Network is a combination of several nodes that create dynamically a specific network without using any base infrastructure. In this study all the mobile nodes can depended upon each other to send any data. Mobile host can pick up data and forwarding to their destination path. Basically MANET depend upon their Quality of Service which is highly constraints to the user. To give better services we need to improve the QOS. In these days MANET QOS requirement to use soft computing techniques. These techniques depend upon their specific requirement and which exists using MANET concepts. Using a soft computing techniques various protocol and algorithms may be considered. In this paper, we provide comparative study review of existing work done in MANET using various kind of soft computing techniques. Our review research is based on their specific protocol or algorithm which provide concern solution of QOS need. We discuss about various protocol through which routing in MANET. In Second section we clear the concepts of Soft Computing and their types. In third section we review the MANET using different kind of soft computing techniques work done before. In forth section we need to understand the concept of QoS requirement which exists in MANET and we done comparative study on different protocol used before and last we conclude the purpose of using MANET with soft computing techniques metrics.Keywords: mobile ad-hoc network, fuzzy improved genetic approach, neural network, routing protocol, wireless mesh network
Procedia PDF Downloads 3511020 Modeling the Philippine Stock Exchange Index Closing Value Using Artificial Neural Network
Authors: Frankie Burgos, Emely Munar, Conrado Basa
Abstract:
This paper aimed at developing an artificial neural network (ANN) model specifically for the Philippine Stock Exchange index closing value. The inputs to the ANN are US Dollar and Philippine Peso(USD-PHP) exchange rate, GDP growth of the country, quarterly inflation rate, 10-year bond yield, credit rating of the country, previous open, high, low, close values and volume of trade of the Philippine Stock Exchange Index (PSEi), gold price of the previous day, National Association of Securities Dealers Automated Quotations (NASDAQ), Standard and Poor’s 500 (S & P 500) and the iShares MSCI Philippines ETF (EPHE) previous closing value. The target is composed of the closing value of the PSEi during the 627 trading days from November 3, 2011, to May 30, 2014. MATLAB’s Neural Network toolbox was employed to create, train and simulate the network using multi-layer feed forward neural network with back-propagation algorithm. The results satisfactorily show that the neural network developed has the ability to model the PSEi, which is affected by both internal and external economic factors. It was found out that the inputs used are the main factors that influence the movement of the PSEi closing value.Keywords: artificial neural networks, artificial intelligence, philippine stocks exchange index, stocks trading
Procedia PDF Downloads 2981019 An Application of Meta-Modeling Methods for Surrogating Lateral Dynamics Simulation in Layout-Optimization for Electric Drivetrains
Authors: Christian Angerer, Markus Lienkamp
Abstract:
Electric vehicles offer a high variety of possible drivetrain topologies with up to 4 motors. Multi-motor-designs can have several advantages regarding traction, vehicle dynamics, safety and even efficiency. With a rising number of motors, the whole drivetrain becomes more complex. All permutations of gearings, drivetrain-layouts, motor-types and –sizes lead up in a very large solution space. Single elements of this solution space can be analyzed by simulation methods. In addition to longitudinal vehicle behavior, which most optimization-approaches are restricted to, also lateral dynamics are important for vehicle dynamics, stability and efficiency. In order to compete large solution spaces and to find an optimal result, genetic algorithm based optimization is state-of-the-art. As lateral dynamics simulation is way more CPU-intensive, optimization takes much more time than in case of longitudinal-only simulation. Therefore, this paper shows an approach how to create meta-models from a 14-degree of freedom vehicle model in order to enable a numerically efficient drivetrain-layout optimization process under consideration of lateral dynamics. Different meta-modelling approaches such as neural networks or DoE are implemented and comparatively discussed.Keywords: driving dynamics, drivetrain layout, genetic optimization, meta-modeling, lateral dynamicx
Procedia PDF Downloads 4181018 Multi-Level Clustering Based Congestion Control Protocol for Cyber Physical Systems
Authors: Manpreet Kaur, Amita Rani, Sanjay Kumar
Abstract:
The Internet of Things (IoT), a cyber-physical paradigm, allows a large number of devices to connect and send the sensory data in the network simultaneously. This tremendous amount of data generated leads to very high network load consequently resulting in network congestion. It further amounts to frequent loss of useful information and depletion of significant amount of nodes’ energy. Therefore, there is a need to control congestion in IoT so as to prolong network lifetime and improve the quality of service (QoS). Hence, we propose a two-level clustering based routing algorithm considering congestion score and packet priority metrics that focus on minimizing the network congestion. In the proposed Priority based Congestion Control (PBCC) protocol the sensor nodes in IoT network form clusters that reduces the amount of traffic and the nodes are prioritized to emphasize important data. Simultaneously, a congestion score determines the occurrence of congestion at a particular node. The proposed protocol outperforms the existing Packet Discard Network Clustering (PDNC) protocol in terms of buffer size, packet transmission range, network region and number of nodes, under various simulation scenarios.Keywords: internet of things, cyber-physical systems, congestion control, priority, transmission rate
Procedia PDF Downloads 3081017 Task Based Functional Connectivity within Reward Network in Food Image Viewing Paradigm Using Functional MRI
Authors: Preetham Shankapal, Jill King, Kori Murray, Corby Martin, Paula Giselman, Jason Hicks, Owen Carmicheal
Abstract:
Activation of reward and satiety networks in the brain while processing palatable food cues, as well as functional connectivity during rest has been studied using functional Magnetic Resonance Imaging of the brain in various obesity phenotypes. However, functional connectivity within the reward and satiety network during food cue processing is understudied. 14 obese individuals underwent two fMRI scans during viewing of Macronutrient Picture System images. Each scan included two blocks of images of High Sugar/High Fat (HSHF), High Carbohydrate/High Fat (HCHF), Low Sugar/Low Fat (LSLF) and also non-food images. Seed voxels within seven food reward relevant ROIs: Insula, putamen and cingulate, precentral, parahippocampal, medial frontal and superior temporal gyri were isolated based on a prior meta-analysis. Beta series correlation for task-related functional connectivity between these seed voxels and the rest of the brain was computed. Voxel-level differences in functional connectivity were calculated between: first and the second scan; individuals who saw novel (N=7) vs. Repeated (N=7) images in the second scan; and between the HC/HF, HSHF blocks vs LSLF and non-food blocks. Computations and analysis showed that during food image viewing, reward network ROIs showed significant functional connectivity with each other and with other regions responsible for attentional and motor control, including inferior parietal lobe and precentral gyrus. These functional connectivity values were heightened among individuals who viewed novel HS/HF images in the second scan. In the second scan session, functional connectivity was reduced within the reward network but increased within attention, memory and recognition regions, suggesting habituation to reward properties and increased recollection of previously viewed images. In conclusion it can be inferred that Functional Connectivity within reward network and between reward and other brain regions, varies by important experimental conditions during food photography viewing, including habituation to shown foods.Keywords: fMRI, functional connectivity, task-based, beta series correlation
Procedia PDF Downloads 2731016 Intrusion Detection in Computer Networks Using a Hybrid Model of Firefly and Differential Evolution Algorithms
Authors: Mohammad Besharatloo
Abstract:
Intrusion detection is an important research topic in network security because of increasing growth in the use of computer network services. Intrusion detection is done with the aim of detecting the unauthorized use or abuse in the networks and systems by the intruders. Therefore, the intrusion detection system is an efficient tool to control the user's access through some predefined regulations. Since, the data used in intrusion detection system has high dimension, a proper representation is required to show the basis structure of this data. Therefore, it is necessary to eliminate the redundant features to create the best representation subset. In the proposed method, a hybrid model of differential evolution and firefly algorithms was employed to choose the best subset of properties. In addition, decision tree and support vector machine (SVM) are adopted to determine the quality of the selected properties. In the first, the sorted population is divided into two sub-populations. These optimization algorithms were implemented on these sub-populations, respectively. Then, these sub-populations are merged to create next repetition population. The performance evaluation of the proposed method is done based on KDD Cup99. The simulation results show that the proposed method has better performance than the other methods in this context.Keywords: intrusion detection system, differential evolution, firefly algorithm, support vector machine, decision tree
Procedia PDF Downloads 941015 Implementation of Inference Fuzzy System as a Valuation Subsidiary is Based Particle Swarm Optimization for Solves the Issue of Decision Making in Middle Size Soccer Robot League
Authors: Zahra Abdolkarimi, Naser Zouri
Abstract:
Nowadays, there is unbelievable growing of Robots created a collection of complex and motivate subject in robotic and intellectual ornate, also it made a mechatronics style base of theoretical and technical way in Robocop. Additionally, robotics system recommended RoboCup factor as a provider of some standardization and testing method in case of computer discussion widely. The actual purpose of RoboCup is creating independent team of robots in 2050 based of FiFa roles to bring the victory in compare of world star team. In addition, decision making of robots depends to environment reaction, self-player and rival player with using inductive Fuzzy system valuation subsidiary to solve issue of robots in land game. The measure of selection in compare with other methods depends to amount of victories percentage in the same team that plays accidently. Consequences, shows method of our discussion is the best way for Particle Swarm Optimization and Fuzzy system compare to other decision of robotics algorithmic.Keywords: PSO algorithm, inference fuzzy system, chaos theory, soccer robot league
Procedia PDF Downloads 4051014 The Impact of the Application of Blockchain Technology in Accounting and Auditing
Authors: Yusuf Adebayo Oduwole
Abstract:
The evaluation of blockchain technology's potential effects on the accounting and auditing fields is the main objective of this essay. It also adds to the existing body of work by examining how these practices alter technological concerns, including cryptocurrency accounting, regulation, governance, accounting practices, and technical challenges. Examples of this advancement include the growth of the concept of blockchain and its application in accounting. This technology is being considered one of the digital revolutions that could disrupt the world and civilization as it can transfer large volumes of virtual currencies like cryptocurrencies with the help of a third party. The basis for this research is a systematic review of the articles using Vosviewer to display and reflect on the bibliometric information of the articles accessible on the Scopus database. Also, as the practice of using blockchain technology in the field of accounting and auditing is still in its infancy, it may be useful to carry out a more thorough analysis of any implications for accounting and auditing regarding aspects of governance, regulation, and cryptocurrency that have not yet been discussed or addressed to any significant extent. The main findings on the relationship between blockchain and accounting show that the application of smart contracts, such as triple-entry accounting, has increased the quality of accounting records as well as reliance on the information available. This results in fewer cyclical assignments, no need for resolution, and real-time accounting, among others. Thereby, to integrate blockchain through a computer system, one must continuously learn and remain naive when using blockchain-integrated accounting software. This includes learning about how cryptocurrencies are accounted for and regulated. In this study, three original and contributed efforts are presented. To offer a transparent view of the state of previous relevant studies and research works in accounting and auditing that focus on blockchain, it begins by using bibliographic visibility analysis and a Scopus narrative analysis. Second, it highlights legislative, governance, and ethical concerns, such as education, where it tackles the use of blockchain in accounting and auditing. Lastly, it examines the impact of blockchain technologies on the accounting recognition of cryptocurrencies. Users of the technology should, therefore, take their time and learn how it works, as well as keep abreast of the different developments. In addition, the accounting industry must integrate blockchain certification and practice, most likely offline or as part of university education for those intending to become auditors or accountants.Keywords: blockchain, crypto assets, governance, regulation & smart contracts
Procedia PDF Downloads 301013 Detecting Venomous Files in IDS Using an Approach Based on Data Mining Algorithm
Authors: Sukhleen Kaur
Abstract:
In security groundwork, Intrusion Detection System (IDS) has become an important component. The IDS has received increasing attention in recent years. IDS is one of the effective way to detect different kinds of attacks and malicious codes in a network and help us to secure the network. Data mining techniques can be implemented to IDS, which analyses the large amount of data and gives better results. Data mining can contribute to improving intrusion detection by adding a level of focus to anomaly detection. So far the study has been carried out on finding the attacks but this paper detects the malicious files. Some intruders do not attack directly, but they hide some harmful code inside the files or may corrupt those file and attack the system. These files are detected according to some defined parameters which will form two lists of files as normal files and harmful files. After that data mining will be performed. In this paper a hybrid classifier has been used via Naive Bayes and Ripper classification methods. The results show how the uploaded file in the database will be tested against the parameters and then it is characterised as either normal or harmful file and after that the mining is performed. Moreover, when a user tries to mine on harmful file it will generate an exception that mining cannot be made on corrupted or harmful files.Keywords: data mining, association, classification, clustering, decision tree, intrusion detection system, misuse detection, anomaly detection, naive Bayes, ripper
Procedia PDF Downloads 4141012 Landsat 8-TIRS NEΔT at Kīlauea Volcano and the Active East Rift Zone, Hawaii
Authors: Flora Paganelli
Abstract:
The radiometric performance of remotely sensed images is important for volcanic monitoring. The Thermal Infrared Sensor (TIRS) on-board Landsat 8 was designed with specific requirements in regard to the noise-equivalent change in temperature (NEΔT) at ≤ 0.4 K at 300 K for the two thermal infrared bands B10 and B11. This study investigated the on-orbit NEΔT of the TIRS two bands from a scene-based method using clear-sky images over the volcanic activity of Kīlauea Volcano and the active East Rift Zone (Hawaii), in order to optimize the use of TIRS data. Results showed that the NEΔTs of the two bands exceeded the design specification by an order of magnitude at 300 K. Both separate bands and split window algorithm were examined to estimate the effect of NEΔT on the land surface temperature (LST) retrieval, and NEΔT contribution to the final LST error. These results were also useful in the current efforts to assess the requirements for volcanology research campaign using the Hyperspectral Infrared Imager (HyspIRI) whose airborne prototype MODIS/ASTER instruments is plan to be flown by NASA as a single campaign to the Hawaiian Islands in support of volcanology and coastal area monitoring in 2016.Keywords: landsat 8, radiometric performance, thermal infrared sensor (TIRS), volcanology
Procedia PDF Downloads 2421011 Multi Object Tracking for Predictive Collision Avoidance
Authors: Bruk Gebregziabher
Abstract:
The safe and efficient operation of Autonomous Mobile Robots (AMRs) in complex environments, such as manufacturing, logistics, and agriculture, necessitates accurate multiobject tracking and predictive collision avoidance. This paper presents algorithms and techniques for addressing these challenges using Lidar sensor data, emphasizing ensemble Kalman filter. The developed predictive collision avoidance algorithm employs the data provided by lidar sensors to track multiple objects and predict their velocities and future positions, enabling the AMR to navigate safely and effectively. A modification to the dynamic windowing approach is introduced to enhance the performance of the collision avoidance system. The overall system architecture encompasses object detection, multi-object tracking, and predictive collision avoidance control. The experimental results, obtained from both simulation and real-world data, demonstrate the effectiveness of the proposed methods in various scenarios, which lays the foundation for future research on global planners, other controllers, and the integration of additional sensors. This thesis contributes to the ongoing development of safe and efficient autonomous systems in complex and dynamic environments.Keywords: autonomous mobile robots, multi-object tracking, predictive collision avoidance, ensemble Kalman filter, lidar sensors
Procedia PDF Downloads 84