Search results for: equation error
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3780

Search results for: equation error

2010 Graphene Supported Nano Cerium Oxides Hybrid as an Electrocatalyst for Oxygen Reduction Reactions

Authors: Siba Soren, Purnendu Parhi

Abstract:

Today, the world is facing a severe challenge due to depletion of traditional fossil fuels. Scientists across the globe are working for a solution that involves a dramatic shift to practical and environmentally sustainable energy sources. High-capacity energy systems, such as metal-air batteries, fuel cells, are highly desirable to meet the urgent requirement of sustainable energies. Among the fuel cells, Direct methanol fuel cells (DMFCs) are recognized as an ideal power source for mobile applications and have received considerable attention in recent past. In this advanced electrochemical energy conversion technologies, Oxygen Reduction Reaction (ORR) is of utmost importance. However, the poor kinetics of cathodic ORR in DMFCs significantly hampers their possibilities of commercialization. The oxygen is reduced in alkaline medium either through a 4-electron (equation i) or a 2-electron (equation ii) reduction pathway at the cathode ((i) O₂ + 2H₂O + 4e⁻ → 4OH⁻, (ii) O₂ + H₂O + 2e⁻ → OH⁻ + HO₂⁻ ). Due to sluggish ORR kinetics the ability to control the reduction of molecular oxygen electrocatalytically is still limited. The electrocatalytic ORR starts with adsorption of O₂ on the electrode surface followed by O–O bond activation/cleavage and oxide removal. The reaction further involves transfer of 4 electrons and 4 protons. The sluggish kinetics of ORR, on the one hand, demands high loading of precious metal-containing catalysts (e.g., Pt), which unfavorably increases the cost of these electrochemical energy conversion devices. Therefore, synthesis of active electrocatalyst with an increase in ORR performance is need of the hour. In the recent literature, there are many reports on transition metal oxide (TMO) based ORR catalysts for their high activity TMOs are also having drawbacks like low electrical conductivity, which seriously affects the electron transfer process during ORR. It was found that 2D graphene layer is having high electrical conductivity, large surface area, and excellent chemical stability, appeared to be an ultimate choice as support material to enhance the catalytic performance of bare metal oxide. g-C₃N₄ is also another candidate that has been used by the researcher for improving the ORR performance of metal oxides. This material provides more active reaction sites than other N containing carbon materials. Rare earth oxide like CeO₂ is also a good candidate for studying the ORR activity as the metal oxide not only possess unique electronic properties but also possess catalytically active sites. Here we will discuss the ORR performance (in alkaline medium) of N-rGO/C₃N₄ supported nano Cerium Oxides hybrid synthesized by microwave assisted Solvothermal method. These materials exhibit superior electrochemical stability and methanol tolerance capability to that of commercial Pt/C.

Keywords: oxygen reduction reaction, electrocatalyst, cerium oxide, graphene

Procedia PDF Downloads 194
2009 When does technology alignment influence supply chain performance

Authors: Joseph Akyeh, Abdul Samed Muntaka, Emmanuel Anin, Dorcas Nuertey

Abstract:

Purpose: This study develops and tests arguments that the relationship between technology alignment and supply chain performance is conditional upon levels of technology championing. Methodology: The proposed relationships are tested on a sample of 217 hospitals in a major sub-Saharan African economy. Findings: Findings from the study indicate that technology alignment has a positive and significant effect on supply chain performance. The study further finds that while technology championing strengthens the direct effects of technology alignment on supply chain performance. Theoretical Contributions: A theoretical contribution from this study is the finding that when technology alignment drives supply chain performance is more complex than previously thought it depends on whether or not technology alignment is first championed by top management. Originality: Though some studies have been conducted on technology alignment and health supply chain performance, to the best of the researcher’s knowledge, no previous study has examined the moderating role of technology championing the link between technology alignment and supply chain performance.

Keywords: technology alignment, supply chain performance, technology championing, structural equation modelling

Procedia PDF Downloads 50
2008 Relationship between Gully Development and Characteristics of Drainage Area in Semi-Arid Region, NW Iran

Authors: Ali Reza Vaezi, Ouldouz Bakhshi Rad

Abstract:

Gully erosion is a widespread and often dramatic form of soil erosion caused by water during and immediately after heavy rainfall. It occurs when flowing surface water is channelled across unprotected land and washes away the soil along the drainage lines. The formation of gully is influenced by various factors, including climate, drainage surface area, slope gradient, vegetation cover, land use, and soil properties. It is a very important problem in semi-arid regions, where soils have lower organic matter and are weakly aggregated. Intensive agriculture and tillage along the slope can accelerate soil erosion by water in the region. There is little information on the development of gully erosion in agricultural rainfed areas. Therefore, this study was carried out to investigate the relationship between gully erosion and morphometric characteristics of the drainage area and the effects of soil properties and soil management factors (land use and tillage method) on gully development. A field study was done in a 900 km2 agricultural area in Hshtroud township located in the south of East Azarbijan province, NW Iran. Toward this, two hundred twenty-two gullies created in rainfed lands were found in the area. Some properties of gullies, consisting of length, width, depth, height difference, cross section area, and volume, were determined. Drainage areas for each or some gullies were determined, and their boundaries were drawn. Additionally, the surface area of each drainage, land use, tillage direction, and soil properties that may affect gully formation were determined. The soil erodibility factor (K) defined in the Universal Soil Loss Equation (USLE) was estimated based on five soil properties (silt and very fine sand, coarse sand, organic matter, soil structure code, and soil permeability). Gully development in each drainage area was quantified using its volume and soil loss. The dependency of gully development on drainage area characteristics (surface area, land use, tillage direction, and soil properties) was determined using correlation matrix analysis. Based on the results, gully length was the most important morphometric characteristic indicating the development of gully erosion in the lands. Gully development in the area was related to slope gradient (r= -0.26), surface area (r= 0.71), the area of rainfed lands (r= 0.23), and the area of rainfed tilled along the slope (r= 0.24). Nevertheless, its correlation with the area of pasture and soil erodibility factor (K) was not significant. Among the characteristics of drainage area, surface area is the major factor controlling gully volume in the agricultural land. No significant correlation was found between gully erosion and soil erodibility factor (K) estimated by the Universal Soil Loss Equation (USLE). It seems the estimated soil erodibility can’t describe the susceptibility of the study soils to the gully erosion process. In these soils, aggregate stability and soil permeability are the two soil physical properties that affect the actual soil erodibility and in consequence, these soil properties can control gully erosion in the rainfed lands.

Keywords: agricultural area, gully properties, soil structure, USLE

Procedia PDF Downloads 77
2007 Information Technology and Business Alignments among Different Divisions: A Comparative Analysis of Japan and South Korea

Authors: Michiko Miyamoto

Abstract:

This paper empirically investigates whether information technology (IT) strategies, business strategies, and divisions are aligned to meet overall business goals for Korean Small and medium-sized enterprises (SMEs), based on structure based Strategic Alignment Model, and make comparison with those of Japanese SMEs. Using 2,869 valid responses of Korean Human Capital Corporate Panel survey, a result of this study suggests that Korean human resources (HR) departments have a major influence over IT strategy, which is the same as Japanese SMEs, even though their management styles are quite different. As for IT strategy, it is not related to other departments at all for Korean SMEs. The Korean management seems to possess a great power over each division, such as Sales/Service, Research and Development/Technical Experts, HR, and Production.

Keywords: IT-business alignment, structured based strategic alignment model, structural equation model, human resources department

Procedia PDF Downloads 271
2006 Media Richness Perspective on Web 2.0 Usage for Knowledge Creation: The Case of the Cocoa Industry in Ghana

Authors: Albert Gyamfi

Abstract:

Cocoa plays critical role in the socio-economic development of Ghana. Meanwhile, smallholder farmers most of whom are illiterate dominate the industry. According to the cocoa-based agricultural knowledge and information system (AKIS) model knowledge is created and transferred to the industry between three key actors: cocoa researchers, extension experts, and cocoa farmers. Dwelling on the SECI model, the media richness theory (MRT), and the AKIS model, a conceptual model of web 2.0-based AKIS model (AKIS 2.0) is developed and used to assess the possible effects of social media usage for knowledge creation in the Ghanaian cocoa industry. A mixed method approach with a survey questionnaire was employed, and a second-order multi-group structural equation model (SEM) was used to analyze the data. The study concludes that the use of web 2.0 applications for knowledge creation would lead to sustainable interactions among the key knowledge actors for effective knowledge creation in the cocoa industry in Ghana.

Keywords: agriculture, cocoa, knowledge, media, web 2.0

Procedia PDF Downloads 333
2005 The Differential Role of Written Corrective Feedback in L2 Students’ Noticing and Its Impact on Writing Scores

Authors: Khaled ElEbyary, Ramy Shabara

Abstract:

L2 research has generally acknowledged the role of noticing in language learning. The role of teacher feedback is to trigger learners’ noticing of errors and direct the writing process. Recently L2 learners are seemingly using computerized applications which provide corrective feedback (CF) at different stages of writing (i.e., during and after writing). This study aimed principally to answer the question, “Is noticing likely to be maximized when feedback on erroneous output is electronically provided either during or after the composing stage, or does teacher annotated feedback have a stronger effect?”. Seventy-five participants were randomly distributed into four groups representing four conditions. These include receiving automated feedback at the composing stage, automated feedback after writing, teacher feedback, and no feedback. Findings demonstrate the impact of CF on writing and the intensity of noticing certain language areas at different writing stages and from different feedback sources.

Keywords: written corrective feedback, error correction, noticing, automated written corrective feedback, L2 acquisition

Procedia PDF Downloads 96
2004 A Proposal for a Combustion Model Considering the Lewis Number and Its Evaluation

Authors: Fujio Akagi, Hiroaki Ito, Shin-Ichi Inage

Abstract:

The aim of this study is to develop a combustion model that can be applied uniformly to laminar and turbulent premixed flames while considering the effect of the Lewis number (Le). The model considers the effect of Le on the transport equations of the reaction progress, which varies with the chemical species and temperature. The distribution of the reaction progress variable is approximated by a hyperbolic tangent function, while the other distribution of the reaction progress variable is estimated using the approximated distribution and transport equation of the reaction progress variable considering the Le. The validity of the model was evaluated under the conditions of propane with Le > 1 and methane with Le = 1 (equivalence ratios of 0.5 and 1). The estimated results were found to be in good agreement with those of previous studies under all conditions. A method of introducing a turbulence model into this model is also described. It was confirmed that conventional turbulence models can be expressed as an approximate theory of this model in a unified manner.

Keywords: combustion model, laminar flame, Lewis number, turbulent flame

Procedia PDF Downloads 123
2003 Frequency Recognition Models for Steady State Visual Evoked Potential Based Brain Computer Interfaces (BCIs)

Authors: Zeki Oralhan, Mahmut Tokmakçı

Abstract:

SSVEP based brain computer interface (BCI) systems have been preferred, because of high information transfer rate (ITR) and practical use. ITR is the parameter of BCI overall performance. For high ITR value, one of specification BCI system is that has high accuracy. In this study, we investigated to recognize SSVEP with shorter time and lower error rate. In the experiment, there were 8 flickers on light crystal display (LCD). Participants gazed to flicker which had 12 Hz frequency and 50% duty cycle ratio on the LCD during 10 seconds. During the experiment, EEG signals were acquired via EEG device. The EEG data was filtered in preprocessing session. After that Canonical Correlation Analysis (CCA), Multiset CCA (MsetCCA), phase constrained CCA (PCCA), and Multiway CCA (MwayCCA) methods were applied on data. The highest average accuracy value was reached when MsetCCA was applied.

Keywords: brain computer interface, canonical correlation analysis, human computer interaction, SSVEP

Procedia PDF Downloads 266
2002 Encephalon-An Implementation of a Handwritten Mathematical Expression Solver

Authors: Shreeyam, Ranjan Kumar Sah, Shivangi

Abstract:

Recognizing and solving handwritten mathematical expressions can be a challenging task, particularly when certain characters are segmented and classified. This project proposes a solution that uses Convolutional Neural Network (CNN) and image processing techniques to accurately solve various types of equations, including arithmetic, quadratic, and trigonometric equations, as well as logical operations like logical AND, OR, NOT, NAND, XOR, and NOR. The proposed solution also provides a graphical solution, allowing users to visualize equations and their solutions. In addition to equation solving, the platform, called CNNCalc, offers a comprehensive learning experience for students. It provides educational content, a quiz platform, and a coding platform for practicing programming skills in different languages like C, Python, and Java. This all-in-one solution makes the learning process engaging and enjoyable for students. The proposed methodology includes horizontal compact projection analysis and survey for segmentation and binarization, as well as connected component analysis and integrated connected component analysis for character classification. The compact projection algorithm compresses the horizontal projections to remove noise and obtain a clearer image, contributing to the accuracy of character segmentation. Experimental results demonstrate the effectiveness of the proposed solution in solving a wide range of mathematical equations. CNNCalc provides a powerful and user-friendly platform for solving equations, learning, and practicing programming skills. With its comprehensive features and accurate results, CNNCalc is poised to revolutionize the way students learn and solve mathematical equations. The platform utilizes a custom-designed Convolutional Neural Network (CNN) with image processing techniques to accurately recognize and classify symbols within handwritten equations. The compact projection algorithm effectively removes noise from horizontal projections, leading to clearer images and improved character segmentation. Experimental results demonstrate the accuracy and effectiveness of the proposed solution in solving a wide range of equations, including arithmetic, quadratic, trigonometric, and logical operations. CNNCalc features a user-friendly interface with a graphical representation of equations being solved, making it an interactive and engaging learning experience for users. The platform also includes tutorials, testing capabilities, and programming features in languages such as C, Python, and Java. Users can track their progress and work towards improving their skills. CNNCalc is poised to revolutionize the way students learn and solve mathematical equations with its comprehensive features and accurate results.

Keywords: AL, ML, hand written equation solver, maths, computer, CNNCalc, convolutional neural networks

Procedia PDF Downloads 122
2001 Modeling and Simulation of Flow Shop Scheduling Problem through Petri Net Tools

Authors: Joselito Medina Marin, Norberto Hernández Romero, Juan Carlos Seck Tuoh Mora, Erick S. Martinez Gomez

Abstract:

The Flow Shop Scheduling Problem (FSSP) is a typical problem that is faced by production planning managers in Flexible Manufacturing Systems (FMS). This problem consists in finding the optimal scheduling to carry out a set of jobs, which are processed in a set of machines or shared resources. Moreover, all the jobs are processed in the same machine sequence. As in all the scheduling problems, the makespan can be obtained by drawing the Gantt chart according to the operations order, among other alternatives. On this way, an FMS presenting the FSSP can be modeled by Petri nets (PNs), which are a powerful tool that has been used to model and analyze discrete event systems. Then, the makespan can be obtained by simulating the PN through the token game animation and incidence matrix. In this work, we present an adaptive PN to obtain the makespan of FSSP by applying PN analytical tools.

Keywords: flow-shop scheduling problem, makespan, Petri nets, state equation

Procedia PDF Downloads 298
2000 The Study of Information Uses Behaviour of Tourists in Songkhla Province, Thailand

Authors: Patraporn Kaewkhanitarak, Suchada Srichuar, Narawat Kanjanapan

Abstract:

This research is the survey research. The purpose of this research is to study information uses behavior and problem of tourists in Songkhla Province. The tool used in this study include structure questioner standardize in 5 levels rating scale. The 400 participants selected by convenience sampling (allowable error 5%) by Taro Yamane method. The collecting data period is 6 months from January-June 2014. The result of this study found that the type of information that the tourists often use to plan their trip is internet (x̅ = 3.81) and the most popular text is restaurant (x̅ = 3.77). The tourists found that booking or buying service from internet provided more affordable price and they could select appropriate plan by themselves. The most convenience source of information that the tourists often use is internet and website (x̅ = 3.69). Nevertheless, they explained that most of tourist information source in Songkhla province are lack and insufficient of tourist organization that provide information and service related to tourism.

Keywords: information, behavior, tourists, Thailand

Procedia PDF Downloads 253
1999 Soil Moisture Regulation in Irrigated Agriculture

Authors: I. Kruashvili, I. Inashvili, K. Bziava, M. Lomishvili

Abstract:

Seepage capillary anomalies in the active layer of soil, related to the soil water movement, often cause variation of soil hydrophysical properties and become one of the main objectives of the hydroecology. It is necessary to mention that all existing equations for computing the seepage flow particularly from soil channels, through dams, bulkheads, and foundations of hydraulic engineering structures are preferable based on the linear seepage law. Regarding the existing beliefs, anomalous seepage is based on postulates according to which the fluid in free volume is characterized by resistance against shear deformation and is presented in the form of initial gradient. According to the above-mentioned information, we have determined: Equation to calculate seepage coefficient when the velocity of transition flow is equal to seepage flow velocity; by means of power function, equations for the calculation of average and maximum velocities of seepage flow have been derived; taking into consideration the fluid continuity condition, average velocity for calculation of average velocity in capillary tube has been received.

Keywords: seepage, soil, velocity, water

Procedia PDF Downloads 462
1998 Magnetohydrodynamic Couette Flow of Fractional Burger’s Fluid in an Annulus

Authors: Sani Isa, Ali Musa

Abstract:

Burgers’ fluid with a fractional derivatives model in an annulus was analyzed. Combining appropriately the basic equations, with the fractionalized fractional Burger’s fluid model allow us to determine the velocity field, temperature and shear stress. The governing partial differential equation was solved using the combine Laplace transformation method and Riemann sum approximation to give velocity field, temperature and shear stress on the fluid flow. The influence of various parameters like fractional parameters, relaxation time and retardation time, are drawn. The results obtained are simulated using Mathcad software and presented graphically. From the graphical results, we observed that the relaxation time and time helps the flow pattern, on the other hand, other material constants resist the fluid flow while fractional parameters effect on fluid flow is opposite to each other.

Keywords: sani isa, Ali musaburger’s fluid, Laplace transform, fractional derivatives, annulus

Procedia PDF Downloads 24
1997 A Numerical Solution Based on Operational Matrix of Differentiation of Shifted Second Kind Chebyshev Wavelets for a Stefan Problem

Authors: Rajeev, N. K. Raigar

Abstract:

In this study, one dimensional phase change problem (a Stefan problem) is considered and a numerical solution of this problem is discussed. First, we use similarity transformation to convert the governing equations into ordinary differential equations with its boundary conditions. The solutions of ordinary differential equation with the associated boundary conditions and interface condition (Stefan condition) are obtained by using a numerical approach based on operational matrix of differentiation of shifted second kind Chebyshev wavelets. The obtained results are compared with existing exact solution which is sufficiently accurate.

Keywords: operational matrix of differentiation, similarity transformation, shifted second kind chebyshev wavelets, stefan problem

Procedia PDF Downloads 403
1996 Deep Neural Network Approach for Navigation of Autonomous Vehicles

Authors: Mayank Raj, V. G. Narendra

Abstract:

Ever since the DARPA challenge on autonomous vehicles in 2005, there has been a lot of buzz about ‘Autonomous Vehicles’ amongst the major tech giants such as Google, Uber, and Tesla. Numerous approaches have been adopted to solve this problem, which can have a long-lasting impact on mankind. In this paper, we have used Deep Learning techniques and TensorFlow framework with the goal of building a neural network model to predict (speed, acceleration, steering angle, and brake) features needed for navigation of autonomous vehicles. The Deep Neural Network has been trained on images and sensor data obtained from the comma.ai dataset. A heatmap was used to check for correlation among the features, and finally, four important features were selected. This was a multivariate regression problem. The final model had five convolutional layers, followed by five dense layers. Finally, the calculated values were tested against the labeled data, where the mean squared error was used as a performance metric.

Keywords: autonomous vehicles, deep learning, computer vision, artificial intelligence

Procedia PDF Downloads 158
1995 Dissolved Gas Analysis Based Regression Rules from Trained ANN for Transformer Fault Diagnosis

Authors: Deepika Bhalla, Raj Kumar Bansal, Hari Om Gupta

Abstract:

Dissolved Gas Analysis (DGA) has been widely used for fault diagnosis in a transformer. Artificial neural networks (ANN) have high accuracy but are regarded as black boxes that are difficult to interpret. For many problems it is desired to extract knowledge from trained neural networks (NN) so that the user can gain a better understanding of the solution arrived by the NN. This paper applies a pedagogical approach for rule extraction from function approximating neural networks (REFANN) with application to incipient fault diagnosis using the concentrations of the dissolved gases within the transformer oil, as the input to the NN. The input space is split into subregions and for each subregion there is a linear equation that is used to predict the type of fault developing within a transformer. The experiments on real data indicate that the approach used can extract simple and useful rules and give fault predictions that match the actual fault and are at times also better than those predicted by the IEC method.

Keywords: artificial neural networks, dissolved gas analysis, rules extraction, transformer

Procedia PDF Downloads 536
1994 Effects of Financial Development on Economic Growth in South Asia

Authors: Anupam Das

Abstract:

Although financial liberalization has been one of the most important policy prescriptions of international organizations like the World Bank and the IMF, the effect of financial liberalization on economic growth in developing countries is far from unanimous. Since the '80s, South Asian countries made a significant development in liberalization the financial sector. However, due to unavailability of a sufficient number of time series observations, the relationship between economic growth and financial development has not been investigated adequately. We aim to fill this gap by examining time series data of five developing countries from the South Asian region: Bangladesh, India, Pakistan, Sri Lanka, and Nepal. Applying the cointegration tests and Granger causality within the vector error correction model (VECM), we do not find unanimous evidence of financial development on positive economic growth. These results are helpful for developing countries which have been trying to liberalize the financial sector in recent decades.

Keywords: economic growth, financial development, Granger causality, South Asia

Procedia PDF Downloads 371
1993 Language Switching Errors of Bilinguals: Role of Top down and Bottom up Process

Authors: Numra Qayyum, Samina Sarwat, Noor ul Ain

Abstract:

Bilingual speakers generally can speak both languages with the same competency without mixing them intentionally and making mistakes, but sometimes errors occur in language selection. This quantitative study particularly deals with the language errors made by Urdu-English bilinguals. In this research, researchers have given special attention to the part played by bottom-up priming and top-down cognitive control in these errors. Unstable Urdu-English bilingual participants termed pictures and were prompted to shift from one language to another under the pressure of time. Different situations were given to manipulate the participants. The long and short runs trials of the same language were also given before switching to another language. The study is concluded with the findings that bilinguals made more errors when switching to the first language from their second language, and these errors are large in number, especially when a speaker is switching from L2 (second language) to L1 (first language) after a long run. When the switching is reversed, i.e., from L2 to LI, it had no effect at all. These results gave the clear responsibility of all these errors to top-down cognitive control.

Keywords: bottom up priming, language error, language switching, top down cognitive control

Procedia PDF Downloads 137
1992 Improving Load Frequency Control of Multi-Area Power System by Considering Uncertainty by Using Optimized Type 2 Fuzzy Pid Controller with the Harmony Search Algorithm

Authors: Mehrdad Mahmudizad, Roya Ahmadi Ahangar

Abstract:

This paper presents the method of designing the type 2 fuzzy PID controllers in order to solve the problem of Load Frequency Control (LFC). The Harmony Search (HS) algorithm is used to regulate the measurement factors and the effect of uncertainty of membership functions of Interval Type 2 Fuzzy Proportional Integral Differential (IT2FPID) controllers in order to reduce the frequency deviation resulted from the load oscillations. The simulation results implicitly show that the performance of the proposed IT2FPID LFC in terms of error, settling time and resistance against different load oscillations is more appropriate and preferred than PID and Type 1 Fuzzy Proportional Integral Differential (T1FPID) controllers.

Keywords: load frequency control, fuzzy-pid controller, type 2 fuzzy system, harmony search algorithm

Procedia PDF Downloads 278
1991 Numerical Computation of Sturm-Liouville Problem with Robin Boundary Condition

Authors: Theddeus T. Akano, Omotayo A. Fakinlede

Abstract:

The modelling of physical phenomena, such as the earth’s free oscillations, the vibration of strings, the interaction of atomic particles, or the steady state flow in a bar give rise to Sturm-Liouville (SL) eigenvalue problems. The boundary applications of some systems like the convection-diffusion equation, electromagnetic and heat transfer problems requires the combination of Dirichlet and Neumann boundary conditions. Hence, the incorporation of Robin boundary condition in the analyses of Sturm-Liouville problem. This paper deals with the computation of the eigenvalues and eigenfunction of generalized Sturm-Liouville problems with Robin boundary condition using the finite element method. Numerical solutions of classical Sturm–Liouville problems are presented. The results show an agreement with the exact solution. High results precision is achieved with higher number of elements.

Keywords: Sturm-Liouville problem, Robin boundary condition, finite element method, eigenvalue problems

Procedia PDF Downloads 362
1990 Electromagnetic Modeling of a MESFET Transistor Using the Moments Method Combined with Generalised Equivalent Circuit Method

Authors: Takoua Soltani, Imen Soltani, Taoufik Aguili

Abstract:

The communications' and radar systems' demands give rise to new developments in the domain of active integrated antennas (AIA) and arrays. The main advantages of AIA arrays are the simplicity of fabrication, low cost of manufacturing, and the combination between free space power and the scanner without a phase shifter. The integrated active antenna modeling is the coupling between the electromagnetic model and the transport model that will be affected in the high frequencies. Global modeling of active circuits is important for simulating EM coupling, interaction between active devices and the EM waves, and the effects of EM radiation on active and passive components. The current review focuses on the modeling of the active element which is a MESFET transistor immersed in a rectangular waveguide. The proposed EM analysis is based on the Method of Moments combined with the Generalised Equivalent Circuit method (MOM-GEC). The Method of Moments which is the most common and powerful software as numerical techniques have been used in resolving the electromagnetic problems. In the class of numerical techniques, MOM is the dominant technique in solving of Maxwell and Transport’s integral equations for an active integrated antenna. In this situation, the equivalent circuit is introduced to the development of an integral method formulation based on the transposition of field problems in a Generalised equivalent circuit that is simpler to treat. The method of Generalised Equivalent Circuit (MGEC) was suggested in order to represent integral equations circuits that describe the unknown electromagnetic boundary conditions. The equivalent circuit presents a true electric image of the studied structures for describing the discontinuity and its environment. The aim of our developed method is to investigate the antenna parameters such as the input impedance and the current density distribution and the electric field distribution. In this work, we propose a global EM modeling of the MESFET AsGa transistor using an integral method. We will begin by describing the modeling structure that allows defining an equivalent EM scheme translating the electromagnetic equations considered. Secondly, the projection of these equations on common-type test functions leads to a linear matrix equation where the unknown variable represents the amplitudes of the current density. Solving this equation resulted in providing the input impedance, the distribution of the current density and the electric field distribution. From electromagnetic calculations, we were able to present the convergence of input impedance for different test function number as a function of the guide mode numbers. This paper presents a pilot study to find the answer to map out the variation of the existing current evaluated by the MOM-GEC. The essential improvement of our method is reducing computing time and memory requirements in order to provide a sufficient global model of the MESFET transistor.

Keywords: active integrated antenna, current density, input impedance, MESFET transistor, MOM-GEC method

Procedia PDF Downloads 198
1989 Predicting Shot Making in Basketball Learnt Fromadversarial Multiagent Trajectories

Authors: Mark Harmon, Abdolghani Ebrahimi, Patrick Lucey, Diego Klabjan

Abstract:

In this paper, we predict the likelihood of a player making a shot in basketball from multiagent trajectories. Previous approaches to similar problems center on hand-crafting features to capture domain-specific knowledge. Although intuitive, recent work in deep learning has shown, this approach is prone to missing important predictive features. To circumvent this issue, we present a convolutional neural network (CNN) approach where we initially represent the multiagent behavior as an image. To encode the adversarial nature of basketball, we use a multichannel image which we then feed into a CNN. Additionally, to capture the temporal aspect of the trajectories, we use “fading.” We find that this approach is superior to a traditional FFN model. By using gradient ascent, we were able to discover what the CNN filters look for during training. Last, we find that a combined FFN+CNN is the best performing network with an error rate of 39%.

Keywords: basketball, computer vision, image processing, convolutional neural network

Procedia PDF Downloads 153
1988 Attention-Based Spatio-Temporal Approach for Fire and Smoke Detection

Authors: Alireza Mirrashid, Mohammad Khoshbin, Ali Atghaei, Hassan Shahbazi

Abstract:

In various industries, smoke and fire are two of the most important threats in the workplace. One of the common methods for detecting smoke and fire is the use of infrared thermal and smoke sensors, which cannot be used in outdoor applications. Therefore, the use of vision-based methods seems necessary. The problem of smoke and fire detection is spatiotemporal and requires spatiotemporal solutions. This paper presents a method that uses spatial features along with temporal-based features to detect smoke and fire in the scene. It consists of three main parts; the task of each part is to reduce the error of the previous part so that the final model has a robust performance. This method also uses transformer modules to increase the accuracy of the model. The results of our model show the proper performance of the proposed approach in solving the problem of smoke and fire detection and can be used to increase workplace safety.

Keywords: attention, fire detection, smoke detection, spatio-temporal

Procedia PDF Downloads 203
1987 Real-Time Image Encryption Using a 3D Discrete Dual Chaotic Cipher

Authors: M. F. Haroun, T. A. Gulliver

Abstract:

In this paper, an encryption algorithm is proposed for real-time image encryption. The scheme employs a dual chaotic generator based on a three dimensional (3D) discrete Lorenz attractor. Encryption is achieved using non-autonomous modulation where the data is injected into the dynamics of the master chaotic generator. The second generator is used to permute the dynamics of the master generator using the same approach. Since the data stream can be regarded as a random source, the resulting permutations of the generator dynamics greatly increase the security of the transmitted signal. In addition, a technique is proposed to mitigate the error propagation due to the finite precision arithmetic of digital hardware. In particular, truncation and rounding errors are eliminated by employing an integer representation of the data which can easily be implemented. The simple hardware architecture of the algorithm makes it suitable for secure real-time applications.

Keywords: chaotic systems, image encryption, non-autonomous modulation, FPGA

Procedia PDF Downloads 506
1986 Reconstruction of Signal in Plastic Scintillator of PET Using Tikhonov Regularization

Authors: L. Raczynski, P. Moskal, P. Kowalski, W. Wislicki, T. Bednarski, P. Bialas, E. Czerwinski, A. Gajos, L. Kaplon, A. Kochanowski, G. Korcyl, J. Kowal, T. Kozik, W. Krzemien, E. Kubicz, Sz. Niedzwiecki, M. Palka, Z. Rudy, O. Rundel, P. Salabura, N.G. Sharma, M. Silarski, A. Slomski, J. Smyrski, A. Strzelecki, A. Wieczorek, M. Zielinski, N. Zon

Abstract:

The J-PET scanner, which allows for single bed imaging of the whole human body, is currently under development at the Jagiellonian University. The J-PET detector improves the TOF resolution due to the use of fast plastic scintillators. Since registration of the waveform of signals with duration times of few nanoseconds is not feasible, a novel front-end electronics allowing for sampling in a voltage domain at four thresholds was developed. To take fully advantage of these fast signals a novel scheme of recovery of the waveform of the signal, based on ideas from the Tikhonov regularization (TR) and Compressive Sensing methods, is presented. The prior distribution of sparse representation is evaluated based on the linear transformation of the training set of waveform of the signals by using the Principal Component Analysis (PCA) decomposition. Beside the advantage of including the additional information from training signals, a further benefit of the TR approach is that the problem of signal recovery has an optimal solution which can be determined explicitly. Moreover, from the Bayes theory the properties of regularized solution, especially its covariance matrix, may be easily derived. This step is crucial to introduce and prove the formula for calculations of the signal recovery error. It has been proven that an average recovery error is approximately inversely proportional to the number of samples at voltage levels. The method is tested using signals registered by means of the single detection module of the J-PET detector built out from the 30 cm long BC-420 plastic scintillator strip. It is demonstrated that the experimental and theoretical functions describing the recovery errors in the J-PET scenario are largely consistent. The specificity and limitations of the signal recovery method in this application are discussed. It is shown that the PCA basis offers high level of information compression and an accurate recovery with just eight samples, from four voltage levels, for each signal waveform. Moreover, it is demonstrated that using the recovered waveform of the signals, instead of samples at four voltage levels alone, improves the spatial resolution of the hit position reconstruction. The experiment shows that spatial resolution evaluated based on information from four voltage levels, without a recovery of the waveform of the signal, is equal to 1.05 cm. After the application of an information from four voltage levels to the recovery of the signal waveform, the spatial resolution is improved to 0.94 cm. Moreover, the obtained result is only slightly worse than the one evaluated using the original raw-signal. The spatial resolution calculated under these conditions is equal to 0.93 cm. It is very important information since, limiting the number of threshold levels in the electronic devices to four, leads to significant reduction of the overall cost of the scanner. The developed recovery scheme is general and may be incorporated in any other investigation where a prior knowledge about the signals of interest may be utilized.

Keywords: plastic scintillators, positron emission tomography, statistical analysis, tikhonov regularization

Procedia PDF Downloads 446
1985 Coarse-Graining in Micromagnetic Simulations of Magnetic Hyperthermia

Authors: Razyeh Behbahani, Martin L. Plumer, Ivan Saika-Voivod

Abstract:

Micromagnetic simulations based on the stochastic Landau-Lifshitz-Gilbert equation are used to calculate dynamic magnetic hysteresis loops relevant to magnetic hyperthermia applications. With the goal to effectively simulate room-temperature loops for large iron-oxide based systems at relatively slow sweep rates on the order of 1 Oe/ns or less, a coarse-graining scheme is proposed and tested. The scheme is derived from a previously developed renormalization-group approach. Loops associated with nanorods, used as building blocks for larger nanoparticles that were employed in preclinical trials (Dennis et al., 2009 Nanotechnology 20 395103), serve as the model test system. The scaling algorithm is shown to produce nearly identical loops over several decades in the model grain sizes. Sweep-rate scaling involving the damping constant alpha is also demonstrated.

Keywords: coarse-graining, hyperthermia, hysteresis loops, micromagnetic simulations

Procedia PDF Downloads 149
1984 Morphological Features Fusion for Identifying INBREAST-Database Masses Using Neural Networks and Support Vector Machines

Authors: Nadia el Atlas, Mohammed el Aroussi, Mohammed Wahbi

Abstract:

In this paper a novel technique of mass characterization based on robust features-fusion is presented. The proposed method consists of mainly four stages: (a) the first phase involves segmenting the masses using edge information’s. (b) The second phase is to calculate and fuse the most relevant morphological features. (c) The last phase is the classification step which allows us to classify the images into benign and malignant masses. In this step we have implemented Support Vectors Machines (SVM) and Artificial Neural Networks (ANN), which were evaluated with the following performance criteria: confusion matrix, accuracy, sensitivity, specificity, receiver operating characteristic ROC, and error histogram. The effectiveness of this new approach was evaluated by a recently developed database: INBREAST database. The fusion of the most appropriate morphological features provided very good results. The SVM gives accuracy to within 64.3%. Whereas the ANN classifier gives better results with an accuracy of 97.5%.

Keywords: breast cancer, mammography, CAD system, features, fusion

Procedia PDF Downloads 599
1983 Performance Improvement of Cooperative Scheme in Wireless OFDM Systems

Authors: Ki-Ro Kim, Seung-Jun Yu, Hyoung-Kyu Song

Abstract:

Recently, the wireless communication systems are required to have high quality and provide high bit rate data services. Researchers have studied various multiple antenna scheme to meet the demand. In practical application, it is difficult to deploy multiple antennas for limited size and cost. Cooperative diversity techniques are proposed to overcome the limitations. Cooperative communications have been widely investigated to improve performance of wireless communication. Among diversity schemes, space-time block code has been widely studied for cooperative communication systems. In this paper, we propose a new cooperative scheme using pre-coding and space-time block code. The proposed cooperative scheme provides improved error performance than a conventional cooperative scheme using space-time block coding scheme.

Keywords: cooperative communication, space-time block coding, pre-coding

Procedia PDF Downloads 359
1982 Estimation of Rare and Clustered Population Mean Using Two Auxiliary Variables in Adaptive Cluster Sampling

Authors: Muhammad Nouman Qureshi, Muhammad Hanif

Abstract:

Adaptive cluster sampling (ACS) is specifically developed for the estimation of highly clumped populations and applied to a wide range of situations like animals of rare and endangered species, uneven minerals, HIV patients and drug users. In this paper, we proposed a generalized semi-exponential estimator with two auxiliary variables under the framework of ACS design. The expressions of approximate bias and mean square error (MSE) of the proposed estimator are derived. Theoretical comparisons of the proposed estimator have been made with existing estimators. A numerical study is conducted on real and artificial populations to demonstrate and compare the efficiencies of the proposed estimator. The results indicate that the proposed generalized semi-exponential estimator performed considerably better than all the adaptive and non-adaptive estimators considered in this paper.

Keywords: auxiliary information, adaptive cluster sampling, clustered populations, Hansen-Hurwitz estimation

Procedia PDF Downloads 238
1981 Fabrication of Tin Oxide and Metal Doped Tin Oxide for Gas Sensor Application

Authors: Goban Kumar Panneer Selvam

Abstract:

In past years, there is lots of death caused due to harmful gases. So its very important to monitor harmful gases for human safety, and semiconductor material play important role in producing effective gas sensors.A novel solvothermal synthesis method based on sol-gel processing was prepared to deposit tin oxide thin films on glass substrate at high temperature for gas sensing application. The structure and morphology of tin oxide were analyzed by X-ray diffraction (XRD), Fourier transforms infrared spectroscopy (FTIR) and scanning electron microscopy (SEM). The SEM analysis of how spheres shape in tin oxide nanoparticles. The structure characterization of tin oxide studied by X-ray diffraction shows 8.95 nm (calculated by sheers equation). The UV visible spectroscopy indicated a maximum absorption band shown at 390 nm. Further dope tin oxide with selected metals to attain maximum sensitivity using dip coating technique with different immersion and sensing characterization are measured.

Keywords: tin oxide, gas sensor, chlorine free, sensitivity, crystalline size

Procedia PDF Downloads 147