Search results for: proposed concept
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7226

Search results for: proposed concept

926 Aspect Oriented Software Architecture

Authors: Pradip Peter Dey, Ronald F. Gonzales, Gordon W. Romney, Mohammad Amin, Bhaskar Raj Sinha

Abstract:

Natural language processing systems pose a unique challenge for software architectural design as system complexity has increased continually and systems cannot be easily constructed from loosely coupled modules. Lexical, syntactic, semantic, and pragmatic aspects of linguistic information are tightly coupled in a manner that requires separation of concerns in a special way in design, implementation and maintenance. An aspect oriented software architecture is proposed in this paper after critically reviewing relevant architectural issues. For the purpose of this paper, the syntactic aspect is characterized by an augmented context-free grammar. The semantic aspect is composed of multiple perspectives including denotational, operational, axiomatic and case frame approaches. Case frame semantics matured in India from deep thematic analysis. It is argued that lexical, syntactic, semantic and pragmatic aspects work together in a mutually dependent way and their synergy is best represented in the aspect oriented approach. The software architecture is presented with an augmented Unified Modeling Language.

Keywords: Language engineering, parsing, software design, user experience.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1743
925 Design and Operation of a Multicarrier Energy System Based On Multi Objective Optimization Approach

Authors: Azadeh Maroufmashat, Sourena Sattari Khavas, Halle Bakhteeyar

Abstract:

Multi-energy systems will enhance the system reliability and power quality. This paper presents an integrated approach for the design and operation of distributed energy resources (DER) systems, based on energy hub modeling. A multi-objective optimization model is developed by considering an integrated view of electricity and natural gas network to analyze the optimal design and operating condition of DER systems, by considering two conflicting objectives, namely, minimization of total cost and the minimization of environmental impact which is assessed in terms of CO2 emissions. The mathematical model considers energy demands of the site, local climate data, and utility tariff structure, as well as technical and financial characteristics of the candidate DER technologies. To provide energy demands, energy systems including photovoltaic, and co-generation systems, boiler, central power grid are considered. As an illustrative example, a hotel in Iran demonstrates potential applications of the proposed method. The results prove that increasing the satisfaction degree of environmental objective leads to increased total cost.

Keywords: Multi objective optimization, DER systems, Energy hub, Cost, CO2 emission.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2466
924 WhatsApp as Part of a Blended Learning Model to Help Programming Novices

Authors: Tlou J. Ramabu

Abstract:

Programming is one of the challenging subjects in the field of computing. In the higher education sphere, some programming novices’ performance, retention rate, and success rate are not improving. Most of the time, the problem is caused by the slow pace of learning, difficulty in grasping the syntax of the programming language and poor logical skills. More importantly, programming forms part of major subjects within the field of computing. As a result, specialized pedagogical methods and innovation are highly recommended. Little research has been done on the potential productivity of the WhatsApp platform as part of a blended learning model. In this article, the authors discuss the WhatsApp group as a part of blended learning model incorporated for a group of programming novices. We discuss possible administrative activities for productive utilisation of the WhatsApp group on the blended learning overview. The aim is to take advantage of the popularity of WhatsApp and the time students spend on it for their educational purpose. We believe that blended learning featuring a WhatsApp group may ease novices’ cognitive load and strengthen their foundational programming knowledge and skills. This is a work in progress as the proposed blended learning model with WhatsApp incorporated is yet to be implemented.

Keywords: Blended learning, higher education, WhatsApp, programming, novices, lecturers.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1177
923 Transient Stability Assessment Using Fuzzy SVM and Modified Preventive Control

Authors: B. Dora Arul Selvi, .N. Kamaraj

Abstract:

Transient Stability is an important issue in power systems planning, operation and extension. The objective of transient stability analysis problem is not satisfied with mere transient instability detection or evaluation and it is most important to complement it by defining fast and efficient control measures in order to ensure system security. This paper presents a new Fuzzy Support Vector Machines (FSVM) to investigate the stability status of power systems and a modified generation rescheduling scheme to bring back the identified unstable cases to a more economical and stable operating point. FSVM improves the traditional SVM (Support Vector Machines) by adding fuzzy membership to each training sample to indicate the degree of membership of this sample to different classes. The preventive control based on economic generator rescheduling avoids the instability of the power systems with minimum change in operating cost under disturbed conditions. Numerical results on the New England 39 bus test system show the effectiveness of the proposed method.

Keywords: Fuzzy Support Vector Machine (FSVM), Incremental Cost, Preventive Control, Transient stability

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1491
922 Seed-Based Region Growing (SBRG) vs Adaptive Network-Based Inference System (ANFIS) vs Fuzzyc-Means (FCM): Brain Abnormalities Segmentation

Authors: Shafaf Ibrahim, Noor Elaiza Abdul Khalid, Mazani Manaf

Abstract:

Segmentation of Magnetic Resonance Imaging (MRI) images is the most challenging problems in medical imaging. This paper compares the performances of Seed-Based Region Growing (SBRG), Adaptive Network-Based Fuzzy Inference System (ANFIS) and Fuzzy c-Means (FCM) in brain abnormalities segmentation. Controlled experimental data is used, which designed in such a way that prior knowledge of the size of the abnormalities are known. This is done by cutting various sizes of abnormalities and pasting it onto normal brain tissues. The normal tissues or the background are divided into three different categories. The segmentation is done with fifty seven data of each category. The knowledge of the size of the abnormalities by the number of pixels are then compared with segmentation results of three techniques proposed. It was proven that the ANFIS returns the best segmentation performances in light abnormalities, whereas the SBRG on the other hand performed well in dark abnormalities segmentation.

Keywords: Seed-Based Region Growing (SBRG), Adaptive Network-Based Fuzzy Inference System (ANFIS), Fuzzy c-Means (FCM), Brain segmentation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2305
921 Development a New Model of EEVC/WG17 Lower Legform for Pedestrian Safety

Authors: Alireza Noorpoor, Akbar Abvabi, Mehdi Saeed Kiasat

Abstract:

Development, calibration and validation of a threedimensional model of the Legform impactor for pedestrian crash with bumper are presented. Lower limb injury is becoming an increasingly important concern in vehicle safety for both occupants and pedestrians. In order to prevent lower extremity injuries to a pedestrian when struck by a car, it is important to elucidate the loadings from car front structures on the lower extremities and the injury mechanism caused by these loadings. An impact test procedure with a legform addressing lower limb injuries in car pedestrian accidents has been proposed by EEVC/WG17. In this study a modified legform impactor is introduced and validated against EEVC/WG17 criteria. The finite element model of this legform is developed using LS-DYNA software. Total mass of legform impactor is 13.4 kg.Technical specifications including the mass and location of the center of gravity and moment of inertia about a horizontal axis through the respective centre of gravity in femur and tibia are determined. The obtained results of legform impactor static and dynamic tests are as specified in the EEVC/WG17.

Keywords: Legform impactor, Pedestrian safety, Finite element model, Knee joint, EEVC/WG17.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3048
920 Web Page Watermarking: XML files using Synonyms and Acronyms

Authors: Nighat Mir, Sayed Afaq Hussain

Abstract:

Advent enhancements in the field of computing have increased massive use of web based electronic documents. Current Copyright protection laws are inadequate to prove the ownership for electronic documents and do not provide strong features against copying and manipulating information from the web. This has opened many channels for securing information and significant evolutions have been made in the area of information security. Digital Watermarking has developed into a very dynamic area of research and has addressed challenging issues for digital content. Watermarking can be visible (logos or signatures) and invisible (encoding and decoding). Many visible watermarking techniques have been studied for text documents but there are very few for web based text. XML files are used to trade information on the internet and contain important information. In this paper, two invisible watermarking techniques using Synonyms and Acronyms are proposed for XML files to prove the intellectual ownership and to achieve the security. Analysis is made for different attacks and amount of capacity to be embedded in the XML file is also noticed. A comparative analysis for capacity is also made for both methods. The system has been implemented using C# language and all tests are made practically to get the results.

Keywords: Watermarking, Extensible Markup Language (XML), Synonyms, Acronyms, Copyright protection.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2281
919 Design and Analysis of MEMS based Accelerometer for Automatic Detection of Railway Wheel Flat

Authors: Rajib Ul Alam Uzzal, Ion Stiharu, Waiz Ahmed

Abstract:

This paper presents the modeling of a MEMS based accelerometer in order to detect the presence of a wheel flat in the railway vehicle. A haversine wheel flat is assigned to one wheel of a 5 DOF pitch plane vehicle model, which is coupled to a 3 layer track model. Based on the simulated acceleration response obtained from the vehicle-track model, an accelerometer is designed that meets all the requirements to detect the presence of a wheel flat. The proposed accelerometer can survive in a dynamic shocking environment with acceleration up to ±150g. The parameters of the accelerometer are calculated in order to achieve the required specifications using lumped element approximation and the results are used for initial design layout. A finite element analysis code (COMSOL) is used to perform simulations of the accelerometer under various operating conditions and to determine the optimum configuration. The simulated results are found within about 2% of the calculated values, which indicates the validity of lumped element approach. The stability of the accelerometer is also determined in the desired range of operation including the condition under shock.

Keywords: MEMS accelerometer, Pitch plane vehicle, wheel flat.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3073
918 Mooring Analysis of Duct-Type Tidal Current Power System in Shallow Water

Authors: Chul H. Jo, Do Y. Kim, Bong K. Cho, Myeong J. Kim

Abstract:

The exhaustion of oil and the environmental pollution from the use of fossil fuel are increasing. Tidal current power (TCP) has been proposed as an alternative energy source because of its predictability and reliability. By applying a duct and single point mooring (SPM) system, a TCP device can amplify the generating power and keep its position properly. Because the generating power is proportional to cube of the current stream velocity, amplifying the current speed by applying a duct to a TCP system is an effective way to improve the efficiency of the power device. An SPM system can be applied at any water depth and is highly cost effective. Simple installation and maintenance procedures are also merits of an SPM system. In this study, we designed an SPM system for a duct-type TCP device for use in shallow water. Motions of the duct are investigated to obtain the response amplitude operator (RAO) as the magnitude of the transfer function. Parameters affecting the stability of the SPM system such as the fairlead departure angle, current velocity, and the number of clamp weights are analyzed and/or optimized. Wadam and OrcaFlex commercial software is used to design the mooring line.

Keywords: Mooring design, parametric analysis, response amplitude operator, single point mooring.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2174
917 A Grey-Fuzzy Controller for Optimization Technique in Wireless Networks

Authors: Yao-Tien Wang, Hsiang-Fu Yu, Dung Chen Chiou

Abstract:

In wireless and mobile communications, this progress provides opportunities for introducing new standards and improving existing services. Supporting multimedia traffic with wireless networks quality of service (QoS). In this paper, a grey-fuzzy controller for radio resource management (GF-RRM) is presented to maximize the number of the served calls and QoS provision in wireless networks. In a wireless network, the call arrival rate, the call duration and the communication overhead between the base stations and the control center are vague and uncertain. In this paper, we develop a method to predict the cell load and to solve the RRM problem based on the GF-RRM, and support the present facility has been built on the application-level of the wireless networks. The GF-RRM exhibits the better adaptability, fault-tolerant capability and performance than other algorithms. Through simulations, we evaluate the blocking rate, update overhead, and channel acquisition delay time of the proposed method. The results demonstrate our algorithm has the lower blocking rate, less updated overhead, and shorter channel acquisition delay.

Keywords: radio resource management, grey prediction, fuzzylogic control, wireless networks, quality of service.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1717
916 Behavior Analysis Based On Nine Degrees-of-Freedom Sensor for Emergency Rescue Evacuation Support System

Authors: Maeng-Hwan Hyun, Dae-Man Do, Young-Bok Choi

Abstract:

Around the world, there are frequent incidents of natural disasters, such as earthquakes, tsunamis, floods, and snowstorms, as well as manmade disasters such as fires, arsons, and acts of terror. These diverse and unpredictable adversities have resulted in a number of fatalities and injuries. If disaster occurrence can be assessed quickly and information such as the exact location of the disaster and evacuation routes can be provided, victims can promptly move to safe locations, minimizing losses. This paper proposes a behavior analysis method based on a nine degrees-of-freedom (9-DOF) sensor that is effective for the emergency rescue evacuation support system (ERESS), which is being researched with an objective of providing evacuation support during disasters. Based on experiments performed using the acceleration sensor and the gyroscope sensor in the 9-DOF sensor, data are analyzed for human behavior regarding stationary position, walking, running, and during emergency situation to suggest guidelines for system judgment. Using the results of the experiments performed to determine disaster occurrence, it was confirmed that the proposed method quickly determines whether a disaster has occurred.

Keywords: Behavior Analysis, Nine degrees-of-freedom sensor, Emergency rescue.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1689
915 Character Segmentation Method for a License Plate with Topological Transform

Authors: Jaedo Kim, Youngjoon Han, Hernsoo Hahn

Abstract:

This paper propose the robust character segmentation method for license plate with topological transform such as twist,rotation. The first step of the proposed method is to find a candidate region for character and license plate. The character or license plate must be appeared as closed loop in the edge image. In the case of detecting candidate for character region, the evaluation of detected region is using topological relationship between each character. When this method decides license plate candidate region, character features in the region with binarization are used. After binarization for the detected candidate region, each character region is decided again. In this step, each character region is fitted more than previous step. In the next step, the method checks other character regions with different scale near the detected character regions, because most license plates have license numbers with some meaningful characters around them. The method uses perspective projection for geometrical normalization. If there is topological distortion in the character region, the method projects the region on a template which is defined as standard license plate using perspective projection. In this step, the method is able to separate each number region and small meaningful characters. The evaluation results are tested with a number of test images.

Keywords: License Plate Detection, Character Segmentation, Perspective Projection, Topological Transform.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1935
914 Identification of Author and Reviewer from Single and Double Blind Paper

Authors: Jatinderkumar R. Saini, Nikita R. Sonthalia, Khushbu A. Dodiya

Abstract:

Research leads to the development of science and technology and hence it leads to the betterment of humankind also. Journals and Conferences provide a platform to receive large number of research papers for publications and presentations before the expert and peer-level scientific community. In order to assure quality of such papers, they are also sent to reviewers for their comments. In order to maintain good ethical standards, the research papers are sent to reviewers in such a way authors and reviewers do not know each other’s identity. This technique is called Double-blind Review Process. It is called Single-blind Review Process, if identity of any one party, generally authors’, is disclosed to the other. This paper presents the techniques by which identity of author as well as reviewer could be found even through Double-blind Review process. It is proposed that the characteristics and techniques presented here will help journals and conferences in assuring intentional or un-intentional disclosure of identity revealing information by the either party. 

Keywords: Author, Conference, Double Blind Paper, Journal, Reviewer, Single Blind Paper.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2447
913 Design of Two-Channel Quincunx Quadrature Mirror Filter Banks Using Digital All-Pass Lattice Filters

Authors: Ju-Hong Lee, Chong-Jia Ciou

Abstract:

This paper deals with the problem of two-dimensional (2-D) recursive two-channel quincunx quadrature mirror filter (QQMF) banks design. The analysis and synthesis filters of the 2-D recursive QQMF bank are composed of 2-D recursive digital allpass lattice filters (DALFs) with symmetric half-plane (SHP) support regions. Using the 2-D doubly complementary half-band (DC-HB) property possessed by the analysis and synthesis filters, we facilitate the design of the proposed QQMF bank. For finding the coefficients of the 2-D recursive SHP DALFs, we present a structure of 2-D recursive digital allpass filters by using 2-D SHP recursive digital all-pass lattice filters (DALFs). The novelty of using 2-D SHP recursive DALFs to construct a 2-D recursive QQMF bank is that the resulting 2-D recursive QQMF bank provides better performance than the existing 2-D recursive QQMF banks. Simulation results are also presented for illustration and comparison.

Keywords: All-pass digital filter, doubly complementary, lattice structure, symmetric half-plane digital filter, quincunx QMF bank.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 886
912 A Hybrid Feature Selection by Resampling, Chi squared and Consistency Evaluation Techniques

Authors: Amir-Massoud Bidgoli, Mehdi Naseri Parsa

Abstract:

In this paper a combined feature selection method is proposed which takes advantages of sample domain filtering, resampling and feature subset evaluation methods to reduce dimensions of huge datasets and select reliable features. This method utilizes both feature space and sample domain to improve the process of feature selection and uses a combination of Chi squared with Consistency attribute evaluation methods to seek reliable features. This method consists of two phases. The first phase filters and resamples the sample domain and the second phase adopts a hybrid procedure to find the optimal feature space by applying Chi squared, Consistency subset evaluation methods and genetic search. Experiments on various sized datasets from UCI Repository of Machine Learning databases show that the performance of five classifiers (Naïve Bayes, Logistic, Multilayer Perceptron, Best First Decision Tree and JRIP) improves simultaneously and the classification error for these classifiers decreases considerably. The experiments also show that this method outperforms other feature selection methods.

Keywords: feature selection, resampling, reliable features, Consistency Subset Evaluation.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2584
911 Thermal Load Calculations of Multilayered Walls

Authors: Bashir M. Suleiman

Abstract:

Thermal load calculations have been performed for multi-layered walls that are composed of three different parts; a common (sand and cement) plaster, and two types of locally produced soft and hard bricks. The masonry construction of these layered walls was based on concrete-backed stone masonry made of limestone bricks joined by mortar. These multilayered walls are forming the outer walls of the building envelope of a typical Libyan house. Based on the periodic seasonal weather conditions, within the Libyan cost region during summer and winter, measured thermal conductivity values were used to implement such seasonal variation of heat flow and the temperature variations through the walls. The experimental measured thermal conductivity values were obtained using the Hot Disk technique. The estimation of the thermal resistance of the wall layers ( R-values) is based on measurements and calculations. The numerical calculations were done using a simplified analytical model that considers two different wall constructions which are characteristics of such houses. According to the obtained results, the R-values were quite low and therefore, several suggestions have been proposed to improve the thermal loading performance that will lead to a reasonable human comfort and reduce energy consumption.

Keywords: Thermal loading, multilayered walls, Libyan bricks, thermal resistance

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2338
910 A Cuckoo Search with Differential Evolution for Clustering Microarray Gene Expression Data

Authors: M. Pandi, K. Premalatha

Abstract:

A DNA microarray technology is a collection of microscopic DNA spots attached to a solid surface. Scientists use DNA microarrays to measure the expression levels of large numbers of genes simultaneously or to genotype multiple regions of a genome. Elucidating the patterns hidden in gene expression data offers a tremendous opportunity for an enhanced understanding of functional genomics. However, the large number of genes and the complexity of biological networks greatly increase the challenges of comprehending and interpreting the resulting mass of data, which often consists of millions of measurements. It is handled by clustering which reveals the natural structures and identifying the interesting patterns in the underlying data. In this paper, gene based clustering in gene expression data is proposed using Cuckoo Search with Differential Evolution (CS-DE). The experiment results are analyzed with gene expression benchmark datasets. The results show that CS-DE outperforms CS in benchmark datasets. To find the validation of the clustering results, this work is tested with one internal and one external cluster validation indexes.

Keywords: DNA, Microarray, genomics, Cuckoo Search, Differential Evolution, Gene expression data, Clustering.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1483
909 Relative Radiometric Correction of Cloudy Multitemporal Satellite Imagery

Authors: Seema Biday, Udhav Bhosle

Abstract:

Repeated observation of a given area over time yields potential for many forms of change detection analysis. These repeated observations are confounded in terms of radiometric consistency due to changes in sensor calibration over time, differences in illumination, observation angles and variation in atmospheric effects. This paper demonstrates applicability of an empirical relative radiometric normalization method to a set of multitemporal cloudy images acquired by Resourcesat1 LISS III sensor. Objective of this study is to detect and remove cloud cover and normalize an image radiometrically. Cloud detection is achieved by using Average Brightness Threshold (ABT) algorithm. The detected cloud is removed and replaced with data from another images of the same area. After cloud removal, the proposed normalization method is applied to reduce the radiometric influence caused by non surface factors. This process identifies landscape elements whose reflectance values are nearly constant over time, i.e. the subset of non-changing pixels are identified using frequency based correlation technique. The quality of radiometric normalization is statistically assessed by R2 value and mean square error (MSE) between each pair of analogous band.

Keywords: Correlation, Frequency domain, Multitemporal, Relative Radiometric Correction

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1981
908 Assertion-Driven Test Repair Based on Priority Criteria

Authors: Ruilian Zhao, Shukai Zhang, Yan Wang, Weiwei Wang

Abstract:

Repairing broken test cases is an expensive and challenging task in evolving software systems. Although an automated repair technique with intent-preservation has been proposed, it does not take into account the association between test repairs and assertions, leading a large number of irrelevant candidates and decreasing the repair capability. This paper proposes a assertion-driven test repair approach. Furthermore, a intent-oriented priority criterion is raised to guide the repair candidate generation, making the repairs closer to the intent of the test. In more detail, repair targets are determined through post-dominance relations between assertions and the methods that directly cause compilation errors. Then, test repairs are generated from the target in a bottom-up way, guided by the the intent-oriented priority criteria. Finally, the generated repair candidates are prioritized to match the original test intent. The approach is implemented and evaluated on the benchmark of 4 open-source programs and 91 broken test cases. The result shows that the approach can fix 89% (81/91) broken test cases, which are more effective than the existing intent-preserved test repair approach, and our intent-oriented priority criteria work well.

Keywords: Test repair, test intent, software test, test case evolution.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 156
907 Enhancing the Effectiveness of Air Defense Systems through Simulation Analysis

Authors: F. Felipe

Abstract:

Air Defense Systems contain high-value assets that are expected to fulfill their mission for several years - in many cases, even decades - while operating in a fast-changing, technology-driven environment. Thus, it is paramount that decision-makers can assess how effective an Air Defense System is in the face of new developing threats, as well as to identify the bottlenecks that could jeopardize the security of the airspace of a country. Given the broad extent of activities and the great variety of assets necessary to achieve the strategic objectives, a systems approach was taken in order to delineate the core requirements and the physical architecture of an Air Defense System. Then, value-focused thinking helped in the definition of the measures of effectiveness. Furthermore, analytical methods were applied to create a formal structure that preliminarily assesses such measures. To validate the proposed methodology, a powerful simulation was also used to determine the measures of effectiveness, now in more complex environments that incorporate both uncertainty and multiple interactions of the entities. The results regarding the validity of this methodology suggest that the approach can support decisions aimed at enhancing the capabilities of Air Defense Systems. In conclusion, this paper sheds some light on how consolidated approaches of Systems Engineering and Operations Research can be used as valid techniques for solving problems regarding a complex and yet vital matter.

Keywords: Air defense, effectiveness, system, simulation, decision-support.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 408
906 A New Approach to Face Recognition Using Dual Dimension Reduction

Authors: M. Almas Anjum, M. Younus Javed, A. Basit

Abstract:

In this paper a new approach to face recognition is presented that achieves double dimension reduction, making the system computationally efficient with better recognition results and out perform common DCT technique of face recognition. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results change with change in face image resolution and provide optimal results when arriving at a certain resolution level. In the proposed model of face recognition, initially image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to increased computational speed and feature extraction potential of Discrete Cosine Transform (DCT), it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A tradeoff between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL , Yale and EME color database.

Keywords: Biometrics, DCT, Face Recognition, Illumination, Computation, Feature extraction.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1686
905 An Energy Aware Dispatch Scheme WSNs

Authors: Siddhartha Chauhan, Kumar S. Pandey, Prateek Chandra

Abstract:

One of the key research issues in wireless sensor networks (WSNs) is how to efficiently deploy sensors to cover an area. In this paper, we present a Fishnet Based Dispatch Scheme (FiBDS) with energy aware mobility and interest based sensing angle. We propose two algorithms, one is FiBDS centralized algorithm and another is FiBDS distributed algorithm. The centralized algorithm is designed specifically for the non-time critical applications, commonly known as non real-time applications while the distributed algorithm is designed specifically for the time critical applications, commonly known as real-time applications. The proposed dispatch scheme works in a phase-selection manner. In this in each phase a specific constraint is dealt with according to the specified priority and then moved onto the next phase and at the end of each only the best suited nodes for the phase are chosen. Simulation results are presented to verify their effectiveness. 

Keywords: Dispatch Scheme, Energy Aware Mobility, Interest based Sensing, Wireless Sensor Networks (WSNs).

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1626
904 The Gravitational Impact of the Sun and the Moon on Heavy Mineral Deposits and Dust Particles in Low Gravity Regions of the Earth

Authors: T. B. Karu Jayasundara

Abstract:

The Earth’s gravity is not uniform. The satellite imageries of the Earth’s surface from NASA reveal a number of different gravity anomaly regions all over the globe. When the moon rotates around the earth, its gravity has a major physical influence on a number of regions on the earth. This physical change can be seen by the tides. The tides make sea levels high and low in coastal regions. During high tide, the gravitational force of the Moon pulls the Earth’s gravity so that the total gravitational intensity of Earth is reduced; it is further reduced in the low gravity regions of Earth. This reduction in gravity helps keep the suspended particles such as dust in the atmosphere, sand grains in the sea water for longer. Dramatic differences can be seen from the floating dust in the low gravity regions when compared with other regions. The above phenomena can be demonstrated from experiments. The experiments have to be done in high and low gravity regions of the earth during high and low tide, which will assist in comparing the final results. One of the experiments that can be done is by using a water filled cylinder about 80 cm tall, a few particles, which have the same density and same diameter (about 1 mm) and a stop watch. The selected particles were dropped from the surface of the water in the cylinder and the time taken for the particles to reach the bottom of the cylinder was measured using the stop watch. The times of high and low tide charts can be obtained from the regional government authorities. This concept is demonstrated by the particle drop times taken at high and low tides. The result of the experiment shows that the particle settlement time is less in low tide and high in high tide. The experiment for dust particles in air can be collected on filters, which are cellulose ester membranes and using a vacuum pump. The dust on filters can be used to make slides according to the NOHSC method. Counting the dust particles on the slides can be done using a phase contrast microscope. The results show that the concentration of dust is high at high tide and low in low tide. As a result of the high tides, a high concentration of heavy minerals deposit on placer deposits and dust particles retain in the atmosphere for longer in low gravity regions. These conditions are remarkably exhibited in the lowest low gravity region of the earth, mainly in the regions of India, Sri Lanka and in the middle part of the Indian Ocean. The biggest heavy mineral placer deposits are found in coastal regions of India and Sri Lanka and heavy dust particles are found in the atmosphere of India, particularly in the Delhi region.

Keywords: Dust particles, high and low tides, heavy minerals. low gravity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 623
903 Using Blockchain Technology to Extend the Vendor Managed Inventory for Sustainability

Authors: Elham Ahmadi, Roshaali Khaturia, Pardis Sahraei, Mohammad Niyayesh, Omid Fatahi Valilai

Abstract:

Nowadays, Information Technology (IT) is changing the way traditional enterprise management concepts work. One of the most dominant IT achievements is the Blockchain Technology. This technology enables the distributed collaboration of stakeholders for their interactions while fulfilling the security and consensus rules among them. This paper has focused on the application of Blockchain technology to enhance one of traditional inventory management models. The Vendor Managed Inventory (VMI) has been considered one of the most efficient mechanisms for vendor inventory planning by the suppliers. While VMI has brought competitive advantages for many industries, however its centralized mechanism limits the collaboration of a pool of suppliers and vendors simultaneously. This paper has studied the recent research for VMI application in industries and also has investigated the applications of Blockchain technology for decentralized collaboration of stakeholders. Focusing on sustainability issue for total supply chain consisting suppliers and vendors, it has proposed a Blockchain based VMI conceptual model. The different capabilities of this model for enabling the collaboration of stakeholders while maintaining the competitive advantages and sustainability issues have been discussed.

Keywords: Vendor Managed Inventory, Blockchain Technology, supply chain planning, sustainability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 803
902 On the Network Packet Loss Tolerance of SVM Based Activity Recognition

Authors: Gamze Uslu, Sebnem Baydere, Alper K. Demir

Abstract:

In this study, data loss tolerance of Support Vector Machines (SVM) based activity recognition model and multi activity classification performance when data are received over a lossy wireless sensor network is examined. Initially, the classification algorithm we use is evaluated in terms of resilience to random data loss with 3D acceleration sensor data for sitting, lying, walking and standing actions. The results show that the proposed classification method can recognize these activities successfully despite high data loss. Secondly, the effect of differentiated quality of service performance on activity recognition success is measured with activity data acquired from a multi hop wireless sensor network, which introduces  high data loss. The effect of number of nodes on the reliability and multi activity classification success is demonstrated in simulation environment. To the best of our knowledge, the effect of data loss in a wireless sensor network on activity detection success rate of an SVM based classification algorithm has not been studied before.

Keywords: Activity recognition, support vector machines, acceleration sensor, wireless sensor networks, packet loss.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2871
901 Coding based Synchronization Algorithm for Secondary Synchronization Channel in WCDMA

Authors: Deng Liao, Dongyu Qiu, Ahmed K. Elhakeem

Abstract:

A new code synchronization algorithm is proposed in this paper for the secondary cell-search stage in wideband CDMA systems. Rather than using the Cyclically Permutable (CP) code in the Secondary Synchronization Channel (S-SCH) to simultaneously determine the frame boundary and scrambling code group, the new synchronization algorithm implements the same function with less system complexity and less Mean Acquisition Time (MAT). The Secondary Synchronization Code (SSC) is redesigned by splitting into two sub-sequences. We treat the information of scrambling code group as data bits and use simple time diversity BCH coding for further reliability. It avoids involved and time-costly Reed-Solomon (RS) code computations and comparisons. Analysis and simulation results show that the Synchronization Error Rate (SER) yielded by the new algorithm in Rayleigh fading channels is close to that of the conventional algorithm in the standard. This new synchronization algorithm reduces system complexities, shortens the average cell-search time and can be implemented in the slot-based cell-search pipeline. By taking antenna diversity and pipelining correlation processes, the new algorithm also shows its flexible application in multiple antenna systems.

Keywords: WCDMA cell-search, synchronization algorithm, secondary synchronization channel, antenna diversity.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2391
900 Visualizing Imaging Pathways after Anatomy-Specific Follow-Up Imaging Recommendations

Authors: Thusitha Mabotuwana, Christopher S. Hall

Abstract:

Radiologists routinely make follow-up imaging recommendations, usually based on established clinical practice guidelines, such as the Fleischner Society guidelines for managing lung nodules. In order to ensure optimal care, it is important to make guideline-compliant recommendations, and also for patients to follow-up on these imaging recommendations in a timely manner. However, determining such compliance rates after a specific finding has been observed usually requires many time-consuming manual steps. To address some of these limitations with current approaches, in this paper we discuss a methodology to automatically detect finding-specific follow-up recommendations from radiology reports and create a visualization for relevant subsequent exams showing the modality transitions. Nearly 5% of patients who had a lung related follow-up recommendation continued to have at least eight subsequent outpatient CT exams during a seven year period following the recommendation. Radiologist and section chiefs can use the proposed tool to better understand how a specific patient population is being managed, identify possible deviations from established guideline recommendations and have a patient-specific graphical representation of the imaging pathways for an abstract view of the overall treatment path thus far.

Keywords: Follow-up recommendations, care pathways, imaging pathway visualization, follow-up tracking.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1061
899 A New Approach to Design an Efficient CIC Decimator Using Signed Digit Arithmetic

Authors: Vishal Awasthi, Krishna Raj

Abstract:

Any digital processing performed on a signal with larger nyquist interval requires more computation than signal processing performed on smaller nyquist interval. The sampling rate alteration generates the unwanted effects in the system such as spectral aliasing and spectral imaging during signal processing. Multirate-multistage implementation of digital filter can result a significant computational saving than single rate filter designed for sample rate conversion. In this paper, we presented an efficient cascaded integrator comb (CIC) decimation filter that perform fast down sampling using signed digit adder algorithm with compensated frequency droop that arises due to aliasing effect during the decimation process. This proposed compensated CIC decimation filter structure with a hybrid signed digit (HSD) fast adder provide an improved performance in terms of down sampling speed by 65.15% than ripple carry adder (RCA) and reduced area and power by 57.5% and 0.01 % than signed digit (SD) adder algorithms respectively.

Keywords: Sampling rate conversion, Multirate Filtering, Compensation Theory, Decimation filter, CIC filter, Redundant signed digit arithmetic, Fast adders.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4892
898 Relay Node Placement for Connectivity Restoration in Wireless Sensor Networks Using Genetic Algorithms

Authors: Hanieh Tarbiat Khosrowshahi, Mojtaba Shakeri

Abstract:

Wireless Sensor Networks (WSNs) consist of a set of sensor nodes with limited capability. WSNs may suffer from multiple node failures when they are exposed to harsh environments such as military zones or disaster locations and lose connectivity by getting partitioned into disjoint segments. Relay nodes (RNs) are alternatively introduced to restore connectivity. They cost more than sensors as they benefit from mobility, more power and more transmission range, enforcing a minimum number of them to be used. This paper addresses the problem of RN placement in a multiple disjoint network by developing a genetic algorithm (GA). The problem is reintroduced as the Steiner tree problem (which is known to be an NP-hard problem) by the aim of finding the minimum number of Steiner points where RNs are to be placed for restoring connectivity. An upper bound to the number of RNs is first computed to set up the length of initial chromosomes. The GA algorithm then iteratively reduces the number of RNs and determines their location at the same time. Experimental results indicate that the proposed GA is capable of establishing network connectivity using a reasonable number of RNs compared to the best existing work.

Keywords: Connectivity restoration, genetic algorithms, multiple-node failure, relay nodes, wireless sensor networks.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1104
897 Uncertainty Propagation and Sensitivity Analysis During Calibration of an Integrated Land Use and Transport Model

Authors: Parikshit Dutta, Mathieu Saujot, Elise Arnaud, Benoit Lefevre, Emmanuel Prados

Abstract:

In this work, propagation of uncertainty during calibration process of TRANUS, an integrated land use and transport model (ILUTM), has been investigated. It has also been examined, through a sensitivity analysis, which input parameters affect the variation of the outputs the most. Moreover, a probabilistic verification methodology of calibration process, which equates the observed and calculated production, has been proposed. The model chosen as an application is the model of the city of Grenoble, France. For sensitivity analysis and uncertainty propagation, Monte Carlo method was employed, and a statistical hypothesis test was used for verification. The parameters of the induced demand function in TRANUS, were assumed as uncertain in the present case. It was found that, if during calibration, TRANUS converges, then with a high probability the calibration process is verified. Moreover, a weak correlation was found between the inputs and the outputs of the calibration process. The total effect of the inputs on outputs was investigated, and the output variation was found to be dictated by only a few input parameters.

Keywords: Uncertainty propagation, sensitivity analysis, calibration under uncertainty, hypothesis testing, integrated land use and transport models, TRANUS, Grenoble.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1521