Search results for: complete orthogonal semantic space
1618 EEG-Based Fractal Analysis of Different Motor Imagery Tasks using Critical Exponent Method
Authors: Montri Phothisonothai, Masahiro Nakagawa
Abstract:
The objective of this paper is to characterize the spontaneous Electroencephalogram (EEG) signals of four different motor imagery tasks and to show hereby a possible solution for the present binary communication between the brain and a machine ora Brain-Computer Interface (BCI). The processing technique used in this paper was the fractal analysis evaluated by the Critical Exponent Method (CEM). The EEG signal was registered in 5 healthy subjects,sampling 15 measuring channels at 1024 Hz.Each channel was preprocessed by the Laplacian space ltering so as to reduce the space blur and therefore increase the spaceresolution. The EEG of each channel was segmented and its Fractaldimension (FD) calculated. The FD was evaluated in the time interval corresponding to the motor imagery and averaged out for all the subjects (each channel). In order to characterize the FD distribution,the linear regression curves of FD over the electrodes position were applied. The differences FD between the proposed mental tasks are quantied and evaluated for each experimental subject. The obtained results of the proposed method are a substantial fractal dimension in the EEG signal of motor imagery tasks and can be considerably utilized as the multiple-states BCI applications.
Keywords: electroencephalogram (EEG), motor imagery tasks, mental tasks, biomedical signals processing, human-machine interface, fractal analysis, critical exponent method (CEM).
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22591617 The Spiral_OWL Model – Towards Spiral Knowledge Engineering
Authors: Hafizullah A. Hashim, Aniza. A
Abstract:
The Spiral development model has been used successfully in many commercial systems and in a good number of defense systems. This is due to the fact that cost-effective incremental commitment of funds, via an analogy of the spiral model to stud poker and also can be used to develop hardware or integrate software, hardware, and systems. To support adaptive, semantic collaboration between domain experts and knowledge engineers, a new knowledge engineering process, called Spiral_OWL is proposed. This model is based on the idea of iterative refinement, annotation and structuring of knowledge base. The Spiral_OWL model is generated base on spiral model and knowledge engineering methodology. A central paradigm for Spiral_OWL model is the concentration on risk-driven determination of knowledge engineering process. The collaboration aspect comes into play during knowledge acquisition and knowledge validation phase. Design rationales for the Spiral_OWL model are to be easy-to-implement, well-organized, and iterative development cycle as an expanding spiral.Keywords: Domain Expert, Knowledge Base, Ontology, Software Process.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17681616 Evaluation of Seismic Damage for Gisha Bridge in Tehran by HAZUS Methodology
Authors: Langroudi B., Salehi E., Keshani S., Baghersad M.
Abstract:
Transportation is of great importance in the current life of human beings. The transportation system plays many roles, from economical development to after-catastrophe aids such as rescue operation in the first hours and days after an earthquake. In after earthquakes response phase, transportation system acts as a basis for ground operations including rescue and relief operation, food providing for victims and etc. It is obvious that partial or complete obstruction of this system results in the stop of these operations. Bridges are one of the most important elements of transportation network. Failure of a bridge, in the most optimistic case, cuts the relation between two regions and in more developed countries, cuts the relation of numerous regions. In this paper, to evaluate the vulnerability and estimate the damage level of Tehran bridges, HAZUS method, developed by Federal Emergency Management Agency (FEMA) with the aid of National Institute of Building Science (NIBS), is used for the first time in Iran. In this method, to evaluate the collapse probability, fragility curves are used. Iran is located on seismic belt and thus, it is vulnerable to earthquakes. Thus, the study of the probability of bridge collapses, as an important part of transportation system, during earthquakes is of great importance. The purpose of this study is to provide fragility curves for Gisha Bridge, one of the longest steel bridges in Tehran, as an important lifeline element. Besides, the damage probability for this bridge during a specific earthquake, introduced as scenario earthquakes, is calculated. The fragility curves show that for the considered scenario, the probability of occurrence of complete collapse for the bridge is 8.6%.Keywords: Bridge, Damage evaluation, Fragility curve, Lifelines, Seismic vulnerability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21421615 A New Class χ2 (M, A,) of the Double Difference Sequences of Fuzzy Numbers
Authors: N.Subramanian, U.K.Misra
Abstract:
The aim of this paper is to introduce and study a new concept of strong double χ2 (M,A, Δ) of fuzzy numbers and also some properties of the resulting sequence spaces of fuzzy numbers were examined.
Keywords: Modulus function, fuzzy number, metric space.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22981614 Modeling “Web of Trust“ with Web 2.0
Authors: Omer Mahmood, Selvakennedy Selvadurai
Abstract:
“Web of Trust" is one of the recognized goals for Web 2.0. It aims to make it possible for the people to take responsibility for what they publish on the web, including organizations, businesses and individual users. These objectives, among others, drive most of the technologies and protocols recently standardized by the governing bodies. One of the great advantages of Web infrastructure is decentralization of publication. The primary motivation behind Web 2.0 is to assist the people to add contents for Collective Intelligence (CI) while providing mechanisms to link content with people for evaluations and accountability of information. Such structure of contents will interconnect users and contents so that users can use contents to find participants and vice versa. This paper proposes conceptual information storage and linking model, based on decentralized information structure, that links contents and people together. The model uses FOAF, Atom, RDF and RDFS and can be used as a blueprint to develop Web 2.0 applications for any e-domain. However, primary target for this paper is online trust evaluation domain. The proposed model targets to assist the individuals to establish “Web of Trust" in online trust domain.Keywords: Web of Trust, Semantic Web, Electronic SocialNetworks, Information Management
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22221613 Demonstration of Land Use Changes Simulation Using Urban Climate Model
Authors: Barbara Vojvodikova, Katerina Jupova, Iva Ticha
Abstract:
Cities in their historical evolution have always adapted their internal structure to the needs of society (for example protective city walls during classicism era lost their defense function, became unnecessary, were demolished and gave space for new features such as roads, museums or parks). Today it is necessary to modify the internal structure of the city in order to minimize the impact of climate changes on the environment of the population. This article discusses the results of the Urban Climate model owned by VITO, which was carried out as part of a project from the European Union's Horizon grant agreement No 730004 Pan-European Urban Climate Services Climate-Fit city. The use of the model was aimed at changes in land use and land cover in cities related to urban heat islands (UHI). The task of the application was to evaluate possible land use change scenarios in connection with city requirements and ideas. Two pilot areas in the Czech Republic were selected. One is Ostrava and the other Hodonín. The paper provides a demonstration of the application of the model for various possible future development scenarios. It contains an assessment of the suitability or inappropriateness of scenarios of future development depending on the temperature increase. Cities that are preparing to reconstruct the public space are interested in eliminating proposals that would lead to an increase in temperature stress as early as in the assignment phase. If they have evaluation on the unsuitability of some type of design, they can limit it into the proposal phases. Therefore, especially in the application of models on Local level - in 1 m spatial resolution, it was necessary to show which type of proposals would create a significant temperature island in its implementation. Such a type of proposal is considered unsuitable. The model shows that the building itself can create a shady place and thus contribute to the reduction of the UHI. If it sensitively approaches the protection of existing greenery, this new construction may not pose a significant problem. More massive interventions leading to the reduction of existing greenery create a new heat island space.
Keywords: Heat islands, land use, urban climate model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8391612 VISMA: A Method for System Analysis in Early Lifecycle Phases
Authors: Walter Sebron, Hans Tschürtz, Peter Krebs
Abstract:
The choice of applicable analysis methods in safety or systems engineering depends on the depth of knowledge about a system, and on the respective lifecycle phase. However, the analysis method chain still shows gaps as it should support system analysis during the lifecycle of a system from a rough concept in pre-project phase until end-of-life. This paper’s goal is to discuss an analysis method, the VISSE Shell Model Analysis (VISMA) method, which aims at closing the gap in the early system lifecycle phases, like the conceptual or pre-project phase, or the project start phase. It was originally developed to aid in the definition of the system boundary of electronic system parts, like e.g. a control unit for a pump motor. Furthermore, it can be also applied to non-electronic system parts. The VISMA method is a graphical sketch-like method that stratifies a system and its parts in inner and outer shells, like the layers of an onion. It analyses a system in a two-step approach, from the innermost to the outermost components followed by the reverse direction. To ensure a complete view of a system and its environment, the VISMA should be performed by (multifunctional) development teams. To introduce the method, a set of rules and guidelines has been defined in order to enable a proper shell build-up. In the first step, the innermost system, named system under consideration (SUC), is selected, which is the focus of the subsequent analysis. Then, its directly adjacent components, responsible for providing input to and receiving output from the SUC, are identified. These components are the content of the first shell around the SUC. Next, the input and output components to the components in the first shell are identified and form the second shell around the first one. Continuing this way, shell by shell is added with its respective parts until the border of the complete system (external border) is reached. Last, two external shells are added to complete the system view, the environment and the use case shell. This system view is also stored for future use. In the second step, the shells are examined in the reverse direction (outside to inside) in order to remove superfluous components or subsystems. Input chains to the SUC, as well as output chains from the SUC are described graphically via arrows, to highlight functional chains through the system. As a result, this method offers a clear and graphical description and overview of a system, its main parts and environment; however, the focus still remains on a specific SUC. It helps to identify the interfaces and interfacing components of the SUC, as well as important external interfaces of the overall system. It supports the identification of the first internal and external hazard causes and causal chains. Additionally, the method promotes a holistic picture and cross-functional understanding of a system, its contributing parts, internal relationships and possible dangers within a multidisciplinary development team.
Keywords: Analysis methods, functional safety, hazard identification, system and safety engineering, system boundary definition, system safety.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11351611 A Context-Sensitive Algorithm for Media Similarity Search
Authors: Guang-Ho Cha
Abstract:
This paper presents a context-sensitive media similarity search algorithm. One of the central problems regarding media search is the semantic gap between the low-level features computed automatically from media data and the human interpretation of them. This is because the notion of similarity is usually based on high-level abstraction but the low-level features do not sometimes reflect the human perception. Many media search algorithms have used the Minkowski metric to measure similarity between image pairs. However those functions cannot adequately capture the aspects of the characteristics of the human visual system as well as the nonlinear relationships in contextual information given by images in a collection. Our search algorithm tackles this problem by employing a similarity measure and a ranking strategy that reflect the nonlinearity of human perception and contextual information in a dataset. Similarity search in an image database based on this contextual information shows encouraging experimental results.
Keywords: Context-sensitive search, image search, media search, similarity ranking, similarity search.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6401610 On the Paradigm Shift of the Overall Urban Design in China
Authors: Gaoyuan Wang, Tian Chen, Junnan Liu
Abstract:
Facing a period of major change that is rarely seen in a century, China formulates the 14th Five-Year Plan and places emphasis on promoting high-quality development. In this context, the overall urban design has become a crucial and systematic tool for high-quality urban development. However, there are bottlenecks in the cognition of nature, content scope and transmission mechanisms of the current overall urban design in China. The paper interprets the emerging demands of the 14th Five-Year Plan on urban design in terms of new value-quality priority, new dynamic-space performance, new target-region coordination and new path-refined governance. Based on the new trend and appeal, the multi-dimensional thinking integrated with the major tasks of urban design are proposed accordingly, which is the biomass thinking in ecological, production and living element, the strategic thinking in spatial structure, the systematic thinking in the cityscape, the low-carbon thinking in urban form, the governance thinking in public space, the user thinking in design implementation. The paper explores the possibility of transforming the value thinking and technical system of urban design in China and provides a breakthrough path for the urban planning and design industry to better respond to the propositions of the country’s 14th Five-Year Plan.
Keywords: China’s 14th five-year plan, overall urban design, urban design thinking, transformation of urban design.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5391609 An Image Enhancement Method Based on Curvelet Transform for CBCT-Images
Authors: Shahriar Farzam, Maryam Rastgarpour
Abstract:
Image denoising plays extremely important role in digital image processing. Enhancement of clinical image research based on Curvelet has been developed rapidly in recent years. In this paper, we present a method for image contrast enhancement for cone beam CT (CBCT) images based on fast discrete curvelet transforms (FDCT) that work through Unequally Spaced Fast Fourier Transform (USFFT). These transforms return a table of Curvelet transform coefficients indexed by a scale parameter, an orientation and a spatial location. Accordingly, the coefficients obtained from FDCT-USFFT can be modified in order to enhance contrast in an image. Our proposed method first uses a two-dimensional mathematical transform, namely the FDCT through unequal-space fast Fourier transform on input image and then applies thresholding on coefficients of Curvelet to enhance the CBCT images. Consequently, applying unequal-space fast Fourier Transform leads to an accurate reconstruction of the image with high resolution. The experimental results indicate the performance of the proposed method is superior to the existing ones in terms of Peak Signal to Noise Ratio (PSNR) and Effective Measure of Enhancement (EME).
Keywords: Curvelet transform, image enhancement, CBCT, image denoising.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12601608 New Method for Determining the Distribution of Birefringence and Linear Dichroism in Polymer Materials Based On Polarization-Holographic Grating
Authors: Barbara Kilosanidze, George Kakauridze, Levan Nadareishvili, Yuri Mshvenieradze
Abstract:
A new method for determining the distribution of birefringence and linear dichroism in optical polymer materials is presented. The method is based on the use of polarizationholographic diffraction grating that forms an orthogonal circular basis in the process of diffraction of probing laser beam on the grating. The intensities ratio of the orders of diffraction on this grating enables the value of birefringence and linear dichroism in the sample to be determined. The distribution of birefringence in the sample is determined by scanning with a circularly polarized beam with a wavelength far from the absorption band of the material. If the scanning is carried out by probing beam with the wavelength near to a maximum of the absorption band of the chromophore then the distribution of linear dichroism can be determined. An appropriate theoretical model of this method is presented. A laboratory setup was created for the proposed method. An optical scheme of the laboratory setup is presented. The results of measurement in polymer films with two-dimensional gradient distribution of birefringence and linear dichroism are discussed.
Keywords: Birefringence, graded oriented polymers, linear dichroism, optical polymers, optical anisotropy, polarization-holographic grating,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15121607 A Model to Study the Effect of Na+ ions on Ca2+diffusion under Rapid Buffering Approximation
Authors: Vikas Tewari, K.R. Pardasani
Abstract:
Calcium is very important for communication among the neurons. It is vital in a number of cell processes such as secretion, cell movement, cell differentiation. To reduce the system of reactiondiffusion equations of [Ca2+] into a single equation, two theories have been proposed one is excess buffer approximation (EBA) other is rapid buffer approximation (RBA). The RBA is more realistic than the EBA as it considers both the mobile and stationary endogenous buffers. It is valid near the mouth of the channel. In this work we have studied the effects of different types of buffers on calcium diffusion under RBA. The novel thing studied is the effect of sodium ions on calcium diffusion. The model has been made realistic by considering factors such as variable [Ca2+], [Na+] sources, sodium-calcium exchange protein(NCX), Sarcolemmal Calcium ATPase pump. The proposed mathematical leads to a system of partial differential equations which has been solved numerically to study the relationships between different parameters such as buffer concentration, buffer disassociation rate, calcium permeability. We have used Forward Time Centred Space (FTCS) approach to solve the system of partial differential equations.Keywords: rapid buffer approximation, sodium-calcium exchangeprotein, Sarcolemmal Calcium ATPase pump, buffer disassociationrate, forward time centred space.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15201606 Human Face Detection and Segmentation using Eigenvalues of Covariance Matrix, Hough Transform and Raster Scan Algorithms
Authors: J. Prakash, K. Rajesh
Abstract:
In this paper we propose a novel method for human face segmentation using the elliptical structure of the human head. It makes use of the information present in the edge map of the image. In this approach we use the fact that the eigenvalues of covariance matrix represent the elliptical structure. The large and small eigenvalues of covariance matrix are associated with major and minor axial lengths of an ellipse. The other elliptical parameters are used to identify the centre and orientation of the face. Since an Elliptical Hough Transform requires 5D Hough Space, the Circular Hough Transform (CHT) is used to evaluate the elliptical parameters. Sparse matrix technique is used to perform CHT, as it squeeze zero elements, and have only a small number of non-zero elements, thereby having an advantage of less storage space and computational time. Neighborhood suppression scheme is used to identify the valid Hough peaks. The accurate position of the circumference pixels for occluded and distorted ellipses is identified using Bresenham-s Raster Scan Algorithm which uses the geometrical symmetry properties. This method does not require the evaluation of tangents for curvature contours, which are very sensitive to noise. The method has been evaluated on several images with different face orientations.Keywords: Circular Hough Transform, Covariance matrix, Eigenvalues, Elliptical Hough Transform, Face segmentation, Raster Scan Algorithm.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 25171605 Investigation of the Flow Characteristics in a Catalytic Muffler with Perforated Inlet Cone
Authors: Gyo Woo Lee, Man Young Kim
Abstract:
Emission regulations for diesel engines are being strengthened and it is impossible to meet the standards without exhaust after-treatment systems. Lack of the space in many diesel vehicles, however, make it difficult to design and install stand-alone catalytic converters such as DOC, DPF, and SCR in the vehicle exhaust systems. Accordingly, those have been installed inside the muffler to save the space, and referred to the catalytic muffler. However, that has complex internal structure with perforated plate and pipe for noise and monolithic catalyst for emission reduction. For this reason, flow uniformity and pressure drop, which affect efficiency of catalyst and engine performance, respectively, should be examined when the catalytic muffler is designed. In this work, therefore, the flow uniformity and pressure drop to improve the performance of the catalytic converter and the engine have been numerically investigated by changing various design parameters such as inlet shape, porosity, and outlet shape of the muffler using the three-dimensional turbulent flow of the incompressible, non-reacting, and steady state inside the catalytic muffler. Finally, it can be found that the shape, in which the muffler has perforated pipe inside the inlet part, has higher uniformity index and lower pressure drop than others considered in this work.
Keywords: Catalytic muffler, Perforated inlet cone, Catalysts, Perforated pipe, Flow uniformity, Pressure drop.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 29021604 Application of Building Information Modeling in Energy Management of Individual Departments Occupying University Facilities
Authors: Kung-Jen Tu, Danny Vernatha
Abstract:
To assist individual departments within universities in their energy management tasks, this study explores the application of Building Information Modeling in establishing the ‘BIM based Energy Management Support System’ (BIM-EMSS). The BIM-EMSS consists of six components: (1) sensors installed for each occupant and each equipment, (2) electricity sub-meters (constantly logging lighting, HVAC, and socket electricity consumptions of each room), (3) BIM models of all rooms within individual departments’ facilities, (4) data warehouse (for storing occupancy status and logged electricity consumption data), (5) building energy management system that provides energy managers with various energy management functions, and (6) energy simulation tool (such as eQuest) that generates real time 'standard energy consumptions' data against which 'actual energy consumptions' data are compared and energy efficiency evaluated. Through the building energy management system, the energy manager is able to (a) have 3D visualization (BIM model) of each room, in which the occupancy and equipment status detected by the sensors and the electricity consumptions data logged are displayed constantly; (b) perform real time energy consumption analysis to compare the actual and standard energy consumption profiles of a space; (c) obtain energy consumption anomaly detection warnings on certain rooms so that energy management corrective actions can be further taken (data mining technique is employed to analyze the relation between space occupancy pattern with current space equipment setting to indicate an anomaly, such as when appliances turn on without occupancy); and (d) perform historical energy consumption analysis to review monthly and annually energy consumption profiles and compare them against historical energy profiles. The BIM-EMSS was further implemented in a research lab in the Department of Architecture of NTUST in Taiwan and implementation results presented to illustrate how it can be used to assist individual departments within universities in their energy management tasks.Keywords: Sensor, electricity sub-meters, database, energy anomaly detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22841603 Graph Codes-2D Projections of Multimedia Feature Graphs for Fast and Effective Retrieval
Authors: Stefan Wagenpfeil, Felix Engel, Paul McKevitt, Matthias Hemmje
Abstract:
Multimedia Indexing and Retrieval is generally de-signed and implemented by employing feature graphs. These graphs typically contain a significant number of nodes and edges to reflect the level of detail in feature detection. A higher level of detail increases the effectiveness of the results but also leads to more complex graph structures. However, graph-traversal-based algorithms for similarity are quite inefficient and computation intensive, espe-cially for large data structures. To deliver fast and effective retrieval, an efficient similarity algorithm, particularly for large graphs, is mandatory. Hence, in this paper, we define a graph-projection into a 2D space (Graph Code) as well as the corresponding algorithms for indexing and retrieval. We show that calculations in this space can be performed more efficiently than graph-traversals due to a simpler processing model and a high level of parallelisation. In consequence, we prove that the effectiveness of retrieval also increases substantially, as Graph Codes facilitate more levels of detail in feature fusion. Thus, Graph Codes provide a significant increase in efficiency and effectiveness (especially for Multimedia indexing and retrieval) and can be applied to images, videos, audio, and text information.
Keywords: indexing, retrieval, multimedia, graph code, graph algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4441602 SLM Using Riemann Sequence Combined with DCT Transform for PAPR Reduction in OFDM Communication Systems
Authors: Pepin Magnangana Zoko Goyoro, Ibrahim James Moumouni, Sroy Abouty
Abstract:
Orthogonal Frequency Division Multiplexing (OFDM) is an efficient method of data transmission for high speed communication systems. However, the main drawback of OFDM systems is that, it suffers from the problem of high Peak-to-Average Power Ratio (PAPR) which causes inefficient use of the High Power Amplifier and could limit transmission efficiency. OFDM consist of large number of independent subcarriers, as a result of which the amplitude of such a signal can have high peak values. In this paper, we propose an effective reduction scheme that combines DCT and SLM techniques. The scheme is composed of the DCT followed by the SLM using the Riemann matrix to obtain phase sequences for the SLM technique. The simulation results show PAPR can be greatly reduced by applying the proposed scheme. In comparison with OFDM, while OFDM had high values of PAPR –about 10.4dB our proposed method achieved about 4.7dB reduction of the PAPR with low complexities computation. This approach also avoids randomness in phase sequence selection, which makes it simpler to decode at the receiver. As an added benefit, the matrices can be generated at the receiver end to obtain the data signal and hence it is not required to transmit side information (SI).Keywords: DCT transform, OFDM, PAPR, Riemann matrix, SLM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26401601 Investigating the Effectiveness of Iranian Architecture on Sustainable Space Creation
Authors: Mansour Nikpour, Mohsen Ghasemi, Elahe Mosavi, Mohd Zin Kandar
Abstract:
lack of convenience condition is one of the problems in open spaces in hot and dry regions. Nowadays parks and green landscapes was designed and constructed without any attention to convenience condition. If this process continues, Citizens will encounter with some problems. Harsh climatic condition decreases the efficiency of people-s activities. However there is hard environment condition in hot and dry regions, Convenience condition has been provided in Iranian traditional architecture by using techniques and methods. In this research at the first step characteristics of Iranian garden that can effect on creating sustainable spaces were investigated through observation method. Pleasure space in cities will be created with using these methods and techniques in future cities. Furthermore the comparison between Iranian garden and landscape in today-s cities demonstrate the effectiveness of Iranian garden characteristics on sustainable spaces. Iranian architects used simple and available methods for creating open architectural spaces. In addition desirable conditions were provided with taking in to account both physically and spiritually. Parks and landscapes in future cities can be designed and constructed with respect to architectural techniques that used in Iranian gardens in hot and arid regions.Keywords: Iranian garden, convenience condition, landscape, sustainable
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18541600 Solving Differential's Equation of Carrier Load on Semiconductor
Authors: Morteza Amirabadi, Vahid Fayaz , Fereshteh Felegary, Hossien Hossienkhani
Abstract:
The most suitable Semiconductor detector, Cadmium Zinc Teloraid , has unique properties because of high Atomic number and wide Brand Gap . It has been tried in this project with different processes such as Lead , Diffusion , Produce and Recombination , effect of Trapping and injection carrier of CdZnTe , to get hole and then present a complete answer of it . Then we should investigate the movement of carrier ( Electron – Hole ) by using above answer.Keywords: Semiconcuctor detector, Trapping, Recommbination, Diffusion
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14531599 Resilient Machine Learning in the Nuclear Industry: Crack Detection as a Case Study
Authors: Anita Khadka, Gregory Epiphaniou, Carsten Maple
Abstract:
There is a dramatic surge in the adoption of Machine Learning (ML) techniques in many areas, including the nuclear industry (such as fault diagnosis and fuel management in nuclear power plants), autonomous systems (including self-driving vehicles), space systems (space debris recovery, for example), medical surgery, network intrusion detection, malware detection, to name a few. Artificial Intelligence (AI) has become a part of everyday modern human life. To date, the predominant focus has been developing underpinning ML algorithms that can improve accuracy, while factors such as resiliency and robustness of algorithms have been largely overlooked. If an adversarial attack is able to compromise the learning method or data, the consequences can be fatal, especially but not exclusively in safety-critical applications. In this paper, we present an in-depth analysis of five adversarial attacks and two defence methods on a crack detection ML model. Our analysis shows that it can be dangerous to adopt ML techniques without rigorous testing, since they may be vulnerable to adversarial attacks, especially in security-critical areas such as the nuclear industry. We observed that while the adopted defence methods can effectively defend against different attacks, none of them could protect against all five adversarial attacks entirely.
Keywords: Resilient Machine Learning, attacks, defences, nuclear industry, crack detection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5011598 The Design of Axisymmetric Ducts for Incompressible Flow with a Parabolic Axial Velocity Inlet Profile
Authors: V.Pavlika
Abstract:
In this paper a numerical algorithm is described for solving the boundary value problem associated with axisymmetric, inviscid, incompressible, rotational (and irrotational) flow in order to obtain duct wall shapes from prescribed wall velocity distributions. The governing equations are formulated in terms of the stream function ψ (x,y)and the function φ (x,y)as independent variables where for irrotational flow φ (x,y)can be recognized as the velocity potential function, for rotational flow φ (x,y)ceases being the velocity potential function but does remain orthogonal to the stream lines. A numerical method based on the finite difference scheme on a uniform mesh is employed. The technique described is capable of tackling the so-called inverse problem where the velocity wall distributions are prescribed from which the duct wall shape is calculated, as well as the direct problem where the velocity distribution on the duct walls are calculated from prescribed duct geometries. The two different cases as outlined in this paper are in fact boundary value problems with Neumann and Dirichlet boundary conditions respectively. Even though both approaches are discussed, only numerical results for the case of the Dirichlet boundary conditions are given. A downstream condition is prescribed such that cylindrical flow, that is flow which is independent of the axial coordinate, exists.
Keywords: Inverse problem, irrotational incompressible flow, Boundary value problem.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16431597 Simultaneous HPAM/SDS Injection in Heterogeneous/Layered Models
Authors: M. H. Sedaghat, A. Zamani, S. Morshedi, R. Janamiri, M. Safdari, I. Mahdavi, A. Hosseini, A. Hatampour
Abstract:
Although lots of experiments have been done in enhanced oil recovery, the number of experiments which consider the effects of local and global heterogeneity on efficiency of enhanced oil recovery based on the polymer-surfactant flooding is low and rarely done. In this research, we have done numerous experiments of water flooding and polymer-surfactant flooding on a five spot glass micromodel in different conditions such as different positions of layers. In these experiments, five different micromodels with three different pore structures are designed. Three models with different layer orientation, one homogenous model and one heterogeneous model are designed. In order to import the effect of heterogeneity of porous media, three types of pore structures are distributed accidentally and with equal ratio throughout heterogeneous micromodel network according to random normal distribution. The results show that maximum EOR recovery factor will happen in a situation where the layers are orthogonal to the path of mainstream and the minimum EOR recovery factor will happen in a situation where the model is heterogeneous. This experiments show that in polymer-surfactant flooding, with increase of angles of layers the EOR recovery factor will increase and this recovery factor is strongly affected by local heterogeneity around the injection zone.
Keywords: Layered Reservoir, Micromodel, Local Heterogeneity, Polymer-Surfactant Flooding, Enhanced Oil Recovery.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22191596 The Effect of Bath Composition for Hot-Dip Aluminizing of AISI 4140 Steel
Authors: Aptullah Karakaş, Murat Baydoğan
Abstract:
In the HDA process, Al or Al-Si molten baths are mostly used. However, in this study, three different Al alloys such as Al4043 (Al-Mg), Al5356 (Al-Si) and Al7020 (Al-Zn) were used as the molten bath in order to see their effects on morphological and mechanical properties of the resulting aluminide layers. AISI 4140 low alloyed steel was used as the substrate. Parameters of the HDA process were bath composition, bath temperature, and dipping time. These parameters were considered within a Taguchi L9 orthogonal array. After the HDA process and subsequent diffusion annealing, coating thickness measurement, microstructural analysis and hardness measurement of the aluminide layers were conducted. The optimum process parameters were evaluated according to coating morphology such as cracks, Kirkendall porosity and hardness of the coatings. According to the results, smooth and clean aluminide layer with less Kirkendall porosity and cracks were observed on the sample, which was aluminized in the molten Al7020 bath at 700 C for 10 minutes, and subsequently diffusion annealed at 750 C. Hardness of the aluminide layer was in between 1100-1300 hardness of Vickers (HV) and the coating thickness was approximately 400 µm. The results were promising such that a hard and thick aluminide layer with less Kirkendall porosity and cracks could be formed. It is therefore, concluded that Al7020 bath may be used in the HDA process of AISI 4140 steel substrate.
Keywords: Aluminum alloys, coating, hot-dip aluminizing, microstructure.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 921595 Minimizing Mutant Sets by Equivalence and Subsumption
Authors: Samia Alblwi, Amani Ayad
Abstract:
Mutation testing is the art of generating syntactic variations of a base program and checking whether a candidate test suite can identify all the mutants that are not semantically equivalent to the base; this technique can be used to assess the quality of test suite. One of the main obstacles to the widespread use of mutation testing is cost, as even small programs (a few dozen lines of code) can give rise to a large number of mutants (up to hundreds); this has created an incentive to seek to reduce the number of mutants while preserving their collective effectiveness. Two criteria have been used to reduce the size of mutant sets: equivalence, which aims to partition the set of mutants into equivalence classes modulo semantic equivalence, and selecting one representative per class; and, subsumption, which aims to define a partial ordering among mutants that ranks mutants by effectiveness and seeks to select maximal elements in this ordering. In this paper, we analyze these two policies using analytical and empirical criteria.
Keywords: Mutation testing, mutant sets, mutant equivalence, mutant subsumption, mutant set minimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1931594 Design and Performance of Adaptive Polarized MIMO MC-SS-CDMA System for Downlink Mobile Communications
Authors: Joseph V. M. Halim, Hesham El-Badawy, Hadia M. El-Hennawy
Abstract:
In this paper, an adaptive polarized Multiple-Input Multiple-Output (MIMO) Multicarrier Spread Spectrum Code Division Multiple Access (MC-SS-CDMA) system is designed for downlink mobile communications. The proposed system will be examined in Frequency Division Duplex (FDD) mode for both macro urban and suburban environments. For the same transmission bandwidth, a performance comparison between both nonoverlapped and orthogonal Frequency Division Multiplexing (FDM) schemes will be presented. Also, the proposed system will be compared with both the closed loop vertical MIMO MC-SS-CDMA system and the synchronous vertical STBC-MIMO MC-SS-CDMA system. As will be shown, the proposed system introduces a significant performance gain as well as reducing the spatial dimensions of the MIMO system and simplifying the receiver implementation. The effect of the polarization diversity characteristics on the BER performance will be discussed. Also, the impact of excluding the cross-polarization MCSS- CDMA blocks in the base station will be investigated. In addition, the system performance will be evaluated under different Feedback Information (FBI) rates for slowly-varying channels. Finally, a performance comparison for vehicular and pedestrian environments will be presentedKeywords: Closed loop technique, MC-SS-CDMA, Polarized MIMO systems, Transmit diversity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16221593 Testing Loaded Programs Using Fault Injection Technique
Authors: S. Manaseer, F. A. Masooud, A. A. Sharieh
Abstract:
Fault tolerance is critical in many of today's large computer systems. This paper focuses on improving fault tolerance through testing. Moreover, it concentrates on the memory faults: how to access the editable part of a process memory space and how this part is affected. A special Software Fault Injection Technique (SFIT) is proposed for this purpose. This is done by sequentially scanning the memory of the target process, and trying to edit maximum number of bytes inside that memory. The technique was implemented and tested on a group of programs in software packages such as jet-audio, Notepad, Microsoft Word, Microsoft Excel, and Microsoft Outlook. The results from the test sample process indicate that the size of the scanned area depends on several factors. These factors are: process size, process type, and virtual memory size of the machine under test. The results show that increasing the process size will increase the scanned memory space. They also show that input-output processes have more scanned area size than other processes. Increasing the virtual memory size will also affect the size of the scanned area but to a certain limit.Keywords: Complex software systems, Error detection, Fault tolerance, Injection and testing methodology, Memory faults, Process and virtual memory.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18861592 An Optimization of Machine Parameters for Modified Horizontal Boring Tool Using Taguchi Method
Authors: Thirasak Panyaphirawat, Pairoj Sapsmarnwong, Teeratas Pornyungyuen
Abstract:
This paper presents the findings of an experimental investigation of important machining parameters for the horizontal boring tool modified to mouth with a horizontal lathe machine to bore an overlength workpiece. In order to verify a usability of a modified tool, design of experiment based on Taguchi method is performed. The parameters investigated are spindle speed, feed rate, depth of cut and length of workpiece. Taguchi L9 orthogonal array is selected for four factors three level parameters in order to minimize surface roughness (Ra and Rz) of S45C steel tubes. Signal to noise ratio analysis and analysis of variance (ANOVA) is performed to study an effect of said parameters and to optimize the machine setting for best surface finish. The controlled factors with most effect are depth of cut, spindle speed, length of workpiece, and feed rate in order. The confirmation test is performed to test the optimal setting obtained from Taguchi method and the result is satisfactory.
Keywords: Design of Experiment, Taguchi Design, Optimization, Analysis of Variance, Machining Parameters, Horizontal Boring Tool.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27061591 Optimization of Control Parameters for EWR in Injection Flushing Type of EDM on Stainless Steel 304 Workpiece
Authors: M. S. Reza, M. Hamdi, S. H. Tomadi, A. R. Ismail
Abstract:
The operating control parameters of injection flushing type of electrical discharge machining process on stainless steel 304 workpiece using copper tools are being optimized according to its individual machining characteristic i.e. Electrode Wear Ratio (EWR). Higher EWR would give bad dimensional precision for the EDM machined workpiece because of high electrode wear. Hence, the quality characteristic for EWR is set to lower-the-better to achieve the optimum dimensional precision for the machined workpiece. Taguchi method has been used for the construction, layout and analysis of the experiment for EWR machining characteristic. The use of Taguchi method in the experiment saves a lot of time and cost of preparing and machining the experiment samples. Therefore, an L18 Orthogonal array which was the fundamental component in the statistical design of experiments has been used to plan the experiments and Analysis of Variance (ANOVA) is used to determine the optimum machining parameters for this machining characteristic. The control parameters selected for this optimization experiments are polarity, pulse on duration, discharge current, discharge voltage, machining depth, machining diameter and dielectric liquid pressure. The result had shown that negative polarity machining parameter setting will decreases EWR.Keywords: ANOVA, EDM, Injection Flushing, L18Orthogonal Array, EWR, Stainless Steel 304
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18481590 Optimization of Process Parameters of Pressure Die Casting using Taguchi Methodology
Authors: Satish Kumar, Arun Kumar Gupta, Pankaj Chandna
Abstract:
The present work analyses different parameters of pressure die casting to minimize the casting defects. Pressure diecasting is usually applied for casting of aluminium alloys. Good surface finish with required tolerances and dimensional accuracy can be achieved by optimization of controllable process parameters such as solidification time, molten temperature, filling time, injection pressure and plunger velocity. Moreover, by selection of optimum process parameters the pressure die casting defects such as porosity, insufficient spread of molten material, flash etc. are also minimized. Therefore, a pressure die casting component, carburetor housing of aluminium alloy (Al2Si2O5) has been considered. The effects of selected process parameters on casting defects and subsequent setting of parameters with the levels have been accomplished by Taguchi-s parameter design approach. The experiments have been performed as per the combination of levels of different process parameters suggested by L18 orthogonal array. Analyses of variance have been performed for mean and signal-to-noise ratio to estimate the percent contribution of different process parameters. Confidence interval has also been estimated for 95% consistency level and three conformational experiments have been performed to validate the optimum level of different parameters. Overall 2.352% reduction in defects has been observed with the help of suggested optimum process parameters.
Keywords: Aluminium Casting, Pressure Die Casting, Taguchi Methodology, Design of Experiments
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 73351589 Performance Analysis in 5th Generation Massive Multiple-Input-Multiple-Output Systems
Authors: Jihad S. Daba, Jean-Pierre Dubois, Georges El Soury
Abstract:
Fifth generation wireless networks guarantee significant capacity enhancement to suit more clients and services at higher information rates with better reliability while consuming less power. The deployment of massive multiple-input-multiple-output technology guarantees broadband wireless networks with the use of base station antenna arrays to serve a large number of users on the same frequency and time-slot channels. In this work, we evaluate the performance of massive multiple-input-multiple-output systems (MIMO) systems in 5th generation cellular networks in terms of capacity and bit error rate. Several cases were considered and analyzed to compare the performance of massive MIMO systems while varying the number of antennas at both transmitting and receiving ends. We found that, unlike classical MIMO systems, reducing the number of transmit antennas while increasing the number of antennas at the receiver end provides a better solution to performance enhancement. In addition, enhanced orthogonal frequency division multiplexing and beam division multiple access schemes further improve the performance of massive MIMO systems and make them more reliable.
Keywords: Beam division multiple access, D2D communication, enhanced OFDM, fifth generation broadband, massive MIMO.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 748