Search results for: domain
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 900

Search results for: domain

180 Numerical Simulation of Supersonic Gas Jet Flows and Acoustics Fields

Authors: Lei Zhang, Wen-jun Ruan, Hao Wang, Peng-xin Wang

Abstract:

The source of the jet noise is generated by rocket exhaust plume during rocket engine testing. A domain decomposition approach is applied to the jet noise prediction in this paper. The aerodynamic noise coupling is based on the splitting into acoustic sources generation and sound propagation in separate physical domains. Large Eddy Simulation (LES) is used to simulate the supersonic jet flow. Based on the simulation results of the flow-fields, the jet noise distribution of the sound pressure level is obtained by applying the Ffowcs Williams-Hawkings (FW-H) acoustics equation and Fourier transform. The calculation results show that the complex structures of expansion waves, compression waves and the turbulent boundary layer could occur due to the strong interaction between the gas jet and the ambient air. In addition, the jet core region, the shock cell and the sound pressure level of the gas jet increase with the nozzle size increasing. Importantly, the numerical simulation results of the far-field sound are in good agreement with the experimental measurements in directivity.

Keywords: Supersonic gas jet, Large Eddy Simulation(LES), acoustic noise, Ffowcs Williams-Hawkings (FW-H) equations, nozzle size.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2619
179 Relative Radiometric Correction of Cloudy Multitemporal Satellite Imagery

Authors: Seema Biday, Udhav Bhosle

Abstract:

Repeated observation of a given area over time yields potential for many forms of change detection analysis. These repeated observations are confounded in terms of radiometric consistency due to changes in sensor calibration over time, differences in illumination, observation angles and variation in atmospheric effects. This paper demonstrates applicability of an empirical relative radiometric normalization method to a set of multitemporal cloudy images acquired by Resourcesat1 LISS III sensor. Objective of this study is to detect and remove cloud cover and normalize an image radiometrically. Cloud detection is achieved by using Average Brightness Threshold (ABT) algorithm. The detected cloud is removed and replaced with data from another images of the same area. After cloud removal, the proposed normalization method is applied to reduce the radiometric influence caused by non surface factors. This process identifies landscape elements whose reflectance values are nearly constant over time, i.e. the subset of non-changing pixels are identified using frequency based correlation technique. The quality of radiometric normalization is statistically assessed by R2 value and mean square error (MSE) between each pair of analogous band.

Keywords: Correlation, Frequency domain, Multitemporal, Relative Radiometric Correction

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1980
178 Fuzzy C-Means Clustering for Biomedical Documents Using Ontology Based Indexing and Semantic Annotation

Authors: S. Logeswari, K. Premalatha

Abstract:

Search is the most obvious application of information retrieval. The variety of widely obtainable biomedical data is enormous and is expanding fast. This expansion makes the existing techniques are not enough to extract the most interesting patterns from the collection as per the user requirement. Recent researches are concentrating more on semantic based searching than the traditional term based searches. Algorithms for semantic searches are implemented based on the relations exist between the words of the documents. Ontologies are used as domain knowledge for identifying the semantic relations as well as to structure the data for effective information retrieval. Annotation of data with concepts of ontology is one of the wide-ranging practices for clustering the documents. In this paper, indexing based on concept and annotation are proposed for clustering the biomedical documents. Fuzzy c-means (FCM) clustering algorithm is used to cluster the documents. The performances of the proposed methods are analyzed with traditional term based clustering for PubMed articles in five different diseases communities. The experimental results show that the proposed methods outperform the term based fuzzy clustering.

Keywords: MeSH Ontology, Concept Indexing, Annotation, semantic relations, Fuzzy c-means.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2303
177 Bayesian Network Model for Students- Laboratory Work Performance Assessment: An Empirical Investigation of the Optimal Construction Approach

Authors: Ifeyinwa E. Achumba, Djamel Azzi, Rinat Khusainov

Abstract:

There are three approaches to complete Bayesian Network (BN) model construction: total expert-centred, total datacentred, and semi data-centred. These three approaches constitute the basis of the empirical investigation undertaken and reported in this paper. The objective is to determine, amongst these three approaches, which is the optimal approach for the construction of a BN-based model for the performance assessment of students- laboratory work in a virtual electronic laboratory environment. BN models were constructed using all three approaches, with respect to the focus domain, and compared using a set of optimality criteria. In addition, the impact of the size and source of the training, on the performance of total data-centred and semi data-centred models was investigated. The results of the investigation provide additional insight for BN model constructors and contribute to literature providing supportive evidence for the conceptual feasibility and efficiency of structure and parameter learning from data. In addition, the results highlight other interesting themes.

Keywords: Bayesian networks, model construction, parameterlearning, structure learning, performance index, model comparison.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1727
176 An Implementation of MacMahon's Partition Analysis in Ordering the Lower Bound of Processing Elements for the Algorithm of LU Decomposition

Authors: Halil Snopce, Ilir Spahiu, Lavdrim Elmazi

Abstract:

A lot of Scientific and Engineering problems require the solution of large systems of linear equations of the form bAx in an effective manner. LU-Decomposition offers good choices for solving this problem. Our approach is to find the lower bound of processing elements needed for this purpose. Here is used the so called Omega calculus, as a computational method for solving problems via their corresponding Diophantine relation. From the corresponding algorithm is formed a system of linear diophantine equalities using the domain of computation which is given by the set of lattice points inside the polyhedron. Then is run the Mathematica program DiophantineGF.m. This program calculates the generating function from which is possible to find the number of solutions to the system of Diophantine equalities, which in fact gives the lower bound for the number of processors needed for the corresponding algorithm. There is given a mathematical explanation of the problem as well. Keywordsgenerating function, lattice points in polyhedron, lower bound of processor elements, system of Diophantine equationsand : calculus.

Keywords: generating function, lattice points in polyhedron, lower bound of processor elements, system of Diophantine equations and calculus.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1473
175 Small Signal Stability Assessment Employing PSO Based TCSC Controller with Comparison to GA Based Design

Authors: D. Mondal, A. Chakrabarti, A. Sengupta

Abstract:

This paper aims to select the optimal location and setting parameters of TCSC (Thyristor Controlled Series Compensator) controller using Particle Swarm Optimization (PSO) and Genetic Algorithm (GA) to mitigate small signal oscillations in a multimachine power system. Though Power System Stabilizers (PSSs) are prime choice in this issue, installation of FACTS device has been suggested here in order to achieve appreciable damping of system oscillations. However, performance of any FACTS devices highly depends upon its parameters and suitable location in the power network. In this paper PSO as well as GA based techniques are used separately and compared their performances to investigate this problem. The results of small signal stability analysis have been represented employing eigenvalue as well as time domain response in face of two common power system disturbances e.g., varying load and transmission line outage. It has been revealed that the PSO based TCSC controller is more effective than GA based controller even during critical loading condition.

Keywords: Genetic Algorithm, Particle Swarm Optimization, Small Signal Stability, Thyristor Controlled Series Compensator.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1956
174 Grid Independence Study of Flow Past a Square Cylinder Using the Multi-Relaxation-Time Lattice Boltzmann Method

Authors: Shams-Ul-Islam, Hamid Rahman, Waqas Sarwar Abbasi

Abstract:

Numerical calculations of flow around a square cylinder are presented using the multi-relaxation-time lattice Boltzmann method at Reynolds number 150. The effects of upstream locations, downstream locations and blockage are investigated systematically. A detail analysis are given in terms of time-trace analysis of drag and lift coefficients, power spectra analysis of lift coefficient, vorticity contours visualizations and phase diagrams. A number of physical quantities mean drag coefficient, drag coefficient, Strouhal number and root-mean-square values of drag and lift coefficients are calculated and compared with the well resolved experimental data and numerical results available in open literature. The results had shown that the upstream, downstream and height of the computational domain are at least 7.5, 37.5 and 12 diameters of the cylinder, respectively.

Keywords: Grid independence, Multi-relaxation-time lattice Boltzmann method, Physical quantities, Square cylinder, Vorticity contours visualizations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3339
173 Application of Central Composite Design Based Response Surface Methodology in Parameter Optimization and on Cellulase Production Using Agricultural Waste

Authors: R.Muthuvelayudham, T.Viruthagiri

Abstract:

Response Surface Methodology (RSM) is a powerful and efficient mathematical approach widely applied in the optimization of cultivation process. Cellulase enzyme production by Trichoderma reesei RutC30 using agricultural waste rice straw and banana fiber as carbon source were investigated. In this work, sequential optimization strategy based statistical design was employed to enhance the production of cellulase enzyme through submerged cultivation. A fractional factorial design (26-2) was applied to elucidate the process parameters that significantly affect cellulase production. Temperature, Substrate concentration, Inducer concentration, pH, inoculum age and agitation speed were identified as important process parameters effecting cellulase enzyme synthesis. The concentration of lignocelluloses and lactose (inducer) in the cultivation medium were found to be most significant factors. The steepest ascent method was used to locate the optimal domain and a Central Composite Design (CCD) was used to estimate the quadratic response surface from which the factor levels for maximum production of cellulase were determined.

Keywords: Banana fiber, Cellulase, Optimization, Rice straw

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2400
172 A Framework for the Design of Green Giga Passive Optical Fiber Access Network in Kuwait

Authors: Ali A. Hammadi

Abstract:

In this work, a practical study on a commissioned Giga Passive Optical Network (GPON) fiber to the home access network in Kuwait is presented. The work covers the framework of the conceptual design of the deployed Passive Optical Networks (PONs), access network, optical fiber cable network distribution, technologies, and standards. The work also describes methodologies applied by system engineers for design of Optical Network Terminals (ONTs) and Optical Line Terminals (OLTs) transceivers with respect to the distance, operating wavelengths, splitting ratios. The results have demonstrated and justified the limitation of transmission distance of a PON link in Fiber to The Premises (FTTP) to not exceed 20 km. Optical Time Domain Reflector (OTDR) test has been carried for this project to confirm compliance with International Telecommunication Union (ITU) specifications regarding the total length of the deployed optical cable, total loss in dB, and loss per km in dB/km with respect to the operating wavelengths. OTDR test results with traces for segments of implemented fiber network will be provided and discussed.

Keywords: Passive optical networks, fiber to the premises, access network, OTDR.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1067
171 Measuring Process Component Design on Achieving Managerial Goals

Authors: Eakong Atiptamvaree, Twittie Senivongse

Abstract:

Process-oriented software development is a new software development paradigm in which software design is modeled by a business process which is in turn translated into a process execution language for execution. The building blocks of this paradigm are software units that are composed together to work according to the flow of the business process. This new paradigm still exhibits the characteristic of the applications built with the traditional software component technology. This paper discusses an approach to apply a traditional technique for software component fabrication to the design of process-oriented software units, called process components. These process components result from decomposing a business process of a particular application domain into subprocesses, and these process components can be reused to design the business processes of other application domains. The decomposition considers five managerial goals, namely cost effectiveness, ease of assembly, customization, reusability, and maintainability. The paper presents how to design or decompose process components from a business process model and measure some technical features of the design that would affect the managerial goals. A comparison between the measurement values from different designs can tell which process component design is more appropriate for the managerial goals that have been set. The proposed approach can be applied in Web Services environment which accommodates process-oriented software development.

Keywords: Business Process Model, Managerial Goals, ProcessComponent.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1513
170 Detecting Interactions between Behavioral Requirements with OWL and SWRL

Authors: Haibo Hu, Dan Yang, Chunxiao Ye, Chunlei Fu, Ren Li

Abstract:

High quality requirements analysis is one of the most crucial activities to ensure the success of a software project, so that requirements verification for software system becomes more and more important in Requirements Engineering (RE) and it is one of the most helpful strategies for improving the quality of software system. Related works show that requirement elicitation and analysis can be facilitated by ontological approaches and semantic web technologies. In this paper, we proposed a hybrid method which aims to verify requirements with structural and formal semantics to detect interactions. The proposed method is twofold: one is for modeling requirements with the semantic web language OWL, to construct a semantic context; the other is a set of interaction detection rules which are derived from scenario-based analysis and represented with semantic web rule language (SWRL). SWRL based rules are working with rule engines like Jess to reason in semantic context for requirements thus to detect interactions. The benefits of the proposed method lie in three aspects: the method (i) provides systematic steps for modeling requirements with an ontological approach, (ii) offers synergy of requirements elicitation and domain engineering for knowledge sharing, and (3)the proposed rules can systematically assist in requirements interaction detection.

Keywords: Requirements Engineering, Semantic Web, OWL, Requirements Interaction Detection, SWRL.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1797
169 Numerical Analysis of a Centrifugal Fan for Improved Performance using Splitter Vanes

Authors: N. Yagnesh Sharma, K. Vasudeva Karanth

Abstract:

The flow field in a centrifugal fan is highly complex with flow reversal taking place on the suction side of impeller and diffuser vanes. Generally performance of the centrifugal fan could be enhanced by judiciously introducing splitter vanes so as to improve the diffusion process. An extensive numerical whole field analysis on the effect of splitter vanes placed in discrete regions of suspected separation points is possible using CFD. This paper examines the effect of splitter vanes corresponding to various geometrical locations on the impeller and diffuser. The analysis shows that the splitter vanes located near the diffuser exit improves the static pressure recovery across the diffusing domain to a larger extent. Also it is found that splitter vanes located at the impeller trailing edge and diffuser leading edge at the mid-span of the circumferential distance between the blades show a marginal improvement in the static pressure recovery across the fan. However, splitters provided near to the suction side of the impeller trailing edge (25% of the circumferential gap between the impeller blades towards the suction side), adversely affect the static pressure recovery of the fan.

Keywords: Splitter vanes, Flow separation, Sliding mesh, Unsteady analysis, Recirculation zone, Jets and wakes.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3081
168 Impulse Response Shortening for Discrete Multitone Transceivers using Convex Optimization Approach

Authors: Ejaz Khan, Conor Heneghan

Abstract:

In this paper we propose a new criterion for solving the problem of channel shortening in multi-carrier systems. In a discrete multitone receiver, a time-domain equalizer (TEQ) reduces intersymbol interference (ISI) by shortening the effective duration of the channel impulse response. Minimum mean square error (MMSE) method for TEQ does not give satisfactory results. In [1] a new criterion for partially equalizing severe ISI channels to reduce the cyclic prefix overhead of the discrete multitone transceiver (DMT), assuming a fixed transmission bandwidth, is introduced. Due to specific constrained (unit morm constraint on the target impulse response (TIR)) in their method, the freedom to choose optimum vector (TIR) is reduced. Better results can be obtained by avoiding the unit norm constraint on the target impulse response (TIR). In this paper we change the cost function proposed in [1] to the cost function of determining the maximum of a determinant subject to linear matrix inequality (LMI) and quadratic constraint and solve the resulting optimization problem. Usefulness of the proposed method is shown with the help of simulations.

Keywords: Equalizer, target impulse response, convex optimization, matrix inequality.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1712
167 A Fast Neural Algorithm for Serial Code Detection in a Stream of Sequential Data

Authors: Hazem M. El-Bakry, Qiangfu Zhao

Abstract:

In recent years, fast neural networks for object/face detection have been introduced based on cross correlation in the frequency domain between the input matrix and the hidden weights of neural networks. In our previous papers [3,4], fast neural networks for certain code detection was introduced. It was proved in [10] that for fast neural networks to give the same correct results as conventional neural networks, both the weights of neural networks and the input matrix must be symmetric. This condition made those fast neural networks slower than conventional neural networks. Another symmetric form for the input matrix was introduced in [1-9] to speed up the operation of these fast neural networks. Here, corrections for the cross correlation equations (given in [13,15,16]) to compensate for the symmetry condition are presented. After these corrections, it is proved mathematically that the number of computation steps required for fast neural networks is less than that needed by classical neural networks. Furthermore, there is no need for converting the input data into symmetric form. Moreover, such new idea is applied to increase the speed of neural networks in case of processing complex values. Simulation results after these corrections using MATLAB confirm the theoretical computations.

Keywords: Fast Code/Data Detection, Neural Networks, Cross Correlation, real/complex values.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1626
166 One-Dimensional Performance Improvement of a Single-Stage Transonic Compressor

Authors: A. Shahsavari, M. Nili-Ahmadabadi

Abstract:

This paper presents an innovative one-dimensional optimization of a transonic compressor based on the radial equilibrium theory by means of increasing blade loading. Firstly, the rotor blade of the transonic compressor is redesigned based on the constant span-wise deHaller number and diffusion. The code is applied to extract compressor meridional plane and blade to blade geometry containing rotor and stator in order to design blade three-dimensional view. A structured grid is generated for the numerical domain of fluid. Finer grids are used for regions near walls to capture boundary layer effects and behavior. RANS equations are solved by finite volume method for rotating zones (rotor) and stationary zones (stator). The experimental data, available for the performance map of NASA Rotor67, is used to validate the results of simulations. Then, the capability of the design method is validated by CFD that is capable of predicting the performance map. The numerical results of new geometry show about 19% increase in pressure ratio and 11% improvement in overall efficiency of the transonic stage; however, the design point mass flow rate of the new compressor is 5.7% less than that of the original compressor.

Keywords: One dimensional design, deHaller number, radial equilibrium, transonic compressor.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1048
165 Design and Implementation of Medium Access Control Based Routing on Real Wireless Sensor Networks Testbed

Authors: Smriti Agarwal, Ashish Payal, B. V. R. Reddy

Abstract:

IEEE 802.15.4 is a Low Rate Wireless Personal Area Networks (LR-WPAN) standard combined with ZigBee, which is going to enable new applications in Wireless Sensor Networks (WSNs) and Internet of Things (IoT) domain. In recent years, it has become a popular standard for WSNs. Wireless communication among sensor motes, enabled by IEEE 802.15.4 standard, is extensively replacing the existing wired technology in a wide range of monitoring and control applications. Researchers have proposed a routing framework and mechanism that interacts with the IEEE 802.15.4 standard using software platform. In this paper, we have designed and implemented MAC based routing (MBR) based on IEEE 802.15.4 standard using a hardware platform “SENSEnuts”. The experimental results include data through light and temperature sensors obtained from communication between PAN coordinator and source node through coordinator, MAC address of some modules used in the experimental setup, topology of the network created for simulation and the remaining battery power of the source node. Our experimental effort on a WSN Testbed has helped us in bridging the gap between theoretical and practical aspect of implementing IEEE 802.15.4 for WSNs applications.

Keywords: IEEE 802.15.4, routing, wireless sensor networks, ZigBee.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1581
164 Design Guidelines for an Enhanced Interaction Experience in the Domain of Smartphone-Based Applications for Sport and Fitness

Authors: Paolo Pilloni, Fabrizio Mulas, Salvatore Carta

Abstract:

Nowadays, several research studies point up that an active lifestyle is essential for physical and mental health benefits. Mobile phones have greatly influenced people’s habits and attitudes also in the way they exercise. Our research work is mainly focused on investigating how to exploit mobile technologies to favour people’s exertion experience. To this end, we developed an exertion framework users can exploit through a real world mobile application, called EverywhereSport Run (EWRun), designed to act as a virtual personal trainer to support runners during their trainings. In this work, inspired by both previous findings in the field of interaction design for people with visual impairments, feedback gathered from real users of our framework, and positive results obtained from two experimentations, we present some new interaction facilities we designed to enhance the interaction experience during a training. The positive obtained results helped us to derive some interaction design recommendations we believe will be a valid support for designers of future mobile systems conceived to be used in circumstances where there are limited possibilities of interaction.

Keywords: Human Computer Interaction, Interaction Design Guidelines, Persuasive Mobile Technologies for Sport and Health.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1946
163 Predicting the Impact of the Defect on the Overall Environment in Function Based Systems

Authors: Parvinder S. Sandhu, Urvashi Malhotra, E. Ardil

Abstract:

There is lot of work done in prediction of the fault proneness of the software systems. But, it is the severity of the faults that is more important than number of faults existing in the developed system as the major faults matters most for a developer and those major faults needs immediate attention. In this paper, we tried to predict the level of impact of the existing faults in software systems. Neuro-Fuzzy based predictor models is applied NASA-s public domain defect dataset coded in C programming language. As Correlation-based Feature Selection (CFS) evaluates the worth of a subset of attributes by considering the individual predictive ability of each feature along with the degree of redundancy between them. So, CFS is used for the selecting the best metrics that have highly correlated with level of severity of faults. The results are compared with the prediction results of Logistic Models (LMT) that was earlier quoted as the best technique in [17]. The results are recorded in terms of Accuracy, Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE). The results show that Neuro-fuzzy based model provide a relatively better prediction accuracy as compared to other models and hence, can be used for the modeling of the level of impact of faults in function based systems.

Keywords: Software Metrics, Fuzzy, Neuro-Fuzzy, Software Faults, Accuracy, MAE, RMSE.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1356
162 Actionable Rules: Issues and New Directions

Authors: Harleen Kaur

Abstract:

Knowledge Discovery in Databases (KDD) is the process of extracting previously unknown, hidden and interesting patterns from a huge amount of data stored in databases. Data mining is a stage of the KDD process that aims at selecting and applying a particular data mining algorithm to extract an interesting and useful knowledge. It is highly expected that data mining methods will find interesting patterns according to some measures, from databases. It is of vital importance to define good measures of interestingness that would allow the system to discover only the useful patterns. Measures of interestingness are divided into objective and subjective measures. Objective measures are those that depend only on the structure of a pattern and which can be quantified by using statistical methods. While, subjective measures depend only on the subjectivity and understandability of the user who examine the patterns. These subjective measures are further divided into actionable, unexpected and novel. The key issues that faces data mining community is how to make actions on the basis of discovered knowledge. For a pattern to be actionable, the user subjectivity is captured by providing his/her background knowledge about domain. Here, we consider the actionability of the discovered knowledge as a measure of interestingness and raise important issues which need to be addressed to discover actionable knowledge.

Keywords: Data Mining Community, Knowledge Discovery inDatabases (KDD), Interestingness, Subjective Measures, Actionability.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1942
161 Semantic Preference across Research Articles: A Corpus-Based Study of Adjectives in English

Authors: Valdênia Carvalho e Almeida

Abstract:

The goal of the present study is to investigate the semantic preference of the most frequent adjectives in research articles through a corpus-based analysis of texts published in journals in Applied Linguistics (AL). The corpus used in this study contains texts published in the period from 2014 to 2018 in the three journals: Language Learning and Technology; English for Academic Purposes, and TESOL Quaterly, totaling more than one million words. A corpus-based analysis was carried out on the corpus to identify the most frequent adjectives that co-occurred in the three journals. By observing the concordance lines of the adjectives and analyzing the words they associated with, the semantic preferences of each adjective were determined. Later, the AL corpus analysis was compared to the investigation of the same adjectives in a corpus of Chemistry. This second part of the study aimed to identify possible differences and similarities between the two corpora in relation to the use of the adjectives in research articles from both areas. The results show that there are some preferences which seem to be closely related not only to the academic genre of the texts but also to the specific domain of the discipline and, to a lesser extent, to the context of research in each journal. This research illustrates a possible contribution of Corpus Linguistics to explore the concept of semantic preference in more detail, considering the complex nature of the phenomenon.

Keywords: Applied linguistics, corpus linguistics, chemistry, research article, semantic preference.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1364
160 RFU Based Computational Unit Design For Reconfigurable Processors

Authors: M. Aqeel Iqbal

Abstract:

Fully customized hardware based technology provides high performance and low power consumption by specializing the tasks in hardware but lacks design flexibility since any kind of changes require re-design and re-fabrication. Software based solutions operate with software instructions due to which a great flexibility is achieved from the easy development and maintenance of the software code. But this execution of instructions introduces a high overhead in performance and area consumption. In past few decades the reconfigurable computing domain has been introduced which overcomes the traditional trades-off between flexibility and performance and is able to achieve high performance while maintaining a good flexibility. The dramatic gains in terms of chip performance and design flexibility achieved through the reconfigurable computing systems are greatly dependent on the design of their computational units being integrated with reconfigurable logic resources. The computational unit of any reconfigurable system plays vital role in defining its strength. In this research paper an RFU based computational unit design has been presented using the tightly coupled, multi-threaded reconfigurable cores. The proposed design has been simulated for VLIW based architectures and a high gain in performance has been observed as compared to the conventional computing systems.

Keywords: Configuration Stream, Configuration overhead, Configuration Controller, Reconfigurable devices.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1621
159 Detection of Cyberattacks on the Metaverse Based on First-Order Logic

Authors: Sulaiman Al Amro

Abstract:

There are currently considerable challenges concerning data security and privacy, particularly in relation to modern technologies. This includes the virtual world known as the Metaverse, which consists of a virtual space that integrates various technologies, and therefore susceptible to cyber threats such as malware, phishing, and identity theft. This has led recent studies to propose the development of Metaverse forensic frameworks and the integration of advanced technologies, including machine learning for intrusion detection and security. In this context, the application of first-order logic offers a formal and systematic approach to defining the conditions of cyberattacks, thereby contributing to the development of effective detection mechanisms. In addition, formalizing the rules and patterns of cyber threats has the potential to enhance the overall security posture of the Metaverse and thus the integrity and safety of this virtual environment. The current paper focuses on the primary actions employed by avatars for potential attacks, including Interval Temporal Logic (ITL) and behavior-based detection to detect an avatar’s abnormal activities within the Metaverse. The research established that the proposed framework attained an accuracy of 92.307%, resulting in the experimental results demonstrating the efficacy of ITL, including its superior performance in addressing the threats posed by avatars within the Metaverse domain.

Keywords: Cyberattacks, detection, first-order logic, Metaverse, privacy, security.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 67
158 Study on Optimization Design of Pressure Hull for Underwater Vehicle

Authors: Qasim Idrees, Gao Liangtian, Liu Bo, Miao Yiran

Abstract:

In order to improve the efficiency and accuracy of the pressure hull structure, optimization of underwater vehicle based on response surface methodology, a method for optimizing the design of pressure hull structure was studied. To determine the pressure shell of five dimensions as a design variable, the application of thin shell theory and the Chinese Classification Society (CCS) specification was carried on the preliminary design. In order to optimize variables of the feasible region, different methods were studied and implemented such as Opt LHD method (to determine the design test sample points in the feasible domain space), parametric ABAQUS solution for each sample point response, and the two-order polynomial response for the surface model of the limit load of structures. Based on the ultimate load of the structure and the quality of the shell, the two-generation genetic algorithm was used to solve the response surface, and the Pareto optimal solution set was obtained. The final optimization result was 41.68% higher than that of the initial design, and the shell quality was reduced by about 27.26%. The parametric method can ensure the accuracy of the test and improve the efficiency of optimization.

Keywords: Parameterization, response surface, structure optimization, pressure hull.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1162
157 Extraction of Fetal Heart Rate and Fetal Heart Rate Variability from Mother's ECG Signal

Authors: Khaldon Lweesy, Luay Fraiwan, Christoph Maier, Hartmut Dickhaus

Abstract:

This paper describes a new method for extracting the fetal heart rate (fHR) and the fetal heart rate variability (fHRV) signal non-invasively using abdominal maternal electrocardiogram (mECG) recordings. The extraction is based on the fundamental frequency (Fourier-s) theorem. The fundamental frequency of the mother-s electrocardiogram signal (fo-m) is calculated directly from the abdominal signal. The heart rate of the fetus is usually higher than that of the mother; as a result, the fundamental frequency of the fetal-s electrocardiogram signal (fo-f) is higher than that of the mother-s (fo-f > fo-m). Notch filters to suppress mother-s higher harmonics were designed; then a bandpass filter to target fo-f and reject fo-m is implemented. Although the bandpass filter will pass some other frequencies (harmonics), we have shown in this study that those harmonics are actually carried on fo-f, and thus have no impact on the evaluation of the beat-to-beat changes (RR intervals). The oscillations of the time-domain extracted signal represent the RR intervals. We have also shown in this study that zero-to-zero evaluation of the periods is more accurate than the peak-to-peak evaluation. This method is evaluated both on simulated signals and on different abdominal recordings obtained at different gestational ages.

Keywords: Aabdominal ECG, fetal heart rate variability, frequency harmonics, fundamental frequency.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2673
156 A Novel Framework for User-Friendly Ontology-Mediated Access to Relational Databases

Authors: Efthymios Chondrogiannis, Vassiliki Andronikou, Efstathios Karanastasis, Theodora Varvarigou

Abstract:

A large amount of data is typically stored in relational databases (DB). The latter can efficiently handle user queries which intend to elicit the appropriate information from data sources. However, direct access and use of this data requires the end users to have an adequate technical background, while they should also cope with the internal data structure and values presented. Consequently the information retrieval is a quite difficult process even for IT or DB experts, taking into account the limited contributions of relational databases from the conceptual point of view. Ontologies enable users to formally describe a domain of knowledge in terms of concepts and relations among them and hence they can be used for unambiguously specifying the information captured by the relational database. However, accessing information residing in a database using ontologies is feasible, provided that the users are keen on using semantic web technologies. For enabling users form different disciplines to retrieve the appropriate data, the design of a Graphical User Interface is necessary. In this work, we will present an interactive, ontology-based, semantically enable web tool that can be used for information retrieval purposes. The tool is totally based on the ontological representation of underlying database schema while it provides a user friendly environment through which the users can graphically form and execute their queries.

Keywords: Ontologies, Relational Databases, SPARQL, Web Interface.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1930
155 Software Maintenance Severity Prediction for Object Oriented Systems

Authors: Parvinder S. Sandhu, Roma Jaswal, Sandeep Khimta, Shailendra Singh

Abstract:

As the majority of faults are found in a few of its modules so there is a need to investigate the modules that are affected severely as compared to other modules and proper maintenance need to be done in time especially for the critical applications. As, Neural networks, which have been already applied in software engineering applications to build reliability growth models predict the gross change or reusability metrics. Neural networks are non-linear sophisticated modeling techniques that are able to model complex functions. Neural network techniques are used when exact nature of input and outputs is not known. A key feature is that they learn the relationship between input and output through training. In this present work, various Neural Network Based techniques are explored and comparative analysis is performed for the prediction of level of need of maintenance by predicting level severity of faults present in NASA-s public domain defect dataset. The comparison of different algorithms is made on the basis of Mean Absolute Error, Root Mean Square Error and Accuracy Values. It is concluded that Generalized Regression Networks is the best algorithm for classification of the software components into different level of severity of impact of the faults. The algorithm can be used to develop model that can be used for identifying modules that are heavily affected by the faults.

Keywords: Neural Network, Software faults, Software Metric.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1575
154 Using HMM-based Classifier Adapted to Background Noises with Improved Sounds Features for Audio Surveillance Application

Authors: Asma Rabaoui, Zied Lachiri, Noureddine Ellouze

Abstract:

Discrimination between different classes of environmental sounds is the goal of our work. The use of a sound recognition system can offer concrete potentialities for surveillance and security applications. The first paper contribution to this research field is represented by a thorough investigation of the applicability of state-of-the-art audio features in the domain of environmental sound recognition. Additionally, a set of novel features obtained by combining the basic parameters is introduced. The quality of the features investigated is evaluated by a HMM-based classifier to which a great interest was done. In fact, we propose to use a Multi-Style training system based on HMMs: one recognizer is trained on a database including different levels of background noises and is used as a universal recognizer for every environment. In order to enhance the system robustness by reducing the environmental variability, we explore different adaptation algorithms including Maximum Likelihood Linear Regression (MLLR), Maximum A Posteriori (MAP) and the MAP/MLLR algorithm that combines MAP and MLLR. Experimental evaluation shows that a rather good recognition rate can be reached, even under important noise degradation conditions when the system is fed by the convenient set of features.

Keywords: Sounds recognition, HMM classifier, Multi-style training, Environmental Adaptation, Feature combinations.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1644
153 Application of Machine Learning Methods to Online Test Error Detection in Semiconductor Test

Authors: Matthias Kirmse, Uwe Petersohn, Elief Paffrath

Abstract:

As in today's semiconductor industries test costs can make up to 50 percent of the total production costs, an efficient test error detection becomes more and more important. In this paper, we present a new machine learning approach to test error detection that should provide a faster recognition of test system faults as well as an improved test error recall. The key idea is to learn a classifier ensemble, detecting typical test error patterns in wafer test results immediately after finishing these tests. Since test error detection has not yet been discussed in the machine learning community, we define central problem-relevant terms and provide an analysis of important domain properties. Finally, we present comparative studies reflecting the failure detection performance of three individual classifiers and three ensemble methods based upon them. As base classifiers we chose a decision tree learner, a support vector machine and a Bayesian network, while the compared ensemble methods were simple and weighted majority vote as well as stacking. For the evaluation, we used cross validation and a specially designed practical simulation. By implementing our approach in a semiconductor test department for the observation of two products, we proofed its practical applicability.

Keywords: Ensemble methods, fault detection, machine learning, semiconductor test.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2274
152 Analysis of Thermal Damping in Si Based Torsional Micromirrors

Authors: R. Resmi, M. R. Baiju

Abstract:

The thermal damping of a dynamic vibrating micromirror is an important factor affecting the design of MEMS based actuator systems. In the development process of new micromirror systems, assessing the extent of energy loss due to thermal damping accurately and predicting the performance of the system is very essential. In this paper, the depth of the thermal penetration layer at different eigenfrequencies and the temperature variation distributions surrounding a vibrating micromirror is analyzed. The thermal penetration depth corresponds to the thermal boundary layer in which energy is lost which is a measure of the thermal damping is found out. The energy is mainly dissipated in the thermal boundary layer and thickness of the layer is an important parameter. The detailed thermoacoustics is used to model the air domain surrounding the micromirror. The thickness of the boundary layer, temperature variations and thermal power dissipation are analyzed for a Si based torsional mode micromirror. It is found that thermal penetration depth decreases with eigenfrequency and hence operating the micromirror at higher frequencies is essential for reducing thermal damping. The temperature variations and thermal power dissipations at different eigenfrequencies are also analyzed. Both frequency-response and eigenfrequency analyses are done using COMSOL Multiphysics software.

Keywords: Eigen frequency analysis, micromirrors, thermal damping, thermoacoustic interactions.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1060
151 Contamination in Industrial Areas and Environmental Management in Latvia

Authors: Juris Burlakovs, Maris Klavins, Raimonds Ernsteins, Armands Ruskulis

Abstract:

Environmental contamination is a common problem in ex-industrial and industrial sites. This article gives a brief description of general applied environmental investigation methodologies and possible remediation applications in Latvia. Most of contaminated areas are situated in former and active industrial, military areas and ports. Industrial and logistic activities very often have been with great impact for more than hundred years thus the contamination level with heavy metals, hydrocarbons, pesticides, persistent organic pollutants is high and is threatening health and environment in general. 242 territories now are numbered as contaminated and fixed in the National Register of contaminated territories in Latvia. Research and remediation of contamination in densely populated areas are of important environmental policy domain. Four different investigation case studies of contaminated areas are given describing the history of use, environmental quality assessment as well as planned environmental management actions. All four case study locations are situated in Riga - the capital of the Republic of Latvia. The aim of this paper is to analyze the situation and problems with management of contaminated areas in Latvia, give description of field research methods and recommendations for remediation industry based on scientific data and innovations.

Keywords: Remediation technology, environmental quality assessment, heavy metals, hydrocarbon contamination, environmental management.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2301