Search results for: input%20shaping
863 Finding the Longest Common Subsequence in Normal DNA and Disease Affected Human DNA Using Self Organizing Map
Authors: G. Tamilpavai, C. Vishnuppriya
Abstract:
Bioinformatics is an active research area which combines biological matter as well as computer science research. The longest common subsequence (LCSS) is one of the major challenges in various bioinformatics applications. The computation of the LCSS plays a vital role in biomedicine and also it is an essential task in DNA sequence analysis in genetics. It includes wide range of disease diagnosing steps. The objective of this proposed system is to find the longest common subsequence which presents in a normal and various disease affected human DNA sequence using Self Organizing Map (SOM) and LCSS. The human DNA sequence is collected from National Center for Biotechnology Information (NCBI) database. Initially, the human DNA sequence is separated as k-mer using k-mer separation rule. Mean and median values are calculated from each separated k-mer. These calculated values are fed as input to the Self Organizing Map for the purpose of clustering. Then obtained clusters are given to the Longest Common Sub Sequence (LCSS) algorithm for finding common subsequence which presents in every clusters. It returns nx(n-1)/2 subsequence for each cluster where n is number of k-mer in a specific cluster. Experimental outcomes of this proposed system produce the possible number of longest common subsequence of normal and disease affected DNA data. Thus the proposed system will be a good initiative aid for finding disease causing sequence. Finally, performance analysis is carried out for different DNA sequences. The obtained values show that the retrieval of LCSS is done in a shorter time than the existing system.Keywords: clustering, k-mers, longest common subsequence, SOM
Procedia PDF Downloads 265862 Impact of Integrated Signals for Doing Human Activity Recognition Using Deep Learning Models
Authors: Milagros Jaén-Vargas, Javier García Martínez, Karla Miriam Reyes Leiva, María Fernanda Trujillo-Guerrero, Francisco Fernandes, Sérgio Barroso Gonçalves, Miguel Tavares Silva, Daniel Simões Lopes, José Javier Serrano Olmedo
Abstract:
Human Activity Recognition (HAR) is having a growing impact in creating new applications and is responsible for emerging new technologies. Also, the use of wearable sensors is an important key to exploring the human body's behavior when performing activities. Hence, the use of these dispositive is less invasive and the person is more comfortable. In this study, a database that includes three activities is used. The activities were acquired from inertial measurement unit sensors (IMU) and motion capture systems (MOCAP). The main objective is differentiating the performance from four Deep Learning (DL) models: Deep Neural Network (DNN), Convolutional Neural Network (CNN), Recurrent Neural Network (RNN) and hybrid model Convolutional Neural Network-Long Short-Term Memory (CNN-LSTM), when considering acceleration, velocity and position and evaluate if integrating the IMU acceleration to obtain velocity and position represent an increment in performance when it works as input to the DL models. Moreover, compared with the same type of data provided by the MOCAP system. Despite the acceleration data is cleaned when integrating, results show a minimal increase in accuracy for the integrated signals.Keywords: HAR, IMU, MOCAP, acceleration, velocity, position, feature maps
Procedia PDF Downloads 96861 XAI Implemented Prognostic Framework: Condition Monitoring and Alert System Based on RUL and Sensory Data
Authors: Faruk Ozdemir, Roy Kalawsky, Peter Hubbard
Abstract:
Accurate estimation of RUL provides a basis for effective predictive maintenance, reducing unexpected downtime for industrial equipment. However, while models such as the Random Forest have effective predictive capabilities, they are the so-called ‘black box’ models, where interpretability is at a threshold to make critical diagnostic decisions involved in industries related to aviation. The purpose of this work is to present a prognostic framework that embeds Explainable Artificial Intelligence (XAI) techniques in order to provide essential transparency in Machine Learning methods' decision-making mechanisms based on sensor data, with the objective of procuring actionable insights for the aviation industry. Sensor readings have been gathered from critical equipment such as turbofan jet engine and landing gear, and the prediction of the RUL is done by a Random Forest model. It involves steps such as data gathering, feature engineering, model training, and evaluation. These critical components’ datasets are independently trained and evaluated by the models. While suitable predictions are served, their performance metrics are reasonably good; such complex models, however obscure reasoning for the predictions made by them and may even undermine the confidence of the decision-maker or the maintenance teams. This is followed by global explanations using SHAP and local explanations using LIME in the second phase to bridge the gap in reliability within industrial contexts. These tools analyze model decisions, highlighting feature importance and explaining how each input variable affects the output. This dual approach offers a general comprehension of the overall model behavior and detailed insight into specific predictions. The proposed framework, in its third component, incorporates the techniques of causal analysis in the form of Granger causality tests in order to move beyond correlation toward causation. This will not only allow the model to predict failures but also present reasons, from the key sensor features linked to possible failure mechanisms to relevant personnel. The causality between sensor behaviors and equipment failures creates much value for maintenance teams due to better root cause identification and effective preventive measures. This step contributes to the system being more explainable. Surrogate Several simple models, including Decision Trees and Linear Models, can be used in yet another stage to approximately represent the complex Random Forest model. These simpler models act as backups, replicating important jobs of the original model's behavior. If the feature explanations obtained from the surrogate model are cross-validated with the primary model, the insights derived would be more reliable and provide an intuitive sense of how the input variables affect the predictions. We then create an iterative explainable feedback loop, where the knowledge learned from the explainability methods feeds back into the training of the models. This feeds into a cycle of continuous improvement both in model accuracy and interpretability over time. By systematically integrating new findings, the model is expected to adapt to changed conditions and further develop its prognosis capability. These components are then presented to the decision-makers through the development of a fully transparent condition monitoring and alert system. The system provides a holistic tool for maintenance operations by leveraging RUL predictions, feature importance scores, persistent sensor threshold values, and autonomous alert mechanisms. Since the system will provide explanations for the predictions given, along with active alerts, the maintenance personnel can make informed decisions on their end regarding correct interventions to extend the life of the critical machinery.Keywords: predictive maintenance, explainable artificial intelligence, prognostic, RUL, machine learning, turbofan engines, C-MAPSS dataset
Procedia PDF Downloads 4860 The Correlation between Political Awareness and Political Participation for University Students’ “Applied Study”
Authors: Rana Mohamed
Abstract:
Despite youth in Egypt were away from political life for a long time; they are able to make a tangible difference in political status. Purpose: This exploratory study aims to determine whether and how much the prevailing political culture influence participatory behavior with a special focus on political awareness factors among university students in Egypt. Methodology: The study employed several data collection methods to ensure the validity of the results, quantitative and qualitative, verifying the positive relationships between the levels of political awareness and political participation and between political values in society and the level of political participation among university students. For achieving the objectives of the paper in the light of the pool of available literature and data, the study adopts system analysis method to apply input-output and conversions associated with the phenomena of political participation to analyze the different factors that have an effect upon the prevailing political culture and the patterns of values in Egyptian society. Findings: The result reveals that the level of political awareness and political participation for students were low, with a statistically significant relationship. In addition, the patterns of values in Egyptian culture significantly influence the levels of student participation. Therefore, the study recommends formulating policies that aim to increase awareness levels and integrate youth into the political process. Originality/Value: The importance of the academic study stems from addressing one of the central issues in political science; this study measures the change in the Egyptian patterns of culture and values among university students.Keywords: political awareness, political participation, civic culture, citizenship, Egyptian universities, political knowledge
Procedia PDF Downloads 249859 Fintech Credit and Bank Efficiency Two-way Relationship: A Comparison Study Across Country Groupings
Authors: Tan Swee Liang
Abstract:
This paper studies the two-way relationship between fintech credit and banking efficiency using the Generalized panel Method of Moment (GMM) estimation in structural equation modeling (SEM). Banking system efficiency, defined as its ability to produce the existing level of outputs with minimal inputs, is measured using input-oriented data envelopment analysis (DEA), where the whole banking system of an economy is treated as a single DMU. Banks are considered an intermediary between depositors and borrowers, utilizing inputs (deposits and overhead costs) to provide outputs (increase credits to the private sector and its earnings). Analysis of the interrelationship between fintech credit and bank efficiency is conducted to determine the impact in different country groupings (ASEAN, Asia and OECD), in particular the banking system response to fintech credit platforms. Our preliminary results show that banks do respond to the greater pressure caused by fintech platforms to enhance their efficiency, but differently across the different groups. The author’s earlier research on ASEAN-5 high bank overhead costs (as a share of total assets) as the determinant of economic growth suggests that expenses may not have been channeled efficiently to income-generating activities. One practical implication of the findings is that policymakers should enable alternative financing, such as fintech credit, as a warning or encouragement for banks to improve their efficiency.Keywords: fintech lending, banking efficiency, data envelopment analysis, structural equation modeling
Procedia PDF Downloads 90858 Delineation of Oil – Polluted Sites in Ibeno LGA, Nigeria, Using Microbiological and Physicochemical Characterization
Authors: Ime R. Udotong, Justina I. R. Udotong, Ofonime U. M. John
Abstract:
Mobil Producing Nigeria Unlimited (MPNU), a subsidiary of ExxonMobil and the highest crude oil & condensate producer in Nigeria has its operational base and an oil terminal, the Qua Iboe terminal (QIT) located at Ibeno, Nigeria. Other oil companies like Network Exploration and Production Nigeria Ltd, Frontier Oil Ltd; Shell Petroleum Development Company Ltd; Elf Petroleum Nigeria Ltd and Nigerian Agip Energy, a subsidiary of the Italian ENI E&P operate onshore, on the continental shelf and in deep offshore of the Atlantic Ocean, respectively with the coastal waters of Ibeno, Nigeria as the nearest shoreline. This study was designed to delineate the oil-polluted sites in Ibeno, Nigeria using microbiological and physico-chemical characterization of soils, sediments and ground and surface water samples from the study area. Results obtained revealed that there have been significant recent hydrocarbon inputs into this environment as observed from the high counts of hydrocarbonoclastic microorganisms in excess of 1% at all the stations sampled. Moreover, high concentrations of THC, BTEX and heavy metals contents in all the samples analyzed corroborate the high recent crude oil input into the study area. The results also showed that the pollution of the different environmental media sampled were of varying degrees, following the trend: Ground water > surface water > sediments > soils.Keywords: microbiological characterization, oil-polluted sites, physico-chemical analyses, total hydrocarbon content
Procedia PDF Downloads 415857 Neural Network Models for Actual Cost and Actual Duration Estimation in Construction Projects: Findings from Greece
Authors: Panagiotis Karadimos, Leonidas Anthopoulos
Abstract:
Predicting the actual cost and duration in construction projects concern a continuous and existing problem for the construction sector. This paper addresses this problem with modern methods and data available from past public construction projects. 39 bridge projects, constructed in Greece, with a similar type of available data were examined. Considering each project’s attributes with the actual cost and the actual duration, correlation analysis is performed and the most appropriate predictive project variables are defined. Additionally, the most efficient subgroup of variables is selected with the use of the WEKA application, through its attribute selection function. The selected variables are used as input neurons for neural network models through correlation analysis. For constructing neural network models, the application FANN Tool is used. The optimum neural network model, for predicting the actual cost, produced a mean squared error with a value of 3.84886e-05 and it was based on the budgeted cost and the quantity of deck concrete. The optimum neural network model, for predicting the actual duration, produced a mean squared error with a value of 5.89463e-05 and it also was based on the budgeted cost and the amount of deck concrete.Keywords: actual cost and duration, attribute selection, bridge construction, neural networks, predicting models, FANN TOOL, WEKA
Procedia PDF Downloads 134856 Identifying Dynamic Structural Parameters of Soil-Structure System Based on Data Recorded during Strong Earthquakes
Authors: Vahidreza Mahmoudabadi, Omid Bahar, Mohammad Kazem Jafari
Abstract:
In many applied engineering problems, structural analysis is usually conducted by assuming a rigid bed, while imposing the effect of structure bed flexibility can affect significantly on the structure response. This article focuses on investigation and evaluation of the effects arising from considering a soil-structure system in evaluation of dynamic characteristics of a steel structure with respect to elastic and inelastic behaviors. The recorded structure acceleration during Taiwan’s strong Chi-Chi earthquake on different floors of the structure was our evaluation criteria. The respective structure is an eight-story steel bending frame structure designed using a displacement-based direct method assuring weak beam - strong column function. The results indicated that different identification methods i.e. reverse Fourier transform or transfer functions, is capable to determine some of the dynamic parameters of the structure precisely, rather than evaluating all of them at once (mode frequencies, mode shapes, structure damping, structure rigidity, etc.). Response evaluation based on the input and output data elucidated that the structure first mode is not significantly affected, even considering the soil-structure interaction effect, but the upper modes have been changed. Also, it was found that the response transfer function of the different stories, in which plastic hinges have occurred in the structure components, provides similar results.Keywords: bending steel frame structure, dynamic characteristics, displacement-based design, soil-structure system, system identification
Procedia PDF Downloads 502855 Design and Manufacture of an Autonomous Agricultural Robot for Pesticide Application
Authors: Caner Koc, Dilara Gerdan Koc, Emrah Saka, H. Ibrahim Karagol
Abstract:
The use of pesticides in agricultural activities is the most harmful to the environment and farmers' health, and it also has the greatest input prices, along with fertilizers. In this study, an electric, electrostatically charged, autonomous agricultural robot was developed, modeled, and prototyped and manufactured. It allows for sensitive pesticide applications with variable levels, has controllable spray nozzles, and uses camera distance sensors to detect and spray into tree canopies. The created prototype was produced with flexibility in mind. Two stages of prototype manufacture were completed. The initial stage involved designing and producing the flexible primary body of the autonomous vehicle. Detachable hanger assemblies are employed so that the main body robot can perform a variety of agricultural tasks. The design of the spraying devices and their fitting to the autonomous vehicle was completed as the second stage of the prototype. The built prototype spraying robot's itinerary was planned using the free, open-source program Mission Planner. PX4, telemetry, and RTK GPS are used to maneuver the autonomous car along the designated path. To avoid potential obstructions, the robot uses ultrasonic and lidar sensors. The developed autonomous vehicle's energy needs are intended to be met entirely by electric batteries. In the event that the batteries run out of power, the sockets are set up to be recharged both by using the generator and the main power source through the specifically constructed panel.Keywords: autonomous agricultural robot, pesticide, smart farming, spraying, variable rate application
Procedia PDF Downloads 82854 Economic Analysis of Cowpea (Unguiculata spp) Production in Northern Nigeria: A Case Study of Kano Katsina and Jigawa States
Authors: Yakubu Suleiman, S. A. Musa
Abstract:
Nigeria is the largest cowpea producer in the world, accounting for about 45%, followed by Brazil with about 17%. Cowpea is grown in Kano, Bauchi, Katsina, Borno in the north, Oyo in the west, and to the lesser extent in Enugu in the east. This study was conducted to determine the input–output relationship of Cowpea production in Kano, Katsina, and Jigawa states of Nigeria. The data were collected with the aid of 1000 structured questionnaires that were randomly distributed to Cowpea farmers in the three states mentioned above of the study area. The data collected were analyzed using regression analysis (Cobb–Douglass production function model). The result of the regression analysis revealed the coefficient of multiple determinations, R2, to be 72.5% and the F ration to be 106.20 and was found to be significant (P < 0.01). The regression coefficient of constant is 0.5382 and is significant (P < 0.01). The regression coefficient with respect to labor and seeds were 0.65554 and 0.4336, respectively, and they are highly significant (P < 0.01). The regression coefficient with respect to fertilizer is 0.26341 which is significant (P < 0.05). This implies that a unit increase of any one of the variable inputs used while holding all other variables inputs constants, will significantly increase the total Cowpea output by their corresponding coefficient. This indicated that farmers in the study area are operating in stage II of the production function. The result revealed that Cowpea farmer in Kano, Jigawa and Katsina States realized a profit of N15,997, N34,016 and N19,788 per hectare respectively. It is hereby recommended that more attention should be given to Cowpea production by government and research institutions.Keywords: coefficient, constant, inputs, regression
Procedia PDF Downloads 408853 Web Service Architectural Style Selection in Multi-Criteria Requirements
Authors: Ahmad Mohsin, Syda Fatima, Falak Nawaz, Aman Ullah Khan
Abstract:
Selection of an appropriate architectural style is vital to the success of target web service under development. The nature of architecture design and selection for service-oriented computing applications is quite different as compared to traditional software. Web Services have complex and rigorous architectural styles to choose. Due to this, selection for accurate architectural style for web services development has become a more complex decision to be made by architects. Architectural style selection is a multi-criteria decision and demands lots of experience in service oriented computing. Decision support systems are good solutions to simplify the selection process of a particular architectural style. Our research suggests a new approach using DSS for selection of architectural styles while developing a web service to cater FRs and NFRs. Our proposed DSS helps architects to select right web service architectural pattern according to the domain and non-functional requirements. In this paper, a rule base DSS has been developed using CLIPS (C Language Integrated Production System) to support decisions using multi-criteria requirements. This DSS takes architectural characteristics, domain requirements and software architect preferences for NFRs as input for different architectural styles in use today in service-oriented computing. Weighted sum model has been applied to prioritize quality attributes and domain requirements. Scores are calculated using multiple criterions to choose the final architecture style.Keywords: software architecture, web-service, rule-based, DSS, multi-criteria requirements, quality attributes
Procedia PDF Downloads 361852 Optimized Real Ground Motion Scaling for Vulnerability Assessment of Building Considering the Spectral Uncertainty and Shape
Authors: Chen Bo, Wen Zengping
Abstract:
Based on the results of previous studies, we focus on the research of real ground motion selection and scaling method for structural performance-based seismic evaluation using nonlinear dynamic analysis. The input of earthquake ground motion should be determined appropriately to make them compatible with the site-specific hazard level considered. Thus, an optimized selection and scaling method are established including the use of not only Monte Carlo simulation method to create the stochastic simulation spectrum considering the multivariate lognormal distribution of target spectrum, but also a spectral shape parameter. Its applications in structural fragility analysis are demonstrated through case studies. Compared to the previous scheme with no consideration of the uncertainty of target spectrum, the method shown here can make sure that the selected records are in good agreement with the median value, standard deviation and spectral correction of the target spectrum, and greatly reveal the uncertainty feature of site-specific hazard level. Meanwhile, it can help improve computational efficiency and matching accuracy. Given the important infection of target spectrum’s uncertainty on structural seismic fragility analysis, this work can provide the reasonable and reliable basis for structural seismic evaluation under scenario earthquake environment.Keywords: ground motion selection, scaling method, seismic fragility analysis, spectral shape
Procedia PDF Downloads 291851 An Efficient Hybrid Feedstock Pretreatment Technique for the Release of Fermentable Sugar from Cassava Peels for Biofuel Production
Authors: Gabriel Sanjo Aruwajoye, E. B. Gueguim Kana
Abstract:
Agricultural residues present a low-cost feedstock for bioenergy production around the world. Cassava peels waste are rich in organic molecules that can be readily converted to value added products such as biomaterials and biofuels. However, due to the presence of high proportion of structural carbohydrates and lignin, the hydrolysis of this feedstock is imperative to achieve maximum substrate utilization and energy yield. This study model and optimises the release of Fermentable Sugar (FS) from cassava peels waste using the Response Surface Methodology. The investigated pretreatment input parameters consisted of soaking temperature (oC), soaking time (hours), autoclave duration (minutes), acid concentration (% v/v), substrate solid loading (% w/v) within the range of 30 to 70, 0 to 24, 5 to 20, 0 to 5 and 2 to 10 respectively. The Box-Behnken design was used to generate 46 experimental runs which were investigated for FS release. The obtained data were used to fit a quadratic model. A coefficient of determination of 0.87 and F value of 8.73 was obtained indicating the good fitness of the model. The predicted optimum pretreatment conditions were 69.62 oC soaking temperature, 2.57 hours soaking duration, 5 minutes autoclave duration, 3.68 % v/v HCl and 9.65 % w/v solid loading corresponding to FS yield of 91.83g/l (0.92 g/g cassava peels) thus 58% improvement on the non-optimised pretreatment. Our findings demonstrate an efficient pretreatment model for fermentable sugar release from cassava peels waste for various bioprocesses.Keywords: feedstock pretreatment, cassava peels, fermentable sugar, response surface methodology
Procedia PDF Downloads 365850 Infrared Thermography as an Informative Tool in Energy Audit and Software Modelling of Historic Buildings: A Case Study of the Sheffield Cathedral
Authors: Ademuyiwa Agbonyin, Stamatis Zoras, Mohammad Zandi
Abstract:
This paper investigates the extent to which building energy modelling can be informed based on preliminary information provided by infrared thermography using a thermal imaging camera in a walkthrough audit. The case-study building is the Sheffield Cathedral, built in the early 1400s. Based on an informative qualitative report generated from the thermal images taken at the site, the regions showing significant heat loss are input into a computer model of the cathedral within the integrated environmental solution (IES) virtual environment software which performs an energy simulation to determine quantitative heat losses through the building envelope. Building data such as material thermal properties and building plans are provided by the architects, Thomas Ford and Partners Ltd. The results of the modelling revealed the portions of the building with the highest heat loss and these aligned with those suggested by the thermal camera. Retrofit options for the building are also considered, however, may not see implementation due to a desire to conserve the architectural heritage of the building. Results show that thermal imaging in a walk-through audit serves as a useful guide for the energy modelling process. Hand calculations were also performed to serve as a 'control' to estimate losses, providing a second set of data points of comparison.Keywords: historic buildings, energy retrofit, thermal comfort, software modelling, energy modelling
Procedia PDF Downloads 169849 Theoretical Analysis of the Optical and Solid State Properties of Thin Film
Authors: E. I. Ugwu
Abstract:
Theoretical analysis of the optical and Solid State properties of ZnS thin film using beam propagation technique in which a scalar wave is propagated through the material thin film deposited on a substrate with the assumption that the dielectric medium is section into a homogenous reference dielectric constant term, and a perturbed dielectric term, representing the deposited thin film medium is presented in this work. These two terms, constitute arbitrary complex dielectric function that describes dielectric perturbation imposed by the medium of for the system. This is substituted into a defined scalar wave equation in which the appropriate Green’s Function was defined on it and solved using series technique. The green’s value obtained from Green’s Function was used in Dyson’s and Lippmann Schwinger equations in conjunction with Born approximation method in computing the propagated field for different input regions of field wavelength during which the influence of the dielectric constants and mesh size of the thin film on the propagating field were depicted. The results obtained from the computed field were used in turn to generate the data that were used to compute the band gaps, solid state and optical properties of the thin film such as reflectance, Transmittance and reflectance with which the band gap obtained was found to be in close approximate to that of experimental value.Keywords: scalar wave, optical and solid state properties, thin film, dielectric medium, perturbation, Lippmann Schwinger equations, Green’s Function, propagation
Procedia PDF Downloads 436848 Predicting the Compressive Strength of Geopolymer Concrete Using Machine Learning Algorithms: Impact of Chemical Composition and Curing Conditions
Authors: Aya Belal, Ahmed Maher Eltair, Maggie Ahmed Mashaly
Abstract:
Geopolymer concrete is gaining recognition as a sustainable alternative to conventional Portland Cement concrete due to its environmentally friendly nature, which is a key goal for Smart City initiatives. It has demonstrated its potential as a reliable material for the design of structural elements. However, the production of Geopolymer concrete is hindered by batch-to-batch variations, which presents a significant challenge to the widespread adoption of Geopolymer concrete. To date, Machine learning has had a profound impact on various fields by enabling models to learn from large datasets and predict outputs accurately. This paper proposes an integration between the current drift to Artificial Intelligence and the composition of Geopolymer mixtures to predict their mechanical properties. This study employs Python software to develop machine learning model in specific Decision Trees. The research uses the percentage oxides and the chemical composition of the Alkali Solution along with the curing conditions as the input independent parameters, irrespective of the waste products used in the mixture yielding the compressive strength of the mix as the output parameter. The results showed 90 % agreement of the predicted values to the actual values having the ratio of the Sodium Silicate to the Sodium Hydroxide solution being the dominant parameter in the mixture.Keywords: decision trees, geopolymer concrete, machine learning, smart cities, sustainability
Procedia PDF Downloads 85847 Client Hacked Server
Authors: Bagul Abhijeet
Abstract:
Background: Client-Server model is the backbone of today’s internet communication. In which normal user can not have control over particular website or server? By using the same processing model one can have unauthorized access to particular server. In this paper, we discussed about application scenario of hacking for simple website or server consist of unauthorized way to access the server database. This application emerges to autonomously take direct access of simple website or server and retrieve all essential information maintain by administrator. In this system, IP address of server given as input to retrieve user-id and password of server. This leads to breaking administrative security of server and acquires the control of server database. Whereas virus helps to escape from server security by crashing the whole server. Objective: To control malicious attack and preventing all government website, and also find out illegal work to do hackers activity. Results: After implementing different hacking as well as non-hacking techniques, this system hacks simple web sites with normal security credentials. It provides access to server database and allow attacker to perform database operations from client machine. Above Figure shows the experimental result of this application upon different servers and provides satisfactory results as required. Conclusion: In this paper, we have presented a to view to hack the server which include some hacking as well as non-hacking methods. These algorithms and methods provide efficient way to hack server database. By breaking the network security allow to introduce new and better security framework. The terms “Hacking” not only consider for its illegal activities but also it should be use for strengthen our global network.Keywords: Hacking, Vulnerabilities, Dummy request, Virus, Server monitoring
Procedia PDF Downloads 250846 Temperature Evolution, Microstructure and Mechanical Properties of Heat-Treatable Aluminum Alloy Welded by Friction Stir Welding: Comparison with Tungsten Inert Gas
Authors: Saliha Gachi, Mouloud Aissani, Fouad Boubenider
Abstract:
Friction Stir Welding (FSW) is a solid-state welding technique that can join material without melting the plates to be welded. In this work, we are interested to demonstrate the potentiality of FSW for joining the heat-treatable aluminum alloy 2024-T3 which is reputed as difficult to be welded by fusion techniques. Thereafter, the FSW joint is compared with another one obtained from a conventional fusion process Tungsten Inert Gas (TIG). FSW welds are made up using an FSW tool mounted on a milling machine. Single pass welding was applied to fabricated TIG joint. The comparison between the two processes has been made on the temperature evolution, mechanical and microstructure behavior. The microstructural examination revealed that FSW weld is composed of four zones: Base metal (BM), Heat affected zone (HAZ), Thermo-mechanical affected zone (THAZ) and the nugget zone (NZ). The NZ exhibits a recrystallized equiaxed refined grains that induce better mechanical properties and good ductility compared to TIG joint where the grains have a larger size in the welded region compared with the BM due to the elevated heat input. The microhardness results show that, in FSW weld, the THAZ contains the lowest microhardness values and increase in the NZ; however, in TIG process, the lowest values are localized on the NZ.Keywords: friction stir welding, tungsten inert gaz, aluminum, microstructure
Procedia PDF Downloads 275845 Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling
Authors: Ali Ben Abbes, ImedRiadh Farah, Vincent Barra
Abstract:
Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.Keywords: multi-temporal satellite image, urban growth, non-stationary, stochastic model
Procedia PDF Downloads 426844 Development of Fault Diagnosis Technology for Power System Based on Smart Meter
Authors: Chih-Chieh Yang, Chung-Neng Huang
Abstract:
In power system, how to improve the fault diagnosis technology of transmission line has always been the primary goal of power grid operators. In recent years, due to the rise of green energy, the addition of all kinds of distributed power also has an impact on the stability of the power system. Because the smart meters are with the function of data recording and bidirectional transmission, the adaptive Fuzzy Neural inference system, ANFIS, as well as the artificial intelligence that has the characteristics of learning and estimation in artificial intelligence. For transmission network, in order to avoid misjudgment of the fault type and location due to the input of these unstable power sources, combined with the above advantages of smart meter and ANFIS, a method for identifying fault types and location of faults is proposed in this study. In ANFIS training, the bus voltage and current information collected by smart meters can be trained through the ANFIS tool in MATLAB to generate fault codes to identify different types of faults and the location of faults. In addition, due to the uncertainty of distributed generation, a wind power system is added to the transmission network to verify the diagnosis correctness of the study. Simulation results show that the method proposed in this study can correctly identify the fault type and location of fault with more efficiency, and can deal with the interference caused by the addition of unstable power sources.Keywords: ANFIS, fault diagnosis, power system, smart meter
Procedia PDF Downloads 136843 The Application of Sequence Stratigraphy to the Sajau (Pliocene) Coal Distribution in Berau Basin, Northeast Kalimantan, Indonesia
Authors: Ahmad Helman Hamdani, Diana Putri Hamdiana
Abstract:
The Sajau coal measures of Berau Basin, northeastern Kalimantan were deposited within a range of facies associations spanning a spectrum of settings from fluvial to marine. The transitional to terrestrial coal measures are dominated by siliciclastics, but they also contain three laterally extensive marine bands (mudstone). These bands act as marker horizons that enable correlation between fully marine and terrestrial facies. Examination of this range of facies and their sedimentology has enabled the development of a high-resolution sequence stratigraphic framework. Set against the established backdrop of third-order Sajau transgression, nine fourth-order sequences are recognized. Results show that, in the composite sequences, peat accumulation predominantly correlates in transitional areas with early transgressive sequence sets (TSS) and highstand sequence set (HSS), while in more landward areas it correlates with the middle TSS to late highstand sequence sets (HSS). Differences in peat accumulation regimes within the sequence stratigraphic framework are attributed to variations in subsidence and background siliciclastic input rates in different depositional settings, with these combining to produce differences in the rate of accommodation change. The preservation of coal resources in the middle to late HSS in this area was most likely related to the rise of the regional base level throughout the Sajau.Keywords: sequence stratigraphy, coal, Pliocene, Berau basin
Procedia PDF Downloads 465842 Tenants Use Less Input on Rented Plots: Evidence from Northern Ethiopia
Authors: Desta Brhanu Gebrehiwot
Abstract:
The study aims to investigate the impact of land tenure arrangements on fertilizer use per hectare in Northern Ethiopia. Household and Plot level data are used for analysis. Land tenure contracts such as sharecropping and fixed rent arrangements have endogeneity. Different unobservable characteristics may affect renting-out decisions. Thus, the appropriate method of analysis was the instrumental variable estimation technic. Therefore, the family of instrumental variable estimation methods two-stage least-squares regression (2SLS, the generalized method of moments (GMM), Limited information maximum likelihood (LIML), and instrumental variable Tobit (IV-Tobit) was used. Besides, a method to handle a binary endogenous variable is applied, which uses a two-step estimation. In the first step probit model includes instruments, and in the second step, maximum likelihood estimation (MLE) (“etregress” command in Stata 14) was used. There was lower fertilizer use per hectare on sharecropped and fixed rented plots relative to owner-operated. The result supports the Marshallian inefficiency principle in sharecropping. The difference in fertilizer use per hectare could be explained by a lack of incentivized detailed contract forms, such as giving more proportion of the output to the tenant under sharecropping contracts, which motivates to use of more fertilizer in rented plots to maximize the production because most sharecropping arrangements share output equally between tenants and landlords.Keywords: tenure-contracts, endogeneity, plot-level data, Ethiopia, fertilizer
Procedia PDF Downloads 84841 Improving Activity Recognition Classification of Repetitious Beginner Swimming Using a 2-Step Peak/Valley Segmentation Method with Smoothing and Resampling for Machine Learning
Authors: Larry Powell, Seth Polsley, Drew Casey, Tracy Hammond
Abstract:
Human activity recognition (HAR) systems have shown positive performance when recognizing repetitive activities like walking, running, and sleeping. Water-based activities are a reasonably new area for activity recognition. However, water-based activity recognition has largely focused on supporting the elite and competitive swimming population, which already has amazing coordination and proper form. Beginner swimmers are not perfect, and activity recognition needs to support the individual motions to help beginners. Activity recognition algorithms are traditionally built around short segments of timed sensor data. Using a time window input can cause performance issues in the machine learning model. The window’s size can be too small or large, requiring careful tuning and precise data segmentation. In this work, we present a method that uses a time window as the initial segmentation, then separates the data based on the change in the sensor value. Our system uses a multi-phase segmentation method that pulls all peaks and valleys for each axis of an accelerometer placed on the swimmer’s lower back. This results in high recognition performance using leave-one-subject-out validation on our study with 20 beginner swimmers, with our model optimized from our final dataset resulting in an F-Score of 0.95.Keywords: time window, peak/valley segmentation, feature extraction, beginner swimming, activity recognition
Procedia PDF Downloads 121840 Constitutive Model for Analysis of Long-Term Municipal Solid Waste Landfill Settlement
Authors: Irena Basaric Ikodinovic, Dragoslav Rakic, Mirjana Vukicevic, Sanja Jockovic, Jovana Jankovic Pantic
Abstract:
Large long-term settlement occurs at the municipal solid waste landfills over an extended period of time which may lead to breakage of the geomembrane, damage of the cover systems, other protective systems or facilities constructed on top of a landfill. Also, municipal solid waste is an extremely heterogeneous material and its properties vary over location and time within a landfill. These material characteristics require the formulation of a new constitutive model to predict the long-term settlement of municipal solid waste. The paper presents a new constitutive model which is formulated to describe the mechanical behavior of municipal solid waste. Model is based on Modified Cam Clay model and the critical state soil mechanics framework incorporating time-dependent components: mechanical creep and biodegradation of municipal solid waste. The formulated constitutive model is optimized and defined with eight input parameters: five Modified Cam Clay parameters, one parameter for mechanical creep and two parameters for biodegradation of municipal solid waste. Thereafter, the constitutive model is implemented in the software suite for finite element analysis (ABAQUS) and numerical analysis of the experimental landfill settlement is performed. The proposed model predicts the total settlement which is in good agreement with field measured settlement at the experimental landfill.Keywords: constitutive model, finite element analysis, municipal solid waste, settlement
Procedia PDF Downloads 229839 High-Speed Imaging and Acoustic Measurements of Dual-frequency Ultrasonic Processing of Graphite in Water
Authors: Justin Morton, Mohammad Khavari, Abhinav Priyadarshi, Nicole Grobert, Dmitry G. Eskin, Jiawei Mi, Kriakos Porfyrakis, Paul Prentice
Abstract:
Ultrasonic cavitation is used for various processes and applications. Recently, ultrasonic assisted liquid phase exfoliation has been implemented to produce two dimensional nanomaterials. Depending on parameters such as input transducer power and the operational frequency used to induce the cavitation, bubble dynamics can be controlled and optimised. Using ultra-high-speed imagining and acoustic pressure measurements, a dual-frequency systemand its effect on bubble dynamics was investigated. A high frequency transducer (1.174 MHz) showed that bubble fragments and satellite bubbles induced from a low frequency transducer (24 kHz) were able to extend their lifecycle. In addition, this combination of ultrasonic frequencies generated higher acoustic emissions (∼24%) than the sum of the individual transducers. The dual-frequency system also produced an increase in cavitation zone size of∼3 times compared to the low frequency sonotrode. Furthermore, the high frequency induced cavitation bubbleswere shown to rapidly oscillate, although remained stable and did not transiently collapse, even in the presence of a low pressure field. Finally, the spatial distribution of satellite and fragment bubbles from the sonotrode were shown to increase, extending the active cavitation zone. These observations elucidated the benefits of using a dual-frequency system for generating nanomaterials with the aid of ultrasound, in deionised water.Keywords: dual-frequency, cavitation, bubble dynamics, graphene
Procedia PDF Downloads 193838 Systematic Examination of Methods Supporting the Social Innovation Process
Authors: Mariann Veresne Somosi, Zoltan Nagy, Krisztina Varga
Abstract:
Innovation is the key element of economic development and a key factor in social processes. Technical innovations can be identified as prerequisites and causes of social change and cannot be created without the renewal of society. The study of social innovation can be characterised as one of the significant research areas of our day. The study’s aim is to identify the process of social innovation, which can be defined by input, transformation, and output factors. This approach divides the social innovation process into three parts: situation analysis, implementation, follow-up. The methods associated with each stage of the process are illustrated by the chronological line of social innovation. In this study, we have sought to present methodologies that support long- and short-term decision-making that is easy to apply, have different complementary content, and are well visualised for different user groups. When applying the methods, the reference objects are different: county, district, settlement, specific organisation. The solution proposed by the study supports the development of a methodological combination adapted to different situations. Having reviewed metric and conceptualisation issues, we wanted to develop a methodological combination along with a change management logic suitable for structured support to the generation of social innovation in the case of a locality or a specific organisation. In addition to a theoretical summary, in the second part of the study, we want to give a non-exhaustive picture of the two counties located in the north-eastern part of Hungary through specific analyses and case descriptions.Keywords: factors of social innovation, methodological combination, social innovation process, supporting decision-making
Procedia PDF Downloads 154837 Recent Advances in Pulse Width Modulation Techniques and Multilevel Inverters
Authors: Satish Kumar Peddapelli
Abstract:
This paper presents advances in pulse width modulation techniques which refers to a method of carrying information on train of pulses and the information be encoded in the width of pulses. Pulse Width Modulation is used to control the inverter output voltage. This is done by exercising the control within the inverter itself by adjusting the ON and OFF periods of inverter. By fixing the DC input voltage we get AC output voltage. In variable speed AC motors the AC output voltage from a constant DC voltage is obtained by using inverter. Recent developments in power electronics and semiconductor technology have lead improvements in power electronic systems. Hence, different circuit configurations namely multilevel inverters have become popular and considerable interest by researcher are given on them. A fast Space-Vector Pulse Width Modulation (SVPWM) method for five-level inverter is also discussed. In this method, the space vector diagram of the five-level inverter is decomposed into six space vector diagrams of three-level inverters. In turn, each of these six space vector diagrams of three-level inverter is decomposed into six space vector diagrams of two-level inverters. After decomposition, all the remaining necessary procedures for the three-level SVPWM are done like conventional two-level inverter. The proposed method reduces the algorithm complexity and the execution time. It can be applied to the multilevel inverters above the five-level also. The experimental setup for three-level diode-clamped inverter is developed using TMS320LF2407 DSP controller and the experimental results are analysed.Keywords: five-level inverter, space vector pulse wide modulation, diode clamped inverter, electrical engineering
Procedia PDF Downloads 386836 A Framework for Chinese Domain-Specific Distant Supervised Named Entity Recognition
Abstract:
The Knowledge Graphs have now become a new form of knowledge representation. However, there is no consensus in regard to a plausible and definition of entities and relationships in the domain-specific knowledge graph. Further, in conjunction with several limitations and deficiencies, various domain-specific entities and relationships recognition approaches are far from perfect. Specifically, named entity recognition in Chinese domain is a critical task for the natural language process applications. However, a bottleneck problem with Chinese named entity recognition in new domains is the lack of annotated data. To address this challenge, a domain distant supervised named entity recognition framework is proposed. The framework is divided into two stages: first, the distant supervised corpus is generated based on the entity linking model of graph attention neural network; secondly, the generated corpus is trained as the input of the distant supervised named entity recognition model to train to obtain named entities. The link model is verified in the ccks2019 entity link corpus, and the F1 value is 2% higher than that of the benchmark method. The re-pre-trained BERT language model is added to the benchmark method, and the results show that it is more suitable for distant supervised named entity recognition tasks. Finally, it is applied in the computer field, and the results show that this framework can obtain domain named entities.Keywords: distant named entity recognition, entity linking, knowledge graph, graph attention neural network
Procedia PDF Downloads 90835 Design and Development of Compact 1KW Floating Battery Discharge Regulator
Authors: A. Sreedevi, G. Anantaramu
Abstract:
The present space research organizations are striving towards the development of lighter, smaller, more efficient, low cost, and highly reliable power supply. Switch mode power supplies (SMPS) overcome the demerits of linear power supplies such as low efficiency, difficulties in thermal management, and in boosting the output voltage. Space applications require a constant DC voltage to supply its load. As the load varies, the battery terminal voltage tends to vary accordingly. To avoid this variation in the load terminal voltage, a DC-DC regulator is required. The conventional regulator for space applications is isolated boost topology. The proposed topology uses an interleaved push-pull converter with a current doubler secondary to reduce the EMI issues and increase efficiency. The proposed topology uses a floating technique where the converter derives power from the battery and generates only the voltage that is required to fill the gap between the bus and the battery voltage. The direct voltage sense and current loop provide tight regulation of output and better stability. Converter is designed with 50 kHz switching frequency using UC 1825 PWM controller employing both voltage and peak current mode control. Experimental tests have been carried out on the converter under different input and load conditions to validate the design. The experimental results showed that the efficiency was greater than 91%. Stability analysis is done using venable stability analyzer.Keywords: push pull converter, current doubler, converter, PWM control
Procedia PDF Downloads 101834 Fuzzy Optimization Multi-Objective Clustering Ensemble Model for Multi-Source Data Analysis
Authors: C. B. Le, V. N. Pham
Abstract:
In modern data analysis, multi-source data appears more and more in real applications. Multi-source data clustering has emerged as a important issue in the data mining and machine learning community. Different data sources provide information about different data. Therefore, multi-source data linking is essential to improve clustering performance. However, in practice multi-source data is often heterogeneous, uncertain, and large. This issue is considered a major challenge from multi-source data. Ensemble is a versatile machine learning model in which learning techniques can work in parallel, with big data. Clustering ensemble has been shown to outperform any standard clustering algorithm in terms of accuracy and robustness. However, most of the traditional clustering ensemble approaches are based on single-objective function and single-source data. This paper proposes a new clustering ensemble method for multi-source data analysis. The fuzzy optimized multi-objective clustering ensemble method is called FOMOCE. Firstly, a clustering ensemble mathematical model based on the structure of multi-objective clustering function, multi-source data, and dark knowledge is introduced. Then, rules for extracting dark knowledge from the input data, clustering algorithms, and base clusterings are designed and applied. Finally, a clustering ensemble algorithm is proposed for multi-source data analysis. The experiments were performed on the standard sample data set. The experimental results demonstrate the superior performance of the FOMOCE method compared to the existing clustering ensemble methods and multi-source clustering methods.Keywords: clustering ensemble, multi-source, multi-objective, fuzzy clustering
Procedia PDF Downloads 188