Search results for: input faults
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2491

Search results for: input faults

871 Hydrological Evaluation of Satellite Precipitation Products Using IHACRES Rainfall-Runoff Model over a Basin in Iran

Authors: Mahmoud Zakeri Niri, Saber Moazami, Arman Abdollahipour, Hossein Ghalkhani

Abstract:

The objective of this research is to hydrological evaluation of four widely-used satellite precipitation products named PERSIANN, TMPA-3B42V7, TMPA-3B42RT, and CMORPH over Zarinehrood basin in Iran. For this aim, at first, daily streamflow of Sarough-cahy river of Zarinehrood basin was simulated using IHACRES rainfall-runoff model with daily rain gauge and temperature as input data from 1988 to 2008. Then, the model was calibrated in two different periods through comparison the simulated discharge with the observed one at hydrometric stations. Moreover, in order to evaluate the performance of satellite precipitation products in streamflow simulation, the calibrated model was validated using daily satellite rainfall estimates from the period of 2003 to 2008. The obtained results indicated that TMPA-3B42V7 with CC of 0.69, RMSE of 5.93 mm/day, MAE of 4.76 mm/day, and RBias of -5.39% performs better simulation of streamflow than those PERSIANN and CMORPH over the study area. It is noteworthy that in Iran, the availability of ground measuring station data is very limited because of the sparse density of hydro-meteorological networks. On the other hand, large spatial and temporal variability of precipitations and lack of a reliable and extensive observing system are the most important challenges to rainfall analysis, flood prediction, and other hydrological applications in this country.

Keywords: hydrological evaluation, IHACRES, satellite precipitation product, streamflow simulation

Procedia PDF Downloads 241
870 A Fuzzy Control System for Reducing Urban Stormwater Runoff by a Stormwater Storage Tank

Authors: Pingping Zhang, Yanpeng Cai, Jianlong Wang

Abstract:

Stormwater storage tank (SST) is a popular low impact development technology for reducing stormwater runoff in the construction of sponge city. At present, it is difficult to perform the automatic control of SST for reducing peak flow. In this paper, fuzzy control was introduced into the peak control of SST to improve the efficiency of reducing stormwater runoff. Firstly, the design of SST was investigated. A catchment area and a return period were assumed, a SST model was manufactured, and then the storage capacity of the SST was verified. Secondly, the control parameters of the SST based on reducing stormwater runoff were analyzed, and a schematic diagram of real-time control (RTC) system based on peak control SST was established. Finally, fuzzy control system of a double input (flow and water level) and double output (inlet and outlet valve) was designed. The results showed that 1) under the different return periods (one year, three years, five years), the SST had the effect of delayed peak control and storage by increasing the detention time, 2) rainfall, pipeline flow, the influent time and the water level in the SST could be used as RTC parameters, and 3) the response curves of flow velocity and water level fluctuated very little and reached equilibrium in a short time. The combination of online monitoring and fuzzy control was feasible to control the SST automatically. This paper provides a theoretical reference for reducing stormwater runoff and improving the operation efficiency of SST.

Keywords: stormwater runoff, stormwater storage tank, real-time control, fuzzy control

Procedia PDF Downloads 202
869 Heat and Mass Transfer Modelling of Industrial Sludge Drying at Different Pressures and Temperatures

Authors: L. Al Ahmad, C. Latrille, D. Hainos, D. Blanc, M. Clausse

Abstract:

A two-dimensional finite volume axisymmetric model is developed to predict the simultaneous heat and mass transfers during the drying of industrial sludge. The simulations were run using COMSOL-Multiphysics 3.5a. The input parameters of the numerical model were acquired from a preliminary experimental work. Results permit to establish correlations describing the evolution of the various parameters as a function of the drying temperature and the sludge water content. The selection and coupling of the equation are validated based on the drying kinetics acquired experimentally at a temperature range of 45-65 °C and absolute pressure range of 200-1000 mbar. The model, incorporating the heat and mass transfer mechanisms at different operating conditions, shows simulated values of temperature and water content. Simulated results are found concordant with the experimental values, only at the first and last drying stages where sludge shrinkage is insignificant. Simulated and experimental results show that sludge drying is favored at high temperatures and low pressure. As experimentally observed, the drying time is reduced by 68% for drying at 65 °C compared to 45 °C under 1 atm. At 65 °C, a 200-mbar absolute pressure vacuum leads to an additional reduction in drying time estimated by 61%. However, the drying rate is underestimated in the intermediate stage. This rate underestimation could be improved in the model by considering the shrinkage phenomena that occurs during sludge drying.

Keywords: industrial sludge drying, heat transfer, mass transfer, mathematical modelling

Procedia PDF Downloads 133
868 Continuous-Time Analysis And Performance Assessment For Digital Control Of High-Frequency Switching Synchronous Dc-Dc Converter

Authors: Rihab Hamdi, Amel Hadri Hamida, Ouafae Bennis, Sakina Zerouali

Abstract:

This paper features a performance analysis and robustness assessment of a digitally controlled DC-DC three-cell buck converter associated in parallel, operating in continuous conduction mode (CCM), facing feeding parameters variation and loads disturbance. The control strategy relies on the continuous-time with an averaged modeling technique for high-frequency switching converter. The methodology is to modulate the complete design procedure, in regard to the existence of an instantaneous current operating point for designing the digital closed-loop, to the same continuous-time domain. Moreover, the adopted approach is to include a digital voltage control (DVC) technique, taking an account for digital control delays and sampling effects, which aims at improving efficiency and dynamic response and preventing generally undesired phenomena. The results obtained under load change, input change, and reference change clearly demonstrates an excellent dynamic response of the proposed technique, also as provide stability in any operating conditions, the effectiveness is fast with a smooth tracking of the specified output voltage. Simulations studies in MATLAB/Simulink environment are performed to verify the concept.

Keywords: continuous conduction mode, digital control, parallel multi-cells converter, performance analysis, power electronics

Procedia PDF Downloads 150
867 Dynamic Risk Identification Using Fuzzy Failure Mode Effect Analysis in Fabric Process Industries: A Research Article as Management Perspective

Authors: A. Sivakumar, S. S. Darun Prakash, P. Navaneethakrishnan

Abstract:

In and around Erode District, it is estimated that more than 1250 chemical and allied textile processing fabric industries are affected, partially closed and shut off for various reasons such as poor management, poor supplier performance, lack of planning for productivity, fluctuation of output, poor investment, waste analysis, labor problems, capital/labor ratio, accumulation of stocks, poor maintenance of resources, deficiencies in the quality of fabric, low capacity utilization, age of plant and equipment, high investment and input but low throughput, poor research and development, lack of energy, workers’ fear of loss of jobs, work force mix and work ethic. The main objective of this work is to analyze the existing conditions in textile fabric sector, validate the break even of Total Productivity (TP), analyze, design and implement fuzzy sets and mathematical programming for improvement of productivity and quality dimensions in the fabric processing industry. It needs to be compatible with the reality of textile and fabric processing industries. The highly risk events from productivity and quality dimension were found by fuzzy systems and results are wrapped up among the textile fabric processing industry.

Keywords: break even point, fuzzy crisp data, fuzzy sets, productivity, productivity cycle, total productive maintenance

Procedia PDF Downloads 338
866 Real-Time Gesture Recognition System Using Microsoft Kinect

Authors: Ankita Wadhawan, Parteek Kumar, Umesh Kumar

Abstract:

Gesture is any body movement that expresses some attitude or any sentiment. Gestures as a sign language are used by deaf people for conveying messages which helps in eliminating the communication barrier between deaf people and normal persons. Nowadays, everybody is using mobile phone and computer as a very important gadget in their life. But there are some physically challenged people who are blind/deaf and the use of mobile phone or computer like device is very difficult for them. So, there is an immense need of a system which works on body gesture or sign language as input. In this research, Microsoft Kinect Sensor, SDK V2 and Hidden Markov Toolkit (HTK) are used to recognize the object, motion of object and human body joints through Touch less NUI (Natural User Interface) in real-time. The depth data collected from Microsoft Kinect has been used to recognize gestures of Indian Sign Language (ISL). The recorded clips are analyzed using depth, IR and skeletal data at different angles and positions. The proposed system has an average accuracy of 85%. The developed Touch less NUI provides an interface to recognize gestures and controls the cursor and click operation in computer just by waving hand gesture. This research will help deaf people to make use of mobile phones, computers and socialize among other persons in the society.

Keywords: gesture recognition, Indian sign language, Microsoft Kinect, natural user interface, sign language

Procedia PDF Downloads 306
865 Learning English from Movies: An Exploratory Study

Authors: Yasamiyan Alolaywi

Abstract:

The sources of second language acquisition vary and depend on a learner’s preferences and choices; however, undoubtedly, the most effective methods provide authentic language input. This current study explores the effectiveness of watching movies as a means of English language acquisition. It explores university students’ views on the impact of this method in improving English language skills. The participants in this study were 74 students (25 males and 49 females) from the Department of English Language and Translation at Qassim University, Saudi Arabia. Data for this research were collected from questionnaires and individual interviews with several selected students. The findings of this study showed that many students watch movies frequently and for various purposes, the most important of which is entertainment. The students also admitted that movies help them acquire a great deal of vocabulary and develop their listening and writing skills. Also, the participants believed that exposure to a target language by native speakers helps enhance language fluency and proficiency. The students learn not only linguistic aspects from films but also other aspects, such as culture, lifestyle, and ways of thinking, in addition to learning other languages such as Spanish. In light of these results, some recommendations are proposed, such as verifying the feasibility of integrating media into a foreign language classroom. While this study covers aspects of the relationship between watching movies and English language acquisition, knowledge gaps remain that need to be filled by further research, such as on incorporating media into the educational process and how movie subtitles can improve learners’ language skills.

Keywords: language acquisition, English movies, EFL learners, perceptions

Procedia PDF Downloads 101
864 A Pipeline for Detecting Copy Number Variation from Whole Exome Sequencing Using Comprehensive Tools

Authors: Cheng-Yang Lee, Petrus Tang, Tzu-Hao Chang

Abstract:

Copy number variations (CNVs) have played an important role in many kinds of human diseases, such as Autism, Schizophrenia and a number of cancers. Many diseases are found in genome coding regions and whole exome sequencing (WES) is a cost-effective and powerful technology in detecting variants that are enriched in exons and have potential applications in clinical setting. Although several algorithms have been developed to detect CNVs using WES and compared with other algorithms for finding the most suitable methods using their own samples, there were not consistent datasets across most of algorithms to evaluate the ability of CNV detection. On the other hand, most of algorithms is using command line interface that may greatly limit the analysis capability of many laboratories. We create a series of simulated WES datasets from UCSC hg19 chromosome 22, and then evaluate the CNV detective ability of 19 algorithms from OMICtools database using our simulated WES datasets. We compute the sensitivity, specificity and accuracy in each algorithm for validation of the exome-derived CNVs. After comparison of 19 algorithms from OMICtools database, we construct a platform to install all of the algorithms in a virtual machine like VirtualBox which can be established conveniently in local computers, and then create a simple script that can be easily to use for detecting CNVs using algorithms selected by users. We also build a table to elaborate on many kinds of events, such as input requirement, CNV detective ability, for all of the algorithms that can provide users a specification to choose optimum algorithms.

Keywords: whole exome sequencing, copy number variations, omictools, pipeline

Procedia PDF Downloads 319
863 Numerical Simulation of a Point Absorber Wave Energy Converter Using OpenFOAM in Indian Scenario

Authors: Pooja Verma, Sumana Ghosh

Abstract:

There is a growing need for alternative way of power generation worldwide. The reason can be attributed to limited resources of fossil fuels, environmental pollution, increasing cost of conventional fuels, and lower efficiency of conversion of energy in existing systems. In this context, one of the potential alternatives for power generation is wave energy. However, it is difficult to estimate the amount of electrical energy generation in an irregular sea condition by experiment and or analytical methods. Therefore in this work, a numerical wave tank is developed using the computational fluid dynamics software Open FOAM. In this software a specific utility known as waves2Foam utility is being used to carry out the simulation work. The computational domain is a tank of dimension: 5m*1.5m*1m with a floating object of dimension: 0.5m*0.2m*0.2m. Regular waves are generated at the inlet of the wave tank according to Stokes second order theory. The main objective of the present study is to validate the numerical model against existing experimental data. It shows a good matching with the existing experimental data of floater displacement. Later the model is exploited to estimate energy extraction due to the movement of such a point absorber in real sea conditions. Scale down the wave properties like wave height, wave length, etc. are used as input parameters. Seasonal variations are also considered.

Keywords: OpenFOAM, numerical wave tank, regular waves, floating object, point absorber

Procedia PDF Downloads 352
862 Economic and Environmental Impact of the Missouri Grazing Schools

Authors: C. A. Roberts, S. L. Mascaro, J. R. Gerrish, J. L. Horner

Abstract:

Management-intensive Grazing (MiG) is a practice that rotates livestock through paddocks in a way that best matches the nutrient requirements of the animal to the yield and quality of the pasture. In the USA, MiG has been taught to livestock producers throughout the state of Missouri in 2- and 3-day workshops called “Missouri Grazing Schools.” The economic impact of these schools was quantified using IMPLAN software. The model included hectares of adoption, animal performance, carrying capacity, and input costs. To date, MiG, as taught in the Missouri Grazing Schools, has been implemented on more than 70,000 hectares in Missouri. The economic impact of these schools is presently $125 million USD per year added to the state economy. This magnitude of impact is the result not only of widespread adoption but also because of increased livestock carrying capacity; in Missouri, a capacity increase of 25 to 30% has been well documented. Additional impacts have been MiG improving forage quality and reducing the cost of feed and fertilizer. The environmental impact of MiG in the state of Missouri is currently being estimated. Environmental impact takes into account the reduction in the application of commercial fertilizers; in MiG systems, nitrogen is supplied by N fixation from legumes, and much of the P and K is recycled naturally by well-distributed manure. The environmental impact also estimates carbon sequestration and methane production; MiG can increase carbon sequestration and reduce methane production in comparison to default grazing practices and feedlot operations in the USA.

Keywords: agricultural education, forage quality, management-intensive grazing, nutrient cycling, stock density, sustainable agriculture

Procedia PDF Downloads 201
861 Potentiality of Biohythane Process for the Gaseous Energy Recovery from Organic Wastes

Authors: Debabrata Das, Preeti Mishra

Abstract:

A two-phase anaerobic process combining biohydrogen followed by biomethane (biohythane technology) serves as an environment-friendly and economically sustainable approach for the improved valorization of organic wastes. Suitability of the pure cultures like Klebsiela pneumonia, C. freundii, B. coagulan, etc. and mixed acidogenic cultures for the biohydrogen production was already studied. The characteristics of organic wastes play a critical role in biohydrogen production. The choice of an appropriate combination of complementary organic wastes can vastly improve the bioenergy generation besides achieving the significant cost reduction. Suitability and economic viability of using the groundnut deoiled cake (GDOC), mustard deoiled cake (MDOC), distillers’ dried grain with soluble (DDGS) and algal biomass (AB) as a co-substrate were studied for a biohythane production. Results show that maximum gaseous energy of 20.7, 9.3, 16.7 and 15.6 % was recovered using GDOC, MDOC, DDGS and AB in the two stage biohythane production, respectively. Both GDOC and DDGS were found to be better co-substrates as compared to MDOC and AB in terms of hythane production, respectively. The maximum cumulative hydrogen and methane production of 150 and 64 mmol/L were achieved using GDOC. Further, 98 % reduction in substrate input cost (SIC) was achieved using the co-supplementation procedure.

Keywords: Biohythane, algal biomass, distillers’ dried grain with soluble (DDGS), groundnut deoiled cake (GDOC), mustard deoiled cake (MDOC)

Procedia PDF Downloads 200
860 Operations Guide Implementation Practice in Information Technology Organizations

Authors: Ziad M. Hejazi, Hani F. Mokhtar, Mohammed S. Bahabri, Mohammed H. Ghafouri, Ahmed S. Bahaitham

Abstract:

This paper demonstrates the efforts taken by an Information Technology (IT) organization at Saudi Aramco to establish Operations Guide in a practical manner. Review of related work and literature revealed several important aspects to be considered when implementing the operation guide including Identify supporting IT groups, specify each group roles and responsibilities, formulate the IT operations in terms of processes (input/output), list each process main steps, provide the details of each process main step, develop the RACI (Responsible, Accountable, Consulted, and Informed) chart, highlight the process KPI’s, utilized systems, and forms. Identified aspects were then addressed in the actual implementation via several practices, including developing the operation guide for all IT supported operations, creating a shared folder for the operations guide, and announcing the implementation to all IT staff. The implementation of the mentioned practice was benchmarked, identified as best in class, and adopted by other internal organizations. Moreover, it was evident and appreciated by IT management. The significance of this study stems from the fact that it might be among the first studies in Saudi Arabia that propose a practical guideline to implement IT operations guide by IT organizations. Additional research significance comes from the study being conducted in Saudi Aramco, one of the world’s biggest integrated energy and petrochemical companies.

Keywords: operations guide, process implementation, Saudi Aramco company, information technology, standard of procedure

Procedia PDF Downloads 95
859 Glaucoma Detection in Retinal Tomography Using the Vision Transformer

Authors: Sushish Baral, Pratibha Joshi, Yaman Maharjan

Abstract:

Glaucoma is a chronic eye condition that causes vision loss that is irreversible. Early detection and treatment are critical to prevent vision loss because it can be asymptomatic. For the identification of glaucoma, multiple deep learning algorithms are used. Transformer-based architectures, which use the self-attention mechanism to encode long-range dependencies and acquire extremely expressive representations, have recently become popular. Convolutional architectures, on the other hand, lack knowledge of long-range dependencies in the image due to their intrinsic inductive biases. The aforementioned statements inspire this thesis to look at transformer-based solutions and investigate the viability of adopting transformer-based network designs for glaucoma detection. Using retinal fundus images of the optic nerve head to develop a viable algorithm to assess the severity of glaucoma necessitates a large number of well-curated images. Initially, data is generated by augmenting ocular pictures. After that, the ocular images are pre-processed to make them ready for further processing. The system is trained using pre-processed images, and it classifies the input images as normal or glaucoma based on the features retrieved during training. The Vision Transformer (ViT) architecture is well suited to this situation, as it allows the self-attention mechanism to utilise structural modeling. Extensive experiments are run on the common dataset, and the results are thoroughly validated and visualized.

Keywords: glaucoma, vision transformer, convolutional architectures, retinal fundus images, self-attention, deep learning

Procedia PDF Downloads 191
858 Finding the Longest Common Subsequence in Normal DNA and Disease Affected Human DNA Using Self Organizing Map

Authors: G. Tamilpavai, C. Vishnuppriya

Abstract:

Bioinformatics is an active research area which combines biological matter as well as computer science research. The longest common subsequence (LCSS) is one of the major challenges in various bioinformatics applications. The computation of the LCSS plays a vital role in biomedicine and also it is an essential task in DNA sequence analysis in genetics. It includes wide range of disease diagnosing steps. The objective of this proposed system is to find the longest common subsequence which presents in a normal and various disease affected human DNA sequence using Self Organizing Map (SOM) and LCSS. The human DNA sequence is collected from National Center for Biotechnology Information (NCBI) database. Initially, the human DNA sequence is separated as k-mer using k-mer separation rule. Mean and median values are calculated from each separated k-mer. These calculated values are fed as input to the Self Organizing Map for the purpose of clustering. Then obtained clusters are given to the Longest Common Sub Sequence (LCSS) algorithm for finding common subsequence which presents in every clusters. It returns nx(n-1)/2 subsequence for each cluster where n is number of k-mer in a specific cluster. Experimental outcomes of this proposed system produce the possible number of longest common subsequence of normal and disease affected DNA data. Thus the proposed system will be a good initiative aid for finding disease causing sequence. Finally, performance analysis is carried out for different DNA sequences. The obtained values show that the retrieval of LCSS is done in a shorter time than the existing system.

Keywords: clustering, k-mers, longest common subsequence, SOM

Procedia PDF Downloads 267
857 Impact of Integrated Signals for Doing Human Activity Recognition Using Deep Learning Models

Authors: Milagros Jaén-Vargas, Javier García Martínez, Karla Miriam Reyes Leiva, María Fernanda Trujillo-Guerrero, Francisco Fernandes, Sérgio Barroso Gonçalves, Miguel Tavares Silva, Daniel Simões Lopes, José Javier Serrano Olmedo

Abstract:

Human Activity Recognition (HAR) is having a growing impact in creating new applications and is responsible for emerging new technologies. Also, the use of wearable sensors is an important key to exploring the human body's behavior when performing activities. Hence, the use of these dispositive is less invasive and the person is more comfortable. In this study, a database that includes three activities is used. The activities were acquired from inertial measurement unit sensors (IMU) and motion capture systems (MOCAP). The main objective is differentiating the performance from four Deep Learning (DL) models: Deep Neural Network (DNN), Convolutional Neural Network (CNN), Recurrent Neural Network (RNN) and hybrid model Convolutional Neural Network-Long Short-Term Memory (CNN-LSTM), when considering acceleration, velocity and position and evaluate if integrating the IMU acceleration to obtain velocity and position represent an increment in performance when it works as input to the DL models. Moreover, compared with the same type of data provided by the MOCAP system. Despite the acceleration data is cleaned when integrating, results show a minimal increase in accuracy for the integrated signals.

Keywords: HAR, IMU, MOCAP, acceleration, velocity, position, feature maps

Procedia PDF Downloads 98
856 XAI Implemented Prognostic Framework: Condition Monitoring and Alert System Based on RUL and Sensory Data

Authors: Faruk Ozdemir, Roy Kalawsky, Peter Hubbard

Abstract:

Accurate estimation of RUL provides a basis for effective predictive maintenance, reducing unexpected downtime for industrial equipment. However, while models such as the Random Forest have effective predictive capabilities, they are the so-called ‘black box’ models, where interpretability is at a threshold to make critical diagnostic decisions involved in industries related to aviation. The purpose of this work is to present a prognostic framework that embeds Explainable Artificial Intelligence (XAI) techniques in order to provide essential transparency in Machine Learning methods' decision-making mechanisms based on sensor data, with the objective of procuring actionable insights for the aviation industry. Sensor readings have been gathered from critical equipment such as turbofan jet engine and landing gear, and the prediction of the RUL is done by a Random Forest model. It involves steps such as data gathering, feature engineering, model training, and evaluation. These critical components’ datasets are independently trained and evaluated by the models. While suitable predictions are served, their performance metrics are reasonably good; such complex models, however obscure reasoning for the predictions made by them and may even undermine the confidence of the decision-maker or the maintenance teams. This is followed by global explanations using SHAP and local explanations using LIME in the second phase to bridge the gap in reliability within industrial contexts. These tools analyze model decisions, highlighting feature importance and explaining how each input variable affects the output. This dual approach offers a general comprehension of the overall model behavior and detailed insight into specific predictions. The proposed framework, in its third component, incorporates the techniques of causal analysis in the form of Granger causality tests in order to move beyond correlation toward causation. This will not only allow the model to predict failures but also present reasons, from the key sensor features linked to possible failure mechanisms to relevant personnel. The causality between sensor behaviors and equipment failures creates much value for maintenance teams due to better root cause identification and effective preventive measures. This step contributes to the system being more explainable. Surrogate Several simple models, including Decision Trees and Linear Models, can be used in yet another stage to approximately represent the complex Random Forest model. These simpler models act as backups, replicating important jobs of the original model's behavior. If the feature explanations obtained from the surrogate model are cross-validated with the primary model, the insights derived would be more reliable and provide an intuitive sense of how the input variables affect the predictions. We then create an iterative explainable feedback loop, where the knowledge learned from the explainability methods feeds back into the training of the models. This feeds into a cycle of continuous improvement both in model accuracy and interpretability over time. By systematically integrating new findings, the model is expected to adapt to changed conditions and further develop its prognosis capability. These components are then presented to the decision-makers through the development of a fully transparent condition monitoring and alert system. The system provides a holistic tool for maintenance operations by leveraging RUL predictions, feature importance scores, persistent sensor threshold values, and autonomous alert mechanisms. Since the system will provide explanations for the predictions given, along with active alerts, the maintenance personnel can make informed decisions on their end regarding correct interventions to extend the life of the critical machinery.

Keywords: predictive maintenance, explainable artificial intelligence, prognostic, RUL, machine learning, turbofan engines, C-MAPSS dataset

Procedia PDF Downloads 6
855 The Correlation between Political Awareness and Political Participation for University Students’ “Applied Study”

Authors: Rana Mohamed

Abstract:

Despite youth in Egypt were away from political life for a long time; they are able to make a tangible difference in political status. Purpose: This exploratory study aims to determine whether and how much the prevailing political culture influence participatory behavior with a special focus on political awareness factors among university students in Egypt. Methodology: The study employed several data collection methods to ensure the validity of the results, quantitative and qualitative, verifying the positive relationships between the levels of political awareness and political participation and between political values in society and the level of political participation among university students. For achieving the objectives of the paper in the light of the pool of available literature and data, the study adopts system analysis method to apply input-output and conversions associated with the phenomena of political participation to analyze the different factors that have an effect upon the prevailing political culture and the patterns of values in Egyptian society. Findings: The result reveals that the level of political awareness and political participation for students were low, with a statistically significant relationship. In addition, the patterns of values in Egyptian culture significantly influence the levels of student participation. Therefore, the study recommends formulating policies that aim to increase awareness levels and integrate youth into the political process. Originality/Value: The importance of the academic study stems from addressing one of the central issues in political science; this study measures the change in the Egyptian patterns of culture and values among university students.

Keywords: political awareness, political participation, civic culture, citizenship, Egyptian universities, political knowledge

Procedia PDF Downloads 250
854 Fintech Credit and Bank Efficiency Two-way Relationship: A Comparison Study Across Country Groupings

Authors: Tan Swee Liang

Abstract:

This paper studies the two-way relationship between fintech credit and banking efficiency using the Generalized panel Method of Moment (GMM) estimation in structural equation modeling (SEM). Banking system efficiency, defined as its ability to produce the existing level of outputs with minimal inputs, is measured using input-oriented data envelopment analysis (DEA), where the whole banking system of an economy is treated as a single DMU. Banks are considered an intermediary between depositors and borrowers, utilizing inputs (deposits and overhead costs) to provide outputs (increase credits to the private sector and its earnings). Analysis of the interrelationship between fintech credit and bank efficiency is conducted to determine the impact in different country groupings (ASEAN, Asia and OECD), in particular the banking system response to fintech credit platforms. Our preliminary results show that banks do respond to the greater pressure caused by fintech platforms to enhance their efficiency, but differently across the different groups. The author’s earlier research on ASEAN-5 high bank overhead costs (as a share of total assets) as the determinant of economic growth suggests that expenses may not have been channeled efficiently to income-generating activities. One practical implication of the findings is that policymakers should enable alternative financing, such as fintech credit, as a warning or encouragement for banks to improve their efficiency.

Keywords: fintech lending, banking efficiency, data envelopment analysis, structural equation modeling

Procedia PDF Downloads 91
853 Delineation of Oil – Polluted Sites in Ibeno LGA, Nigeria, Using Microbiological and Physicochemical Characterization

Authors: Ime R. Udotong, Justina I. R. Udotong, Ofonime U. M. John

Abstract:

Mobil Producing Nigeria Unlimited (MPNU), a subsidiary of ExxonMobil and the highest crude oil & condensate producer in Nigeria has its operational base and an oil terminal, the Qua Iboe terminal (QIT) located at Ibeno, Nigeria. Other oil companies like Network Exploration and Production Nigeria Ltd, Frontier Oil Ltd; Shell Petroleum Development Company Ltd; Elf Petroleum Nigeria Ltd and Nigerian Agip Energy, a subsidiary of the Italian ENI E&P operate onshore, on the continental shelf and in deep offshore of the Atlantic Ocean, respectively with the coastal waters of Ibeno, Nigeria as the nearest shoreline. This study was designed to delineate the oil-polluted sites in Ibeno, Nigeria using microbiological and physico-chemical characterization of soils, sediments and ground and surface water samples from the study area. Results obtained revealed that there have been significant recent hydrocarbon inputs into this environment as observed from the high counts of hydrocarbonoclastic microorganisms in excess of 1% at all the stations sampled. Moreover, high concentrations of THC, BTEX and heavy metals contents in all the samples analyzed corroborate the high recent crude oil input into the study area. The results also showed that the pollution of the different environmental media sampled were of varying degrees, following the trend: Ground water > surface water > sediments > soils.

Keywords: microbiological characterization, oil-polluted sites, physico-chemical analyses, total hydrocarbon content

Procedia PDF Downloads 416
852 Neural Network Models for Actual Cost and Actual Duration Estimation in Construction Projects: Findings from Greece

Authors: Panagiotis Karadimos, Leonidas Anthopoulos

Abstract:

Predicting the actual cost and duration in construction projects concern a continuous and existing problem for the construction sector. This paper addresses this problem with modern methods and data available from past public construction projects. 39 bridge projects, constructed in Greece, with a similar type of available data were examined. Considering each project’s attributes with the actual cost and the actual duration, correlation analysis is performed and the most appropriate predictive project variables are defined. Additionally, the most efficient subgroup of variables is selected with the use of the WEKA application, through its attribute selection function. The selected variables are used as input neurons for neural network models through correlation analysis. For constructing neural network models, the application FANN Tool is used. The optimum neural network model, for predicting the actual cost, produced a mean squared error with a value of 3.84886e-05 and it was based on the budgeted cost and the quantity of deck concrete. The optimum neural network model, for predicting the actual duration, produced a mean squared error with a value of 5.89463e-05 and it also was based on the budgeted cost and the amount of deck concrete.

Keywords: actual cost and duration, attribute selection, bridge construction, neural networks, predicting models, FANN TOOL, WEKA

Procedia PDF Downloads 134
851 Identifying Dynamic Structural Parameters of Soil-Structure System Based on Data Recorded during Strong Earthquakes

Authors: Vahidreza Mahmoudabadi, Omid Bahar, Mohammad Kazem Jafari

Abstract:

In many applied engineering problems, structural analysis is usually conducted by assuming a rigid bed, while imposing the effect of structure bed flexibility can affect significantly on the structure response. This article focuses on investigation and evaluation of the effects arising from considering a soil-structure system in evaluation of dynamic characteristics of a steel structure with respect to elastic and inelastic behaviors. The recorded structure acceleration during Taiwan’s strong Chi-Chi earthquake on different floors of the structure was our evaluation criteria. The respective structure is an eight-story steel bending frame structure designed using a displacement-based direct method assuring weak beam - strong column function. The results indicated that different identification methods i.e. reverse Fourier transform or transfer functions, is capable to determine some of the dynamic parameters of the structure precisely, rather than evaluating all of them at once (mode frequencies, mode shapes, structure damping, structure rigidity, etc.). Response evaluation based on the input and output data elucidated that the structure first mode is not significantly affected, even considering the soil-structure interaction effect, but the upper modes have been changed. Also, it was found that the response transfer function of the different stories, in which plastic hinges have occurred in the structure components, provides similar results.

Keywords: bending steel frame structure, dynamic characteristics, displacement-based design, soil-structure system, system identification

Procedia PDF Downloads 503
850 Design and Manufacture of an Autonomous Agricultural Robot for Pesticide Application

Authors: Caner Koc, Dilara Gerdan Koc, Emrah Saka, H. Ibrahim Karagol

Abstract:

The use of pesticides in agricultural activities is the most harmful to the environment and farmers' health, and it also has the greatest input prices, along with fertilizers. In this study, an electric, electrostatically charged, autonomous agricultural robot was developed, modeled, and prototyped and manufactured. It allows for sensitive pesticide applications with variable levels, has controllable spray nozzles, and uses camera distance sensors to detect and spray into tree canopies. The created prototype was produced with flexibility in mind. Two stages of prototype manufacture were completed. The initial stage involved designing and producing the flexible primary body of the autonomous vehicle. Detachable hanger assemblies are employed so that the main body robot can perform a variety of agricultural tasks. The design of the spraying devices and their fitting to the autonomous vehicle was completed as the second stage of the prototype. The built prototype spraying robot's itinerary was planned using the free, open-source program Mission Planner. PX4, telemetry, and RTK GPS are used to maneuver the autonomous car along the designated path. To avoid potential obstructions, the robot uses ultrasonic and lidar sensors. The developed autonomous vehicle's energy needs are intended to be met entirely by electric batteries. In the event that the batteries run out of power, the sockets are set up to be recharged both by using the generator and the main power source through the specifically constructed panel.

Keywords: autonomous agricultural robot, pesticide, smart farming, spraying, variable rate application

Procedia PDF Downloads 84
849 Economic Analysis of Cowpea (Unguiculata spp) Production in Northern Nigeria: A Case Study of Kano Katsina and Jigawa States

Authors: Yakubu Suleiman, S. A. Musa

Abstract:

Nigeria is the largest cowpea producer in the world, accounting for about 45%, followed by Brazil with about 17%. Cowpea is grown in Kano, Bauchi, Katsina, Borno in the north, Oyo in the west, and to the lesser extent in Enugu in the east. This study was conducted to determine the input–output relationship of Cowpea production in Kano, Katsina, and Jigawa states of Nigeria. The data were collected with the aid of 1000 structured questionnaires that were randomly distributed to Cowpea farmers in the three states mentioned above of the study area. The data collected were analyzed using regression analysis (Cobb–Douglass production function model). The result of the regression analysis revealed the coefficient of multiple determinations, R2, to be 72.5% and the F ration to be 106.20 and was found to be significant (P < 0.01). The regression coefficient of constant is 0.5382 and is significant (P < 0.01). The regression coefficient with respect to labor and seeds were 0.65554 and 0.4336, respectively, and they are highly significant (P < 0.01). The regression coefficient with respect to fertilizer is 0.26341 which is significant (P < 0.05). This implies that a unit increase of any one of the variable inputs used while holding all other variables inputs constants, will significantly increase the total Cowpea output by their corresponding coefficient. This indicated that farmers in the study area are operating in stage II of the production function. The result revealed that Cowpea farmer in Kano, Jigawa and Katsina States realized a profit of N15,997, N34,016 and N19,788 per hectare respectively. It is hereby recommended that more attention should be given to Cowpea production by government and research institutions.

Keywords: coefficient, constant, inputs, regression

Procedia PDF Downloads 409
848 Web Service Architectural Style Selection in Multi-Criteria Requirements

Authors: Ahmad Mohsin, Syda Fatima, Falak Nawaz, Aman Ullah Khan

Abstract:

Selection of an appropriate architectural style is vital to the success of target web service under development. The nature of architecture design and selection for service-oriented computing applications is quite different as compared to traditional software. Web Services have complex and rigorous architectural styles to choose. Due to this, selection for accurate architectural style for web services development has become a more complex decision to be made by architects. Architectural style selection is a multi-criteria decision and demands lots of experience in service oriented computing. Decision support systems are good solutions to simplify the selection process of a particular architectural style. Our research suggests a new approach using DSS for selection of architectural styles while developing a web service to cater FRs and NFRs. Our proposed DSS helps architects to select right web service architectural pattern according to the domain and non-functional requirements. In this paper, a rule base DSS has been developed using CLIPS (C Language Integrated Production System) to support decisions using multi-criteria requirements. This DSS takes architectural characteristics, domain requirements and software architect preferences for NFRs as input for different architectural styles in use today in service-oriented computing. Weighted sum model has been applied to prioritize quality attributes and domain requirements. Scores are calculated using multiple criterions to choose the final architecture style.

Keywords: software architecture, web-service, rule-based, DSS, multi-criteria requirements, quality attributes

Procedia PDF Downloads 364
847 Optimized Real Ground Motion Scaling for Vulnerability Assessment of Building Considering the Spectral Uncertainty and Shape

Authors: Chen Bo, Wen Zengping

Abstract:

Based on the results of previous studies, we focus on the research of real ground motion selection and scaling method for structural performance-based seismic evaluation using nonlinear dynamic analysis. The input of earthquake ground motion should be determined appropriately to make them compatible with the site-specific hazard level considered. Thus, an optimized selection and scaling method are established including the use of not only Monte Carlo simulation method to create the stochastic simulation spectrum considering the multivariate lognormal distribution of target spectrum, but also a spectral shape parameter. Its applications in structural fragility analysis are demonstrated through case studies. Compared to the previous scheme with no consideration of the uncertainty of target spectrum, the method shown here can make sure that the selected records are in good agreement with the median value, standard deviation and spectral correction of the target spectrum, and greatly reveal the uncertainty feature of site-specific hazard level. Meanwhile, it can help improve computational efficiency and matching accuracy. Given the important infection of target spectrum’s uncertainty on structural seismic fragility analysis, this work can provide the reasonable and reliable basis for structural seismic evaluation under scenario earthquake environment.

Keywords: ground motion selection, scaling method, seismic fragility analysis, spectral shape

Procedia PDF Downloads 292
846 An Efficient Hybrid Feedstock Pretreatment Technique for the Release of Fermentable Sugar from Cassava Peels for Biofuel Production

Authors: Gabriel Sanjo Aruwajoye, E. B. Gueguim Kana

Abstract:

Agricultural residues present a low-cost feedstock for bioenergy production around the world. Cassava peels waste are rich in organic molecules that can be readily converted to value added products such as biomaterials and biofuels. However, due to the presence of high proportion of structural carbohydrates and lignin, the hydrolysis of this feedstock is imperative to achieve maximum substrate utilization and energy yield. This study model and optimises the release of Fermentable Sugar (FS) from cassava peels waste using the Response Surface Methodology. The investigated pretreatment input parameters consisted of soaking temperature (oC), soaking time (hours), autoclave duration (minutes), acid concentration (% v/v), substrate solid loading (% w/v) within the range of 30 to 70, 0 to 24, 5 to 20, 0 to 5 and 2 to 10 respectively. The Box-Behnken design was used to generate 46 experimental runs which were investigated for FS release. The obtained data were used to fit a quadratic model. A coefficient of determination of 0.87 and F value of 8.73 was obtained indicating the good fitness of the model. The predicted optimum pretreatment conditions were 69.62 oC soaking temperature, 2.57 hours soaking duration, 5 minutes autoclave duration, 3.68 % v/v HCl and 9.65 % w/v solid loading corresponding to FS yield of 91.83g/l (0.92 g/g cassava peels) thus 58% improvement on the non-optimised pretreatment. Our findings demonstrate an efficient pretreatment model for fermentable sugar release from cassava peels waste for various bioprocesses.

Keywords: feedstock pretreatment, cassava peels, fermentable sugar, response surface methodology

Procedia PDF Downloads 366
845 Infrared Thermography as an Informative Tool in Energy Audit and Software Modelling of Historic Buildings: A Case Study of the Sheffield Cathedral

Authors: Ademuyiwa Agbonyin, Stamatis Zoras, Mohammad Zandi

Abstract:

This paper investigates the extent to which building energy modelling can be informed based on preliminary information provided by infrared thermography using a thermal imaging camera in a walkthrough audit. The case-study building is the Sheffield Cathedral, built in the early 1400s. Based on an informative qualitative report generated from the thermal images taken at the site, the regions showing significant heat loss are input into a computer model of the cathedral within the integrated environmental solution (IES) virtual environment software which performs an energy simulation to determine quantitative heat losses through the building envelope. Building data such as material thermal properties and building plans are provided by the architects, Thomas Ford and Partners Ltd. The results of the modelling revealed the portions of the building with the highest heat loss and these aligned with those suggested by the thermal camera. Retrofit options for the building are also considered, however, may not see implementation due to a desire to conserve the architectural heritage of the building. Results show that thermal imaging in a walk-through audit serves as a useful guide for the energy modelling process. Hand calculations were also performed to serve as a 'control' to estimate losses, providing a second set of data points of comparison.

Keywords: historic buildings, energy retrofit, thermal comfort, software modelling, energy modelling

Procedia PDF Downloads 170
844 Theoretical Analysis of the Optical and Solid State Properties of Thin Film

Authors: E. I. Ugwu

Abstract:

Theoretical analysis of the optical and Solid State properties of ZnS thin film using beam propagation technique in which a scalar wave is propagated through the material thin film deposited on a substrate with the assumption that the dielectric medium is section into a homogenous reference dielectric constant term, and a perturbed dielectric term, representing the deposited thin film medium is presented in this work. These two terms, constitute arbitrary complex dielectric function that describes dielectric perturbation imposed by the medium of for the system. This is substituted into a defined scalar wave equation in which the appropriate Green’s Function was defined on it and solved using series technique. The green’s value obtained from Green’s Function was used in Dyson’s and Lippmann Schwinger equations in conjunction with Born approximation method in computing the propagated field for different input regions of field wavelength during which the influence of the dielectric constants and mesh size of the thin film on the propagating field were depicted. The results obtained from the computed field were used in turn to generate the data that were used to compute the band gaps, solid state and optical properties of the thin film such as reflectance, Transmittance and reflectance with which the band gap obtained was found to be in close approximate to that of experimental value.

Keywords: scalar wave, optical and solid state properties, thin film, dielectric medium, perturbation, Lippmann Schwinger equations, Green’s Function, propagation

Procedia PDF Downloads 438
843 Predicting the Compressive Strength of Geopolymer Concrete Using Machine Learning Algorithms: Impact of Chemical Composition and Curing Conditions

Authors: Aya Belal, Ahmed Maher Eltair, Maggie Ahmed Mashaly

Abstract:

Geopolymer concrete is gaining recognition as a sustainable alternative to conventional Portland Cement concrete due to its environmentally friendly nature, which is a key goal for Smart City initiatives. It has demonstrated its potential as a reliable material for the design of structural elements. However, the production of Geopolymer concrete is hindered by batch-to-batch variations, which presents a significant challenge to the widespread adoption of Geopolymer concrete. To date, Machine learning has had a profound impact on various fields by enabling models to learn from large datasets and predict outputs accurately. This paper proposes an integration between the current drift to Artificial Intelligence and the composition of Geopolymer mixtures to predict their mechanical properties. This study employs Python software to develop machine learning model in specific Decision Trees. The research uses the percentage oxides and the chemical composition of the Alkali Solution along with the curing conditions as the input independent parameters, irrespective of the waste products used in the mixture yielding the compressive strength of the mix as the output parameter. The results showed 90 % agreement of the predicted values to the actual values having the ratio of the Sodium Silicate to the Sodium Hydroxide solution being the dominant parameter in the mixture.

Keywords: decision trees, geopolymer concrete, machine learning, smart cities, sustainability

Procedia PDF Downloads 87
842 Client Hacked Server

Authors: Bagul Abhijeet

Abstract:

Background: Client-Server model is the backbone of today’s internet communication. In which normal user can not have control over particular website or server? By using the same processing model one can have unauthorized access to particular server. In this paper, we discussed about application scenario of hacking for simple website or server consist of unauthorized way to access the server database. This application emerges to autonomously take direct access of simple website or server and retrieve all essential information maintain by administrator. In this system, IP address of server given as input to retrieve user-id and password of server. This leads to breaking administrative security of server and acquires the control of server database. Whereas virus helps to escape from server security by crashing the whole server. Objective: To control malicious attack and preventing all government website, and also find out illegal work to do hackers activity. Results: After implementing different hacking as well as non-hacking techniques, this system hacks simple web sites with normal security credentials. It provides access to server database and allow attacker to perform database operations from client machine. Above Figure shows the experimental result of this application upon different servers and provides satisfactory results as required. Conclusion: In this paper, we have presented a to view to hack the server which include some hacking as well as non-hacking methods. These algorithms and methods provide efficient way to hack server database. By breaking the network security allow to introduce new and better security framework. The terms “Hacking” not only consider for its illegal activities but also it should be use for strengthen our global network.

Keywords: Hacking, Vulnerabilities, Dummy request, Virus, Server monitoring

Procedia PDF Downloads 251