Search results for: transient analysis
27014 Performance Assessment Of An Existing Multi-effect Desalination System Driven By Solar Energy
Authors: B. Shahzamanian, S. Varga, D. C. Alarcón-Padilla
Abstract:
Desalination is considered the primary alternative to increase water supply for domestic, agricultural and industrial use. Sustainable desalination is only possible in places where renewable energy resources are available. Solar energy is the most relevant type of renewable energy to driving desalination systems since most of the areas suffering from water scarcity are characterized by a high amount of available solar radiation during the year. Multi-Effect Desalination (MED) technology integrated with solar thermal concentrators is a suitable combination for heat-driven desalination. It can also be coupled with thermal vapour compressors or absorption heat pumps to boost overall system performance. The most interesting advantage of MED is the suitability to be used with a transient source of energy like solar. An experimental study was carried out to assess the performance of the most important life-size multi-effect desalination plant driven by solar energy located in the Plataforma Solar de Almería (PSA). The MED plant is used as a reference in many studies regarding multi-effect distillation. The system consists of a 14-effect MED plant coupled with a double-effect absorption heat pump. The required thermal energy to run the desalination system is supplied by means of hot water generated from 60 static flat-plate solar collectors with a total aperture area of 606 m2. In order to compensate for the solar energy variation, a thermal storage system with two interconnected tanks and an overall volume of 40 m3 is coupled to the MED unit. The multi-effect distillation unit is built in a forward feed configuration, and the last effect is connected to a double-effect LiBr-H2O absorption heat pump. The heat pump requires steam at 180 ºC (10 bar a) that is supplied by a small-aperture parabolic trough solar field with a total aperture area of 230 m2. When needed, a gas boiler is used as an auxiliary heat source for operating the heat pump and the MED plant when solar energy is not available. A set of experiments was carried out for evaluating the impact of the heating water temperature (Th), top brine temperature (TBT) and temperature difference between effects (ΔT) on the performance ratio of the MED plant. The considered range for variation of Th, TBT and ΔT was 60-70°C, 54-63°C and 1.1-1.6°C, respectively. The performance ratio (PR), defined as kg of distillate produced for every 2326 kJ of thermal energy supplied to the MED system, was almost independent of the applied variables with a variation of less than 5% for all the cases. The maximum recorded PR was 12.4. The results indicated that the system demonstrated robustness for the whole range of operating conditions considered. Author gratitude is expressed to the PSA for providing access to its installations, the support of its scientific and technical staff, and the financial support of the SFERA-III project (Grant Agreement No 823802). Special thanks to the access provider staff members who ensured the access support.Keywords: multi-effect distillation, performance ratio, robustness, solar energy
Procedia PDF Downloads 19027013 Reassembling a Fragmented Border Landscape at Crossroads: Indigenous Rights, Rural Sustainability, Regional Integration and Post-Colonial Justice in Hong Kong
Authors: Chiu-Yin Leung
Abstract:
This research investigates a complex assemblage among indigenous identities, socio-political organization and national apparatus in the border landscape of post-colonial Hong Kong. This former British colony had designated a transient mode of governance in its New Territories and particularly the northernmost borderland in 1951-2012. With a discriminated system of land provisions for the indigenous villagers, the place has been inherited with distinctive village-based culture, historic monuments and agrarian practices until its sovereignty return into the People’s Republic of China. In its latest development imperatives by the national strategic planning, the frontier area of Hong Kong has been identified as a strategy site for regional economic integration in South China, with cross-border projects of innovation and technology zones, mega-transport infrastructure and inter-jurisdictional arrangement. Contemporary literature theorizes borders as the material and discursive production of territoriality, which manifest in state apparatus and the daily lives of its citizens and condense in the contested articulations of power, security and citizenship. Drawing on the concept of assemblage, this paper attempts to tract how the border regime and infrastructure in Hong Kong as a city are deeply ingrained in the everyday lived spaces of the local communities but also the changing urban and regional strategies across different longitudinal moments. Through an intensive ethnographic fieldwork among the borderland villages since 2008 and the extensive analysis of colonial archives, new development plans and spatial planning frameworks, the author navigates the genealogy of the border landscape in Ta Kwu Ling frontier area and its implications as the milieu for new state space, covering heterogeneous fields particularly in indigenous rights, heritage preservation, rural sustainability and regional economy. Empirical evidence suggests an apparent bias towards indigenous power and colonial representation in classifying landscape values and conserving historical monuments. Squatter and farm tenants are often deprived of property rights, statutory participation and livelihood option in the planning process. The postcolonial bureaucracies have great difficulties in mobilizing resources to catch up with the swift, political-first approach of the mainland counterparts. Meanwhile, the cultural heritage, lineage network and memory landscape are not protected altogether with any holistic view or collaborative effort across the border. The enactment of land resumption and compensation scheme is furthermore disturbed by lineage-based customary law, technocratic bureaucracy, intra-community conflicts and multi-scalar political mobilization. As many traces of colonial misfortune and tyranny have been whitewashed without proper management, the author argues that postcolonial justice is yet reconciled in this fragmented border landscape. The assemblage of border in mainstream representation has tended to oversimplify local struggles as a collective mist and setup a wider production of schizophrenia experiences in the discussion of further economic integration among Hong Kong and other mainland cities in the Pearl River Delta Region. The research is expected to shed new light on the theorizing of border regions and postcolonialism beyond Eurocentric perspectives. In reassembling the borderland experiences with other arrays in state governance, village organization and indigenous identities, the author also suggests an alternative epistemology in reconciling socio-spatial differences and opening up imaginaries for positive interventions.Keywords: heritage conservation, indigenous communities, post-colonial borderland, regional development, rural sustainability
Procedia PDF Downloads 21027012 Understanding the Semantic Network of Tourism Studies in Taiwan by Using Bibliometrics Analysis
Authors: Chun-Min Lin, Yuh-Jen Wu, Ching-Ting Chung
Abstract:
The formulation of tourism policies requires objective academic research and evidence as support, especially research from local academia. Taiwan is a small island, and its economic growth relies heavily on tourism revenue. Taiwanese government has been devoting to the promotion of the tourism industry over the past few decades. Scientific research outcomes by Taiwanese scholars may and will help lay the foundations for drafting future tourism policy by the government. In this study, a total of 120 full journal articles published between 2008 and 2016 from the Journal of Tourism and Leisure Studies (JTSL) were examined to explore the scientific research trend of tourism study in Taiwan. JTSL is one of the most important Taiwanese journals in the tourism discipline which focuses on tourism-related issues and uses traditional Chinese as the study language. The method of co-word analysis from bibliometrics approaches was employed for semantic analysis in this study. When analyzing Chinese words and phrases, word segmentation analysis is a crucial step. It must be carried out initially and precisely in order to obtain meaningful word or word chunks for further frequency calculation. A word segmentation system basing on N-gram algorithm was developed in this study to conduct semantic analysis, and 100 groups of meaningful phrases with the highest recurrent rates were located. Subsequently, co-word analysis was employed for semantic classification. The results showed that the themes of tourism research in Taiwan in recent years cover the scope of tourism education, environmental protection, hotel management, information technology, and senior tourism. The results can give insight on the related issues and serve as a reference for tourism-related policy making and follow-up research.Keywords: bibliometrics, co-word analysis, word segmentation, tourism research, policy
Procedia PDF Downloads 23127011 NOx Emission and Computational Analysis of Jatropha Curcus Fuel and Crude Oil
Authors: Vipan Kumar Sohpal, Rajesh K Sharma
Abstract:
Diminishing of conventional fuels and hysterical vehicles emission leads to deterioration of the environment, which emphasize the research to work on biofuels. Biofuels from different sources attract the attention of research due to low emission and biodegradability. Emission of carbon monoxide, carbon dioxide and H-C reduced drastically using Biofuels (B-20) combustion. Contrary to the conventional fuel, engine emission results indicated that nitrous oxide emission is higher in Biofuels. So this paper examines and compares the nitrogen oxide emission of Jatropha Curcus (JCO) B-20% blends with the vegetable oil. In addition to that computational analysis of crude non edible oil performed to assess the impact of composition on emission quality. In conclusion, JCO have the potential feedstock for the biodiesel production after the genetic modification in the plant.Keywords: jatropha curcus, computational analysis, emissions, NOx biofuels
Procedia PDF Downloads 59027010 Understanding Mathematics Achievements among U. S. Middle School Students: A Bayesian Multilevel Modeling Analysis with Informative Priors
Authors: Jing Yuan, Hongwei Yang
Abstract:
This paper aims to understand U.S. middle school students’ mathematics achievements by examining relevant student and school-level predictors. Through a variance component analysis, the study first identifies evidence supporting the use of multilevel modeling. Then, a multilevel analysis is performed under Bayesian statistical inference where prior information is incorporated into the modeling process. During the analysis, independent variables are entered sequentially in the order of theoretical importance to create a hierarchy of models. By evaluating each model using Bayesian fit indices, a best-fit and most parsimonious model is selected where Bayesian statistical inference is performed for the purpose of result interpretation and discussion. The primary dataset for Bayesian modeling is derived from the Program for International Student Assessment (PISA) in 2012 with a secondary PISA dataset from 2003 analyzed under the traditional ordinary least squares method to provide the information needed to specify informative priors for a subset of the model parameters. The dependent variable is a composite measure of mathematics literacy, calculated from an exploratory factor analysis of all five PISA 2012 mathematics achievement plausible values for which multiple evidences are found supporting data unidimensionality. The independent variables include demographics variables and content-specific variables: mathematics efficacy, teacher-student ratio, proportion of girls in the school, etc. Finally, the entire analysis is performed using the MCMCpack and MCMCglmm packages in R.Keywords: Bayesian multilevel modeling, mathematics education, PISA, multilevel
Procedia PDF Downloads 33627009 Copula-Based Estimation of Direct and Indirect Effects in Path Analysis Models
Authors: Alam Ali, Ashok Kumar Pathak
Abstract:
Path analysis is a statistical technique used to evaluate the direct and indirect effects of variables in path models. One or more structural regression equations are used to estimate a series of parameters in path models to find the better fit of data. However, sometimes the assumptions of classical regression models, such as ordinary least squares (OLS), are violated by the nature of the data, resulting in insignificant direct and indirect effects of exogenous variables. This article aims to explore the effectiveness of a copula-based regression approach as an alternative to classical regression, specifically when variables are linked through an elliptical copula.Keywords: path analysis, copula-based regression models, direct and indirect effects, k-fold cross validation technique
Procedia PDF Downloads 4327008 The Analysis of a Reactive Hydromagnetic Internal Heat Generating Poiseuille Fluid Flow through a Channel
Authors: Anthony R. Hassan, Jacob A. Gbadeyan
Abstract:
In this paper, the analysis of a reactive hydromagnetic Poiseuille fluid flow under each of sensitized, Arrhenius and bimolecular chemical kinetics through a channel in the presence of heat source is carried out. An exothermic reaction is assumed while the concentration of the material is neglected. Adomian Decomposition Method (ADM) together with Pade Approximation is used to obtain the solutions of the governing nonlinear non – dimensional differential equations. Effects of various physical parameters on the velocity and temperature fields of the fluid flow are investigated. The entropy generation analysis and the conditions for thermal criticality are also presented.Keywords: chemical kinetics, entropy generation, thermal criticality, adomian decomposition method (ADM) and pade approximation
Procedia PDF Downloads 46427007 Mitigation of Size Effects in Woven Fabric Composites Using Finite Element Analysis Approach
Authors: Azeez Shaik, Yagnik Kalariya, Amit Salvi
Abstract:
High-performance requirements and emission norms were forcing the automobile industry to opt for lightweight materials which improve the fuel efficiency and absorb energy during crash applications. In such scenario, the woven fabric composites are providing better energy absorption compared to metals. Woven fabric composites have a repetitive unit cell (RUC) and the mechanical properties of these materials are highly dependent on RUC. This work investigates the importance of detailed modelling of the RUC, the size effects associated and the mitigation techniques to avoid them using Finite element analysis approach.Keywords: repetitive unit cell, representative volume element, size effects, cohesive zone, finite element analysis
Procedia PDF Downloads 25727006 The Moderation Effect of Critical Item on the Strategic Purchasing: Quality Performance Relationship
Authors: Kwong Yeung
Abstract:
Theories about strategic purchasing and quality performance are underdeveloped. Understanding the evolving role of purchasing from reactive to proactive is a pressing strategic issue. Using survey responses from 176 manufacturing and electronics industry professionals, we study the relationships between strategic purchasing and supply chain partners’ quality performance to answer the following questions: Can transaction cost economics be used to elucidate the strategic purchasing-quality performance relationship? Is this strategic purchasing-quality performance relationship moderated by critical item analysis? The findings indicate that critical item analysis positively and significantly moderates the strategic purchasing-quality performance relationship.Keywords: critical item analysis, moderation, quality performance, strategic purchasing, transaction cost economics
Procedia PDF Downloads 56527005 Coordinated Voltage Control in a Radial Distribution System
Authors: Shivarudraswamy, Anubhav Shrivastava, Lakshya Bhat
Abstract:
Distributed generation has indeed become a major area of interest in recent years. Distributed Generation can address large number of loads in a power line and hence has better efficiency over the conventional methods. However there are certain drawbacks associated with it, increase in voltage being the major one. This paper addresses the voltage control at the buses for an IEEE 30 bus system by regulating reactive power. For carrying out the analysis, the suitable location for placing distributed generators (DG) is identified through load flow analysis and seeing where the voltage profile is dipping. MATLAB programming is used to regulate the voltage at all buses within +/-5% of the base value even after the introduction of DG’s. Three methods for regulation of voltage are discussed. A sensitivity based analysis is later carried out to determine the priority among the various methods listed in the paper.Keywords: distributed generators, distributed system, reactive power, voltage control
Procedia PDF Downloads 50127004 An Analysis of Discourse Markers Awareness in Writing Undergraduate Thesis of English Education Student in Sebelas Maret University
Authors: Oktanika Wahyu Nurjanah, Anggun Fitriana Dewi
Abstract:
An undergraduate thesis is one of the academic writings which should fulfill some characteristics, one of them is coherency. Moreover, a coherence of a text depends on the usage of discourse markers. In other word, discourse markers take an essential role in writing. Therefore, the researchers aim to know the awareness of the discourse markers usage in writing the under-graduate thesis of an English Education student at Sebelas Maret University. This research uses a qualitative case study in order to obtain a deep analysis. The sample of this research is an under-graduate thesis of English Education student in Sebelas Maret University which chosen based on some criteria. Additionally, the researchers were guided by some literature attempted to group the discourse markers based on their functions. Afterward, the analysis was held based on it. From the analysis, it found that the awareness of discourse markers usage is moderate. The last point, the researcher suggest undergraduate students to familiarize themselves with discourse markers, especially for those who want to write thesis.Keywords: discourse markers, English education, thesis writing, undergraduate student
Procedia PDF Downloads 35927003 Resource Constrained Time-Cost Trade-Off Analysis in Construction Project Planning and Control
Authors: Sangwon Han, Chengquan Jin
Abstract:
Time-cost trade-off (TCTO) is one of the most significant part of construction project management. Despite the significance, current TCTO analysis, based on the Critical Path Method, does not consider resource constraint, and accordingly sometimes generates an impractical and/or infeasible schedule planning in terms of resource availability. Therefore, resource constraint needs to be considered when doing TCTO analysis. In this research, genetic algorithms (GA) based optimization model is created in order to find the optimal schedule. This model is utilized to compare four distinct scenarios (i.e., 1) initial CPM, 2) TCTO without considering resource constraint, 3) resource allocation after TCTO, and 4) TCTO with considering resource constraint) in terms of duration, cost, and resource utilization. The comparison results identify that ‘TCTO with considering resource constraint’ generates the optimal schedule with the respect of duration, cost, and resource. This verifies the need for consideration of resource constraint when doing TCTO analysis. It is expected that the proposed model will produce more feasible and optimal schedule.Keywords: time-cost trade-off, genetic algorithms, critical path, resource availability
Procedia PDF Downloads 18927002 MXene Mediated Layered 2D-3D-2D g-C3N4@WO3@Ti3C2 Multijunctional Heterostructure with Enhanced Photoelectrochemical and Photocatalytic Properties
Authors: Lekgowa Collen Makola, Cecil Naphtaly Moro Ouma, Sharon Moeno, Langelihle Dlamini
Abstract:
In recent years, advancement in the field of nanotechnology has evolved new strategies to address energy and environmental issues. Amongst the developing technologies, visible-light-driven photocatalysis is regarded as a sustainable approach for energy production and environmental detoxifications, where transition metal oxides (TMOs) and metal-free carbon-based semiconductors such as graphitic carbon nitride (CN) evidenced notable potential in this matter. Herein, g-C₃N₄@WO₃@Ti₃C₂Tx three-component multijunction photocatalyst was fabricated via facile ultrasonic-assisted self-assembly, followed by calcination to facilitate extensive integrations of the materials. A series of different Ti₃C₂ wt% loading in the g-C₃N4@WO₃@Ti₃C₂Tx were prepared and represented as 1-CWT, 3-CWT, 5-CWT, and 7-CWT corresponding to 1, 3, 5, and 7wt%, respectively. Systematic characterization using spectroscopic and microscopic techniques were employed to validate the successful preparation of the photocatalysts. Enhanced optoelectronic and photoelectrochemical properties were observed for the WO₃@Ti₃C2@g-C₃N4 heterostructure with respect to the individual materials. Photoluminescence spectra and Nyquist plots show restrained recombination rates and improved photocarrier conductivities, respectively, and this was credited to the synergistic coupling effect and the presence of highly conductive Ti₃C2 MXene. The strong interfacial contact surfaces upon the formation of the composite were confirmed using XPS. Multiple charge transfer mechanisms were proposed for the WO3@Ti3C₂@g-C3N4, which couples Z-scheme and Schottky-junction mediated with Ti3C2 MXene. Bode phase plots show improved charge carrier life-times upon the formation of the multijunctional photocatalyst. Moreover, transient photocurrent density of 7-CWT is 40 and seven (7) times higher compared to that of g-C₃N4 and WO3, correspondingly. Unlike in the traditional Z-Scheme, the formed ternary heterostructure possesses interfaces through the metallic 2D Ti₃C₂ MXene, which provided charge transfer channels for efficient photocarrier transfers with carrier concentrations (ND) of 17.49×1021 cm-3 and 4.86% photo-to-chemical conversion efficiency. The as-prepared ternary g-C₃N₄@WO₃@Ti₃C₂Tx exhibited excellent photoelectrochemical properties with reserved redox band potential potencies to facilitate efficient photo-oxidation and -reduction reactions. The fabricated multijunction photocatalyst exhibits potentials to be used in an extensive range of photocatalytic process vis., production of valuable hydrocarbons from CO₂, production of H₂, and degradation of a plethora of pollutants from wastewater.Keywords: photocatalysis, Z-scheme, multijunction heterostructure, Ti₃C₂ MXene, g-C₃N₄
Procedia PDF Downloads 12627001 Integrating Data Envelopment Analysis and Variance Inflation Factor to Measure the Efficiency of Decision Making Units
Authors: Mostafa Kazemi, Zahra N. Farkhani
Abstract:
This paper proposes an integrated Data Envelopment Analysis (DEA) and Variance Inflation Factor (VIF) model for measuring the technical efficiency of decision making units. The model is validated using a set of 69% sales representatives’ dairy products. The analysis is done in two stages, in the first stage, VIF technique is used to distinguish independent effective factors of resellers, and in the second stage we used DEA for measuring efficiency for both constant and variable return to scales status. Further DEA is used to examine the utilization of environmental factors on efficiency. Results of this paper indicated an average managerial efficiency of 83% in the whole sales representatives’ dairy products. In addition, technical and scale efficiency were counted 96% and 80% respectively. 38% of sales representative have the technical efficiency of 100% and 72% of the sales representative in terms of managerial efficiency are quite efficient.High levels of relative efficiency indicate a good condition for sales representative efficiency.Keywords: data envelopment analysis (DEA), relative efficiency, sales representatives’ dairy products, variance inflation factor (VIF)
Procedia PDF Downloads 57027000 Effects of Small Amount of Poly(D-Lactic Acid) on the Properties of Poly(L-Lactic Acid)/Microcrystalline Cellulose/Poly(D-Lactic Acid) Blends
Authors: Md. Hafezur Rahaman, Md. Sagor Hosen, Md. Abdul Gafur, Rasel Habib
Abstract:
This research is a systematic study of effects of poly(D-lactic acid) (PDLA) on the properties of poly(L-lactic acid)(PLLA)/microcrystalline cellulose (MCC)/PDLA blends by stereo complex crystallization. Blends were prepared with constant percentage of (3 percent) MCC and different percentage of PDLA by solution casting methods. These blends were characterized by Fourier Transform Infrared Spectroscopy (FTIR) for the confirmation of blends compatibility, Wide-Angle X-ray Scattering (WAXS) and scanning electron microscope (SEM) for the analysis of morphology, thermo-gravimetric analysis (TGA) and differential thermal analysis (DTA) for thermal properties measurement. FTIR Analysis results confirm no new characteristic absorption peaks appeared in the spectrum instead shifting of peaks due to hydrogen bonding help to have compatibility of blends component. Development of three new peaks from XRD analysis indicates strongly the formation of stereo complex crystallinity in the PLLA structure with the addition of PDLA. TGA and DTG results indicate that PDLA can improve the heat resistivity of the PLLA/MCC blends by increasing its degradation temperature. Comparison of DTA peaks also ensure developed thermal properties. Image of SEM shows the improvement of surface morphology.Keywords: microcrystalline cellulose, poly(l-lactic acid), stereocomplex crystallization, thermal stability
Procedia PDF Downloads 13826999 Identify the Renewable Energy Potential through Sustainability Indicators and Multicriteria Analysis
Authors: Camila Lima, Murilo Andrade Valle, Patrícia Teixeira Leite Asano
Abstract:
The growth in demand for electricity, caused by human development, depletion and environmental impacts caused by traditional sources of electricity generation have made new energy sources are increasingly encouraged and necessary for companies in the electricity sector. Based on this scenario, this paper assesses the negative environmental impacts associated with thermoelectric power plants in Brazil, pointing out the importance of using renewable energy sources, reducing environmental aggression. This article points out the existence of an energy alternative, wind energy, of the municipalities of São Paulo, represented by georeferenced maps with the help of GIS, using as a premise the indicators of sustainability and multicriteria analysis in the decision-making process.Keywords: GIS (geographic information systems), multicriteria analysis, sustainability, wind energy
Procedia PDF Downloads 36726998 The Non-Linear Analysis of Brain Response to Visual Stimuli
Authors: H. Namazi, H. T. N. Kuan
Abstract:
Brain activity can be measured by acquiring and analyzing EEG signals from an individual. In fact, the human brain response to external and internal stimuli is mapped in his EEG signals. During years some methods such as Fourier transform, wavelet transform, empirical mode decomposition, etc. have been used to analyze the EEG signals in order to find the effect of stimuli, especially external stimuli. But each of these methods has some weak points in analysis of EEG signals. For instance, Fourier transform and wavelet transform methods are linear signal analysis methods which are not good to be used for analysis of EEG signals as nonlinear signals. In this research we analyze the brain response to visual stimuli by extracting information in the form of various measures from EEG signals using a software developed by our research group. The used measures are Jeffrey’s measure, Fractal dimension and Hurst exponent. The results of these analyses are useful not only for fundamental understanding of brain response to visual stimuli but provide us with very good recommendations for clinical purposes.Keywords: visual stimuli, brain response, EEG signal, fractal dimension, hurst exponent, Jeffrey’s measure
Procedia PDF Downloads 56326997 Static and Dynamical Analysis on Clutch Discs on Different Material and Geometries
Authors: Jairo Aparecido Martins, Estaner Claro Romão
Abstract:
This paper presents the static and cyclic stresses in combination with fatigue analysis resultant of loads applied on the friction discs usually utilized on industrial clutches. The material chosen to simulate the friction discs under load is aluminum. The numerical simulation was done by software COMSOLTM Multiphysics. The results obtained for static loads showed enough stiffness for both geometries and the material utilized. On the other hand, in the fatigue standpoint, failure is clearly verified, what demonstrates the importance of both approaches, mainly dynamical analysis. The results and the conclusion are based on the stresses on disc, counted stress cycles, and fatigue usage factor.Keywords: aluminum, industrial clutch, static and dynamic loading, numerical simulation
Procedia PDF Downloads 19126996 Hydrology and Hydraulics Analysis of Beko Abo Dam and Appurtenant Structre Design, Ethiopia
Authors: Azazhu Wassie
Abstract:
This study tried to evaluate the maximum design flood for appurtenance structure design using the given climatological and hydrological data analysis on the referenced study area. The maximum design flood is determined by using flood frequency analysis. Using this method, the peak discharge is 32,583.67 m3/s, but the data is transferred because the dam site is not on the gauged station. Then the peak discharge becomes 38,115 m3/s. The study was conducted in June 2023. This dam is built across a river to create a reservoir on its upstream side for impounding water. The water stored in the reservoir is used for various purposes, such as irrigation, hydropower, navigation, fishing, etc. The total average volume of annual runoff is estimated to be 115.1 billion m3. The total potential of the land for irrigation development can go beyond 3 million ha.Keywords: dam design, flow duration curve, peak flood, rainfall, reservoir capacity, risk and reliability
Procedia PDF Downloads 3026995 Study of Mobile Game Addiction Using Electroencephalography Data Analysis
Authors: Arsalan Ansari, Muhammad Dawood Idrees, Maria Hafeez
Abstract:
Use of mobile phones has been increasing considerably over the past decade. Currently, it is one of the main sources of communication and information. Initially, mobile phones were limited to calls and messages, but with the advent of new technology smart phones were being used for many other purposes including video games. Despite of positive outcomes, addiction to video games on mobile phone has become a leading cause of psychological and physiological problems among many people. Several researchers examined the different aspects of behavior addiction with the use of different scales. Objective of this study is to examine any distinction between mobile game addicted and non-addicted players with the use of electroencephalography (EEG), based upon psycho-physiological indicators. The mobile players were asked to play a mobile game and EEG signals were recorded by BIOPAC equipment with AcqKnowledge as data acquisition software. Electrodes were places, following the 10-20 system. EEG was recorded at sampling rate of 200 samples/sec (12,000samples/min). EEG recordings were obtained from the frontal (Fp1, Fp2), parietal (P3, P4), and occipital (O1, O2) lobes of the brain. The frontal lobe is associated with behavioral control, personality, and emotions. The parietal lobe is involved in perception, understanding logic, and arithmetic. The occipital lobe plays a role in visual tasks. For this study, a 60 second time window was chosen for analysis. Preliminary analysis of the signals was carried out with Acqknowledge software of BIOPAC Systems. From the survey based on CGS manual study 2010, it was concluded that five participants out of fifteen were in addictive category. This was used as prior information to group the addicted and non-addicted by physiological analysis. Statistical analysis showed that by applying clustering analysis technique authors were able to categorize the addicted and non-addicted players specifically on theta frequency range of occipital area.Keywords: mobile game, addiction, psycho-physiology, EEG analysis
Procedia PDF Downloads 16726994 Post Pandemic Mobility Analysis through Indexing and Sharding in MongoDB: Performance Optimization and Insights
Authors: Karan Vishavjit, Aakash Lakra, Shafaq Khan
Abstract:
The COVID-19 pandemic has pushed healthcare professionals to use big data analytics as a vital tool for tracking and evaluating the effects of contagious viruses. To effectively analyze huge datasets, efficient NoSQL databases are needed. The analysis of post-COVID-19 health and well-being outcomes and the evaluation of the effectiveness of government efforts during the pandemic is made possible by this research’s integration of several datasets, which cuts down on query processing time and creates predictive visual artifacts. We recommend applying sharding and indexing technologies to improve query effectiveness and scalability as the dataset expands. Effective data retrieval and analysis are made possible by spreading the datasets into a sharded database and doing indexing on individual shards. Analysis of connections between governmental activities, poverty levels, and post-pandemic well being is the key goal. We want to evaluate the effectiveness of governmental initiatives to improve health and lower poverty levels. We will do this by utilising advanced data analysis and visualisations. The findings provide relevant data that supports the advancement of UN sustainable objectives, future pandemic preparation, and evidence-based decision-making. This study shows how Big Data and NoSQL databases may be used to address problems with global health.Keywords: big data, COVID-19, health, indexing, NoSQL, sharding, scalability, well being
Procedia PDF Downloads 7326993 Ergonomical Study of Hand-Arm Vibrational Exposure in a Gear Manufacturing Plant in India
Authors: Santosh Kumar, M. Muralidhar
Abstract:
The term ‘ergonomics’ is derived from two Greek words: ‘ergon’, meaning work and ‘nomoi’, meaning natural laws. Ergonomics is the study of how working conditions, machines and equipment can be arranged in order that people can work with them more efficiently. In this research communication an attempt has been made to study the effect of hand-arm vibrational exposure on the workers of a gear manufacturing plant by comparison of potential Carpal Tunnel Syndrome (CTS) symptoms and effect of different exposure levels of vibration on occurrence of CTS in actual industrial environment. Chi square test and correlation analysis have been considered for statistical analysis. From Chi square test, it has been found that the potential CTS symptoms occurrence is significantly dependent on the level of vibrational exposure. Data analysis indicates that 40.51% workers having potential CTS symptoms are exposed to vibration. Correlation analysis reveals that potential CTS symptoms are significantly correlated with exposure to level of vibration from handheld tools and to repetitive wrist movements.Keywords: CTS symptoms, hand-arm vibration, ergonomics, physical tests
Procedia PDF Downloads 37326992 Holomorphic Prioritization of Sets within Decagram of Strategic Decision Making of POSM Using Operational Research (OR): Analytic Hierarchy Process (AHP) Analysis
Authors: Elias Ogutu Azariah Tembe, Hussain Abdullah Habib Al-Salamin
Abstract:
There is decagram of strategic decisions of operations and production/service management (POSM) within operational research (OR) which must collate, namely: design, inventory, quality, location, process and capacity, layout, scheduling, maintain ace, and supply chain. This paper presents an architectural configuration conceptual framework of a decagram of sets decisions in a form of mathematical complete graph and abelian graph. Mathematically, a complete graph is undirected (UDG), and directed (DG) a relationship where every pair of vertices are connected, collated, confluent, and holomorphic. There has not been any study conducted which, however, prioritizes the holomorphic sets which of POMS within OR field of study. The study utilizes OR structured technique known as The Analytic Hierarchy Process (AHP) analysis for organizing, sorting and prioritizing (ranking) the sets within the decagram of POMS according to their attribution (propensity), and provides an analysis how the prioritization has real-world application within the 21st century.Keywords: holomorphic, decagram, decagon, confluent, complete graph, AHP analysis, SCM, HRM, OR, OM, abelian graph
Procedia PDF Downloads 40426991 The Use of Random Set Method in Reliability Analysis of Deep Excavations
Authors: Arefeh Arabaninezhad, Ali Fakher
Abstract:
Since the deterministic analysis methods fail to take system uncertainties into account, probabilistic and non-probabilistic methods are suggested. Geotechnical analyses are used to determine the stress and deformation caused by construction; accordingly, many input variables which depend on ground behavior are required for geotechnical analyses. The Random Set approach is an applicable reliability analysis method when comprehensive sources of information are not available. Using Random Set method, with relatively small number of simulations compared to fully probabilistic methods, smooth extremes on system responses are obtained. Therefore random set approach has been proposed for reliability analysis in geotechnical problems. In the present study, the application of random set method in reliability analysis of deep excavations is investigated through three deep excavation projects which were monitored during the excavating process. A finite element code is utilized for numerical modeling. Two expected ranges, from different sources of information, are established for each input variable, and a specific probability assignment is defined for each range. To determine the most influential input variables and subsequently reducing the number of required finite element calculations, sensitivity analysis is carried out. Input data for finite element model are obtained by combining the upper and lower bounds of the input variables. The relevant probability share of each finite element calculation is determined considering the probability assigned to input variables present in these combinations. Horizontal displacement of the top point of excavation is considered as the main response of the system. The result of reliability analysis for each intended deep excavation is presented by constructing the Belief and Plausibility distribution function (i.e. lower and upper bounds) of system response obtained from deterministic finite element calculations. To evaluate the quality of input variables as well as applied reliability analysis method, the range of displacements extracted from models has been compared to the in situ measurements and good agreement is observed. The comparison also showed that Random Set Finite Element Method applies to estimate the horizontal displacement of the top point of deep excavation. Finally, the probability of failure or unsatisfactory performance of the system is evaluated by comparing the threshold displacement with reliability analysis results.Keywords: deep excavation, random set finite element method, reliability analysis, uncertainty
Procedia PDF Downloads 26926990 Deformation Analysis of Pneumatized Sphenoid Bone Caused Due to Elevated Intracranial Pressure Using Finite Element Analysis
Authors: Dilesh Mogre, Jitendra Toravi, Saurabh Joshi, Prutha Deshpande, Aishwarya Kura
Abstract:
In earlier days of technology, it was not possible to understand the nature of complex biomedical problems and were only left to clinical postulations. With advancement in science today, we have tools like Finite Element Modelling and simulation to solve complex biomedical problems. This paper presents how ANSYS WORKBENCH can be used to study deformation of pneumatized sphenoid bone caused by increased intracranial pressure. Intracranial pressure refers to the pressure inside the skull. The increase in the pressure above the normal range of 15mmhg can lead to serious conditions due to developed stresses and deformation. One of the areas where the deformation is suspected to occur is Sphenoid Bone. Moreover, the varying degree of pneumatization increases the complexity of the conditions. It is necessary to study deformation patterns on pneumatized sphenoid bone model at elevated intracranial pressure. Finite Element Analysis plays a major role in developing and analyzing model and give quantitative results.Keywords: intracranial pressure, pneumatized sphenoid bone, deformation, finite element analysis
Procedia PDF Downloads 19626989 Elastic Stress Analysis of Annular Bi-Material Discs with Variable Thickness under Mechanical and Thermomechanical Loads
Authors: Erhan Çetin, Ali Kurşun, Şafak Aksoy, Merve Tunay Çetin
Abstract:
The closed form study deal with elastic stress analysis of annular bi-material discs with variable thickness subjected to the mechanical and termomechanical loads. Those discs have many applications in the aerospace industry, such as gas turbines and gears. Those discs normally work under thermal and mechanical loads. Their life cycle can increase when stress components are minimized. Each material property is assumed to be isotropic. The results show that material combinations and thickness profiles play an important role in determining the responses of bi-material discs and an optimal design of those structures. Stress distribution is investigated and results are shown as graphs.Keywords: bi-material discs, elastic stress analysis, mechanical loads, rotating discs
Procedia PDF Downloads 32926988 Reliability-Based Method for Assessing Liquefaction Potential of Soils
Authors: Mehran Naghizaderokni, Asscar Janalizadechobbasty
Abstract:
This paper explores probabilistic method for assessing the liquefaction potential of sandy soils. The current simplified methods for assessing soil liquefaction potential use a deterministic safety factor in order to determine whether liquefaction will occur or not. However, these methods are unable to determine the liquefaction probability related to a safety factor. A solution to this problem can be found by reliability analysis.This paper presents a reliability analysis method based on the popular certain liquefaction analysis method. The proposed probabilistic method is formulated based on the results of reliability analyses of 190 field records and observations of soil performance against liquefaction. The results of the present study show that confidence coefficient greater and smaller than 1 does not mean safety and/or liquefaction in cadence for liquefaction, and for assuring liquefaction probability, reliability based method analysis should be used. This reliability method uses the empirical acceleration attenuation law in the Chalos area to derive the probability density distribution function and the statistics for the earthquake-induced cyclic shear stress ratio (CSR). The CSR and CRR statistics are used in continuity with the first order and second moment method to calculate the relation between the liquefaction probability, the safety factor and the reliability index. Based on the proposed method, the liquefaction probability related to a safety factor can be easily calculated. The influence of some of the soil parameters on the liquefaction probability can be quantitatively evaluated.Keywords: liquefaction, reliability analysis, chalos area, civil and structural engineering
Procedia PDF Downloads 47126987 Computational Fluid Dynamics Analysis of a Biomass Burner Gas Chamber in OpenFOAM
Authors: Óscar Alfonso Gómez Sepúlveda, Julián Ernesto Jaramillo, Diego Camilo Durán
Abstract:
The global climate crisis has affected different aspects of human life, and in an effort to reverse the effects generated, we seek to optimize and improve the equipment and plants that produce high emissions of CO₂, being possible to achieve this through numerical simulations. These equipments include biomass combustion chambers. The objective of this research is to visualize the thermal behavior of a gas chamber that is used in the process of obtaining vegetable extracts. The simulation is carried out with OpenFOAM taking into account the conservation of energy, turbulence, and radiation; for the purposes of the simulation, combustion is omitted and replaced by heat generation. Within the results, the streamlines generated by the primary and secondary flows are analyzed in order to visualize whether they generate the expected effect, and the energy is used to the maximum. The inclusion of radiation seeks to compare its influence and also simplify the computational times to perform mesh analysis. An analysis is carried out with simplified geometries and with experimental data to corroborate the selection of the models to be used, and it is obtained that for turbulence, the appropriate one is the standard k - w. As a means of verification, a general energy balance is made and compared with the results of the numerical analysis, where the error is 1.67%, which is considered acceptable. From the approach to improvement options, it was found that with the implementation of fins, heat can be increased by up to 7.3%.Keywords: CFD analysis, biomass, heat transfer, radiation, OpenFOAM
Procedia PDF Downloads 12026986 Analysis of Dynamics Underlying the Observation Time Series by Using a Singular Spectrum Approach
Authors: O. Delage, H. Bencherif, T. Portafaix, A. Bourdier
Abstract:
The main purpose of time series analysis is to learn about the dynamics behind some time ordered measurement data. Two approaches are used in the literature to get a better knowledge of the dynamics contained in observation data sequences. The first of these approaches concerns time series decomposition, which is an important analysis step allowing patterns and behaviors to be extracted as components providing insight into the mechanisms producing the time series. As in many cases, time series are short, noisy, and non-stationary. To provide components which are physically meaningful, methods such as Empirical Mode Decomposition (EMD), Empirical Wavelet Transform (EWT) or, more recently, Empirical Adaptive Wavelet Decomposition (EAWD) have been proposed. The second approach is to reconstruct the dynamics underlying the time series as a trajectory in state space by mapping a time series into a set of Rᵐ lag vectors by using the method of delays (MOD). Takens has proved that the trajectory obtained with the MOD technic is equivalent to the trajectory representing the dynamics behind the original time series. This work introduces the singular spectrum decomposition (SSD), which is a new adaptive method for decomposing non-linear and non-stationary time series in narrow-banded components. This method takes its origin from singular spectrum analysis (SSA), a nonparametric spectral estimation method used for the analysis and prediction of time series. As the first step of SSD is to constitute a trajectory matrix by embedding a one-dimensional time series into a set of lagged vectors, SSD can also be seen as a reconstruction method like MOD. We will first give a brief overview of the existing decomposition methods (EMD-EWT-EAWD). The SSD method will then be described in detail and applied to experimental time series of observations resulting from total columns of ozone measurements. The results obtained will be compared with those provided by the previously mentioned decomposition methods. We will also compare the reconstruction qualities of the observed dynamics obtained from the SSD and MOD methods.Keywords: time series analysis, adaptive time series decomposition, wavelet, phase space reconstruction, singular spectrum analysis
Procedia PDF Downloads 10726985 Ripple Effect Analysis of Government Investment for Research and Development by the Artificial Neural Networks
Authors: Hwayeon Song
Abstract:
The long-term purpose of research and development (R&D) programs is to strengthen national competitiveness by developing new knowledge and technologies. Thus, it is important to determine a proper budget for government programs to maintain the vigor of R&D when the total funding is tight due to the national deficit. In this regard, a ripple effect analysis for the budgetary changes in R&D programs is necessary as well as an investigation of the current status. This study proposes a new approach using Artificial Neural Networks (ANN) for both tasks. It particularly focuses on R&D programs related to Construction and Transportation (C&T) technology in Korea. First, key factors in C&T technology are explored to draw impact indicators in three areas: economy, society, and science and technology (S&T). Simultaneously, ANN is employed to evaluate the relationship between data variables. From this process, four major components in R&D including research personnel, expenses, management, and equipment are assessed. Then the ripple effect analysis is performed to see the changes in the hypothetical future by modifying current data. Any research findings can offer an alternative strategy about R&D programs as well as a new analysis tool.Keywords: Artificial Neural Networks, construction and transportation technology, Government Research and Development, Ripple Effect
Procedia PDF Downloads 249