Search results for: nonlinear analysis
26534 Iot Device Cost Effective Storage Architecture and Real-Time Data Analysis/Data Privacy Framework
Authors: Femi Elegbeleye, Omobayo Esan, Muienge Mbodila, Patrick Bowe
Abstract:
This paper focused on cost effective storage architecture using fog and cloud data storage gateway and presented the design of the framework for the data privacy model and data analytics framework on a real-time analysis when using machine learning method. The paper began with the system analysis, system architecture and its component design, as well as the overall system operations. The several results obtained from this study on data privacy model shows that when two or more data privacy model is combined we tend to have a more stronger privacy to our data, and when fog storage gateway have several advantages over using the traditional cloud storage, from our result shows fog has reduced latency/delay, low bandwidth consumption, and energy usage when been compare with cloud storage, therefore, fog storage will help to lessen excessive cost. This paper dwelt more on the system descriptions, the researchers focused on the research design and framework design for the data privacy model, data storage, and real-time analytics. This paper also shows the major system components and their framework specification. And lastly, the overall research system architecture was shown, its structure, and its interrelationships.Keywords: IoT, fog, cloud, data analysis, data privacy
Procedia PDF Downloads 10526533 Asymptotic Analysis of the Viscous Flow through a Pipe and the Derivation of the Darcy-Weisbach Law
Authors: Eduard Marusic-Paloka
Abstract:
The Darcy-Weisbach formula is used to compute the pressure drop of the fluid in the pipe, due to the friction against the wall. Because of its simplicity, the Darcy-Weisbach formula became widely accepted by engineers and is used for laminar as well as the turbulent flows through pipes, once the method to compute the mysterious friction coefficient was derived. Particularly in the second half of the 20th century. Formula is empiric, and our goal is to derive it from the basic conservation law, via rigorous asymptotic analysis. We consider the case of the laminar flow but with significant Reynolds number. In case of the perfectly smooth pipe, the situation is trivial, as the Navier-Stokes system can be solved explicitly via the Poiseuille formula leading to the friction coefficient in the form 64/Re. For the rough pipe, the situation is more complicated and some effects of the roughness appear in the friction coefficient. We start from the Navier-Stokes system in the pipe with periodically corrugated wall and derive an asymptotic expansion for the pressure and for the velocity. We use the homogenization techniques and the boundary layer analysis. The approximation derived by formal analysis is then justified by rigorous error estimate in the norm of the appropriate Sobolev space, using the energy formulation and classical a priori estimates for the Navier-Stokes system. Our method leads to the formula for the friction coefficient. The formula involves resolution of the appropriate boundary layer problems, namely the boundary value problems for the Stokes system in an infinite band, that needs to be done numerically. However, theoretical analysis characterising their nature can be done without solving them.Keywords: Darcy-Weisbach law, pipe flow, rough boundary, Navier law
Procedia PDF Downloads 35526532 A Hierarchical Bayesian Calibration of Data-Driven Models for Composite Laminate Consolidation
Authors: Nikolaos Papadimas, Joanna Bennett, Amir Sakhaei, Timothy Dodwell
Abstract:
Composite modeling of consolidation processes is playing an important role in the process and part design by indicating the formation of possible unwanted prior to expensive experimental iterative trial and development programs. Composite materials in their uncured state display complex constitutive behavior, which has received much academic interest, and this with different models proposed. Errors from modeling and statistical which arise from this fitting will propagate through any simulation in which the material model is used. A general hyperelastic polynomial representation was proposed, which can be readily implemented in various nonlinear finite element packages. In our case, FEniCS was chosen. The coefficients are assumed uncertain, and therefore the distribution of parameters learned using Markov Chain Monte Carlo (MCMC) methods. In engineering, the approach often followed is to select a single set of model parameters, which on average, best fits a set of experiments. There are good statistical reasons why this is not a rigorous approach to take. To overcome these challenges, A hierarchical Bayesian framework was proposed in which population distribution of model parameters is inferred from an ensemble of experiments tests. The resulting sampled distribution of hyperparameters is approximated using Maximum Entropy methods so that the distribution of samples can be readily sampled when embedded within a stochastic finite element simulation. The methodology is validated and demonstrated on a set of consolidation experiments of AS4/8852 with various stacking sequences. The resulting distributions are then applied to stochastic finite element simulations of the consolidation of curved parts, leading to a distribution of possible model outputs. With this, the paper, as far as the authors are aware, represents the first stochastic finite element implementation in composite process modelling.Keywords: data-driven , material consolidation, stochastic finite elements, surrogate models
Procedia PDF Downloads 14926531 Sensitivity Analysis of Prestressed Post-Tensioned I-Girder and Deck System
Authors: Tahsin A. H. Nishat, Raquib Ahsan
Abstract:
Sensitivity analysis of design parameters of the optimization procedure can become a significant factor while designing any structural system. The objectives of the study are to analyze the sensitivity of deck slab thickness parameter obtained from both the conventional and optimum design methodology of pre-stressed post-tensioned I-girder and deck system and to compare the relative significance of slab thickness. For analysis on conventional method, the values of 14 design parameters obtained by the conventional iterative method of design of a real-life I-girder bridge project have been considered. On the other side for analysis on optimization method, cost optimization of this system has been done using global optimization methodology 'Evolutionary Operation (EVOP)'. The problem, by which optimum values of 14 design parameters have been obtained, contains 14 explicit constraints and 46 implicit constraints. For both types of design parameters, sensitivity analysis has been conducted on deck slab thickness parameter which can become too sensitive for the obtained optimum solution. Deviations of slab thickness on both the upper and lower side of its optimum value have been considered reflecting its realistic possible ranges of variations during construction. In this procedure, the remaining parameters have been kept unchanged. For small deviations from the optimum value, compliance with the explicit and implicit constraints has been examined. Variations in the cost have also been estimated. It is obtained that without violating any constraint deck slab thickness obtained by the conventional method can be increased up to 25 mm whereas slab thickness obtained by cost optimization can be increased only up to 0.3 mm. The obtained result suggests that slab thickness becomes less sensitive in case of conventional method of design. Therefore, for realistic design purpose sensitivity should be conducted for any of the design procedure of girder and deck system.Keywords: sensitivity analysis, optimum design, evolutionary operations, PC I-girder, deck system
Procedia PDF Downloads 14026530 Air Quality Analysis Using Machine Learning Models Under Python Environment
Authors: Salahaeddine Sbai
Abstract:
Air quality analysis using machine learning models is a method employed to assess and predict air pollution levels. This approach leverages the capabilities of machine learning algorithms to analyze vast amounts of air quality data and extract valuable insights. By training these models on historical air quality data, they can learn patterns and relationships between various factors such as weather conditions, pollutant emissions, and geographical features. The trained models can then be used to predict air quality levels in real-time or forecast future pollution levels. This application of machine learning in air quality analysis enables policymakers, environmental agencies, and the general public to make informed decisions regarding health, environmental impact, and mitigation strategies. By understanding the factors influencing air quality, interventions can be implemented to reduce pollution levels, mitigate health risks, and enhance overall air quality management. Climate change is having significant impacts on Morocco, affecting various aspects of the country's environment, economy, and society. In this study, we use some machine learning models under python environment to predict and analysis air quality change over North of Morocco to evaluate the climate change impact on agriculture.Keywords: air quality, machine learning models, pollution, pollutant emissions
Procedia PDF Downloads 9626529 Evaluation of Applicability of High Strength Stirrup for Prestressed Concrete Members
Authors: J.-Y. Lee, H.-S. Lim, S.-E. Kim
Abstract:
Recently, the use of high-strength materials is increasing as the construction of large structures and high-rise structures increases. This paper presents an analysis of the shear behavior of prestressed concrete members with various types of materials by simulating a finite element (FE) analysis. The analytical results indicated that the shear strength and shear failure mode were strongly influenced by not only the shear reinforcement ratio but also the yield strength of shear reinforcement and the compressive strength of concrete. Though the yield strength of shear reinforcement increased the shear strength of prestressed concrete members, there was a limit to the increase in strength because of the change of shear failure modes. According to the results of FE analysis on various parameters, the maximum yield strength of the steel stirrup that can be applied to prestressed concrete members was about 860 MPa.Keywords: prestressed concrete members, high strength reinforcing bars, high strength concrete, shear behavior
Procedia PDF Downloads 30326528 Value Analysis of Islamic Banking and Conventional Banking to Measure Value Co-Creation
Authors: Amna Javed, Hisashi Masuda, Youji Kohda
Abstract:
This study examines the value analysis in Islamic and conventional banking services in Pakistan. Many scholars have focused on co-creation of values in services but mainly economic values not non-economic. As Islamic banking is based on Islamic principles that are more concerned with non-economic values (well-being, partnership, fairness, trust worthy, and justice) than economic values as money in terms of interest. This study is important to know the providers point of view about the co-created values, because, it may be more sustainable and appropriate for today’s unpredictable socioeconomic environment. Data were collected from 4 banks (2 Islamic and 2 conventional banks). Text mining technique is applied for data analysis, and values with 100% occurrences in Islamic banking are chosen. The results reflect that Islamic banking is more centric towards non-economic values than economic values and it promotes team work and partnership concept by applying Islamic spirit and trust worthiness concept.Keywords: economic values, Islamic banking, non-economic values, value system
Procedia PDF Downloads 46926527 Design Patterns for Emergency Management Processes
Authors: Tomáš Ludík, Jiří Barta, Josef Navrátil
Abstract:
Natural or human made disasters have a significant negative impact on the environment. At the same time there is an extensive effort to support management and decision making in emergency situations by information technologies. Therefore the purpose of the paper is to propose a design patterns applicable in emergency management, enabling better analysis and design of emergency management processes and therefore easier development and deployment of information systems in the field of emergency management. It will be achieved by detailed analysis of existing emergency management legislation, contingency plans, and information systems. The result is a set of design patterns focused at emergency management processes that enable easier design of emergency plans or development of new information system. These results will have a major impact on the development of new information systems as well as to more effective and faster solving of emergencies.Keywords: analysis and design, Business Process Modelling Notation, contingency plans, design patterns, emergency management
Procedia PDF Downloads 48926526 Adapting Tools for Text Monitoring and for Scenario Analysis Related to the Field of Social Disasters
Authors: Svetlana Cojocaru, Mircea Petic, Inga Titchiev
Abstract:
Humanity faces more and more often with different social disasters, which in turn can generate new accidents and catastrophes. To mitigate their consequences, it is important to obtain early possible signals about the events which are or can occur and to prepare the corresponding scenarios that could be applied. Our research is focused on solving two problems in this domain: identifying signals related that an accident occurred or may occur and mitigation of some consequences of disasters. To solve the first problem, methods of selecting and processing texts from global network Internet are developed. Information in Romanian is of special interest for us. In order to obtain the mentioned tools, we should follow several steps, divided into preparatory stage and processing stage. Throughout the first stage, we manually collected over 724 news articles and classified them into 10 categories of social disasters. It constitutes more than 150 thousand words. Using this information, a controlled vocabulary of more than 300 keywords was elaborated, that will help in the process of classification and identification of the texts related to the field of social disasters. To solve the second problem, the formalism of Petri net has been used. We deal with the problem of inhabitants’ evacuation in useful time. The analysis methods such as reachability or coverability tree and invariants technique to determine dynamic properties of the modeled systems will be used. To perform a case study of properties of extended evacuation system by adding time, the analysis modules of PIPE such as Generalized Stochastic Petri Nets (GSPN) Analysis, Simulation, State Space Analysis, and Invariant Analysis have been used. These modules helped us to obtain the average number of persons situated in the rooms and the other quantitative properties and characteristics related to its dynamics.Keywords: lexicon of disasters, modelling, Petri nets, text annotation, social disasters
Procedia PDF Downloads 20126525 Pilot Study of Determining the Impact of Surface Subsidence at The Intersection of Cave Mining with the Surface Using an Electrical Impedance Tomography
Authors: Ariungerel Jargal
Abstract:
: Cave mining is a bulk underground mining method, which allows large low-grade deposits to be mined underground. This method involves undermining the orebody to make it collapse under its own weight into a series of chambers from which the ore extracted. It is a useful technique to extend the life of large deposits previously mined by open pits, and it is a method increasingly proposed for new mines around the world. We plan to conduct a feasibility study using Electrical impedance tomography (EIT) technology to show how much subsidence there is at the intersection with the cave mining surface. EIT is an imaging technique which uses electrical measurements at electrodes attached on the body surface to yield a cross-sectional image of conductivity changes within the object. EIT has been developed in several different applications areas as a simpler, cheaper alternative to many other imaging methods. A low frequency current is injected between pairs of electrodes while voltage measurements are collected at all other electrode pairs. In the difference EIT, images are reconstructed of the change in conductivity distribution (σ) between the acquisition of the two sets of measurements. Image reconstruction in EIT requires the solution of an ill-conditioned nonlinear inverse problem on noisy data, typically requiring make simpler assumptions or regularization. It is noted that the ratio of current to voltage represents a complex value according to Ohm’s law, and that it is theoretically possible to re-express EIT. The results of the experiment were presented on the simulation, and it was concluded that it is possible to conduct further real experiments. Drill a certain number of holes in the top wall of the cave to attach the electrodes, flow a current through them, and measure and acquire the potential through these electrodes. Appropriate values should be selected depending on the distance between the holes, the frequency and duration of the measurements, the surface characteristics and the size of the study area using an EIT device.Keywords: impedance tomography, cave mining, soil, EIT device
Procedia PDF Downloads 13026524 Thermal Analysis of Friction Stir Welded Dissimilar Materials with Different Preheating Conditions
Authors: Prashant S. Humnabad
Abstract:
The objective of this work is to carry out a thermal heat transfer analysis to obtain the time dependent temperature field in welding process friction stir welded dissimilar materials with different preheating temperature. A series of joints were made on four mm thick aluminum and steel plates. The temperature used was 100ºC, 150ºC and 200ºC. The welding operation was performed with different rotational speeds and traverse speed (1000, 1400 and 2000 rmp and 16, 20 and 25 mm/min..). In numerical model, the welded plate was modeled as the weld line is the symmetric line. The work-piece has dimensions of 100x100x4 mm. The obtained result was compared with experimental result, which shows good agreement and within the acceptable limit. The peak temperature at the weld zone increases significantly with respect to increase in process time.Keywords: FEA, thermal analysis, preheating, friction stir welding
Procedia PDF Downloads 19426523 Multivariate Genome-Wide Association Studies for Identifying Additional Loci for Myopia
Authors: Qiao Fan, Xiaobo Guo, Junxian Zhu, Xiaohu Ding, Ching-Yu Cheng, Tien-Yin Wong, Mingguang He, Heping Zhang, Xueqin Wang
Abstract:
A systematic, simultaneous analysis of multiple phenotypes in genome-wide association studies (GWASs) draws a great attention to integrate the signals from single phenotypes with increased power. However, lacking an interpretable and efficient multivariate GWAS analysis impede the application of such approach. In this study, we propose to decompose the multivariate model into a series of simple univariate models. This transformation illuminates what exactly the individual trait contributes to the significant signals from the multivariate analyses. By employing our approach in the analysis of three myopia-related endophenotypes from the Singapore Malay Eye Study (SIMES), we identify novel candidate loci which were successfully validated in an independent Guangzhou Twin Eye Study (GTES).Keywords: GWAS multivariate, multiple traits, myopia, association
Procedia PDF Downloads 22826522 Prevalence Of Listeria And Salmonella Contamination In Fda Recalled Foods
Authors: Oluwatofunmi Musa-Ajakaiye, Paul Olorunfemi M.D MPH, John Obafaiye
Abstract:
Introduction: The U.S Food and Drug Administration (FDA) reports the public notices for recalled FDA-regulated products over periods of time. It study reviewed the primary reasons for recalls of products of various types over a period of 7 years. Methods: The study analyzed data provided in the FDA’s archived recalls for the years 2010-2017. It identified the various reasons for product recalls in the categories of foods, beverages, drugs, medical devices, animal and veterinary products, and dietary supplements. Using SPSS version 29, descriptive statistics and chi-square analysis of the data were performed. Results (numbers, percentages, p-values, chi-square): Over the period of analysis, a total of 931 recalls were reported. The most frequent reason for recalls was undeclared products (36.7%). The analysis showed that the most recalled product type in the data set was foods and beverages, representing 591 of all recalled products (63.5%).In addition, it was observed that foods and beverages represent 77.2% of products recalled due to the presence of microorganisms. Also, a sub-group analysis of recall reasons of food and beverages found that the most prevalent reason for such recalls was undeclared products (50.1%) followed by Listeria (17.3%) then Salmonella (13.2%). Conclusion: This analysis shows that foods and beverages have the greatest percentages of total recalls due to the presence of undeclared products listeria contamination and Salmonella contamination. The prevalence of Salmonella and Listeria contamination suggests that there is a high risk of microbial contamination in FDA-approved products and further studies on the effects of such contamination must be conducted to ensure consumer safety.Keywords: food, beverages, listeria, salmonella, FDA, contamination, microbial
Procedia PDF Downloads 6926521 Empirical Investigation of Bullwhip Effect with Sensitivity Analysis in Supply Chain
Authors: Shoaib Yousaf
Abstract:
The main purpose of this research is to the empirical investigation of the bullwhip effect under sensitivity analysis in the two-tier supply chain. The simulation modeling technique has been applied in this research as a research methodology to see the sensitivity analysis of the bullwhip effect in the rice industry of Pakistan. The research comprises two case studies that have been chosen as a sample. The results of this research have confirmed that reduction in production delay reduces the bullwhip effect, which conforms to the time compressing paradigm and the significance of the reduction in production delay to lessen demand amplification. The result of this research also indicates that by increasing the value of time to adjust inventory decreases the bullwhip effect. Furthermore, by decreasing the value of alpha increases the damping effect of the exponential smoother, it is not surprising that it also reduces the bullwhip effect. Moreover, by reducing the value of time to work in progress also reduces the bullwhip effect. This research will help practitioners and operation managers to reduces the major costs of their products in three ways. They can reduce their i) inventory levels, ii) better utilize their capacity and iii) improve their forecasting techniques. However, this study is based on two tier supply chain, while in reality the supply chain has got many tiers. Hence, future work will be extended across more than two-tier supply chains.Keywords: bullwhip effect, rice industry, supply chain dynamics, simulation, sensitivity analysis
Procedia PDF Downloads 14826520 An Analysis System for Integrating High-Throughput Transcript Abundance Data with Metabolic Pathways in Green Algae
Authors: Han-Qin Zheng, Yi-Fan Chiang-Hsieh, Chia-Hung Chien, Wen-Chi Chang
Abstract:
As the most important non-vascular plants, algae have many research applications, including high species diversity, biofuel sources, adsorption of heavy metals and, following processing, health supplements. With the increasing availability of next-generation sequencing (NGS) data for algae genomes and transcriptomes, an integrated resource for retrieving gene expression data and metabolic pathway is essential for functional analysis and systems biology in algae. However, gene expression profiles and biological pathways are displayed separately in current resources, and making it impossible to search current databases directly to identify the cellular response mechanisms. Therefore, this work develops a novel AlgaePath database to retrieve gene expression profiles efficiently under various conditions in numerous metabolic pathways. AlgaePath, a web-based database, integrates gene information, biological pathways, and next-generation sequencing (NGS) datasets in Chlamydomonasreinhardtii and Neodesmus sp. UTEX 2219-4. Users can identify gene expression profiles and pathway information by using five query pages (i.e. Gene Search, Pathway Search, Differentially Expressed Genes (DEGs) Search, Gene Group Analysis, and Co-Expression Analysis). The gene expression data of 45 and 4 samples can be obtained directly on pathway maps in C. reinhardtii and Neodesmus sp. UTEX 2219-4, respectively. Genes that are differentially expressed between two conditions can be identified in Folds Search. Furthermore, the Gene Group Analysis of AlgaePath includes pathway enrichment analysis, and can easily compare the gene expression profiles of functionally related genes in a map. Finally, Co-Expression Analysis provides co-expressed transcripts of a target gene. The analysis results provide a valuable reference for designing further experiments and elucidating critical mechanisms from high-throughput data. More than an effective interface to clarify the transcript response mechanisms in different metabolic pathways under various conditions, AlgaePath is also a data mining system to identify critical mechanisms based on high-throughput sequencing.Keywords: next-generation sequencing (NGS), algae, transcriptome, metabolic pathway, co-expression
Procedia PDF Downloads 41026519 Studying Projection Distance and Flow Properties by Shape Variations of Foam Monitor
Authors: Hyun-Kyu Cho, Jun-Su Kim, Choon-Geun Huh, Geon Lee Young-Chul Park
Abstract:
In this study, the relationship between flow properties and fluid projection distance look into connection for shape variations of foam monitor. A numerical analysis technique for fluid analysis of a foam monitor was developed for the prediction. Shape of foam monitor the flow path of fluid flow according to the shape, The fluid losses were calculated from flow analysis result.. The modified model used the length increase model of the flow path, and straight line of the model. Inlet pressure was 7 [bar] and external was atmosphere codition. am. The results showed that the length increase model of the flow path and straight line of the model was improved in the nozzle projection distance.Keywords: injection performance, finite element method, foam monitor, Projection distance
Procedia PDF Downloads 35026518 Modelling, Assessment, and Optimisation of Rules for Selected Umgeni Water Distribution Systems
Authors: Khanyisile Mnguni, Muthukrishnavellaisamy Kumarasamy, Jeff C. Smithers
Abstract:
Umgeni Water is a water board that supplies most parts of KwaZulu Natal with bulk portable water. Currently, Umgeni Water is running its distribution system based on required reservoir levels and demands and does not consider the energy cost at different times of the day, number of pump switches, and background leakages. Including these constraints can reduce operational cost, energy usage, leakages, and increase performance. Optimising pump schedules can reduce energy usage and costs while adhering to hydraulic and operational constraints. Umgeni Water has installed an online hydraulic software, WaterNet Advisor, that allows running different operational scenarios prior to implementation in order to optimise the distribution system. This study will investigate operation scenarios using optimisation techniques and WaterNet Advisor for a local water distribution system. Based on studies reported in the literature, introducing pump scheduling optimisation can reduce energy usage by approximately 30% without any change in infrastructure. Including tariff structures in an optimisation problem can reduce pumping costs by 15%, while including leakages decreases cost by 10%, and pressure drop in the system can be up to 12 m. Genetical optimisation algorithms are widely used due to their ability to solve nonlinear, non-convex, and mixed-integer problems. Other methods such as branch and bound linear programming have also been successfully used. A suitable optimisation method will be chosen based on its efficiency. The objective of the study is to reduce energy usage, operational cost, and leakages, and the feasibility of optimal solution will be checked using the Waternet Advisor. This study will provide an overview of the optimisation of hydraulic networks and progress made to date in multi-objective optimisation for a selected sub-system operated by Umgeni Water.Keywords: energy usage, pump scheduling, WaterNet Advisor, leakages
Procedia PDF Downloads 9926517 The Prevalence and Impact of Anxiety Among Medical Students in the MENA Region: A Systematic Review, Meta-Analysis, and Meta-Regression
Authors: Kawthar F. Albasri, Abdullah M. AlHudaithi, Dana B. AlTurairi, Abdullaziz S. AlQuraini, Adoub Y. AlDerazi, Reem A. Hubail, Haitham A. Jahrami
Abstract:
Several studies have found that medical students have a significant prevalence of anxiety. The purpose of this review paper is to carefully evaluate the current research on anxiety among medical students in the MENA region and, as a result, estimate the prevalence of these disturbances. Multiple databases, including the CINAHL (Cumulative Index to Nursing and Allied Health Literature), Cochrane Library, Embase, MEDLINE (Medical Literature Analysis and Retrieval System Online), PubMed, PsycINFO (Psychological Information Database), Scopus, Web of Science, UpToDate, ClinicalTrials.gov, WHO Global Health Library, EbscoHost, ProQuest, JAMA Network, and ScienceDirect, were searched. The retrieved article reference lists were rigorously searched and rated for quality. A random effects meta-analysis was performed to compute estimates. The current meta-analysis revealed an alarming estimated pooled prevalence of anxiety (K = 46, N = 27023) of 52.5% [95%CI: 43.3%–61.6%]. A total of 62.0% [95% CI 42.9%; 78.0%] of the students (K = 18, N = 16466) suffered from anxiety during the COVID-19 pandemic, while 52.5% [95% CI 43.3%; 61.6%] had anxiety before COVID-19. Based on the GAD-7 measure, a total of 55.7% [95%CI 30.5%; 78.3%] of the students (K = 10, N = 5830) had anxiety, and a total of 54.7% of the students (K = 18, N = 12154) [95%CI 42.8%; 66.0%] had anxiety using the DASS-21 or 42 measure. Anxiety is a common issue among medical students, making it a genuine problem. Further research should be conducted post-COVD 19, with a focus on anxiety prevention and intervention initiatives for medical students.Keywords: anxiety, medical students, MENA, meta-analysis, prevalence
Procedia PDF Downloads 7626516 Crack Propagation in Concrete Gravity Dam
Authors: Faramarz Khoshnoudian
Abstract:
A seismic stability assessment of the concrete gravity dam was performed. Initially (Phase 1), a linear response spectrum analysis was performed to verify the potential for crack formation. The result shows the possibility of developing cracks in the upstream face of the dam close to the lowest gallery, which were sufficiently long that the dam would not be stable following the earthquake. The results show the dam has potentially inadequate seismic and post-earthquake resistance and recommended an update of the stability analysis.Keywords: crack propgation, concrete gravity dam, seismic, assesment
Procedia PDF Downloads 7626515 Intersubjectivity of Forensic Handwriting Analysis
Authors: Marta Nawrocka
Abstract:
In each of the legal proceedings, in which expert evidence is carried out, a major concern is the assessment of the evidential value of expert reports. Judicial institutions, while making decisions, rely heavily on the expert reports, because they usually do not possess 'special knowledge' from a certain fields of science which makes it impossible for them to verify the results presented in the processes. In handwriting studies, the standards of analysis are developed. They unify procedures used by experts in comparing signs and in constructing expert reports. However, the methods used by experts are usually of a qualitative nature. They rely on the application of knowledge and experience of expert and in effect give significant range of margin in the assessment. Moreover, the standards used by experts are still not very precise and the process of reaching the conclusions is poorly understood. The above-mentioned circumstances indicate that expert opinions in the field of handwriting analysis, for many reasons, may not be sufficiently reliable. It is assumed that this state of affairs has its source in a very low level of intersubjectivity of measuring scales and analysis procedures, which consist elements of this kind of analysis. Intersubjectivity is a feature of cognition which (in relation to methods) indicates the degree of consistency of results that different people receive using the same method. The higher the level of intersubjectivity is, the more reliable and credible the method can be considered. The aim of the conducted research was to determine the degree of intersubjectivity of the methods used by the experts from the scope of handwriting analysis. 30 experts took part in the study and each of them received two signatures, with varying degrees of readability, for analysis. Their task was to distinguish graphic characteristics in the signature, estimate the evidential value of the found characteristics and estimate the evidential value of the signature. The obtained results were compared with each other using the Alpha Krippendorff’s statistic, which numerically determines the degree of compatibility of the results (assessments) that different people receive under the same conditions using the same method. The estimation of the degree of compatibility of the experts' results for each of these tasks allowed to determine the degree of intersubjectivity of the studied method. The study showed that during the analysis, the experts identified different signature characteristics and attributed different evidential value to them. In this scope, intersubjectivity turned out to be low. In addition, it turned out that experts in various ways called and described the same characteristics, and the language used was often inconsistent and imprecise. Thus, significant differences have been noted on the basis of language and applied nomenclature. On the other hand, experts attributed a similar evidential value to the entire signature (set of characteristics), which indicates that in this range, they were relatively consistent.Keywords: forensic sciences experts, handwriting analysis, inter-rater reliability, reliability of methods
Procedia PDF Downloads 15026514 Analysis of Exponential Nonuniform Transmission Line Parameters
Authors: Mounir Belattar
Abstract:
In this paper the Analysis of voltage waves that propagate along a lossless exponential nonuniform line is presented. For this analysis the parameters of this line are assumed to be varying function of the distance x along the line from the source end. The approach is based on the tow-port networks cascading presentation to derive the ABDC parameters of transmission using Picard-Carson Method which is a powerful method in getting a power series solution for distributed network because it is easy to calculate poles and zeros and solves differential equations such as telegrapher equations by an iterative sequence. So the impedance, admittance voltage and current along the line are expanded as a Taylor series in x/l where l is the total length of the line to obtain at the end, the main transmission line parameters such as voltage response and transmission and reflexion coefficients represented by scattering parameters in frequency domain.Keywords: ABCD parameters, characteristic impedance exponential nonuniform transmission line, Picard-Carson's method, S parameters, Taylor's series
Procedia PDF Downloads 44626513 A Methodology for Investigating Public Opinion Using Multilevel Text Analysis
Authors: William Xiu Shun Wong, Myungsu Lim, Yoonjin Hyun, Chen Liu, Seongi Choi, Dasom Kim, Kee-Young Kwahk, Namgyu Kim
Abstract:
Recently, many users have begun to frequently share their opinions on diverse issues using various social media. Therefore, numerous governments have attempted to establish or improve national policies according to the public opinions captured from various social media. In this paper, we indicate several limitations of the traditional approaches to analyze public opinion on science and technology and provide an alternative methodology to overcome these limitations. First, we distinguish between the science and technology analysis phase and the social issue analysis phase to reflect the fact that public opinion can be formed only when a certain science and technology is applied to a specific social issue. Next, we successively apply a start list and a stop list to acquire clarified and interesting results. Finally, to identify the most appropriate documents that fit with a given subject, we develop a new logical filter concept that consists of not only mere keywords but also a logical relationship among the keywords. This study then analyzes the possibilities for the practical use of the proposed methodology thorough its application to discover core issues and public opinions from 1,700,886 documents comprising SNS, blogs, news, and discussions.Keywords: big data, social network analysis, text mining, topic modeling
Procedia PDF Downloads 30026512 Identifying Principle Components Affecting Competitiveness of Thai Automotive Parts Industry
Authors: Thanatip Lerttanaporn, Tuanjai Somboonwiwat, Charoenchai Khompatraporn
Abstract:
The automotive parts industry is one of the vital sectors in Thai economy and now is facing a greater competition from ASEAN Economic Community (AEC). This article identifies important factors that impact the competitiveness of Thai automotive parts industry. There are eight groups of factors with a total of 58 factors. Due to a variety of factors, the Exploratory Factor Analysis and Principle Component Analysis have been applied to classify factors into groups or principle components. The results show that there are 15 groups and four of them are critical, covering 80% of important value. These four critical groups are then used to formulate strategies to improve the competitiveness of the Thai automotive parts industry.Keywords: factor analysis, Thai automotive parts, principle components, exploratory factor, ASEAN economic community
Procedia PDF Downloads 25526511 Big Data Analysis with Rhipe
Authors: Byung Ho Jung, Ji Eun Shin, Dong Hoon Lim
Abstract:
Rhipe that integrates R and Hadoop environment made it possible to process and analyze massive amounts of data using a distributed processing environment. In this paper, we implemented multiple regression analysis using Rhipe with various data sizes of actual data. Experimental results for comparing the performance of our Rhipe with stats and biglm packages available on bigmemory, showed that our Rhipe was more fast than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases. We also compared the computing speeds of pseudo-distributed and fully-distributed modes for configuring Hadoop cluster. The results showed that fully-distributed mode was faster than pseudo-distributed mode, and computing speeds of fully-distributed mode were faster as the number of data nodes increases.Keywords: big data, Hadoop, Parallel regression analysis, R, Rhipe
Procedia PDF Downloads 50126510 Electric Field Analysis of XLPE, Cross-Linked Polyethylene Covered Aerial Line and Insulator Lashing
Authors: Jyh-Cherng Gu, Ming-Ta Yang, Dai-Ling Tsai
Abstract:
Both sparse lashing and dense lashing are applied to secure overhead XLPE (cross-linked polyethylene) covered power lines on ceramic insulators or HDPE polymer insulators. The distribution of electric field in and among the lashing wires, the XLPE power lines and insulators in normal clean condition and when conducting materials such as salt, metal particles, dust, smoke or acidic smog are present is studied in this paper. The ANSYS Maxwell commercial software is used in this study for electric field analysis. Although the simulation analysis is performed assuming ideal conditions due to the constraints of the simulation software, the result may not be the same as in real situation but still be of sufficient practical values.Keywords: electric field intensity, insulator, XLPE covered aerial line, empty
Procedia PDF Downloads 26926509 Price Effect Estimation of Tobacco on Low-wage Male Smokers: A Causal Mediation Analysis
Authors: Kawsar Ahmed, Hong Wang
Abstract:
The study's goal was to estimate the causal mediation impact of tobacco tax before and after price hikes among low-income male smokers, with a particular emphasis on the effect estimating pathways framework for continuous and dichotomous variables. From July to December 2021, a cross-sectional investigation of observational data (n=739) was collected from Bangladeshi low-wage smokers. The Quasi-Bayesian technique, binomial probit model, and sensitivity analysis using a simulation of the computational tools R mediation package had been used to estimate the effect. After a price rise for tobacco products, the average number of cigarettes or bidis sticks taken decreased from 6.7 to 4.56. Tobacco product rising prices have a direct effect on low-income people's decisions to quit or lessen their daily smoking habits of Average Causal Mediation Effect (ACME) [effect=2.31, 95 % confidence interval (C.I.) = (4.71-0.00), p<0.01], Average Direct Effect (ADE) [effect=8.6, 95 percent (C.I.) = (6.8-0.11), p<0.001], and overall significant effects (p<0.001). Tobacco smoking choice is described by the mediated proportion of income effect, which is 26.1% less of following price rise. The curve of ACME and ADE is based on observational figures of the coefficients of determination that asses the model of hypothesis as the substantial consequence after price rises in the sensitivity analysis. To reduce smoking product behaviors, price increases through taxation have a positive causal mediation with income that affects the decision to limit tobacco use and promote low-income men's healthcare policy.Keywords: causal mediation analysis, directed acyclic graphs, tobacco price policy, sensitivity analysis, pathway estimation
Procedia PDF Downloads 11726508 Characterization of Nanostructured and Conventional TiAlN and AlCrN Coated ASTM-SA213-T-11 Boiler Steel
Authors: Vikas Chawla, Buta Singh Sidhu, Amita Rani, Amit Handa
Abstract:
The main objective of the present work is microstructural and mechanical characterization of the conventional and nanostructured TiAlN and AlCrN coatings deposited on T-11 boiler steel. In case of conventional coatings, Al-Cr and Ti-Al metallic powders were deposited using plasma spray process followed by gas nitriding of the surface which was done in the lab with optimized parameters after conducting several trials on plasma-sprayed coated specimens. The physical vapor deposition process (PAPVD) was employed for depositing nanostructured TiAlN and AlCrN coatings. The field emission scanning electron microscopy (FE-SEM) with energy dispersive X-ray analysis (EDAX) attachment, X-ray diffraction (XRD) analysis, atomic force microscopy (AFM) analysis and the X-Ray mapping analysis techniques have been used to study surface and cross-sectional morphology of the coatings. The surface roughness and micro-hardness were also measured. A good adhesion of the conventional thick TiAlN and AlCrN coatings was found. The coatings under study are recommended for the applications to super-heater and re-heater tubes of the boilers based upon the outcomes of the research work.Keywords: nanostructure, physical vapour deposition, oxides, thin films, electron microscopy
Procedia PDF Downloads 14326507 One-Class Classification Approach Using Fukunaga-Koontz Transform and Selective Multiple Kernel Learning
Authors: Abdullah Bal
Abstract:
This paper presents a one-class classification (OCC) technique based on Fukunaga-Koontz Transform (FKT) for binary classification problems. The FKT is originally a powerful tool to feature selection and ordering for two-class problems. To utilize the standard FKT for data domain description problem (i.e., one-class classification), in this paper, a set of non-class samples which exist outside of positive class (target class) describing boundary formed with limited training data has been constructed synthetically. The tunnel-like decision boundary around upper and lower border of target class samples has been designed using statistical properties of feature vectors belonging to the training data. To capture higher order of statistics of data and increase discrimination ability, the proposed method, termed one-class FKT (OC-FKT), has been extended to its nonlinear version via kernel machines and referred as OC-KFKT for short. Multiple kernel learning (MKL) is a favorable family of machine learning such that tries to find an optimal combination of a set of sub-kernels to achieve a better result. However, the discriminative ability of some of the base kernels may be low and the OC-KFKT designed by this type of kernels leads to unsatisfactory classification performance. To address this problem, the quality of sub-kernels should be evaluated, and the weak kernels must be discarded before the final decision making process. MKL/OC-FKT and selective MKL/OC-FKT frameworks have been designed stimulated by ensemble learning (EL) to weight and then select the sub-classifiers using the discriminability and diversities measured by eigenvalue ratios. The eigenvalue ratios have been assessed based on their regions on the FKT subspaces. The comparative experiments, performed on various low and high dimensional data, against state-of-the-art algorithms confirm the effectiveness of our techniques, especially in case of small sample size (SSS) conditions.Keywords: ensemble methods, fukunaga-koontz transform, kernel-based methods, multiple kernel learning, one-class classification
Procedia PDF Downloads 2826506 Model-Based Software Regression Test Suite Reduction
Authors: Shiwei Deng, Yang Bao
Abstract:
In this paper, we present a model-based regression test suite reducing approach that uses EFSM model dependence analysis and probability-driven greedy algorithm to reduce software regression test suites. The approach automatically identifies the difference between the original model and the modified model as a set of elementary model modifications. The EFSM dependence analysis is performed for each elementary modification to reduce the regression test suite, and then the probability-driven greedy algorithm is adopted to select the minimum set of test cases from the reduced regression test suite that cover all interaction patterns. Our initial experience shows that the approach may significantly reduce the size of regression test suites.Keywords: dependence analysis, EFSM model, greedy algorithm, regression test
Procedia PDF Downloads 43226505 A Study on the Relationship between Nonverbal Communication, Perceived Expertise, Service Quality, and Trust: Focusing on Cosmetic Stores in Korea
Authors: Park Sung-Min, Chan Kwon Park, Kim Chae-Bogk
Abstract:
This study aims to analyze the relationship between nonverbal communication, perceived expertise, service quality, and trust. The study was conducted with clients using cosmetic stores in the Daegu area of Korea. Based on the prior study, the measurement questions are correctly amended and organized in this study. The exploration factor analysis was performed using SPSS 22 for the configured measurement questions. And PLS 2.0 was used to perform a confirmatory factor analysis and path analysis. As a result of the analysis, nonverbal communication has been categorized as physical appearance, kinesics, vocal behavior and proxemics. It has been shown that all of the factors in nonverbal communication have a (+) significant effect on perceived expertise. The degree of impact has been analyzed to influence the perceived expertise in the order in which physical appearance, vocal behavior, kinesics and proxemics. The perceived expertise was analyzed to have a (+) significant effect on the service quality. It has been shown that the service quality has a (+) significant effect on trust.Keywords: nonverbal communication, perceived expertise, service quality, trust
Procedia PDF Downloads 287