Search results for: cloud data privacy and integrity
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25493

Search results for: cloud data privacy and integrity

22553 Annual Water Level Simulation Using Support Vector Machine

Authors: Maryam Khalilzadeh Poshtegal, Seyed Ahmad Mirbagheri, Mojtaba Noury

Abstract:

In this paper, by application of the input yearly data of rainfall, temperature and flow to the Urmia Lake, the simulation of water level fluctuation were applied by means of three models. According to the climate change investigation the fluctuation of lakes water level are of high interest. This study investigate data-driven models, support vector machines (SVM), SVM method which is a new regression procedure in water resources are applied to the yearly level data of Lake Urmia that is the biggest and the hyper saline lake in Iran. The evaluated lake levels are found to be in good correlation with the observed values. The results of SVM simulation show better accuracy and implementation. The mean square errors, mean absolute relative errors and determination coefficient statistics are used as comparison criteria.

Keywords: simulation, water level fluctuation, urmia lake, support vector machine

Procedia PDF Downloads 355
22552 Dynamic Mode Decomposition and Wake Flow Modelling of a Wind Turbine

Authors: Nor Mazlin Zahari, Lian Gan, Xuerui Mao

Abstract:

The power production in wind farms and the mechanical loads on the turbines are strongly impacted by the wake of the wind turbine. Thus, there is a need for understanding and modelling the turbine wake dynamic in the wind farm and the layout optimization. Having a good wake model is important in predicting plant performance and understanding fatigue loads. In this paper, the Dynamic Mode Decomposition (DMD) was applied to the simulation data generated by a Direct Numerical Simulation (DNS) of flow around a turbine, perturbed by upstream inflow noise. This technique is useful in analyzing the wake flow, to predict its future states and to reflect flow dynamics associated with the coherent structures behind wind turbine wake flow. DMD was employed to describe the dynamic of the flow around turbine from the DNS data. Since the DNS data comes with the unstructured meshes and non-uniform grid, the interpolation of each occurring within each element in the data to obtain an evenly spaced mesh was performed before the DMD was applied. DMD analyses were able to tell us characteristics of the travelling waves behind the turbine, e.g. the dominant helical flow structures and the corresponding frequencies. As the result, the dominant frequency will be detected, and the associated spatial structure will be identified. The dynamic mode which represented the coherent structure will be presented.

Keywords: coherent structure, Direct Numerical Simulation (DNS), dominant frequency, Dynamic Mode Decomposition (DMD)

Procedia PDF Downloads 332
22551 Graph-Oriented Summary for Optimized Resource Description Framework Graphs Streams Processing

Authors: Amadou Fall Dia, Maurras Ulbricht Togbe, Aliou Boly, Zakia Kazi Aoul, Elisabeth Metais

Abstract:

Existing RDF (Resource Description Framework) Stream Processing (RSP) systems allow continuous processing of RDF data issued from different application domains such as weather station measuring phenomena, geolocation, IoT applications, drinking water distribution management, and so on. However, processing window phase often expires before finishing the entire session and RSP systems immediately delete data streams after each processed window. Such mechanism does not allow optimized exploitation of the RDF data streams as the most relevant and pertinent information of the data is often not used in a due time and almost impossible to be exploited for further analyzes. It should be better to keep the most informative part of data within streams while minimizing the memory storage space. In this work, we propose an RDF graph summarization system based on an explicit and implicit expressed needs through three main approaches: (1) an approach for user queries (SPARQL) in order to extract their needs and group them into a more global query, (2) an extension of the closeness centrality measure issued from Social Network Analysis (SNA) to determine the most informative parts of the graph and (3) an RDF graph summarization technique combining extracted user query needs and the extended centrality measure. Experiments and evaluations show efficient results in terms of memory space storage and the most expected approximate query results on summarized graphs compared to the source ones.

Keywords: centrality measures, RDF graphs summary, RDF graphs stream, SPARQL query

Procedia PDF Downloads 184
22550 Assessing the Impact of Climate Change on Pulses Production in Khyber Pakhtunkhwa, Pakistan

Authors: Khuram Nawaz Sadozai, Rizwan Ahmad, Munawar Raza Kazmi, Awais Habib

Abstract:

Climate change and crop production are intrinsically associated with each other. Therefore, this research study is designed to assess the impact of climate change on pulses production in Southern districts of Khyber Pakhtunkhwa (KP) Province of Pakistan. Two pulses (i.e. chickpea and mung bean) were selected for this research study with respect to climate change. Climatic variables such as temperature, humidity and precipitation along with pulses production and area under cultivation of pulses were encompassed as the major variables of this study. Secondary data of climatic variables and crop variables for the period of thirty four years (1986-2020) were obtained from Pakistan Metrological Department and Agriculture Statistics of KP respectively. Panel data set of chickpea and mung bean crops was estimated separately. The analysis validate that both data sets were a balanced panel data. The Hausman specification test was run separately for both the panel data sets whose findings had suggested the fixed effect model can be deemed as an appropriate model for chickpea panel data, however random effect model was appropriate for estimation of the panel data of mung bean. Major findings confirm that maximum temperature is statistically significant for the chickpea yield. This implies if maximum temperature increases by 1 0C, it can enhance the chickpea yield by 0.0463 units. However, the impact of precipitation was reported insignificant. Furthermore, the humidity was statistically significant and has a positive association with chickpea yield. In case of mung bean the minimum temperature was significantly contributing in the yield of mung bean. This study concludes that temperature and humidity can significantly contribute to enhance the pulses yield. It is recommended that capacity building of pulses growers may be made to adapt the climate change strategies. Moreover, government may ensure the availability of climate change resistant varieties of pulses to encourage the pulses cultivation.

Keywords: climate change, pulses productivity, agriculture, Pakistan

Procedia PDF Downloads 36
22549 Rasagiline Improves Metabolic Function and Reduces Tissue Injury in the Substantia Nigra in Parkinson's Disease: A Longitudinal In-Vivo Advanced MRI Study

Authors: Omar Khan, Shana Krstevska, Edwin George, Veronica Gorden, Fen Bao, Christina Caon, NP-C, Carla Santiago, Imad Zak, Navid Seraji-Bozorgzad

Abstract:

Objective: To quantify cellular injury in the substantia nigra (SN) in patients with Parkinson's disease (PD) and to examine the effect of rasagiline of tissue injury in the SN in patients with PD. Background: N-acetylaspartate (NAA) quantified with MRS is a reliable marker of neuronal metabolic function. Fractional anisotropy (FA) and mean diffusivity (MD) obtained with DTI, characterize tissue alignment and integrity. Rasagline, has been shown to exert anti-apototic effect. We applied these advanced MRI techniques to examine: (i) the effect of rasagiline on cellular injury and metabolism in patients with early PD, and (ii) longitudinal changes seen over time in PD. Methods: We conducted a prospective longitudinal study in patients with mild PD, naive to dopaminergic treatment. The imaging protocol included multi-voxel proton-MRS and DTI of the SN, acquired on a 3T scanner. Scans were performed at baseline and month 3, during which the patient was on no treatment. At that point, rasagiline 1 mg orally daily was initiated and MRI scans are were obtained at 6 and 12 months after starting rasagiline. The primary objective was to compare changes during the 3-month period of “no treatment” to the changes observed “on treatment” with rasagiline at month 12. Age-matched healthy controls were also imaged. Image analysis was performed blinded to treatment allocation and period. Results: 25 patients were enrolled in this study. Compared to the period of “no treatment”, there was significant increase in the NAA “on treatment” period (-3.04 % vs +10.95 %, p= 0.0006). Compared to the period of “no treatment”, there was significant increase in following 12 month in the FA “on treatment” (-4.8% vs +15.3%, p<0.0001). The MD increased during “no treatment” and decreased in “on treatment” (+2.8% vs -7.5%, p=0.0056). Further analysis and clinical correlation are ongoing. Conclusions: Advanced MRI techniques quantifying cellular injury in the SN in PD is a feasible approach to investigate dopaminergic neuronal injury and could be developed as an outcome in exploratory studies. Rasagiline appears to have a stabilizing effect on dopaminergic cell loss and metabolism in the SN in PD, that warrants further investigation in long-term studies.

Keywords: substantia nigra, Parkinson's disease, MRI, neuronal loss, biomarker

Procedia PDF Downloads 306
22548 Geothermal Energy Evaluation of Lower Benue Trough Using Spectral Analysis of Aeromagnetic Data

Authors: Stella C. Okenu, Stephen O. Adikwu, Martins E. Okoro

Abstract:

The geothermal energy resource potential of the Lower Benue Trough (LBT) in Nigeria was evaluated in this study using spectral analysis of high-resolution aeromagnetic (HRAM) data. The reduced to the equator aeromagnetic data was divided into sixteen (16) overlapping blocks, and each of the blocks was analyzed to obtain the radial averaged power spectrum which enabled the computation of the top and centroid depths to magnetic sources. The values were then used to assess the Curie Point Depth (CPD), geothermal gradients, and heat flow variations in the study area. Results showed that CPD varies from 7.03 to 18.23 km, with an average of 12.26 km; geothermal gradient values vary between 31.82 and 82.50°C/km, with an average of 51.21°C/km, while heat flow variations range from 79.54 to 206.26 mW/m², with an average of 128.02 mW/m². Shallow CPD zones that run from the eastern through the western and southwestern parts of the study area correspond to zones of high geothermal gradient values and high subsurface heat flow distributions. These areas signify zones associated with anomalous subsurface thermal conditions and are therefore recommended for detailed geothermal energy exploration studies.

Keywords: geothermal energy, curie-point depth, geothermal gradient, heat flow, aeromagnetic data, LBT

Procedia PDF Downloads 63
22547 Detailed Depositional Resolutions in Upper Miocene Sands of HT-3X Well, Nam Con Son Basin, Vietnam

Authors: Vo Thi Hai Quan

Abstract:

Nam Con Son sedimentary basin is one of the very important oil and gas basins in offshore Vietnam. Hai Thach field of block 05-2 contains mostly gas accumulations in fine-grained, sand/mud-rich turbidite system, which was deposited in a turbidite channel and fan environment. Major Upper Miocene reservoir of HT-3X lies above a well-developed unconformity. The main objectives of this study are to reconstruct depositional environment and to assess the reservoir quality using data from 14 meters of core samples and digital wireline data of the well HT-3X. The wireline log and core data showed that the vertical sequences of representative facies of the well mainly range from Tb to Te divisions of Bouma sequences with predominance of Tb and Tc compared to Td and Te divisions. Sediments in this well were deposited in a submarine fan association with very fine to fine-grained, homogeneous sandstones that have high porosity and permeability, high- density turbidity currents with longer transport route from the sediment source to the basin, indicating good quality of reservoir. Sediments are comprised mainly of the following sedimentary structures: massive, laminated sandstones, convoluted bedding, laminated ripples, cross-laminated ripples, deformed sandstones, contorted bedding.

Keywords: Hai Thach field, Miocene sand, turbidite, wireline data

Procedia PDF Downloads 287
22546 Impediments to Female Sports Management and Participation: The Experience in the Selected Nigeria South West Colleges of Education

Authors: Saseyi Olaitan Olaoluwa, Osifeko Olalekan Remigious

Abstract:

The study was meant to identify the impediments to female sports management and participation in the selected colleges. Seven colleges of education in the south west parts of the country were selected for the study. A total of one hundred and five subjects were sampled to supply data. Only one hundred adequately completed and returned, copies of the questionnaire were used for data analysis. The collected data were analysed descriptively. The result of the study showed that inadequate fund, personnel, facilities equipment, supplies, management of sports, supervision and coaching were some of the impediments to female sports management and participation. Athletes were not encouraged to participate. Based on the findings, it was recommended that the government should come to the aid of the colleges by providing fund and other needs that will make sports attractive for enhanced participation.

Keywords: female sports, impediments, management, Nigeria, south west, colleges

Procedia PDF Downloads 399
22545 Radiology Information System’s Mechanisms: HL7-MHS & HL7/DICOM Translation

Authors: Kulwinder Singh Mann

Abstract:

The innovative features of information system, known as Radiology Information System (RIS), for electronic medical records has shown a good impact in the hospital. The objective is to help and make their work easier; such as for a physician to access the patient’s data and for a patient to check their bill transparently. The interoperability of RIS with the other intra-hospital information systems it interacts with, dealing with the compatibility and open architecture issues, are accomplished by two novel mechanisms. The first one is the particular message handling system that is applied for the exchange of information, according to the Health Level Seven (HL7) protocol’s specifications and serves the transfer of medical and administrative data among the RIS applications and data store unit. The second one implements the translation of information between the formats that HL7 and Digital Imaging and Communication in Medicine (DICOM) protocols specify, providing the communication between RIS and Picture and Archive Communication System (PACS) which is used for the increasing incorporation of modern medical imaging equipment.

Keywords: RIS, PACS, HIS, HL7, DICOM, messaging service, interoperability, digital images

Procedia PDF Downloads 292
22544 Energy Management System and Interactive Functions of Smart Plug for Smart Home

Authors: Win Thandar Soe, Innocent Mpawenimana, Mathieu Di Fazio, Cécile Belleudy, Aung Ze Ya

Abstract:

Intelligent electronic equipment and automation network is the brain of high-tech energy management systems in critical role of smart homes dominance. Smart home is a technology integration for greater comfort, autonomy, reduced cost, and energy saving as well. These services can be provided to home owners for managing their home appliances locally or remotely and consequently allow them to automate intelligently and responsibly their consumption by individual or collective control systems. In this study, three smart plugs are described and one of them tested on typical household appliances. This article proposes to collect the data from the wireless technology and to extract some smart data for energy management system. This smart data is to quantify for three kinds of load: intermittent load, phantom load and continuous load. Phantom load is a waste power that is one of unnoticed power of each appliance while connected or disconnected to the main. Intermittent load and continuous load take in to consideration the power and using time of home appliances. By analysing the classification of loads, this smart data will be provided to reduce the communication of wireless sensor network for energy management system.

Keywords: energy management, load profile, smart plug, wireless sensor network

Procedia PDF Downloads 263
22543 The Connection between the Schwartz Theory of Basic Values and Ethical Principles in Clinical Psychology

Authors: Matej Stritesky

Abstract:

The research deals with the connection between the Schwartz Theory of Basic Values and the ethical principles in psychology, on which the meta-code of ethics the European Federation of Psychological Associations is based. The research focuses on ethically problematic situations in clinical psychology in the Czech Republic. Based on the analysis of papers that identified ethically problematic situations faced by clinical psychologists, a questionnaire of ethically problematic situations in clinical psychology (EPSCP) was created for the purposes of the research. The questionnaire was created to represent situations that correspond to the 4 principles on which the meta-code of ethics the European Federation of Psychological Associations is based. The questionnaire EPSCP consists of descriptions of 32 situations that respondents evaluate on a scale from 1 (psychologist's behaviour is ethically perfectly fine) to 10 (psychologist's behaviour is ethically completely unacceptable). The EPSCP questionnaire, together with Schwartz's PVQ questionnaire, will be presented to 60 psychology students. The relationship between principles in clinical psychology and the values on Schwartz´s value continuum will be described using multidimensional scaling. A positive correlation is assumed between the higher-order value of openness to change and problematic ethical situations related to the principle of integrity; a positive correlation between the value of the higher order of self-transcendence and the principle of respect and responsibility; a positive correlation between the value of the higher order of conservation and the principle of competence; and negative correlation between the value of the higher order of ego strengthening and sensitivity to ethically problematic situations. The research also includes an experimental part. The first half of the students are presented with the code of ethics of the Czech Association of Clinical Psychologists before completing the questionnaires, and to the second half of the students is the code of ethics presented after completing the questionnaires. In addition to reading the code of ethics, students describe the three rules of the code of ethics that they consider most important and state why they chose these rules. The output of the experimental part will be to determine whether the presentation of the code of ethics leads to greater sensitivity to ethically problematic situations.

Keywords: clinical psychology, ethically problematic situations in clinical psychology, ethical principles in psychology, Schwartz theory of basic values

Procedia PDF Downloads 107
22542 Time Series Simulation by Conditional Generative Adversarial Net

Authors: Rao Fu, Jie Chen, Shutian Zeng, Yiping Zhuang, Agus Sudjianto

Abstract:

Generative Adversarial Net (GAN) has proved to be a powerful machine learning tool in image data analysis and generation. In this paper, we propose to use Conditional Generative Adversarial Net (CGAN) to learn and simulate time series data. The conditions include both categorical and continuous variables with different auxiliary information. Our simulation studies show that CGAN has the capability to learn different types of normal and heavy-tailed distributions, as well as dependent structures of different time series. It also has the capability to generate conditional predictive distributions consistent with training data distributions. We also provide an in-depth discussion on the rationale behind GAN and the neural networks as hierarchical splines to establish a clear connection with existing statistical methods of distribution generation. In practice, CGAN has a wide range of applications in market risk and counterparty risk analysis: it can be applied to learn historical data and generate scenarios for the calculation of Value-at-Risk (VaR) and Expected Shortfall (ES), and it can also predict the movement of the market risk factors. We present a real data analysis including a backtesting to demonstrate that CGAN can outperform Historical Simulation (HS), a popular method in market risk analysis to calculate VaR. CGAN can also be applied in economic time series modeling and forecasting. In this regard, we have included an example of hypothetical shock analysis for economic models and the generation of potential CCAR scenarios by CGAN at the end of the paper.

Keywords: conditional generative adversarial net, market and credit risk management, neural network, time series

Procedia PDF Downloads 129
22541 Enhancing Large Language Models' Data Analysis Capability with Planning-and-Execution and Code Generation Agents: A Use Case for Southeast Asia Real Estate Market Analytics

Authors: Kien Vu, Jien Min Soh, Mohamed Jahangir Abubacker, Piyawut Pattamanon, Soojin Lee, Suvro Banerjee

Abstract:

Recent advances in Generative Artificial Intelligence (GenAI), in particular Large Language Models (LLMs) have shown promise to disrupt multiple industries at scale. However, LLMs also present unique challenges, notably, these so-called "hallucination" which is the generation of outputs that are not grounded in the input data that hinders its adoption into production. Common practice to mitigate hallucination problem is utilizing Retrieval Agmented Generation (RAG) system to ground LLMs'response to ground truth. RAG converts the grounding documents into embeddings, retrieve the relevant parts with vector similarity between user's query and documents, then generates a response that is not only based on its pre-trained knowledge but also on the specific information from the retrieved documents. However, the RAG system is not suitable for tabular data and subsequent data analysis tasks due to multiple reasons such as information loss, data format, and retrieval mechanism. In this study, we have explored a novel methodology that combines planning-and-execution and code generation agents to enhance LLMs' data analysis capabilities. The approach enables LLMs to autonomously dissect a complex analytical task into simpler sub-tasks and requirements, then convert them into executable segments of code. In the final step, it generates the complete response from output of the executed code. When deployed beta version on DataSense, the property insight tool of PropertyGuru, the approach yielded promising results, as it was able to provide market insights and data visualization needs with high accuracy and extensive coverage by abstracting the complexities for real-estate agents and developers from non-programming background. In essence, the methodology not only refines the analytical process but also serves as a strategic tool for real estate professionals, aiding in market understanding and enhancement without the need for programming skills. The implication extends beyond immediate analytics, paving the way for a new era in the real estate industry characterized by efficiency and advanced data utilization.

Keywords: large language model, reasoning, planning and execution, code generation, natural language processing, prompt engineering, data analysis, real estate, data sense, PropertyGuru

Procedia PDF Downloads 69
22540 Neuro-Connectivity Analysis Using Abide Data in Autism Study

Authors: Dulal Bhaumik, Fei Jie, Runa Bhaumik, Bikas Sinha

Abstract:

Human brain is an amazingly complex network. Aberrant activities in this network can lead to various neurological disorders such as multiple sclerosis, Parkinson’s disease, Alzheimer’s disease and autism. fMRI has emerged as an important tool to delineate the neural networks affected by such diseases, particularly autism. In this paper, we propose mixed-effects models together with an appropriate procedure for controlling false discoveries to detect disrupted connectivities in whole brain studies. Results are illustrated with a large data set known as Autism Brain Imaging Data Exchange or ABIDE which includes 361 subjects from 8 medical centers. We believe that our findings have addressed adequately the small sample inference problem, and thus are more reliable for therapeutic target for intervention. In addition, our result can be used for early detection of subjects who are at high risk of developing neurological disorders.

Keywords: ABIDE, autism spectrum disorder, fMRI, mixed-effects model

Procedia PDF Downloads 275
22539 Impact of Transitioning to Renewable Energy Sources on KPIs and AI Modules of data centre

Authors: Ahmed Hossam El Molla, Mohamed Hatem Saleh, Hamza Mostafa, Lara Mamdouh, Yassin Wael

Abstract:

Artificial intelligence (AI) is reshaping industries, and its potential to revolutionize renewable energy and data center operations is immense. By harnessing AI's capabilities, we can optimize energy consumption, predict fluctuations in renewable energy generation, and improve the efficiency of data center infrastructure. This convergence of technologies promises a future where energy is managed more intelligently, sustainably, and cost-effectively. The integration of AI into renewable energy systems unlocks a wealth of opportunities. Ma-chine learning algorithms can analyze vast amounts of data to forecast weather patterns, solar irradiance, and wind speeds, enabling more accurate energy production planning. AI-powered systems can optimize energy storage and grid management, ensuring a stable power supply even during intermittent renewable genera-tion. Moreover, AI can identify maintenance needs for renewable energy infrastructure, preventing costly breakdowns and maximizing system lifespan. Data centers, which consume substantial amounts of energy, are prime candidates for AI-driven optimization. AI can analyze energy consumption patterns, identify inefficien-cies, and recommend adjustments to cooling systems, server utilization, and power distribution. Predictive maintenance using AI can prevent equipment failures, reducing energy waste and downtime. Additionally, AI can optimize data placement and retrieval, minimizing energy consumption associated with data transfer. As AI transforms renewable energy and data center operations, new Key Performance Indicators (KPIs) will emerge. Traditional metrics like energy efficiency and cost-per-megawatt-hour will continue to be relevant, but additional KPIs focused on AI's impact will be essential. These might include AI-driven cost savings, predic-tive accuracy of energy generation and consumption, and the reduction of carbon emissions attributed to AI-optimized operations. By tracking these KPIs, organizations can measure the success of their AI initiatives and identify areas for improvement. Ultimately, the synergy between AI, renewable energy, and data centers holds the potential to create a more sustainable and resilient future. By embracing these technologies, we can build smarter, greener, and more efficient systems that benefit both the environment and the economy.

Keywords: data center, artificial intelligence (AI), renewable energy, energy efficiency, sustainability, optimization, predictive analytics, energy consumption, energy storage, grid management, data center optimization, key performance indicators (KPIs), carbon emissions, resiliency

Procedia PDF Downloads 5
22538 Environment Situation Analysis of Germany

Authors: K. Y. Chen, H. Chua, C. W. Kan

Abstract:

In this study, we will analyze Germany’s environmental situation such as water and air quality and review its environmental policy. In addition, we will collect the yearly environmental data as well as information concerning public environmental investment. Based on the data collect, we try to find out the relationship between public environmental investment and sustainable development in Germany. In addition, after comparing the trend of environmental quality and situation of environmental policy and investment, we may have some conclusions and learnable aspects to refer to. Based upon the data collected, it was revealed that Germany has established a well-developed institutionalization of environmental education. And the ecological culture at school is dynamic and continuous renewal. The booming of green markets in Germany is a very successful experience for learning. The green market not only creates a number of job opportunities, but also helps the government to improve and protect the environment. Acknowledgement: Authors would like to thank the financial support from the Hong Kong Polytechnic University for this work.

Keywords: Germany, public environmental investment, environment quality, sustainable development

Procedia PDF Downloads 241
22537 A Prediction Model of Adopting IPTV

Authors: Jeonghwan Jeon

Abstract:

With the advent of IPTV in the fierce competition with existing broadcasting system, it is emerged as an important issue to predict how much the adoption of IPTV service will be. This paper aims to suggest a prediction model for adopting IPTV using classification and Ranking Belief Simplex (CaRBS). A simplex plot method of representing data allows a clear visual representation to the degree of interaction of the support from the variables to the prediction of the objects. CaRBS is applied to the survey data on the IPTV adoption.

Keywords: prediction, adoption, IPTV, CaRBS

Procedia PDF Downloads 404
22536 Systematic Mapping Study of Digitization and Analysis of Manufacturing Data

Authors: R. Clancy, M. Ahern, D. O’Sullivan, K. Bruton

Abstract:

The manufacturing industry is currently undergoing a digital transformation as part of the mega-trend Industry 4.0. As part of this phase of the industrial revolution, traditional manufacturing processes are being combined with digital technologies to achieve smarter and more efficient production. To successfully digitally transform a manufacturing facility, the processes must first be digitized. This is the conversion of information from an analogue format to a digital format. The objective of this study was to explore the research area of digitizing manufacturing data as part of the worldwide paradigm, Industry 4.0. The formal methodology of a systematic mapping study was utilized to capture a representative sample of the research area and assess its current state. Specific research questions were defined to assess the key benefits and limitations associated with the digitization of manufacturing data. Research papers were classified according to the type of research and type of contribution to the research area. Upon analyzing 54 papers identified in this area, it was noted that 23 of the papers originated in Germany. This is an unsurprising finding as Industry 4.0 is originally a German strategy with supporting strong policy instruments being utilized in Germany to support its implementation. It was also found that the Fraunhofer Institute for Mechatronic Systems Design, in collaboration with the University of Paderborn in Germany, was the most frequent contributing Institution of the research papers with three papers published. The literature suggested future research directions and highlighted one specific gap in the area. There exists an unresolved gap between the data science experts and the manufacturing process experts in the industry. The data analytics expertise is not useful unless the manufacturing process information is utilized. A legitimate understanding of the data is crucial to perform accurate analytics and gain true, valuable insights into the manufacturing process. There lies a gap between the manufacturing operations and the information technology/data analytics departments within enterprises, which was borne out by the results of many of the case studies reviewed as part of this work. To test the concept of this gap existing, the researcher initiated an industrial case study in which they embedded themselves between the subject matter expert of the manufacturing process and the data scientist. Of the papers resulting from the systematic mapping study, 12 of the papers contributed a framework, another 12 of the papers were based on a case study, and 11 of the papers focused on theory. However, there were only three papers that contributed a methodology. This provides further evidence for the need for an industry-focused methodology for digitizing and analyzing manufacturing data, which will be developed in future research.

Keywords: analytics, digitization, industry 4.0, manufacturing

Procedia PDF Downloads 100
22535 Grammatically Coded Corpus of Spoken Lithuanian: Methodology and Development

Authors: L. Kamandulytė-Merfeldienė

Abstract:

The paper deals with the main issues of methodology of the Corpus of Spoken Lithuanian which was started to be developed in 2006. At present, the corpus consists of 300,000 grammatically annotated word forms. The creation of the corpus consists of three main stages: collecting the data, the transcription of the recorded data, and the grammatical annotation. Collecting the data was based on the principles of balance and naturality. The recorded speech was transcribed according to the CHAT requirements of CHILDES. The transcripts were double-checked and annotated grammatically using CHILDES. The development of the Corpus of Spoken Lithuanian has led to the constant increase in studies on spontaneous communication, and various papers have dealt with a distribution of parts of speech, use of different grammatical forms, variation of inflectional paradigms, distribution of fillers, syntactic functions of adjectives, the mean length of utterances.

Keywords: CHILDES, corpus of spoken Lithuanian, grammatical annotation, grammatical disambiguation, lexicon, Lithuanian

Procedia PDF Downloads 225
22534 Disaggregation the Daily Rainfall Dataset into Sub-Daily Resolution in the Temperate Oceanic Climate Region

Authors: Mohammad Bakhshi, Firas Al Janabi

Abstract:

High resolution rain data are very important to fulfill the input of hydrological models. Among models of high-resolution rainfall data generation, the temporal disaggregation was chosen for this study. The paper attempts to generate three different rainfall resolutions (4-hourly, hourly and 10-minutes) from daily for around 20-year record period. The process was done by DiMoN tool which is based on random cascade model and method of fragment. Differences between observed and simulated rain dataset are evaluated with variety of statistical and empirical methods: Kolmogorov-Smirnov test (K-S), usual statistics, and Exceedance probability. The tool worked well at preserving the daily rainfall values in wet days, however, the generated data are cumulated in a shorter time period and made stronger storms. It is demonstrated that the difference between generated and observed cumulative distribution function curve of 4-hourly datasets is passed the K-S test criteria while in hourly and 10-minutes datasets the P-value should be employed to prove that their differences were reasonable. The results are encouraging considering the overestimation of generated high-resolution rainfall data.

Keywords: DiMoN Tool, disaggregation, exceedance probability, Kolmogorov-Smirnov test, rainfall

Procedia PDF Downloads 193
22533 Analysis of the Statistical Characterization of Significant Wave Data Exceedances for Designing Offshore Structures

Authors: Rui Teixeira, Alan O’Connor, Maria Nogal

Abstract:

The statistical theory of extreme events is progressively a topic of growing interest in all the fields of science and engineering. The changes currently experienced by the world, economic and environmental, emphasized the importance of dealing with extreme occurrences with improved accuracy. When it comes to the design of offshore structures, particularly offshore wind turbines, the importance of efficiently characterizing extreme events is of major relevance. Extreme events are commonly characterized by extreme values theory. As an alternative, the accurate modeling of the tails of statistical distributions and the characterization of the low occurrence events can be achieved with the application of the Peak-Over-Threshold (POT) methodology. The POT methodology allows for a more refined fit of the statistical distribution by truncating the data with a minimum value of a predefined threshold u. For mathematically approximating the tail of the empirical statistical distribution the Generalised Pareto is widely used. Although, in the case of the exceedances of significant wave data (H_s) the 2 parameters Weibull and the Exponential distribution, which is a specific case of the Generalised Pareto distribution, are frequently used as an alternative. The Generalized Pareto, despite the existence of practical cases where it is applied, is not completely recognized as the adequate solution to model exceedances over a certain threshold u. References that set the Generalised Pareto distribution as a secondary solution in the case of significant wave data can be identified in the literature. In this framework, the current study intends to tackle the discussion of the application of statistical models to characterize exceedances of wave data. Comparison of the application of the Generalised Pareto, the 2 parameters Weibull and the Exponential distribution are presented for different values of the threshold u. Real wave data obtained in four buoys along the Irish coast was used in the comparative analysis. Results show that the application of the statistical distributions to characterize significant wave data needs to be addressed carefully and in each particular case one of the statistical models mentioned fits better the data than the others. Depending on the value of the threshold u different results are obtained. Other variables of the fit, as the number of points and the estimation of the model parameters, are analyzed and the respective conclusions were drawn. Some guidelines on the application of the POT method are presented. Modeling the tail of the distributions shows to be, for the present case, a highly non-linear task and, due to its growing importance, should be addressed carefully for an efficient estimation of very low occurrence events.

Keywords: extreme events, offshore structures, peak-over-threshold, significant wave data

Procedia PDF Downloads 261
22532 Nurse-Led Codes: Practical Application in the Emergency Department during a Global Pandemic

Authors: F. DelGaudio, H. Gill

Abstract:

Resuscitation during cardiopulmonary (CPA) arrest is dynamic, high stress, high acuity situation, which can easily lead to communication breakdown, and errors. The care of these high acuity patients has also been shown to increase physiologic stress and task saturation of providers, which can negatively impact the care being provided. These difficulties are further complicated during a global pandemic and pose a significant safety risk to bedside providers. Nurse-led codes are a relatively new concept that may be a potential solution for alleviating some of these difficulties. An experienced nurse who has completed advanced cardiac life support (ACLS), and additional training, assumed the responsibility of directing the mechanics of the appropriate ACLS algorithm. This was done in conjunction with a physician who also acted as a physician leader. The additional nurse-led code training included a multi-disciplinary in situ simulation of a CPA on a suspected COVID-19 patient. During the CPA, the nurse leader’s responsibilities include: ensuring adequate compression depth and rate, minimizing interruptions in chest compressions, the timing of rhythm/pulse checks, and appropriate medication administration. In addition, the nurse leader also functions as a last line safety check for appropriate personal protective equipment and limiting exposure of staff. The use of nurse-led codes for CPA has shown to decrease the cognitive overload and task saturation for the physician, as well as limiting the number of staff being exposed to a potentially infectious patient. The real-world application has allowed physicians to perform and oversee high-risk procedures such as intubation, line placement, and point of care ultrasound, without sacrificing the integrity of the resuscitation. Nurse-led codes have also given the physician the bandwidth to review pertinent medical history, advanced directives, determine reversible causes, and have the end of life conversations with family. While there is a paucity of research on the effectiveness of nurse-led codes, there are many potentially significant benefits. In addition to its value during a pandemic, it may also be beneficial during complex circumstances such as extracorporeal cardiopulmonary resuscitation.

Keywords: cardiopulmonary arrest, COVID-19, nurse-led code, task saturation

Procedia PDF Downloads 139
22531 Modeling Acceptability of a Personalized and Contextualized Radio Embedded in Vehicles

Authors: Ludivine Gueho, Sylvain Fleury, Eric Jamet

Abstract:

Driver distraction is known to be a major contributing factor of car accidents. Since many years, constructors have been designing embedded technologies to face this problem and reduce distraction. Being able to predict user acceptance would further be helpful in the development process to build appropriate systems. The present research aims at modelling the acceptability of a specific system, an innovative personalized and contextualized embedded radio, through an online survey of 202 people in France that assessed the psychological variables determining intentions to use the system. The questionnaire instantiated the dimensions of the extended version of the UTAUT acceptability model. Because of the specific features of the system assessed, we added 4 dimensions: perceived security, anxiety, trust and privacy concerns. Results showed that hedonic motivation, i.e., the fun or pleasure derived from using a technology, and performance expectancy, i.e., the degree to which individuals believe that the characteristics of the system meet their needs, are the most important dimensions in determining behavioral intentions about the innovative radio. To a lesser extent, social influence, i.e., the degree to which individuals think they can use the system while respecting their social group’s norms and while giving a positive image of themselves, had an effect on behavioral intentions. Moreover, trust, that is, the positive belief about the perceived reliability of, dependability of, and confidence in a person, object or process, had a significant effect, mediated by performance expectancy. In an applicative way, the present research reveals that, to be accepted, in-car embedded new technology has to address individual needs, for instance by facilitating the driving activity or by providing useful information. If it shows hedonic qualities by being entertaining, pretty or comfortable, this may improve the intentions to use it. Therefore, it is clearly important to include reflection about user experience in the design process. Finally, the users have to be reassured on the system’s reliability. For example, improving the transparency of the system by providing information about the system functioning, could improve trust. These results bring some highlights on determinant of acceptance of an in-vehicle technology and are useful for constructors to design acceptable systems.

Keywords: acceptability, innovative embedded radio, structural equation, user-centric evaluation, UTAUT

Procedia PDF Downloads 261
22530 Cadmium Separation from Aqueous Solutions by Natural Biosorbents

Authors: Z. V. P. Murthy, Preeti Arunachalam, Sangeeta Balram

Abstract:

Removal of metal ions from different wastewaters has become important due to their effects on living beings. Cadmium is one of the heavy metals found in different industrial wastewaters. There are many conventional methods available to remove heavy metals from wastewaters like adsorption, membrane separations, precipitation, electrolytic methods, etc. and all of them have their own advantages and disadvantages. The present work deals with the use of natural biosorbents (chitin and chitosan) to separate cadmium ions from aqueous solutions. The adsorption data were fitted with different isotherms and kinetics models. Amongst different adsorption isotherms used to fit the adsorption data, the Freundlich isotherm showed better fits for both the biosorbents. The kinetics data of adsorption of cadmium showed better fit with pseudo-second order model for both the biosorbents. Chitosan, the derivative from chitin, showed better performance than chitin. The separation results are encouraging.

Keywords: chitin, chitosan, cadmium, isotherm, kinetics

Procedia PDF Downloads 399
22529 Analysis of Airborne Data Using Range Migration Algorithm for the Spotlight Mode of Synthetic Aperture Radar

Authors: Peter Joseph Basil Morris, Chhabi Nigam, S. Ramakrishnan, P. Radhakrishna

Abstract:

This paper brings out the analysis of the airborne Synthetic Aperture Radar (SAR) data using the Range Migration Algorithm (RMA) for the spotlight mode of operation. Unlike in polar format algorithm (PFA), space-variant defocusing and geometric distortion effects are mitigated in RMA since it does not assume that the illuminating wave-fronts are planar. This facilitates the use of RMA for imaging scenarios involving severe differential range curvatures enabling the imaging of larger scenes at fine resolution and at shorter ranges with low center frequencies. The RMA algorithm for the spotlight mode of SAR is analyzed in this paper using the airborne data. Pre-processing operations viz: - range de-skew and motion compensation to a line are performed on the raw data before being fed to the RMA component. Various stages of the RMA viz:- 2D Matched Filtering, Along Track Fourier Transform and Slot Interpolation are analyzed to find the performance limits and the dependence of the imaging geometry on the resolution of the final image. The ability of RMA to compensate for severe differential range curvatures in the two-dimensional spatial frequency domain are also illustrated in this paper.

Keywords: range migration algorithm, spotlight SAR, synthetic aperture radar, matched filtering, slot interpolation

Procedia PDF Downloads 232
22528 Prevention of Student Radicalism in School through Civic Education

Authors: Triyanto

Abstract:

Radicalism poses a real threat to Indonesia's future. The target of radicalism is the youth of Indonesia. This is proven by the majority of terrorists are young people. Radicalization is not only a repressive act but also requires educational action. One of the educational efforts is civic education. This study discusses the prevention of radicalism for students through civic education and its constraints. This is qualitative research. Data were collected through literature studies, observations and in-depth interviews. Data were validated by triangulation. The sample of this research is 30 high school students in Surakarta. Data were analyzed by the interactive model of analysis from Miles & Huberman. The results show that (1) civic education can be a way of preventing student radicalism in schools in the form of cultivating the values of education through learning in the classroom and outside the classroom; (2) The obstacles encountered include the lack of learning facilities, the limited ability of teachers and the low attention of students to the civic education.

Keywords: prevention, radicalism, senior high school student, civic education

Procedia PDF Downloads 219
22527 Two-Channels Thermal Energy Storage Tank: Experiments and Short-Cut Modelling

Authors: M. Capocelli, A. Caputo, M. De Falco, D. Mazzei, V. Piemonte

Abstract:

This paper presents the experimental results and the related modeling of a thermal energy storage (TES) facility, ideated and realized by ENEA and realizing the thermocline with an innovative geometry. Firstly, the thermal energy exchange model of an equivalent shell & tube heat exchanger is described and tested to reproduce the performance of the spiral exchanger installed in the TES. Through the regression of the experimental data, a first-order thermocline model was also validated to provide an analytical function of the thermocline, useful for the performance evaluation and the comparison with other systems and implementation in simulations of integrated systems (e.g. power plants). The experimental data obtained from the plant start-up and the short-cut modeling of the system can be useful for the process analysis, for the scale-up of the thermal storage system and to investigate the feasibility of its implementation in actual case-studies.

Keywords: CSP plants, thermal energy storage, thermocline, mathematical modelling, experimental data

Procedia PDF Downloads 319
22526 Effect of Tissue Preservation Chemicals on Decomposition in Different Soil Types

Authors: Onyekachi Ogbonnaya Iroanya, Taiye Abdullahi Gegele, Frank Tochukwu Egwuatu

Abstract:

Introduction: Forensic taphonomy is a multifaceted area that incorporates decomposition, chemical and biological cadaver exposure in post-mortem event chronology and reconstruction to predict the Post Mortem Interval (PMI). The aim of this study was to evaluate the integrity of DNA extracted from the remains of embalmed decomposed Sus domesticus tissues buried in different soil types. Method: A total of 12 limbs of Sus domesticus weighing between 0.7-1.4 kg were used. Each of the samples across the groups was treated with 10% formaldehyde, absolute methanol and 50% Pine oil for 24 hours before burial except the control samples, which were buried immediately. All samples were buried in shallow simulated Clay, Sandy and Loamy soil graves for 12 months. The DNA for each sample was extracted and quantified with Nanodrop Spectrophotometer (6305 JENWAY spectrometers). The rate of decomposition was examined through the modified qualitative decomposition analysis. Extracted DNA was amplified through PCR and bands visualized via gel electrophoresis. A biochemical enzyme assay was done for each burial grave soil. Result: The limbs in all burial groups had lost weight over the burial period. There was a significant increase in the soil urease level in the samples preserved in formaldehyde across the 3 soil type groups (p≤0.01). Also, the control grave soils recorded significantly higher alkaline phosphatase, dehydrogenase and calcium carbonate values compared to experimental grave soils (p≤0.01). The experimental samples showed a significant decrease in DNA concentration and purity when compared to the control groups (p≤0.01). Obtained findings of the soil biochemical analysis showed the embalming treatment altered the relationship between organic matter decomposition and soil biochemical properties as observed in the fluctuations that were recorded in the soil biochemical parameters. The PCR amplified DNA showed no bands on the gel electrophoresis plates. Conclusion: In criminal investigations, factors such as burial grave soil, grave soil biochemical properties, antemortem exposure to embalming chemicals should be considered in post-mortem interval (PMI) determination.

Keywords: forensic taphonomy, post-mortem interval (PMI), embalmment, decomposition, grave soil

Procedia PDF Downloads 151
22525 Approach Based on Fuzzy C-Means for Band Selection in Hyperspectral Images

Authors: Diego Saqui, José H. Saito, José R. Campos, Lúcio A. de C. Jorge

Abstract:

Hyperspectral images and remote sensing are important for many applications. A problem in the use of these images is the high volume of data to be processed, stored and transferred. Dimensionality reduction techniques can be used to reduce the volume of data. In this paper, an approach to band selection based on clustering algorithms is presented. This approach allows to reduce the volume of data. The proposed structure is based on Fuzzy C-Means (or K-Means) and NWHFC algorithms. New attributes in relation to other studies in the literature, such as kurtosis and low correlation, are also considered. A comparison of the results of the approach using the Fuzzy C-Means and K-Means with different attributes is performed. The use of both algorithms show similar good results but, particularly when used attributes variance and kurtosis in the clustering process, however applicable in hyperspectral images.

Keywords: band selection, fuzzy c-means, k-means, hyperspectral image

Procedia PDF Downloads 390
22524 Development of a Remote Testing System for Performance of Gas Leakage Detectors

Authors: Gyoutae Park, Woosuk Kim, Sangguk Ahn, Seungmo Kim, Minjun Kim, Jinhan Lee, Youngdo Jo, Jongsam Moon, Hiesik Kim

Abstract:

In this research, we designed a remote system to test parameters of gas detectors such as gas concentration and initial response time. This testing system is available to measure two gas instruments simultaneously. First of all, we assembled an experimental jig with a square structure. Those parts are included with a glass flask, two high-quality cameras, and two Ethernet modems for transmitting data. This remote gas detector testing system extracts numerals from videos with continually various gas concentrations while LCDs show photographs from cameras. Extracted numeral data are received to a laptop computer through Ethernet modem. And then, the numerical data with gas concentrations and the measured initial response speeds are recorded and graphed. Our remote testing system will be diversely applied on gas detector’s test and will be certificated in domestic and international countries.

Keywords: gas leak detector, inspection instrument, extracting numerals, concentration

Procedia PDF Downloads 370