Search results for: robust estimation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3164

Search results for: robust estimation

1604 Performance of Photovoltaic Thermal Greenhouse Dryer in Composite Climate of India

Authors: G. N. Tiwari, Shyam

Abstract:

Photovoltaic thermal (PVT) roof type greenhouse dryer installed above the wind tower of SODHA BERS COMPLEX, Varanasi has been analyzed for all types of weather conditions. The product to be dried has been kept at three different trays. The upper tray receives energy from the PV cover while the bottom tray receives thermal energy from the hot air of the wind tower. The annual energy estimation has been done for the all types of weather condition of composite climate of northern India. It has been found that maximum energy saving is observed for c type of weather condition whereas minimum energy saving is observed for a type of weather condition. The energy saving on overall thermal energy basis and exergy basis are 1206.8 kWh and 360 kWh respectively for c type of weather condition. The energy saving from all types of weather condition are found to be 3175.3 kWh and 957.6 kWh on overall thermal energy and overall exergy basis respectively.

Keywords: exergy, greenhouse, photovoltaic thermal, solar dryer

Procedia PDF Downloads 401
1603 Automated Adaptions of Semantic User- and Service Profile Representations by Learning the User Context

Authors: Nicole Merkle, Stefan Zander

Abstract:

Ambient Assisted Living (AAL) describes a technological and methodological stack of (e.g. formal model-theoretic semantics, rule-based reasoning and machine learning), different aspects regarding the behavior, activities and characteristics of humans. Hence, a semantic representation of the user environment and its relevant elements are required in order to allow assistive agents to recognize situations and deduce appropriate actions. Furthermore, the user and his/her characteristics (e.g. physical, cognitive, preferences) need to be represented with a high degree of expressiveness in order to allow software agents a precise evaluation of the users’ context models. The correct interpretation of these context models highly depends on temporal, spatial circumstances as well as individual user preferences. In most AAL approaches, model representations of real world situations represent the current state of a universe of discourse at a given point in time by neglecting transitions between a set of states. However, the AAL domain currently lacks sufficient approaches that contemplate on the dynamic adaptions of context-related representations. Semantic representations of relevant real-world excerpts (e.g. user activities) help cognitive, rule-based agents to reason and make decisions in order to help users in appropriate tasks and situations. Furthermore, rules and reasoning on semantic models are not sufficient for handling uncertainty and fuzzy situations. A certain situation can require different (re-)actions in order to achieve the best results with respect to the user and his/her needs. But what is the best result? To answer this question, we need to consider that every smart agent requires to achieve an objective, but this objective is mostly defined by domain experts who can also fail in their estimation of what is desired by the user and what not. Hence, a smart agent has to be able to learn from context history data and estimate or predict what is most likely in certain contexts. Furthermore, different agents with contrary objectives can cause collisions as their actions influence the user’s context and constituting conditions in unintended or uncontrolled ways. We present an approach for dynamically updating a semantic model with respect to the current user context that allows flexibility of the software agents and enhances their conformance in order to improve the user experience. The presented approach adapts rules by learning sensor evidence and user actions using probabilistic reasoning approaches, based on given expert knowledge. The semantic domain model consists basically of device-, service- and user profile representations. In this paper, we present how this semantic domain model can be used in order to compute the probability of matching rules and actions. We apply this probability estimation to compare the current domain model representation with the computed one in order to adapt the formal semantic representation. Our approach aims at minimizing the likelihood of unintended interferences in order to eliminate conflicts and unpredictable side-effects by updating pre-defined expert knowledge according to the most probable context representation. This enables agents to adapt to dynamic changes in the environment which enhances the provision of adequate assistance and affects positively the user satisfaction.

Keywords: ambient intelligence, machine learning, semantic web, software agents

Procedia PDF Downloads 268
1602 Subband Coding and Glottal Closure Instant (GCI) Using SEDREAMS Algorithm

Authors: Harisudha Kuresan, Dhanalakshmi Samiappan, T. Rama Rao

Abstract:

In modern telecommunication applications, Glottal Closure Instants location finding is important and is directly evaluated from the speech waveform. Here, we study the GCI using Speech Event Detection using Residual Excitation and the Mean Based Signal (SEDREAMS) algorithm. Speech coding uses parameter estimation using audio signal processing techniques to model the speech signal combined with generic data compression algorithms to represent the resulting modeled in a compact bit stream. This paper proposes a sub-band coder SBC, which is a type of transform coding and its performance for GCI detection using SEDREAMS are evaluated. In SBCs code in the speech signal is divided into two or more frequency bands and each of these sub-band signal is coded individually. The sub-bands after being processed are recombined to form the output signal, whose bandwidth covers the whole frequency spectrum. Then the signal is decomposed into low and high-frequency components and decimation and interpolation in frequency domain are performed. The proposed structure significantly reduces error, and precise locations of Glottal Closure Instants (GCIs) are found using SEDREAMS algorithm.

Keywords: SEDREAMS, GCI, SBC, GOI

Procedia PDF Downloads 343
1601 Lower Risk of Ischemic Stroke in Hormone Therapy Users with Use of Chinese Herbal Medicine

Authors: Shu-Hui Wen, Wei-Chuan Chang, Hsien-Chang Wu

Abstract:

Background: Little is known about the benefits and risks of use of Chinese herbal medicine (CHM) in conditions related to hormone therapy (HT) use on the risk of ischemic stroke (IS). The aim of this study is to explore the risk of IS in menopausal women treated with HT and CHM. Materials and methods: A total of 32,441 menopausal women without surgical menopause aged 40- 65 years were selected from 2003 to 2010 using the 2-million random samples of the National Health Insurance Research Database in Taiwan. According to the medication usage of HT and CHM, we divided the current and recent users into two groups: an HT use-only group (n = 4,989) and an HT/CHM group (n = 9,265). Propensity-score matching samples (4,079 pairs) were further created to deal with confounding by indication. The adjusted hazard ratios (HR) of IS during HT or CHM treatment were estimated by the robust Cox proportional hazards model. Results: The incidence rate of IS in the HT/CHM group was significantly lower than in the HT group (4.5 vs. 12.8 per 1000 person-year, p < 0.001). Multivariate analysis results indicated that additional CHM use was significant with a lower risk of IS (HR = 0.3; 95% confidence interval, 0.21-0.43). Further subgroup analyses and sensitivity analyses had similar findings. Conclusion: We found that combined use of HT and CHM was associated with a lower risk for IS than HT use only. Further study is needed to examine possible mechanism underlying this association.

Keywords: Chinese herbal medicine, hormone therapy, ischemic stroke, menopause

Procedia PDF Downloads 347
1600 Refinery Sulfur as an Alternative Agent to Decrease Pesticide Exposure in Pistachio Orchards and Common Pistachio Psylla’s Control

Authors: Mehdi Basirat, Mohammad Rouhani, Shahla Borzouei, Majid Zarangi, Asma Abolghasemi, Mohammad Fazel Soltani, Mohammad Gorji, Mohammad Amin Samih

Abstract:

The common pistachio psylla, Agonoscena pistaciae Burckhardt and Lauterer (Hemiptera: Aphalaridae), as one of the most detrimental pests in all pistachio producing regions, causes great economic damages to pistachio trees. Nowadays, various pesticides are used to control the common pistachio psylla and robust pesticide exposure has occurred in orchards. In this study, field experiments were conducted during 2018–2021 to assess the effects of sulfur on A. pistaciae. This study compared sulfur with asafoetida extract and pesticide (acetamiprid) on A. pistaciae based on complete randomized blocks with three replications. The analysis results of variance showed that the effect of treatments on egg (F2,24 = 17.61, P = 0.00) and nymphs (F2,24 = 18.29, P = 0.00) had a significant difference at a 1% level. The results demonstrated that sulfur had the highest measure of control on eggs and nymphs significantly compared to the plant extract and pesticide (negative control). These results provide support to the potential use of sulfur as an alternative pest management tool against A. pistaciae. The results clearly indicated that sulfur could control the common pistachio psylla population for six weeks at least.

Keywords: Agonoscena pistaciae, pesticide exposure, pistachio, sulfur

Procedia PDF Downloads 152
1599 Colour Quick Response Code with High Damage Resistance Capability

Authors: Minh Nguyen

Abstract:

Today, QR or Quick Response Codes are prevalent, and mobile/smart devices can efficiently read and understand them. Therefore, we can see their appearance in many areas, such as storing web pages/websites, business phone numbers, redirecting to an app download, business location, social media. The popularity of the QR Code is mainly because of its many advantages, such as it can hold a good amount of information, is small, easy to scan and read by a general RGB camera, and it can still work with some damages on its surface. However, there are still some issues. For instance, some areas needed to be kept untouched for its successful decode (e.g., the “Finder Patterns,” the “Quiet Zone,” etc.), the capability of built-in auto-correction is not robust enough, and it is not flexible enough for many application such as Augment Reality (AR). We proposed a new Colour Quick Response Code that has several advantages over the original ones: (1) there is no untouchable area, (2) it allows up to 40% of the entire code area to be damaged, (3) it is more beneficial for Augmented Reality applications, and (4) it is back-compatible and readable by available QR Code scanners such as Pyzbar. From our experience, our Colour Quick Response Code is significantly more flexible on damage compared to the original QR Code. Our code is believed to be suitable in situations where standard 2D Barcodes fail to work, such as curved and shiny surfaces, for instance, medical blood test sample tubes and syringes.

Keywords: QR code, computer vision, image processing, 2D barcode

Procedia PDF Downloads 104
1598 International Criminal Prosecution and Core International Crimes

Authors: Ikediobi Lottanna Samuel

Abstract:

Days are gone when perpetrators of core international crimes hide under the cloak of sovereignty to go with impunity. The principle of international criminal responsibility is a reality. This move to end impunity for violation of human rights has led to the creation of international and hybrid tribunals, a permanent international criminal court, and increased prosecution of human rights violations in domestic courts. This article examines the attempts by the international community to bring perpetrators of heinous crimes to book. The work reveals the inadequacy of the current international mechanism for prosecuting core international crimes in order to end the culture of impunity and entrench the culture of accountability. It also identifies that ad hoc international criminal tribunals and the international criminal court face similar challenges ranging from lack of cooperation by nation states, non-existence of hierarchy of crimes, lack of effective enforcement mechanism, limited prosecutorial capacity and agenda, difficulty in apprehending suspects, difficulty in blending different legal tradition, absence of a coherent sentencing guideline, distant location of courts, selective indictment, etc. These challenges adversely affect the functioning of these courts. It is suggested that a more helpful way to end impunity would be to have a more robust and synergistic relationship between national, regional, and international approaches to prosecuting core international crimes.

Keywords: prosecution, criminal, international, tribunal, justice, ad hoc

Procedia PDF Downloads 206
1597 A Simple Method for Evaluation of Uniformity for Gafchromic Sheets for Film Dosimetry

Authors: Fayzan Ahmed, Saad Bin Saeed, Abdul Qadir Jangda

Abstract:

Gafchromic™ sheet are extensively used for the QA of intensity modulated radiation therapy and other in-vivo dosimetry. Intra-sheet Non-uniformity of scanner as well as film causes undesirable fluctuations which are reflected in dosimetry The aim of this study is to define a systematic and robust method to investigate the intra-sheet uniformity of the unexposed Gafchromic Sheets and the region of interest (ROI) of the scanner. Sheets of lot No#: A05151201 were scanned before and after the expiry period with the EPSON™ XL10000 scanner in the transmission mode, landscape orientation, and 72 dpi resolution. ROI of (8’x 10’ inches) equal to the sheet dimension in the center of the scanner is used to acquire images with full transmission, block transmission and with sheets in place. 500 virtual grids, created in MATALB® are imported as a macros in ImageJ (1.49m Wayne Rasband) to analyze the images. In order to remove the edge effects, the outer 86 grids are excluded from the analysis. The standard deviation of the block transmission and full transmission are 0.38% and 0.66% confirming a higher uniformity of the scanner. Expired and non-expired sheets have standard deviations of 2.18% and 1.29%, show that uniformity decreases after expiry. The results are promising and indicate a good potential of this method to be used as a uniformity check for scanner and unexposed Gafchromic sheets.

Keywords: IMRT, film dosimetry, virtual grids, uniformity

Procedia PDF Downloads 416
1596 Optimal Placement of Phasor Measurement Units (PMU) Using Mixed Integer Programming (MIP) for Complete Observability in Power System Network

Authors: Harshith Gowda K. S, Tejaskumar N, Shubhanga R. B, Gowtham N, Deekshith Gowda H. S

Abstract:

Phasor measurement units (PMU) are playing an important role in the current power system for state estimation. It is necessary to have complete observability of the power system while minimizing the cost. For this purpose, the optimal location of the phasor measurement units in the power system is essential. In a bus system, zero injection buses need to be evaluated to minimize the number of PMUs. In this paper, the optimization problem is formulated using mixed integer programming to obtain the optimal location of the PMUs with increased observability. The formulation consists of with and without zero injection bus as constraints. The formulated problem is simulated using a CPLEX solver in the GAMS software package. The proposed method is tested on IEEE 30, IEEE 39, IEEE 57, and IEEE 118 bus systems. The results obtained show that the number of PMUs required is minimal with increased observability.

Keywords: PMU, observability, mixed integer programming (MIP), zero injection buses (ZIB)

Procedia PDF Downloads 153
1595 Identification of CLV for Online Shoppers Using RFM Matrix: A Case Based on Features of B2C Architecture

Authors: Riktesh Srivastava

Abstract:

Online Shopping have established an astonishing evolution in the last few years. And it is now apparent that B2C architecture is becoming progressively imperative channel for even traditional brick and mortar type traders as well. In this completion knowing customers and predicting behavior are extremely important. More important, when any customer logs onto the B2C architecture, the traces of their buying patterns can be stored and used for future predictions. Such a prediction is called Customer Lifetime Value (CLV). Earlier, we used Net Present Value to do so, however, it ignores two important aspects of B2C architecture, “market risks” and “big amount of customer data”. Now, we use RFM- Recency, Frequency and Monetary Value to estimate the CLV, and as the term exemplifies, market risks, is well sheltered. Big Data Analysis is also roofed in RFM, which gives real exploration of the Big Data and lead to a better estimation for future cash flow from customers. In the present paper, 6 factors (collected from varied sources) are used to determine as to what attracts the customers to the B2C architecture. For these 6 factors, RFM is computed for 3 years (2013, 2014 and 2015) respectively. CLV and Revenue are the two parameters defined using RFM analysis, which gives the clear picture of the future predictions.

Keywords: CLV, RFM, revenue, recency, frequency, monetary value

Procedia PDF Downloads 212
1594 The Different Improvement of Numerical Magnitude and Spatial Representation of Numbers to Symbolic Approximate Arithmetic: A Training Study of Preschooler

Authors: Yu Liang, Wei Wei

Abstract:

Spatial representation of numbers and numerical magnitude are important for preschoolers’ mathematical ability. Mental number line, a typical index to measure numbers spatial representation, and numerical comparison are both related to arithmetic obviously. However, they seem to rely on different mechanisms and probably influence arithmetic through different mechanisms. In line with this idea, preschool children were trained with two tasks to investigate which one is more important for approximate arithmetic. The training of numerical processing and number line estimation were proved to be effective. They both improved the ability of approximate arithmetic. When the difficulty of approximate arithmetic was taken into account, the performance in number line training group was not significantly different among three levels. However, two harder levels achieved significance in numerical comparison training group. Thus, comparing spatial representation ability, symbolic approximation arithmetic relies more on numerical magnitude. Educational implications of the study were discussed.

Keywords: approximate arithmetic, mental number line, numerical magnitude, preschooler

Procedia PDF Downloads 236
1593 Change Point Detection Using Random Matrix Theory with Application to Frailty in Elderly Individuals

Authors: Malika Kharouf, Aly Chkeir, Khac Tuan Huynh

Abstract:

Detecting change points in time series data is a challenging problem, especially in scenarios where there is limited prior knowledge regarding the data’s distribution and the nature of the transitions. We present a method designed for detecting changes in the covariance structure of high-dimensional time series data, where the number of variables closely matches the data length. Our objective is to achieve unbiased test statistic estimation under the null hypothesis. We delve into the utilization of Random Matrix Theory to analyze the behavior of our test statistic within a high-dimensional context. Specifically, we illustrate that our test statistic converges pointwise to a normal distribution under the null hypothesis. To assess the effectiveness of our proposed approach, we conduct evaluations on a simulated dataset. Furthermore, we employ our method to examine changes aimed at detecting frailty in the elderly.

Keywords: change point detection, hypothesis tests, random matrix theory, frailty in elderly

Procedia PDF Downloads 28
1592 Periodicity Analysis of Long-Term Waterquality Data Series of the Hungarian Section of the River Tisza Using Morlet Wavelet Spectrum Estimation

Authors: Péter Tanos, József Kovács, Angéla Anda, Gábor Várbíró, Sándor Molnár, István Gábor Hatvani

Abstract:

The River Tisza is the second largest river in Central Europe. In this study, Morlet wavelet spectrum (periodicity) analysis was used with chemical, biological and physical water quality data for the Hungarian section of the River Tisza. In the research 15, water quality parameters measured at 14 sampling sites in the River Tisza and 4 sampling sites in the main artificial changes were assessed for the time period 1993 - 2005. Results show that annual periodicity was not always to be found in the water quality parameters, at least at certain sampling sites. Periodicity was found to vary over space and time, but in general, an increase was observed in the company of higher trophic states of the river heading downstream.

Keywords: annual periodicity water quality, spatiotemporal variability of periodic behavior, Morlet wavelet spectrum analysis, River Tisza

Procedia PDF Downloads 333
1591 Effects of Cell Phone Electromagnetic Radiation on the Brain System

Authors: A. Alao Olumuyiwa

Abstract:

Health hazards reported to be associated with exposure to electromagnetic radiations which include brain tumors, genotoxic effects, neurological effects, immune system deregulation, allergic responses and some cardiovascular effects are discussed under a closed tabular model in this study. This review however showed that there is strong and robust evidence that chronic exposures to electromagnetic frequency across the spectrum, through strength, consistency, biological plausibility and many dose-response relationships, may result in brain cancer and other carcinogenic disease symptoms. There is therefore no safe threshold because of the genotoxic nature of the mechanism that may however be involved. The discussed study explains that the cell phone has induced effects upon the blood –brain barrier permeability and the cerebellum exposure to continuous long hours RF radiation may result in significant increase in albumin extravasations. A physical Biomodeling approach is however employed to review this health effects using Specific Absorption Rate (SAR) of different GSM machines to critically examine the symptoms such as a decreased loco motor activity, increased grooming and reduced memory functions in a variety of animal spices in classified grouped and sub grouped models.

Keywords: brain cancer, electromagnetic radiations, physical biomodeling, specific absorption rate (SAR)

Procedia PDF Downloads 338
1590 Horizontal Stress Magnitudes Using Poroelastic Model in Upper Assam Basin, India

Authors: Jenifer Alam, Rima Chatterjee

Abstract:

Upper Assam sedimentary basin is one of the oldest commercially producing basins of India. Being in a tectonically active zone, estimation of tectonic strain and stress magnitudes has vast application in hydrocarbon exploration and exploitation. This East North East –West South West trending shelf-slope basin encompasses the Bramhaputra valley extending from Mikir Hills in the southwest to the Naga foothills in the northeast. Assam Shelf lying between the Main Boundary Thrust (MBT) and Naga Thrust area is comparatively free from thrust tectonics and depicts normal faulting mechanism. The study area is bounded by the MBT and Main Central Thrust in the northwest. The Belt of Schuppen in the southeast, is bordered by Naga and Disang thrust marking the lower limit of the study area. The entire Assam basin shows low-level seismicity compared to other regions of northeast India. Pore pressure (PP), vertical stress magnitude (SV) and horizontal stress magnitudes have been estimated from two wells - N1 and T1 located in Upper Assam. N1 is located in the Assam gap below the Bramhaputra river while T1, lies in the Belt of Schuppen. N1 penetrates geological formations from top Alluvial through Dhekiajuli, Girujan, Tipam, Barail, Kopili, Sylhet and Langpur to the granitic basement while T1 in trusted zone crosses through Girujan Suprathrust, Tipam Suprathrust, Barail Suprathrust to reach Naga Thrust. Normal compaction trend is drawn through shale points through both wells for estimation of PP using the conventional Eaton sonic equation with an exponent of 1.0 which is validated with Modular Dynamic Tester and mud weight. Observed pore pressure gradient ranges from 10.3 MPa/km to 11.1 MPa/km. The SV has a gradient from 22.20 to 23.80 MPa/km. Minimum and maximum horizontal principal stress (Sh and SH) magnitudes under isotropic conditions are determined using poroelastic model. This approach determines biaxial tectonic strain utilizing static Young’s Modulus, Poisson’s Ratio, SV, PP, leak off test (LOT) and SH derived from breakouts using prior information on unconfined compressive strength. Breakout derived SH information is used for obtaining tectonic strain due to lack of measured SH data from minifrac or hydrofracturing. Tectonic strain varies from 0.00055 to 0.00096 along x direction and from -0.0010 to 0.00042 along y direction. After obtaining tectonic strains at each well, the principal horizontal stress magnitudes are calculated from linear poroelastic model. The magnitude of Sh and SH gradient in normal faulting region are 12.5 and 16.0 MPa/km while in thrust faulted region the gradients are 17.4 and 20.2 MPa/km respectively. Model predicted Sh and SH matches well with the LOT data and breakout derived SH data in both wells. It is observed from this study that the stresses SV>SH>Sh prevailing in the shelf region while near the Naga foothills the regime changes to SH≈SV>Sh area corresponds to normal faulting regime. Hence this model is a reliable tool for predicting stress magnitudes from well logs under active tectonic regime in Upper Assam Basin.

Keywords: Eaton, strain, stress, poroelastic model

Procedia PDF Downloads 198
1589 Image Segmentation Techniques: Review

Authors: Lindani Mbatha, Suvendi Rimer, Mpho Gololo

Abstract:

Image segmentation is the process of dividing an image into several sections, such as the object's background and the foreground. It is a critical technique in both image-processing tasks and computer vision. Most of the image segmentation algorithms have been developed for gray-scale images and little research and algorithms have been developed for the color images. Most image segmentation algorithms or techniques vary based on the input data and the application. Nearly all of the techniques are not suitable for noisy environments. Most of the work that has been done uses the Markov Random Field (MRF), which involves the computations and is said to be robust to noise. In the past recent years' image segmentation has been brought to tackle problems such as easy processing of an image, interpretation of the contents of an image, and easy analysing of an image. This article reviews and summarizes some of the image segmentation techniques and algorithms that have been developed in the past years. The techniques include neural networks (CNN), edge-based techniques, region growing, clustering, and thresholding techniques and so on. The advantages and disadvantages of medical ultrasound image segmentation techniques are also discussed. The article also addresses the applications and potential future developments that can be done around image segmentation. This review article concludes with the fact that no technique is perfectly suitable for the segmentation of all different types of images, but the use of hybrid techniques yields more accurate and efficient results.

Keywords: clustering-based, convolution-network, edge-based, region-growing

Procedia PDF Downloads 77
1588 Integration GIS–SCADA Power Systems to Enclosure Air Dispersion Model

Authors: Ibrahim Shaker, Amr El Hossany, Moustafa Osman, Mohamed El Raey

Abstract:

This paper will explore integration model between GIS–SCADA system and enclosure quantification model to approach the impact of failure-safe event. There are real demands to identify spatial objects and improve control system performance. Nevertheless, the employed methodology is predicting electro-mechanic operations and corresponding time to environmental incident variations. Open processing, as object systems technology, is presented for integration enclosure database with minimal memory size and computation time via connectivity drivers such as ODBC:JDBC during main stages of GIS–SCADA connection. The function of Geographic Information System is manipulating power distribution in contrast to developing issues. In other ward, GIS-SCADA systems integration will require numerical objects of process to enable system model calibration and estimation demands, determine of past events for analysis and prediction of emergency situations for response training.

Keywords: air dispersion model, environmental management, SCADA systems, GIS system, integration power system

Procedia PDF Downloads 354
1587 BIM Data and Digital Twin Framework: Preserving the Past and Predicting the Future

Authors: Mazharuddin Syed Ahmed

Abstract:

This research presents a framework used to develop The Ara Polytechnic College of Architecture Studies building “Kahukura” which is Green Building certified. This framework integrates the development of a smart building digital twin by utilizing Building Information Modelling (BIM) and its BIM maturity levels, including Levels of Development (LOD), eight dimensions of BIM, Heritage-BIM (H-BIM) and Facility Management BIM (FM BIM). The research also outlines a structured approach to building performance analysis and integration with the circular economy, encapsulated within a five-level digital twin framework. Starting with Level 1, the Descriptive Twin provides a live, editable visual replica of the built asset, allowing for specific data inclusion and extraction. Advancing to Level 2, the Informative Twin integrates operational and sensory data, enhancing data verification and system integration. At Level 3, the Predictive Twin utilizes operational data to generate insights and proactive management suggestions. Progressing to Level 4, the Comprehensive Twin simulates future scenarios, enabling robust “what-if” analyses. Finally, Level 5, the Autonomous Twin, represents the pinnacle of digital twin evolution, capable of learning and autonomously acting on behalf of users.

Keywords: building information modelling, circular economy integration, digital twin, predictive analytics

Procedia PDF Downloads 33
1586 Government (Big) Data Ecosystem: Definition, Classification of Actors, and Their Roles

Authors: Syed Iftikhar Hussain Shah, Vasilis Peristeras, Ioannis Magnisalis

Abstract:

Organizations, including governments, generate (big) data that are high in volume, velocity, veracity, and come from a variety of sources. Public Administrations are using (big) data, implementing base registries, and enforcing data sharing within the entire government to deliver (big) data related integrated services, provision of insights to users, and for good governance. Government (Big) data ecosystem actors represent distinct entities that provide data, consume data, manipulate data to offer paid services, and extend data services like data storage, hosting services to other actors. In this research work, we perform a systematic literature review. The key objectives of this paper are to propose a robust definition of government (big) data ecosystem and a classification of government (big) data ecosystem actors and their roles. We showcase a graphical view of actors, roles, and their relationship in the government (big) data ecosystem. We also discuss our research findings. We did not find too much published research articles about the government (big) data ecosystem, including its definition and classification of actors and their roles. Therefore, we lent ideas for the government (big) data ecosystem from numerous areas that include scientific research data, humanitarian data, open government data, industry data, in the literature.

Keywords: big data, big data ecosystem, classification of big data actors, big data actors roles, definition of government (big) data ecosystem, data-driven government, eGovernment, gaps in data ecosystems, government (big) data, public administration, systematic literature review

Procedia PDF Downloads 149
1585 Comparison of Unit Hydrograph Models to Simulate Flood Events at the Field Scale

Authors: Imene Skhakhfa, Lahbaci Ouerdachi

Abstract:

To ensure the overall coherence of simulated results, it is necessary to develop a robust validation process. In many applications, it is no longer content to calibrate and validate the model only in relation to the hydro graph measured at the outlet, but we try to better simulate the functioning of the watershed in space. Therefore the timing also performs compared to other variables such as water level measurements in intermediate stations or groundwater levels. As part of this work, we limit ourselves to modeling flood of short duration for which the process of evapotranspiration is negligible. The main parameters to identify the models are related to the method of unit hydro graph (HU). Three different models were tested: SNYDER, CLARK and SCS. These models differ in their mathematical structure and parameters to be calibrated while hydrological data are the same, the initial water content and precipitation. The models are compared on the basis of their performance in terms six objective criteria, three global criteria and three criteria representing volume, peak flow, and the mean square error. The first type of criteria gives more weight to strong events whereas the second considers all events to be of equal weight. The results show that the calibrated parameter values are dependent and also highlight the problems associated with the simulation of low flow events and intermittent precipitation.

Keywords: model calibration, intensity, runoff, hydrograph

Procedia PDF Downloads 477
1584 Estimation of Coefficient of Discharge of Side Trapezoidal Labyrinth Weir Using Group Method of Data Handling Technique

Authors: M. A. Ansari, A. Hussain, A. Uddin

Abstract:

A side weir is a flow diversion structure provided in the side wall of a channel to divert water from the main channel to a branch channel. The trapezoidal labyrinth weir is a special type of weir in which crest length of the weir is increased to pass higher discharge. Experimental and numerical studies related to the coefficient of discharge of trapezoidal labyrinth weir in an open channel have been presented in the present study. Group Method of Data Handling (GMDH) with the transfer function of quadratic polynomial has been used to predict the coefficient of discharge for the side trapezoidal labyrinth weir. A new model is developed for coefficient of discharge of labyrinth weir by regression method. Generalized models for predicting the coefficient of discharge for labyrinth weir using Group Method of Data Handling (GMDH) network have also been developed. The prediction based on GMDH model is more satisfactory than those given by traditional regression equations.

Keywords: discharge coefficient, group method of data handling, open channel, side labyrinth weir

Procedia PDF Downloads 150
1583 A Joint Possibilistic-Probabilistic Tool for Load Flow Uncertainty Assessment-Part II: Case Studies

Authors: Morteza Aien, Masoud Rashidinejad, Mahmud Fotuhi-Firuzabad

Abstract:

Power systems are innately uncertain systems. To face with such uncertain systems, robust uncertainty assessment tools are appealed. This paper inspects the uncertainty assessment formulation of the load flow (LF) problem considering different kinds of uncertainties, developed in its companion paper through some case studies. The proposed methodology is based on the evidence theory and joint propagation of possibilistic and probabilistic uncertainties. The load and wind power generation are considered as probabilistic uncertain variables and the electric vehicles (EVs) and gas turbine distributed generation (DG) units are considered as possibilistic uncertain variables. The cumulative distribution function (CDF) of the system output parameters obtained by the pure probabilistic method lies within the belief and plausibility functions obtained by the joint propagation approach. Furthermore, the imprecision in the DG parameters is explicitly reflected by the gap between the belief and plausibility functions. This gap, due to the epistemic uncertainty on the DG resources parameters grows as the penetration level increases.

Keywords: electric vehicles, joint possibilistic- probabilistic uncertainty modeling, uncertain load flow, wind turbine generator

Procedia PDF Downloads 419
1582 Optimizing Pediatric Pneumonia Diagnosis with Lightweight MobileNetV2 and VAE-GAN Techniques in Chest X-Ray Analysis

Authors: Shriya Shukla, Lachin Fernando

Abstract:

Pneumonia, a leading cause of mortality in young children globally, presents significant diagnostic challenges, particularly in resource-limited settings. This study presents an approach to diagnosing pediatric pneumonia using Chest X-Ray (CXR) images, employing a lightweight MobileNetV2 model enhanced with synthetic data augmentation. Addressing the challenge of dataset scarcity and imbalance, the study used a Variational Autoencoder-Generative Adversarial Network (VAE-GAN) to generate synthetic CXR images, improving the representation of normal cases in the pediatric dataset. This approach not only addresses the issues of data imbalance and scarcity prevalent in medical imaging but also provides a more accessible and reliable diagnostic tool for early pneumonia detection. The augmented data improved the model’s accuracy and generalization, achieving an overall accuracy of 95% in pneumonia detection. These findings highlight the efficacy of the MobileNetV2 model, offering a computationally efficient yet robust solution well-suited for resource-constrained environments such as mobile health applications. This study demonstrates the potential of synthetic data augmentation in enhancing medical image analysis for critical conditions like pediatric pneumonia.

Keywords: pneumonia, MobileNetV2, image classification, GAN, VAE, deep learning

Procedia PDF Downloads 70
1581 ECG Based Reliable User Identification Using Deep Learning

Authors: R. N. Begum, Ambalika Sharma, G. K. Singh

Abstract:

Identity theft has serious ramifications beyond data and personal information loss. This necessitates the implementation of robust and efficient user identification systems. Therefore, automatic biometric recognition systems are the need of the hour, and ECG-based systems are unquestionably the best choice due to their appealing inherent characteristics. The CNNs are the recent state-of-the-art techniques for ECG-based user identification systems. However, the results obtained are significantly below standards, and the situation worsens as the number of users and types of heartbeats in the dataset grows. As a result, this study proposes a highly accurate and resilient ECG-based person identification system using CNN's dense learning framework. The proposed research explores explicitly the calibre of dense CNNs in the field of ECG-based human recognition. The study tests four different configurations of dense CNN which are trained on a dataset of recordings collected from eight popular ECG databases. With the highest FAR of 0.04 percent and the highest FRR of 5%, the best performing network achieved an identification accuracy of 99.94 percent. The best network is also tested with various train/test split ratios. The findings show that DenseNets are not only extremely reliable but also highly efficient. Thus, they might also be implemented in real-time ECG-based human recognition systems.

Keywords: Biometrics, Dense Networks, Identification Rate, Train/Test split ratio

Procedia PDF Downloads 152
1580 Reducing Uncertainty of Monte Carlo Estimated Fatigue Damage in Offshore Wind Turbines Using FORM

Authors: Jan-Tore H. Horn, Jørgen Juncher Jensen

Abstract:

Uncertainties related to fatigue damage estimation of non-linear systems are highly dependent on the tail behaviour and extreme values of the stress range distribution. By using a combination of the First Order Reliability Method (FORM) and Monte Carlo simulations (MCS), the accuracy of the fatigue estimations may be improved for the same computational efforts. The method is applied to a bottom-fixed, monopile-supported large offshore wind turbine, which is a non-linear and dynamically sensitive system. Different curve fitting techniques to the fatigue damage distribution have been used depending on the sea-state dependent response characteristics, and the effect of a bi-linear S-N curve is discussed. Finally, analyses are performed on several environmental conditions to investigate the long-term applicability of this multistep method. Wave loads are calculated using state-of-the-art theory, while wind loads are applied with a simplified model based on rotor thrust coefficients.

Keywords: fatigue damage, FORM, monopile, Monte Carlo, simulation, wind turbine

Procedia PDF Downloads 247
1579 Financial Investment of a Wine Cavein Greece

Authors: Stamataki Erofili Nellie, Benardos Andreas

Abstract:

Winemaking and aging in Greece has been performed so far in special facilities, designed either as above ground or shallow underground buildings. The latter are well-known in Santorini as “canaves,” dating back to the 1700s. Canaves were mainly used for wine storage and aging, although occasionally, they included a winepress to complete there the whole wine production. On the other hand, wine caves are subterranean caves of the same use as canaves in the wine manufacturing industry, but they are excavated at a much greater depth of more than 53 meters or 175 feet. Whereas canaves or a typical wine cellar is around 10 feet deep, with is equivalent to almost 3 meters. This paper discusses the advantages and the disadvantages of creating a wine cave for the vinification of a winery in Greece and the financial investment or risk that has to be taken. The data presented and analysed are given from wineries in Greece and especially from those located in Santorini island. The estimation of the cost for the excavation of the model selected as a wine cave will be compared with the financial budget of the existing premises and facilities above ground in Greek wineries. In order to show whether it is viable for a greek winery to invest in a wine cave.

Keywords: underground space use, subterranean winery, wine cave, underground winery, greece

Procedia PDF Downloads 171
1578 Suitability of the Sport Motivation Scale–II for Use in Jr. High Physical Education: A Confirmatory Factor Analysis

Authors: Keven A. Prusak, William F. Christensen, Zack Beddoes

Abstract:

Background: For more than a decade, the Sport Motivation Scale (SMS) has been used to measure contextual motivation across a variety of sporting and physical education (PE) settings but not without criticism as to its tenability. Consequently, a new version of the sport motivation scale (SMS-II) was created to address certain weakness of the original SMS. Purpose: The purpose of this study is to assess the suitability of the SMS-II in the secondary PE setting. Methods: Three hundred and twenty (204 females, and 116 males; grades 7-9) completed the 18-item, six-subscale SMS-II at the end of a required PE class. Factor means, standard deviations, and correlations were calculated and further examined via confirmatory factor analysis (CFA). Model parameters were estimated maximum likelihood function. Results: Results indicate that participants held generally positive perceptions toward PE as a context (more so for males than females). Reliability analysis yielded adequate alphas (rα = 0.71 to 0.87, Mα = 0.78) with the exception of the AM subscale (αAM = .64). Correlation analysis indicated some support for the SIMPLEX pattern, but distal ends of the motivation continuum displayed no relationship. CFA yielded robust fit indices to the proposed structure of the SMS-II for PE. A small but significant variance across genders was noted and discussed. Conclusions: In all, the SMS-II suitably accesses PE context-specific motivational indices within Jr. High PE.

Keywords: motivation, self-determination theory, physical education, confirmatory factor analysis

Procedia PDF Downloads 316
1577 An Alternative Framework of Multi-Resolution Nested Weighted Essentially Non-Oscillatory Schemes for Solving Euler Equations with Adaptive Order

Authors: Zhenming Wang, Jun Zhu, Yuchen Yang, Ning Zhao

Abstract:

In the present paper, an alternative framework is proposed to construct a class of finite difference multi-resolution nested weighted essentially non-oscillatory (WENO) schemes with an increasingly higher order of accuracy for solving inviscid Euler equations. These WENO schemes firstly obtain a set of reconstruction polynomials by a hierarchy of nested central spatial stencils, and then recursively achieve a higher order approximation through the lower-order precision WENO schemes. The linear weights of such WENO schemes can be set as any positive numbers with a requirement that their sum equals one and they will not pollute the optimal order of accuracy in smooth regions and could simultaneously suppress spurious oscillations near discontinuities. Numerical results obtained indicate that these alternative finite-difference multi-resolution nested WENO schemes with different accuracies are very robust with low dissipation and use as few reconstruction stencils as possible while maintaining the same efficiency, achieving the high-resolution property without any equivalent multi-resolution representation. Besides, its finite volume form is easier to implement in unstructured grids.

Keywords: finite-difference, WENO schemes, high order, inviscid Euler equations, multi-resolution

Procedia PDF Downloads 133
1576 Asset Pricing Model: A Quality Paradigm

Authors: Urmi Khatri

Abstract:

Capital asset pricing model (CAPM) draws a direct relationship between the risk and the expected rate of return. There was a criticism on the beta and the assumptions of CAPM, as they are not applicable in the real world. Fama French Three Factor Model and Fama French Five Factor Model have given different factors, which have an impact on the return of any asset like size, value, investment and profitability. This study proposes to see Capital Asset pricing Model through the lenses of the quality aspect. In the study, the six factors are studied. The Fama French Five Factor Model and addition of the quality dimension are studied. Here, Graham’s seven quality and quantity criteria are measured to determine the score of the sample firms. Thus, this study tries to check the model fit. The beta coefficient of the quality dimension and the R square value is seen to determine validity of the proposed model. The sample is drawn from the firms listed on Indian Stock Exchange (BSE). For the study, only nonfinancial firms are been selected. The time period of the study is from January 1999 to December 2019. Hence, the primary objective of the study is to check how robust the model becomes after giving the quality dimension to the capital asset pricing model in addition to the size, value, profitability and investment.

Keywords: asset pricing model, CAPM, Graham’s score, G-score, multifactor model, quality

Procedia PDF Downloads 149
1575 Multi-Tooled Robotic Hand for Tele-Operation of Explosive Devices

Authors: Faik Derya Ince, Ugur Topgul, Alp Gunay, Can Bayoglu, Dante J. Dorantes-Gonzalez

Abstract:

Explosive attacks are arguably the most lethal threat that may occur in terrorist attacks. In order to counteract this issue, explosive ordnance disposal operators put their lives on the line to dispose of a possible improvised explosive device. Robots can make the disposal process more accurately and saving human lives. For this purpose, there is a demand for more accurate and dexterous manipulating robotic hands that can be teleoperated from a distance. The aim of this project is to design a robotic hand that contains two active and two passive DOF for each finger, as well as a minimum set of tools for mechanical cutting and screw driving within the same robotic hand. Both hand and toolset, are teleoperated from a distance from a haptic robotic glove in order to manipulate dangerous objects such as improvised explosive devices. SolidWorks® Computer-Aided Design, computerized dynamic simulation, and MATLAB® kinematic and static analysis were used for the robotic hand and toolset design. Novel, dexterous and robust solutions for the fingers were obtained, and six servo motors are used in total to remotely control the multi-tooled robotic hand. This project is still undergoing and presents currents results. Future research steps are also presented.

Keywords: Explosive Manipulation, Robotic Hand, Tele-Operation, Tool Integration

Procedia PDF Downloads 128