Search results for: data interpolating empirical orthogonal function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 29453

Search results for: data interpolating empirical orthogonal function

26873 Dataset Quality Index:Development of Composite Indicator Based on Standard Data Quality Indicators

Authors: Sakda Loetpiparwanich, Preecha Vichitthamaros

Abstract:

Nowadays, poor data quality is considered one of the majority costs for a data project. The data project with data quality awareness almost as much time to data quality processes while data project without data quality awareness negatively impacts financial resources, efficiency, productivity, and credibility. One of the processes that take a long time is defining the expectations and measurements of data quality because the expectation is different up to the purpose of each data project. Especially, big data project that maybe involves with many datasets and stakeholders, that take a long time to discuss and define quality expectations and measurements. Therefore, this study aimed at developing meaningful indicators to describe overall data quality for each dataset to quick comparison and priority. The objectives of this study were to: (1) Develop a practical data quality indicators and measurements, (2) Develop data quality dimensions based on statistical characteristics and (3) Develop Composite Indicator that can describe overall data quality for each dataset. The sample consisted of more than 500 datasets from public sources obtained by random sampling. After datasets were collected, there are five steps to develop the Dataset Quality Index (SDQI). First, we define standard data quality expectations. Second, we find any indicators that can measure directly to data within datasets. Thirdly, each indicator aggregates to dimension using factor analysis. Next, the indicators and dimensions were weighted by an effort for data preparing process and usability. Finally, the dimensions aggregate to Composite Indicator. The results of these analyses showed that: (1) The developed useful indicators and measurements contained ten indicators. (2) the developed data quality dimension based on statistical characteristics, we found that ten indicators can be reduced to 4 dimensions. (3) The developed Composite Indicator, we found that the SDQI can describe overall datasets quality of each dataset and can separate into 3 Level as Good Quality, Acceptable Quality, and Poor Quality. The conclusion, the SDQI provide an overall description of data quality within datasets and meaningful composition. We can use SQDI to assess for all data in the data project, effort estimation, and priority. The SDQI also work well with Agile Method by using SDQI to assessment in the first sprint. After passing the initial evaluation, we can add more specific data quality indicators into the next sprint.

Keywords: data quality, dataset quality, data quality management, composite indicator, factor analysis, principal component analysis

Procedia PDF Downloads 121
26872 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm

Authors: Ameur Abdelkader, Abed Bouarfa Hafida

Abstract:

Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.

Keywords: predictive analysis, big data, predictive analysis algorithms, CART algorithm

Procedia PDF Downloads 130
26871 Canopy Temperature Acquired from Daytime and Nighttime Aerial Data as an Indicator of Trees’ Health Status

Authors: Agata Zakrzewska, Dominik Kopeć, Adrian Ochtyra

Abstract:

The growing number of new cameras, sensors, and research methods allow for a broader application of thermal data in remote sensing vegetation studies. The aim of this research was to check whether it is possible to use thermal infrared data with a spectral range (3.6-4.9 μm) obtained during the day and the night to assess the health condition of selected species of deciduous trees in an urban environment. For this purpose, research was carried out in the city center of Warsaw (Poland) in 2020. During the airborne data acquisition, thermal data, laser scanning, and orthophoto map images were collected. Synchronously with airborne data, ground reference data were obtained for 617 studied species (Acer platanoides, Acer pseudoplatanus, Aesculus hippocastanum, Tilia cordata, and Tilia × euchlora) in different health condition states. The results were as follows: (i) healthy trees are cooler than trees in poor condition and dying both in the daytime and nighttime data; (ii) the difference in the canopy temperatures between healthy and dying trees was 1.06oC of mean value on the nighttime data and 3.28oC of mean value on the daytime data; (iii) condition classes significantly differentiate on both daytime and nighttime thermal data, but only on daytime data all condition classes differed statistically significantly from each other. In conclusion, the aerial thermal data can be considered as an alternative to hyperspectral data, a method of assessing the health condition of trees in an urban environment. Especially data obtained during the day, which can differentiate condition classes better than data obtained at night. The method based on thermal infrared and laser scanning data fusion could be a quick and efficient solution for identifying trees in poor health that should be visually checked in the field.

Keywords: middle wave infrared, thermal imagery, tree discoloration, urban trees

Procedia PDF Downloads 101
26870 Trigonelline: A Promising Compound for The Treatment of Alzheimer's Disease

Authors: Mai M. Farid, Ximeng Yang, Tomoharu Kuboyama, Chihiro Tohda

Abstract:

Trigonelline is a major alkaloid component derived from Trigonella foenum-graecum L. (fenugreek) and has been reported before as a potential neuroprotective agent, especially in Alzheimer’s disease (AD). However, the previous data were unclear and used model mice were not well established. In the present study, the effect of trigonelline on memory function was investigated in Alzheimer’s disease transgenic model mouse, 5XFAD which overexpresses the mutated APP and PS1 genes. Oral administration of trigonelline for 14 days significantly enhanced object recognition and object location memories. Plasma and cerebral cortex were isolated at 30 min, 1h, 3h, and 6 h after oral administration of trigonelline. LC-MS/MS analysis indicated that trigonelline was detected in both plasma and cortex from 30 min after, suggesting good penetration of trigonelline into the brain. In addition, trigonelline significantly ameliorated axonal and dendrite atrophy in Amyloid β-treated cortical neurons. These results suggest that trigonelline could be a promising therapeutic candidate for AD.

Keywords: alzheimer’s disease, cortical neurons, LC-MS/MS analysis, trigonelline

Procedia PDF Downloads 133
26869 Parametric Influence and Optimization of Wire-EDM on Oil Hardened Non-Shrinking Steel

Authors: Nixon Kuruvila, H. V. Ravindra

Abstract:

Wire-cut Electro Discharge Machining (WEDM) is a special form of conventional EDM process in which electrode is a continuously moving conductive wire. The present study aims at determining parametric influence and optimum process parameters of Wire-EDM using Taguchi’s Technique and Genetic algorithm. The variation of the performance parameters with machining parameters was mathematically modeled by Regression analysis method. The objective functions are Dimensional Accuracy (DA) and Material Removal Rate (MRR). Experiments were designed as per Taguchi’s L16 Orthogonal Array (OA) where in Pulse-on duration, Pulse-off duration, Current, Bed-speed and Flushing rate have been considered as the important input parameters. The matrix experiments were conducted for the material Oil Hardened Non Shrinking Steel (OHNS) having the thickness of 40 mm. The results of the study reveals that among the machining parameters it is preferable to go in for lower pulse-off duration for achieving over all good performance. Regarding MRR, OHNS is to be eroded with medium pulse-off duration and higher flush rate. Finally, the validation exercise performed with the optimum levels of the process parameters. The results confirm the efficiency of the approach employed for optimization of process parameters in this study.

Keywords: dimensional accuracy (DA), regression analysis (RA), Taguchi method (TM), volumetric material removal rate (VMRR)

Procedia PDF Downloads 395
26868 Bayesian Value at Risk Forecast Using Realized Conditional Autoregressive Expectiel Mdodel with an Application of Cryptocurrency

Authors: Niya Chen, Jennifer Chan

Abstract:

In the financial market, risk management helps to minimize potential loss and maximize profit. There are two ways to assess risks; the first way is to calculate the risk directly based on the volatility. The most common risk measurements are Value at Risk (VaR), sharp ratio, and beta. Alternatively, we could look at the quantile of the return to assess the risk. Popular return models such as GARCH and stochastic volatility (SV) focus on modeling the mean of the return distribution via capturing the volatility dynamics; however, the quantile/expectile method will give us an idea of the distribution with the extreme return value. It will allow us to forecast VaR using return which is direct information. The advantage of using these non-parametric methods is that it is not bounded by the distribution assumptions from the parametric method. But the difference between them is that expectile uses a second-order loss function while quantile regression uses a first-order loss function. We consider several quantile functions, different volatility measures, and estimates from some volatility models. To estimate the expectile of the model, we use Realized Conditional Autoregressive Expectile (CARE) model with the bayesian method to achieve this. We would like to see if our proposed models outperform existing models in cryptocurrency, and we will test it by using Bitcoin mainly as well as Ethereum.

Keywords: expectile, CARE Model, CARR Model, quantile, cryptocurrency, Value at Risk

Procedia PDF Downloads 97
26867 The Impact of Research and Development Cooperation Partner Diversity, Knowledge Source Diversity and Knowledge Source Network Embeddedness on Radical Innovation: Direct Relationships and Interaction with Non-Price Competition

Authors: Natalia Strobel, Jan Kratzer

Abstract:

In this paper, we test whether different types of research and development (R&D) alliances positively impact the radical innovation performance of firms. We differentiate between the R&D alliances without extern R&D orders and embeddedness in knowledge source network. We test the differences between the domestically diversified R&D alliances and R&D alliances diversified abroad. Moreover, we test how non-price competition influences the impact of domestically diversified R&D alliances, and R&D alliance diversified abroad on radical innovation performance. Our empirical analysis is based on the comprehensive Swiss innovation panel, which allowed us to study 3520 firms between the years between 1996 and 2011 in 3 years intervals. We analyzed the data with a linear estimation with Swamy-Aurora transformation using plm package in R software. Our results show as hypothesized a positive impact of R&D alliances diversity abroad as well as domestically on radical innovation performance. The effect of non-price interaction is in contrast to our hypothesis, not significant. This suggests that diversity of R&D alliances is highly advantageous independent of non-price competition.

Keywords: R&D alliances, partner diversity, knowledge source diversity, non-price competition, absorptive capacity

Procedia PDF Downloads 349
26866 Judicial Review of Indonesia's Position as the First Archipelagic State to implement the Traffic Separation Scheme to Establish Maritime Safety and Security

Authors: Rosmini Yanti, Safira Aviolita, Marsetio

Abstract:

Indonesia has several straits that are very important as a shipping lane, including the Sunda Strait and the Lombok Strait, which are the part of the Indonesian Archipelagic Sea Lane (IASL). An increase in traffic on the Marine Archipelago makes the task of monitoring sea routes increasingly difficult. Indonesia has proposed the establishment of a Traffic Separation Scheme (TSS) in the Sunda Strait and the Lombok Strait and the country now has the right to be able to conceptualize the TSS as well as the obligation to regulate it. Indonesia has the right to maintain national safety and sovereignty. In setting the TSS, Indonesia needs to issue national regulations that are in accordance with international law and the general provisions of the IMO (International Maritime Organization) can then be used as guidelines for maritime safety and security in the Sunda Strait and the Lombok Strait. The research method used is a qualitative method with the concept of linguistic and visual data collection. The source of the data is the analysis of documents and regulations. The results show that the determination of TSS was justified by International Law, in accordance with article 22, article 41, and article 53 of the United Nations Convention on the Law of the Sea (UNCLOS) 1982. The determination of TSS by the Indonesian government would be in accordance with COLREG (International Convention on Preventing Collisions at Sea) 10, which has been designed to follow IASL. Thus, TSS can provide a function as a safety and monitoring medium to minimize ship accidents or collisions, including the warship and aircraft of other countries that cross the IASL.

Keywords: archipelago state, maritime law, maritime security, traffic separation scheme

Procedia PDF Downloads 114
26865 Application of the Quantile Regression Approach to the Heterogeneity of the Fine Wine Prices

Authors: Charles-Olivier Amédée-Manesme, Benoit Faye, Eric Le Fur

Abstract:

In this paper, the heterogeneity of the Bordeaux Legends 50 wine market price segment is addressed. For this purpose, quantile regression is applied – with market segmentation based on wine bottle price quantile – and the hedonic price of wine attributes is computed for various price segments of the market. The approach is applied to a major privately held data set which consists of approximately 30,000 transactions over the 2003–2014 period. The findings suggest that the relative hedonic prices of several wine attributes differ significantly among deciles. In particular, the elasticity coefficient of the expert ratings shows strong variation among prices. If - as suggested in the literature - expert ratings have a positive influence on wine price on average, they have a clearly decreasing impact over the quantiles. Finally, the lower the wine price, the higher the potential for price appreciation over time. Other variables such as chateaux or vintage are also shown to vary across the distribution of wine prices. While enhancing our understanding of the complex market dynamics that underlie Bordeaux wines’ price, this research provides empirical evidence that the QR approach adequately captures heterogeneity among wine price ranges, which simultaneously applies to wine stock, vintage and auctions’ house.

Keywords: hedonics, market segmentation, quantile regression, heterogeneity, wine economics

Procedia PDF Downloads 323
26864 Reasons to Redesign: Teacher Education for a Brighter Tomorrow

Authors: Deborah L. Smith

Abstract:

To review our program and determine the best redesign options, department members gathered feedback and input through focus groups, analysis of data, and a review of the current research to ensure that the changes proposed were not based solely on the state’s new professional standards. In designing course assignments and assessments, we listened to a variety of constituents, including students, other institutions of higher learning, MDE webinars, host teachers, literacy clinic personnel, and other disciplinary experts. As a result, we are designing a program that is more inclusive of a variety of field experiences for growth. We have determined ways to improve our program by connecting academic disciplinary knowledge, educational psychology, and community building both inside and outside the classroom for professional learning communities. The state’s release of new professional standards led my department members to question what is working and what needs improvement in our program. One aspect of our program that continues to be supported by research and data analysis is the function of supervised field experiences with meaningful feedback. We seek to expand in this area. Other data indicate that we have strengths in modeling a variety of approaches such as cooperative learning, discussions, literacy strategies, and workshops. In the new program, field assignments will be connected to multiple courses, and efforts to scaffold student learning to guide them toward best evidence-based practices will be continuous. Despite running a program that meets multiple sets of standards, there are areas of need that we directly address in our redesign proposal. Technology is ever-changing, so it’s inevitable that improving digital skills is a focus. In addition, scaffolding procedures for English Language Learners (ELL) or other students who struggle is imperative. Diversity, equity, and inclusion (DEI) has been an integral part of our curriculum, but the research indicates that more self-reflection and a deeper understanding of culturally relevant practices would help the program improve. Connections with professional learning communities will be expanded, as will leadership components, so that teacher candidates understand their role in changing the face of education. A pilot program will run in academic year 22/23, and additional data will be collected each semester through evaluations and continued program review.

Keywords: DEI, field experiences, program redesign, teacher preparation

Procedia PDF Downloads 153
26863 Virtual Reality as a Tool in Modern Education

Authors: Łukasz Bis

Abstract:

The author is going to discuss virtual reality and its importance for new didactic methods. It has been known for years that experience-based education gives much better results in terms of long-term memory than theoretical study. However, practice is expensive - virtual reality allows the use of an empirical approach to learning, with minimized production costs. The author defines what makes a given VR experience appropriate (adequate) for the didactic and cognitive process. The article is a kind of a list of guidelines and their importance for the VR experience under development.

Keywords: virtual reality, education, universal design, guideline

Procedia PDF Downloads 92
26862 Advancements in Truss Design for High-Performance Facades and Roof System: A Structural Analysis

Authors: Milind Anurag

Abstract:

This study investigates cutting-edge truss design improvements, which are specifically adapted to satisfy the structural demands and difficulties associated with high-performance facades and roofs in modern architectural environments. With a growing emphasis on sustainability, energy efficiency, and eye-catching architectural aesthetics, the structural components that support these characteristics play an important part in attaining the right balance of form and function. The paper seeks to contribute to the evolution of truss design methods by combining data from these investigations, giving significant insights for architects, engineers, and researchers interested in the creation of high-performance building envelopes. The findings of this study are meant to inform future design standards and practices, promoting the development of structures that seamlessly integrate architectural innovation with structural robustness and environmental responsibility.

Keywords: truss design, high-performance, facades, finite element analysis, structural efficiency

Procedia PDF Downloads 29
26861 Hierarchical Clustering Algorithms in Data Mining

Authors: Z. Abdullah, A. R. Hamdan

Abstract:

Clustering is a process of grouping objects and data into groups of clusters to ensure that data objects from the same cluster are identical to each other. Clustering algorithms in one of the areas in data mining and it can be classified into partition, hierarchical, density based, and grid-based. Therefore, in this paper, we do a survey and review for four major hierarchical clustering algorithms called CURE, ROCK, CHAMELEON, and BIRCH. The obtained state of the art of these algorithms will help in eliminating the current problems, as well as deriving more robust and scalable algorithms for clustering.

Keywords: clustering, unsupervised learning, algorithms, hierarchical

Procedia PDF Downloads 870
26860 DNA Methylation Score Development for In utero Exposure to Paternal Smoking Using a Supervised Machine Learning Approach

Authors: Cristy Stagnar, Nina Hubig, Diana Ivankovic

Abstract:

The epigenome is a compelling candidate for mediating long-term responses to environmental effects modifying disease risk. The main goal of this research is to develop a machine learning-based DNA methylation score, which will be valuable in delineating the unique contribution of paternal epigenetic modifications to the germline impacting childhood health outcomes. It will also be a useful tool in validating self-reports of nonsmoking and in adjusting epigenome-wide DNA methylation association studies for this early-life exposure. Using secondary data from two population-based methylation profiling studies, our DNA methylation score is based on CpG DNA methylation measurements from cord blood gathered from children whose fathers smoked pre- and peri-conceptually. Each child’s mother and father fell into one of three class labels in the accompanying questionnaires -never smoker, former smoker, or current smoker. By applying different machine learning algorithms to the accessible resource for integrated epigenomic studies (ARIES) sub-study of the Avon longitudinal study of parents and children (ALSPAC) data set, which we used for training and testing of our model, the best-performing algorithm for classifying the father smoker and mother never smoker was selected based on Cohen’s κ. Error in the model was identified and optimized. The final DNA methylation score was further tested and validated in an independent data set. This resulted in a linear combination of methylation values of selected probes via a logistic link function that accurately classified each group and contributed the most towards classification. The result is a unique, robust DNA methylation score which combines information on DNA methylation and early life exposure of offspring to paternal smoking during pregnancy and which may be used to examine the paternal contribution to offspring health outcomes.

Keywords: epigenome, health outcomes, paternal preconception environmental exposures, supervised machine learning

Procedia PDF Downloads 176
26859 Effect of Social Media on Online Buyer Behavior

Authors: Zebider Asire Munyelet, Yibeltal Chanie Manie

Abstract:

In the modern digital landscape, the increase of social media platforms has become identical to the evolution of online consumer behavior. This study investigates the complicated relationship between social media and the purchasing decisions of online buyers. Through an extensive review of existing literature and empirical research, the aim is to comprehensively analyze the multidimensional impact that social media exerts on the various stages of the online buyer's journey. The investigation encompasses the exploration of how social media platforms serve as influential channels for information dissemination, product discovery, and consumer engagement. Additionally, the study investigates the psychological aspects underlying the role of social media in shaping buyer preferences, perceptions, and trust in online transactions. The methodologies employed include both quantitative and qualitative analyses, incorporating surveys, interviews, and data analytics to derive meaningful insights. Statistical models are applied to distinguish patterns in online buyer behavior concerning product awareness, brand loyalty, and decision-making processes. The expected outcomes of this research contribute not only to the academic understanding of the dynamic interplay between social media and online buyer behavior but also offer practical implications for marketers, e-commerce platforms, and policymakers.

Keywords: social platforms, buyer behavior, consumer behavior, digital era

Procedia PDF Downloads 59
26858 A Patient Passport Application for Adults with Cystic Fibrosis

Authors: Tamara Vagg, Cathy Shortt, Claire Hickey, Joseph A. Eustace, Barry J. Plant, Sabin Tabirca

Abstract:

Introduction: Paper-based patient passports have been used advantageously for older patients, patients with diabetes, and patients with learning difficulties. However, these passports can experience issues with data security, patients forgetting to bring the passport, patients being over encumbered, and uncertainty with who is responsible for entering and managing data in this passport. These issues could be resolved by transferring the paper-based system to a convenient platform such as a smartphone application (app). Background: Life expectancy for some Cystic Fibrosis (CF) patients are rising and as such new complications and procedures are predicted. Subsequently, there is a need for education and management interventions that can benefit CF adults. This research proposes a CF patient passport to record basic medical information through a smartphone app which will allow CF adults access to their basic medical information. Aim: To provide CF patients with their basic medical information via mobile multimedia so that they can receive care when traveling abroad or between CF centres. Moreover, by recording their basic medical information, CF patients may become more aware of their own condition and more active in their health care. Methods: This app is designed by a CF multidisciplinary team to be a lightweight reflection of a hospital patient file. The passport app is created using PhoneGap so that it can be deployed for both Android and iOS devices. Data entered into the app is encrypted and stored locally only. The app is password protected and includes the ability to set reminders and a graph to visualise weight and lung function over time. The app is introduced to seven participants as part of a stress test. The participants are asked to test the performance and usability of the app and report any issues identified. Results: Feedback and suggestions received via this testing include the ability to reorder the list of clinical appointments via date, an open format of recording dates (in the event specifics are unknown), and a drop down menu for data which is difficult to enter (such as bugs found in mucus). The app is found to be usable and accessible and is now being prepared for a pilot study with adult CF patients. Conclusions: It is anticipated that such an app will be beneficial to CF adult patients when travelling abroad and between CF centres.

Keywords: Cystic Fibrosis, digital patient passport, mHealth, self management

Procedia PDF Downloads 235
26857 End to End Monitoring in Oracle Fusion Middleware for Data Verification

Authors: Syed Kashif Ali, Usman Javaid, Abdullah Chohan

Abstract:

In large enterprises multiple departments use different sort of information systems and databases according to their needs. These systems are independent and heterogeneous in nature and sharing information/data between these systems is not an easy task. The usage of middleware technologies have made data sharing between systems very easy. However, monitoring the exchange of data/information for verification purposes between target and source systems is often complex or impossible for maintenance department due to security/access privileges on target and source systems. In this paper, we are intended to present our experience of an end to end data monitoring approach at middle ware level implemented in Oracle BPEL for data verification without any help of monitoring tool.

Keywords: service level agreement, SOA, BPEL, oracle fusion middleware, web service monitoring

Procedia PDF Downloads 465
26856 Patient Perspectives on Telehealth During the Pandemic in the United States

Authors: Manal Sultan Alhussein, Xiang Michelle Liu

Abstract:

Telehealth is an advanced technology using digital information and telecommunication facilities that provide access to health services from a distance. It slows the transmission factor of COVID-19, especially for elderly patients and patients with chronic diseases during the pandemic. Therefore, understanding patient perspectives on telehealth services and the factors impacting their option of telehealth service will shed light on the measures that healthcare providers can take to improve the quality of telehealth services. This study aimed to evaluate perceptions of telehealth services among different patient groups and explore various aspects of telehealth utilization in the United States during the COVID-19 pandemic. An online survey distributed via social media platforms was used to collect research data. In addition to the descriptive statistics, both correlation and regression analyses were conducted to test research hypotheses. The empirical results highlighted that the factors such as accessibility to telehealth services and the type of specialty clinics that the patients required play important roles in the effectiveness of telehealth services they received. However, the results found that patients’ waiting time to receive telehealth services and their annual income did not significantly influence their desire to select receiving healthcare services via telehealth. The limitations of the study and future research directions are discussed.

Keywords: telehealth, patient satisfaction, pandemic, healthcare, survey

Procedia PDF Downloads 99
26855 Experimental Study on the Variation of Young's Modulus of Hollow Clay Brick Obtained from Static and Dynamic Tests

Authors: M. Aboudalle, Le Btth, M. Sari, F. Meftah

Abstract:

In parallel with the appearance of new materials, brick masonry had and still has an essential part of the construction market today, with new technical challenges in designing bricks to meet additional requirements. Being used in structural applications, predicting the performance of clay brick masonry allows a significant cost reduction, in terms of practical experimentation. The behavior of masonry walls depends on the behavior of their elementary components, such as bricks, joints, and coatings. Therefore, it is necessary to consider it at different scales (from the scale of the intrinsic material to the real scale of the wall) and then to develop appropriate models, using numerical simulations. The work presented in this paper focuses on the mechanical characterization of the terracotta material at ambient temperature. As a result, the static Young’s modulus obtained from the flexural test shows different values in comparison with the compression test, as well as with the dynamic Young’s modulus obtained from the Impulse excitation of vibration test. Moreover, the Young's modulus varies according to the direction in which samples are extracted, where the values in the extrusion direction diverge from the ones in the orthogonal directions. Based on these results, hollow bricks can be considered as transversely isotropic bimodulus material.

Keywords: bimodulus material, hollow clay brick, ımpulse excitation of vibration, transversely isotropic material, young’s modulus

Procedia PDF Downloads 186
26854 Post-bladder Catheter Infection

Authors: Mahla Azimi

Abstract:

Introduction: Post-bladder catheter infection is a common and significant healthcare-associated infection that affects individuals with indwelling urinary catheters. These infections can lead to various complications, including urinary tract infections (UTIs), bacteremia, sepsis, and increased morbidity and mortality rates. This article aims to provide a comprehensive review of post-bladder catheter infections, including their causes, risk factors, clinical presentation, diagnosis, treatment options, and preventive measures. Causes and Risk Factors: Post-bladder catheter infections primarily occur due to the colonization of microorganisms on the surface of the urinary catheter. The most common pathogens involved are Escherichia coli, Klebsiella pneumoniae, Pseudomonas aeruginosa, and Enterococcus species. Several risk factors contribute to the development of these infections, such as prolonged catheterization duration, improper insertion technique, poor hygiene practices during catheter care, compromised immune system function in patients with underlying conditions or immunosuppressive therapy. Clinical Presentation: Patients with post-bladder catheter infections may present with symptoms such as fever, chills, malaise, suprapubic pain or tenderness, and cloudy or foul-smelling urine. In severe cases or when left untreated for an extended period of time, patients may develop more severe symptoms like hematuria or signs of systemic infection. Diagnosis: The diagnosis of post-bladder catheter infection involves a combination of clinical evaluation and laboratory investigations. Urinalysis is crucial in identifying pyuria (presence of white blood cells) and bacteriuria (presence of bacteria). A urine culture is performed to identify the causative organism(s) and determine its antibiotic susceptibility profile. Treatment Options: Prompt initiation of appropriate antibiotic therapy is essential in managing post-bladder catheter infections. Empirical treatment should cover common pathogens until culture results are available. The choice of antibiotics should be guided by local antibiogram data to ensure optimal therapy. In some cases, catheter removal may be necessary, especially if the infection is recurrent or associated with severe complications. Preventive Measures: Prevention plays a vital role in reducing the incidence of post-bladder catheter infections. Strategies include proper hand hygiene, aseptic technique during catheter insertion and care, regular catheter maintenance, and timely removal of unnecessary catheters. Healthcare professionals should also promote patient education regarding self-care practices and signs of infection. Conclusion: Post-bladder catheter infections are a significant healthcare concern that can lead to severe complications and increased healthcare costs. Early recognition, appropriate diagnosis, and prompt treatment are crucial in managing these infections effectively. Implementing preventive measures can significantly reduce the incidence of post-bladder catheter infections and improve patient outcomes. Further research is needed to explore novel strategies for prevention and management in this field.

Keywords: post-bladder catheter infection, urinary tract infection, bacteriuria, indwelling urinary catheters, prevention

Procedia PDF Downloads 69
26853 Victim and Active Subject of the Crime of Violence in Family Reflected in the Criminal Code of the Republic of Moldova

Authors: Nastas Andrei

Abstract:

Ensuring accessible and functional justice is one of the priority objectives of judicial reform, and protecting the family against any acts that may harm its existence is one of the first priorities that have determined the need to defend the social order. In this context, the correlative approach of the victim and the aggressor becomes relevant as a subject of the crime of domestic violence. Domestic violence is a threat of physical, moral, or material harm, externalized now or in the past, or its provocation, which is characterized by a constant tendency to escalate and a high probability of repetitiveness in the relationship between the social partners, regardless of their legal status or domicile.Studying the legal support to identify the particularities of the victim and the subject of the crime of domestic violence facilitates the identification of the determinants of this crime, therefore, the development of means to prevent domestic violence. The scientific research has been effectuated on the base of the proper and authentic empirical data obtained from the analysis of the judicial practice in the matter of domestic violence, as well as being based on the most recent scientific issues in the field of the Substantive Criminal Law and other branches of science (criminology, psychology, sociology, pedagogy). As a result of the study performed, there have been formulated conclusions and interpretations able to be used in the science of the Substantive Criminal law, as well as in the practice of application of the legal norm in the matter of domestic violence.

Keywords: family violence, victim, crime, violence

Procedia PDF Downloads 96
26852 Lateralisation of Visual Function in Yellow-Eyed Mullet (Aldrichetta forsteri) and Its Role in Schooling Behaviour

Authors: Karen L. Middlemiss, Denham G. Cook, Peter Jaksons, Alistair Jerrett, William Davison

Abstract:

Lateralisation of cognitive function is a common phenomenon found throughout the animal kingdom. Strong biases in functional behaviours have evolved from asymmetrical brain hemispheres which differ in structure and/or cognitive function. In fish, lateralisation is involved in visually mediated behaviours such as schooling, predator avoidance, and foraging, and is considered to have a direct impact on species fitness. Currently, there is very little literature on the role of lateralisation in fish schools. The yellow-eyed mullet (Aldrichetta forsteri), is an estuarine and coastal species found commonly throughout temperate regions of Australia and New Zealand. This study sought to quantify visually mediated behaviours in yellow-eyed mullet to identify the significance of lateralisation, and the factors which influence functional behaviours in schooling fish. Our approach to study design was to conduct a series of tank based experiments investigating; a) individual and population level lateralisation, b) schooling behaviour, and d) optic lobe anatomy. Yellow-eyed mullet showed individual variation in direction and strength of lateralisation in juveniles, and trait specific spatial positioning within the school was evidenced in strongly lateralised fish. In combination with observed differences in schooling behaviour, the possibility of ontogenetic plasticity in both behavioural lateralisation and optic lobe morphology in adults is suggested. These findings highlight the need for research into the genetic and environmental factors (epigenetics) which drive functional behaviours such as schooling, feeding and aggression. Improved knowledge on collective behaviour could have significant benefits to captive rearing programmes through improved culture techniques and will add to the limited body of knowledge on the complex ecophysiological interactions present in our inshore fisheries.

Keywords: cerebral asymmetry, fisheries, schooling, visual bias

Procedia PDF Downloads 203
26851 On the Internal Structure of the ‘Enigmatic Electrons’

Authors: Natarajan Tirupattur Srinivasan

Abstract:

Quantum mechanics( QM) and (special) relativity (SR) have indeed revolutionized the very thinking of physicists, and the spectacular successes achieved over a century due to these two theories are mind-boggling. However, there is still a strong disquiet among some physicists. While the mathematical structure of these two theories has been established beyond any doubt, their physical interpretations are still being contested by many. Even after a hundred years of their existence, we cannot answer a very simple question, “What is an electron”? Physicists are struggling even now to come to grips with the different interpretations of quantum mechanics with all their ramifications. However, it is indeed strange that the (special) relativity theory of Einstein enjoys many orders of magnitude of “acceptance”, though both theories have their own stocks of weirdness in the results, like time dilation, mass increase with velocity, the collapse of the wave function, quantum jump, tunnelling, etc. Here, in this paper, it would be shown that by postulating an intrinsic internal motion to these enigmatic electrons, one can build a fairly consistent picture of reality, revealing a very simple picture of nature. This is also evidenced by Schrodinger’s ‘Zitterbewegung’ motion, about which so much has been written. This leads to a helical trajectory of electrons when they move in a laboratory frame. It will be shown that the helix is a three-dimensional wave having all the characteristics of our familiar 2D wave. Again, the helix, being a geodesic on an imaginary cylinder, supports ‘quantization’, and its representation is just the complex exponentials matching with the wave function of quantum mechanics. By postulating the instantaneous velocity of the electrons to be always ‘c’, the velocity of light, the entire relativity comes alive, and we can interpret the ‘time dilation’, ‘mass increase with velocity’, etc., in a very simple way. Thus, this model unifies both QM and SR without the need for a counterintuitive postulate of Einstein about the constancy of the velocity of light for all inertial observers. After all, if the motion of an inertial frame cannot affect the velocity of light, the converse that this constant also cannot affect the events in the frame must be true. But entire relativity is about how ‘c’ affects time, length, mass, etc., in different frames.

Keywords: quantum reconstruction, special theory of relativity, quantum mechanics, zitterbewegung, complex wave function, helix, geodesic, Schrodinger’s wave equations

Procedia PDF Downloads 57
26850 A Study on the Relationships among Teacher Empowerment, Professional Commitment and School Effectiveness

Authors: S. C. Lin, W. F. Hung, W. W. Cheng

Abstract:

Teacher empowerment was regarded as investing teachers with the right to participate in the determination of school goals and policies and to exercise professional judgment about what and how to teach. Professional commitment was considered as a person’s belief in and acceptance of the values of his or her chosen occupation or line of work, and a willingness to maintain membership in that occupation. An effective school has been defined as one in which students’ progress further than might be expected from consideration of its intake. An effective school thus adds extra value to its students' outcomes, in comparison with other schools serving similar intakes. A number of literature from various countries explored that teacher empowerment and professional commitment significantly influenced school effectiveness. However, there lacked more empirical studies to examine the relationships among them. Hence, this study was to explore the relationships among teacher empowerment, professional commitment and school effectiveness in junior high schools in Taiwan. Samples were seven hundred and five junior high school teachers selected from Taichung City, Changhua County and Nantou County. Questionnaire was applied to collect data. Data were analyzed by using descriptive statistics, t-test, one-way ANOVA, Pearson’s product-moment correlation, and multiple regression analysis. The findings of this study were as follows: First, the overall performances of teachers’ perceptions of teacher empowerment, teacher professional commitment and school effectiveness were above average. Second, the teachers’ perceptions of teacher empowerment were significant different in gender, designated duty, and school size. Third, the teachers’ perceptions of teacher professional commitment were significant different in gender, designated duty, and school size. Fourth, the teachers’ perceptions of school effectiveness were significant different in designated duty. Fifth, teacher empowerment was mid-positively correlation by teacher professional commitment. Sixth, there was mid-positively correlation between teacher empowerment and school effectiveness. Seventh, there was mid-positively correlation between teacher professional commitment and school effectiveness. Eighth, Teacher empowerment and professional commitment could significantly predict school effectiveness. Based on the findings of this study, the study proposed some suggestions for educational authorities, schools, teachers, and future studies as well.

Keywords: junior high school teacher, teacher empowerment, teacher professional commitment, school effectiveness

Procedia PDF Downloads 448
26849 Dissimilarity Measure for General Histogram Data and Its Application to Hierarchical Clustering

Authors: K. Umbleja, M. Ichino

Abstract:

Symbolic data mining has been developed to analyze data in very large datasets. It is also useful in cases when entry specific details should remain hidden. Symbolic data mining is quickly gaining popularity as datasets in need of analyzing are becoming ever larger. One type of such symbolic data is a histogram, which enables to save huge amounts of information into a single variable with high-level of granularity. Other types of symbolic data can also be described in histograms, therefore making histogram a very important and general symbolic data type - a method developed for histograms - can also be applied to other types of symbolic data. Due to its complex structure, analyzing histograms is complicated. This paper proposes a method, which allows to compare two histogram-valued variables and therefore find a dissimilarity between two histograms. Proposed method uses the Ichino-Yaguchi dissimilarity measure for mixed feature-type data analysis as a base and develops a dissimilarity measure specifically for histogram data, which allows to compare histograms with different number of bins and bin widths (so called general histogram). Proposed dissimilarity measure is then used as a measure for clustering. Furthermore, linkage method based on weighted averages is proposed with the concept of cluster compactness to measure the quality of clustering. The method is then validated with application on real datasets. As a result, the proposed dissimilarity measure is found producing adequate and comparable results with general histograms without the loss of detail or need to transform the data.

Keywords: dissimilarity measure, hierarchical clustering, histograms, symbolic data analysis

Procedia PDF Downloads 150
26848 Deep Neural Networks for Restoration of Sky Images Affected by Static and Anisotropic Aberrations

Authors: Constanza A. Barriga, Rafael Bernardi, Amokrane Berdja, Christian D. Guzman

Abstract:

Most image restoration methods in astronomy rely upon probabilistic tools that infer the best solution for a deconvolution problem. They achieve good performances when the point spread function (PSF) is spatially invariable in the image plane. However, this latter condition is not always satisfied with real optical systems. PSF angular variations cannot be evaluated directly from the observations, neither be corrected at a pixel resolution. We have developed a method for the restoration of images affected by static and anisotropic aberrations using deep neural networks that can be directly applied to sky images. The network is trained using simulated sky images corresponding to the T-80 telescope optical system, an 80 cm survey imager at Cerro Tololo (Chile), which are synthesized using a Zernike polynomial representation of the optical system. Once trained, the network can be used directly on sky images, outputting a corrected version of the image, which has a constant and known PSF across its field-of-view. The method was tested with the T-80 telescope, achieving better results than with PSF deconvolution techniques. We present the method and results on this telescope.

Keywords: aberrations, deep neural networks, image restoration, variable point spread function, wide field images

Procedia PDF Downloads 122
26847 Use of SUDOKU Design to Assess the Implications of the Block Size and Testing Order on Efficiency and Precision of Dulce De Leche Preference Estimation

Authors: Jéssica Ferreira Rodrigues, Júlio Silvio De Sousa Bueno Filho, Vanessa Rios De Souza, Ana Carla Marques Pinheiro

Abstract:

This study aimed to evaluate the implications of the block size and testing order on efficiency and precision of preference estimation for Dulce de leche samples. Efficiency was defined as the inverse of the average variance of pairwise comparisons among treatments. Precision was defined as the inverse of the variance of treatment means (or effects) estimates. The experiment was originally designed to test 16 treatments as a series of 8 Sudoku 16x16 designs being 4 randomized independently and 4 others in the reverse order, to yield balance in testing order. Linear mixed models were assigned to the whole experiment with 112 testers and all their grades, as well as their partially balanced subgroups, namely: a) experiment with the four initial EU; b) experiment with EU 5 to 8; c) experiment with EU 9 to 12; and b) experiment with EU 13 to 16. To record responses we used a nine-point hedonic scale, it was assumed a mixed linear model analysis with random tester and treatments effects and with fixed test order effect. Analysis of a cumulative random effects probit link model was very similar, with essentially no different conclusions and for simplicity, we present the results using Gaussian assumption. R-CRAN library lme4 and its function lmer (Fit Linear Mixed-Effects Models) was used for the mixed models and libraries Bayesthresh (default Gaussian threshold function) and ordinal with the function clmm (Cumulative Link Mixed Model) was used to check Bayesian analysis of threshold models and cumulative link probit models. It was noted that the number of samples tested in the same session can influence the acceptance level, underestimating the acceptance. However, proving a large number of samples can help to improve the samples discrimination.

Keywords: acceptance, block size, mixed linear model, testing order, testing order

Procedia PDF Downloads 310
26846 Pathologies in the Left Atrium Reproduced Using a Low-Order Synergistic Numerical Model of the Cardiovascular System

Authors: Nicholas Pearce, Eun-jin Kim

Abstract:

Pathologies of the cardiovascular (CV) system remain a serious and deadly health problem for human society. Computational modelling provides a relatively accessible tool for diagnosis, treatment, and research into CV disorders. However, numerical models of the CV system have largely focused on the function of the ventricles, frequently overlooking the behaviour of the atria. Furthermore, in the study of the pressure-volume relationship of the heart, which is a key diagnosis of cardiac vascular pathologies, previous works often evoke popular yet questionable time-varying elastance (TVE) method that imposes the pressure-volume relationship instead of calculating it consistently. Despite the convenience of the TVE method, there have been various indications of its limitations and the need for checking its validity in different scenarios. A model of the combined left ventricle (LV) and left atrium (LA) is presented, which consistently considers various feedback mechanisms in the heart without having to use the TVE method. Specifically, a synergistic model of the left ventricle is extended and modified to include the function of the LA. The synergy of the original model is preserved by modelling the electro-mechanical and chemical functions of the micro-scale myofiber for the LA and integrating it with the microscale and macro-organ-scale heart dynamics of the left ventricle and CV circulation. The atrioventricular node function is included and forms the conduction pathway for electrical signals between the atria and ventricle. The model reproduces the essential features of LA behaviour, such as the two-phase pressure-volume relationship and the classic figure of eight pressure-volume loops. Using this model, disorders in the internal cardiac electrical signalling are investigated by recreating the mechano-electric feedback (MEF), which is impossible where the time-varying elastance method is used. The effects of AV node block and slow conduction are then investigated in the presence of an atrial arrhythmia. It is found that electrical disorders and arrhythmia in the LA degrade the CV system by reducing the cardiac output, power, and heart rate.

Keywords: cardiovascular system, left atrium, numerical model, MEF

Procedia PDF Downloads 102
26845 Dissolved Gas Analysis Based Regression Rules from Trained ANN for Transformer Fault Diagnosis

Authors: Deepika Bhalla, Raj Kumar Bansal, Hari Om Gupta

Abstract:

Dissolved Gas Analysis (DGA) has been widely used for fault diagnosis in a transformer. Artificial neural networks (ANN) have high accuracy but are regarded as black boxes that are difficult to interpret. For many problems it is desired to extract knowledge from trained neural networks (NN) so that the user can gain a better understanding of the solution arrived by the NN. This paper applies a pedagogical approach for rule extraction from function approximating neural networks (REFANN) with application to incipient fault diagnosis using the concentrations of the dissolved gases within the transformer oil, as the input to the NN. The input space is split into subregions and for each subregion there is a linear equation that is used to predict the type of fault developing within a transformer. The experiments on real data indicate that the approach used can extract simple and useful rules and give fault predictions that match the actual fault and are at times also better than those predicted by the IEC method.

Keywords: artificial neural networks, dissolved gas analysis, rules extraction, transformer

Procedia PDF Downloads 521
26844 A Multivariate 4/2 Stochastic Covariance Model: Properties and Applications to Portfolio Decisions

Authors: Yuyang Cheng, Marcos Escobar-Anel

Abstract:

This paper introduces a multivariate 4/2 stochastic covariance process generalizing the one-dimensional counterparts presented in Grasselli (2017). Our construction permits stochastic correlation not only among stocks but also among volatilities, also known as co-volatility movements, both driven by more convenient 4/2 stochastic structures. The parametrization is flexible enough to separate these types of correlation, permitting their individual study. Conditions for proper changes of measure and closed-form characteristic functions under risk-neutral and historical measures are provided, allowing for applications of the model to risk management and derivative pricing. We apply the model to an expected utility theory problem in incomplete markets. Our analysis leads to closed-form solutions for the optimal allocation and value function. Conditions are provided for well-defined solutions together with a verification theorem. Our numerical analysis highlights and separates the impact of key statistics on equity portfolio decisions, in particular, volatility, correlation, and co-volatility movements, with the latter being the least important in an incomplete market.

Keywords: stochastic covariance process, 4/2 stochastic volatility model, stochastic co-volatility movements, characteristic function, expected utility theory, veri cation theorem

Procedia PDF Downloads 140