Search results for: extended labelled dependency graph
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 2089

Search results for: extended labelled dependency graph

1819 Cytotoxic Activity against MCF-7 Breast Cancer Cells and Antioxidant Property of Aqueous Tempe Extracts from Extended Fermentation

Authors: Zatil Athaillah, Anastasia Devi, Dian Muzdalifah, Wirasuwasti Nugrahani, Linar Udin

Abstract:

During tempe fermentation, some chemical changes occurred and they contributed to sensory, appearance, and health benefits of soybeans. Many studies on health properties of tempe have specialized on their isoflavones. In this study, other components of tempe, particularly water soluble chemicals, was investigated for their biofunctionality. The study was focused on the ability to suppress MCF-7 breast cancer cell growth and antioxidant activity, as expressed by DPPH radical scavenging activity, total phenols and total flavonoids, of the water extracts. Fermentation time of tempe was extended up to 120 hr to increase the possibility to find the functional components. Extraction yield and soluble nitrogen content were also quantified as accompanying data. Our findings suggested that yield of water extraction of tempe increased as fermentation was extended up to 120 hr, except for a slight decrease at 72 hr. Water extracts of tempe showed inhibition of MCF-7 breast cancer cell growth, as shown by lower IC50 values when compared to control (unfermented soybeans). Among the varied fermentation timescales, 60-hr period showed the highest activity (IC50 of 8.7 ± 4.95 µg/ml). The anticancer activity of extracts obtained from different fermentation time was positively correlated with total soluble nitrogens, but less relevant with antioxidant data. During 48-72 hr fermentation, at which cancer suppression activity was significant, the antioxidant properties from the three assays were not higher than control. These findings indicated that water extracts of tempe from extended fermentation could inhibit breast cancer cell growth but further study to determine the mechanism and compounds that play important role in the activity should be conducted.

Keywords: tempe, anticancer, antioxidant, phenolic compounds

Procedia PDF Downloads 239
1818 Time Series Analysis on the Production of Fruit Juice: A Case Study of National Horticultural Research Institute (Nihort) Ibadan, Oyo State

Authors: Abiodun Ayodele Sanyaolu

Abstract:

The research was carried out to investigate the time series analysis on quarterly production of fruit juice at the National Horticultural Research Institute Ibadan from 2010 to 2018. Documentary method of data collection was used, and the method of least square and moving average were used in the analysis. From the calculation and the graph, it was glaring that there was increase, decrease, and uniform movements in both the graph of the original data and the tabulated quarter values of the original data. Time series analysis was used to detect the trend in the highest number of fruit juice and it appears to be good over a period of time and the methods used to forecast are additive and multiplicative models. Since it was observed that the production of fruit juice is usually high in January of every year, it is strongly advised that National Horticultural Research Institute should make more provision for fruit juice storage outside this period of the year.

Keywords: fruit juice, least square, multiplicative models, time series

Procedia PDF Downloads 138
1817 Unlocking the Future of Grocery Shopping: Graph Neural Network-Based Cold Start Item Recommendations with Reverse Next Item Period Recommendation (RNPR)

Authors: Tesfaye Fenta Boka, Niu Zhendong

Abstract:

Recommender systems play a crucial role in connecting individuals with the items they require, as is particularly evident in the rapid growth of online grocery shopping platforms. These systems predominantly rely on user-centered recommendations, where items are suggested based on individual preferences, garnering considerable attention and adoption. However, our focus lies on the item-centered recommendation task within the grocery shopping context. In the reverse next item period recommendation (RNPR) task, we are presented with a specific item and challenged to identify potential users who are likely to consume it in the upcoming period. Despite the ever-expanding inventory of products on online grocery platforms, the cold start item problem persists, posing a substantial hurdle in delivering personalized and accurate recommendations for new or niche grocery items. To address this challenge, we propose a Graph Neural Network (GNN)-based approach. By capitalizing on the inherent relationships among grocery items and leveraging users' historical interactions, our model aims to provide reliable and context-aware recommendations for cold-start items. This integration of GNN technology holds the promise of enhancing recommendation accuracy and catering to users' individual preferences. This research contributes to the advancement of personalized recommendations in the online grocery shopping domain. By harnessing the potential of GNNs and exploring item-centered recommendation strategies, we aim to improve the overall shopping experience and satisfaction of users on these platforms.

Keywords: recommender systems, cold start item recommendations, online grocery shopping platforms, graph neural networks

Procedia PDF Downloads 86
1816 Evaluation of High Damping Rubber Considering Initial History through Dynamic Loading Test and Program Analysis

Authors: Kyeong Hoon Park, Taiji Mazuda

Abstract:

High damping rubber (HDR) bearings are dissipating devices mainly used in seismic isolation systems and have a great damping performance. Although many studies have been conducted on the dynamic model of HDR bearings, few models can reflect phenomena such as dependency of experienced shear strain on initial history. In order to develop a model that can represent the dependency of experienced shear strain of HDR by Mullins effect, dynamic loading test was conducted using HDR specimen. The reaction of HDR was measured by applying a horizontal vibration using a hybrid actuator under a constant vertical load. Dynamic program analysis was also performed after dynamic loading test. The dynamic model applied in program analysis is a bilinear type double-target model. This model is modified from typical bilinear model. This model can express the nonlinear characteristics related to the initial history of HDR bearings. Based on the dynamic loading test and program analysis results, equivalent stiffness and equivalent damping ratio were calculated to evaluate the mechanical properties of HDR and the feasibility of the bilinear type double-target model was examined.

Keywords: base-isolation, bilinear model, high damping rubber, loading test

Procedia PDF Downloads 118
1815 Extended Boolean Petri Nets Generating N-Ary Trees

Authors: Riddhi Jangid, Gajendra Pratap Singh

Abstract:

Petri nets, a mathematical tool, is used for modeling in different areas of computer sciences, biological networks, chemical systems and many other disciplines. A Petri net model of a given system is created by the graphical representation that describes the properties and behavior of the system. While looking for the behavior of any system, 1-safe Petri nets are of particular interest to many in the application part. Boolean Petri nets correspond to those class in 1- safe Petri nets that generate all the binary n-vectors in their reachability analysis. We study the class by changing different parameters like the token counts in the places and how the structure of the tree changes in the reachability analysis. We discuss here an extended class of Boolean Petri nets that generates n-ary trees in their reachability-based analysis.

Keywords: marking vector, n-vector, petri nets, reachability

Procedia PDF Downloads 76
1814 Application of the Extended Kantorovich Method to Size-Dependent Vibrational Analysis of Fully Clamped Rectangular Micro-Plates

Authors: Amir R. Askari, Masoud Tahani

Abstract:

The objective of the present paper is to investigate the effect of size on the vibrational behavior of fully clamped rectangular micro-plates based on the modified couple stress theory (MCST). To this end, a size-dependent Kirchhoff plate model is considered and the equation of motion which accounts for the effect of residual and couple stress components is derived using the Hamilton's principle. The eigenvalue problem associated with the free vibrations of fully clamped micro-plates is extracted and solved analytically using the extended Kantorovich method (EKM). The present findings are compared and validated by available results in the literature and an excellent agreement between them is observed. A parametric study is also conducted to show the significant effects of couple stress components on natural frequencies of fully clamped micro-plates. It is found that the ratio of MCST natural frequencies to those obtained by the classical theory (CT) only depends on the Poisson's ratio of the plate and is totally independent of plate's aspect ratio for cases with no residual stresses.

Keywords: vibrational analysis, modified couple stress theory, fully clamped rectangular micro-plates, extended Kantorovich method.

Procedia PDF Downloads 384
1813 On Disaggregation and Consolidation of Imperfect Quality Shipments in an Extended EPQ Model

Authors: Hung-Chi Chang

Abstract:

For an extended EPQ model with random yield, the existent study revealed that both the disaggregating and consolidating shipment policies for the imperfect quality items are independent of holding cost, and recommended a model with economic benefit by comparing the least total cost for each of the three models investigated. To better capture the real situation, we generalize the existent study to include different holding costs for perfect and imperfect quality items. Through analysis, we show that the above shipment policies are dependent on holding costs. Furthermore, we derive a simple decision rule solely based on the thresholds of problem parameters to select a superior model. The results are illustrated analytically and numerically.

Keywords: consolidating shipments, disaggregating shipments, EPQ, imperfect quality, inventory

Procedia PDF Downloads 371
1812 Real-Time Scheduling and Control of Supply Chain Networks: Challenges and Graph-Based Solution Approach

Authors: Jens Ehm

Abstract:

Manufacturing in supply chains requires an efficient organisation of production and transport processes in order to guarantee the supply of all partners within the chain with the material that is needed for the reliable fulfilment of tasks. If one partner is not able to supply products for a certain period, these products might be missing as the working material for the customer to perform the next manufacturing step, potentially as supply for further manufacturing steps. This way, local disruptions can influence the whole supply chain. In order to avoid material shortages, an efficient scheduling of tasks is necessary. However, the occurrence of unexpected disruptions cannot be eliminated, so that a modification of the schedule should be arranged as fast as possible. This paper discusses the challenges for the implementation of real-time scheduling and control methods and presents a graph-based approach that enables the integrated scheduling of production and transport processes for multiple supply chain partners and offers the potential for quick adaptations to parts of the initial schedule.

Keywords: production, logistics, integrated scheduling, real-time scheduling

Procedia PDF Downloads 370
1811 A Method for Identifying Unusual Transactions in E-commerce Through Extended Data Flow Conformance Checking

Authors: Handie Pramana Putra, Ani Dijah Rahajoe

Abstract:

The proliferation of smart devices and advancements in mobile communication technologies have permeated various facets of life with the widespread influence of e-commerce. Detecting abnormal transactions holds paramount significance in this realm due to the potential for substantial financial losses. Moreover, the fusion of data flow and control flow assumes a critical role in the exploration of process modeling and data analysis, contributing significantly to the accuracy and security of business processes. This paper introduces an alternative approach to identify abnormal transactions through a model that integrates both data and control flows. Referred to as the Extended Data Petri net (DPNE), our model encapsulates the entire process, encompassing user login to the e-commerce platform and concluding with the payment stage, including the mobile transaction process. We scrutinize the model's structure, formulate an algorithm for detecting anomalies in pertinent data, and elucidate the rationale and efficacy of the comprehensive system model. A case study validates the responsive performance of each system component, demonstrating the system's adeptness in evaluating every activity within mobile transactions. Ultimately, the results of anomaly detection are derived through a thorough and comprehensive analysis.

Keywords: database, data analysis, DPNE, extended data flow, e-commerce

Procedia PDF Downloads 51
1810 Hydrogen Production Using an Anion-Exchange Membrane Water Electrolyzer: Mathematical and Bond Graph Modeling

Authors: Hugo Daneluzzo, Christelle Rabbat, Alan Jean-Marie

Abstract:

Water electrolysis is one of the most advanced technologies for producing hydrogen and can be easily combined with electricity from different sources. Under the influence of electric current, water molecules can be split into oxygen and hydrogen. The production of hydrogen by water electrolysis favors the integration of renewable energy sources into the energy mix by compensating for their intermittence through the storage of the energy produced when production exceeds demand and its release during off-peak production periods. Among the various electrolysis technologies, anion exchange membrane (AEM) electrolyser cells are emerging as a reliable technology for water electrolysis. Modeling and simulation are effective tools to save time, money, and effort during the optimization of operating conditions and the investigation of the design. The modeling and simulation become even more important when dealing with multiphysics dynamic systems. One of those systems is the AEM electrolysis cell involving complex physico-chemical reactions. Once developed, models may be utilized to comprehend the mechanisms to control and detect flaws in the systems. Several modeling methods have been initiated by scientists. These methods can be separated into two main approaches, namely equation-based modeling and graph-based modeling. The former approach is less user-friendly and difficult to update as it is based on ordinary or partial differential equations to represent the systems. However, the latter approach is more user-friendly and allows a clear representation of physical phenomena. In this case, the system is depicted by connecting subsystems, so-called blocks, through ports based on their physical interactions, hence being suitable for multiphysics systems. Among the graphical modelling methods, the bond graph is receiving increasing attention as being domain-independent and relying on the energy exchange between the components of the system. At present, few studies have investigated the modelling of AEM systems. A mathematical model and a bond graph model were used in previous studies to model the electrolysis cell performance. In this study, experimental data from literature were simulated using OpenModelica using bond graphs and mathematical approaches. The polarization curves at different operating conditions obtained by both approaches were compared with experimental ones. It was stated that both models predicted satisfactorily the polarization curves with error margins lower than 2% for equation-based models and lower than 5% for the bond graph model. The activation polarization of hydrogen evolution reactions (HER) and oxygen evolution reactions (OER) were behind the voltage loss in the AEM electrolyzer, whereas ion conduction through the membrane resulted in the ohmic loss. Therefore, highly active electro-catalysts are required for both HER and OER while high-conductivity AEMs are needed for effectively lowering the ohmic losses. The bond graph simulation of the polarisation curve for operating conditions at various temperatures has illustrated that voltage increases with temperature owing to the technology of the membrane. Simulation of the polarisation curve can be tested virtually, hence resulting in reduced cost and time involved due to experimental testing and improved design optimization. Further improvements can be made by implementing the bond graph model in a real power-to-gas-to-power scenario.

Keywords: hydrogen production, anion-exchange membrane, electrolyzer, mathematical modeling, multiphysics modeling

Procedia PDF Downloads 87
1809 Public Transport Planning System by Dijkstra Algorithm: Case Study Bangkok Metropolitan Area

Authors: Pimploi Tirastittam, Phutthiwat Waiyawuththanapoom

Abstract:

Nowadays the promotion of the public transportation system in the Bangkok Metropolitan Area is increased such as the “Free Bus for Thai Citizen” Campaign and the prospect of the several MRT routes to increase the convenient and comfortable to the Bangkok Metropolitan area citizens. But citizens do not make full use of them it because the citizens are lack of the data and information and also the confident to the public transportation system of Thailand especially in the time and safety aspects. This research is the Public Transport Planning System by Dijkstra Algorithm: Case Study Bangkok Metropolitan Area by focusing on buses, BTS and MRT schedules/routes to give the most information to passengers. They can choose the way and the routes easily by using Dijkstra STAR Algorithm of Graph Theory which also shows the fare of the trip. This Application was evaluated by 30 normal users to find the mean and standard deviation of the developed system. Results of the evaluation showed that system is at a good level of satisfaction (4.20 and 0.40). From these results we can conclude that the system can be used properly and effectively according to the objective.

Keywords: Dijkstra algorithm, graph theory, public transport, Bangkok metropolitan area

Procedia PDF Downloads 245
1808 Demographic Dividend Explained by Infrastructure Costs of Population Growth Rate, Distinct from Age Dependency

Authors: Jane N. O'Sullivan

Abstract:

Although it is widely believed that fertility decline has benefitted economic advancement, particularly in East and South-East Asian countries, the causal mechanisms for this stimulus are contested. Since the turn of this century, demographic dividend theory has been increasingly recognised, hypothesising that higher proportions of working-age people can contribute to economic expansion if conditions are met to employ them productively. Population growth rate, as a systemic condition distinct from age composition, has not been similar attention since the 1970s and has lacked methodology for quantitative assessment. This paper explores conceptual and empirical quantification of the burden of expanding physical capital to accommodate a growing population. In proof-of-concept analyses of Australia and the United Kingdom, actual expenditure on gross fixed capital formation was compiled over four decades and apportioned to maintenance/turnover or expansion to accommodate population growth, based on lifespan of capital assets and population growth rate. In both countries, capital expansion was estimated to cost 6.5-7.0% of GDP per 1% population growth rate. This opportunity cost impedes the improvement of per capita capacity needed to realise the potential of the working-age population. Economic modelling of demographic scenarios have to date omitted this channel of influence; the implications of its inclusion are discussed.

Keywords: age dependency, demographic dividend, infrastructure, population growth rate

Procedia PDF Downloads 136
1807 Scene Classification Using Hierarchy Neural Network, Directed Acyclic Graph Structure, and Label Relations

Authors: Po-Jen Chen, Jian-Jiun Ding, Hung-Wei Hsu, Chien-Yao Wang, Jia-Ching Wang

Abstract:

A more accurate scene classification algorithm using label relations and the hierarchy neural network was developed in this work. In many classification algorithms, it is assumed that the labels are mutually exclusive. This assumption is true in some specific problems, however, for scene classification, the assumption is not reasonable. Because there are a variety of objects with a photo image, it is more practical to assign multiple labels for an image. In this paper, two label relations, which are exclusive relation and hierarchical relation, were adopted in the classification process to achieve more accurate multiple label classification results. Moreover, the hierarchy neural network (hierarchy NN) is applied to classify the image and the directed acyclic graph structure is used for predicting a more reasonable result which obey exclusive and hierarchical relations. Simulations show that, with these techniques, a much more accurate scene classification result can be achieved.

Keywords: convolutional neural network, label relation, hierarchy neural network, scene classification

Procedia PDF Downloads 453
1806 An Investigation to Study the Moisture Dependency of Ground Enhancement Compound

Authors: Arunima Shukla, Vikas Almadi, Devesh Jaiswal, Sunil Saini, Bhusan S. Patil

Abstract:

Lightning protection consists of three main parts; mainly air termination system, down conductor, and earth termination system. Earth termination system is the most important part as earth is the sink and source of charges. Therefore, even when the charges are captured and delivered to the ground, and an easy path is not provided to the charges, earth termination system would lead to problems. Soil has significantly different resistivities ranging from 10 Ωm for wet organic soil to 10000 Ωm for bedrock. Different methods have been discussed and used conventionally such as deep-ground-well method and altering the length of the rod. Those methods are not considered economical. Therefore, it was a general practice to use charcoal along with salt to reduce the soil resistivity. Bentonite is worldwide acceptable material, that had led our interest towards study of bentonite at first. It was concluded that bentonite is a clay which is non-corrosive, environment friendly. Whereas bentonite is suitable only when there is moisture present in the soil, as in the absence of moisture, cracks will appear on the surface which will provide an open passage to the air, resulting into increase in the resistivity. Furthermore, bentonite without moisture does not have enough bonding property, moisture retention, conductivity, and non-leachability. Therefore, bentonite was used along with the other backfill material to overcome the dependency of bentonite on moisture. Different experiments were performed to get the best ratio of bentonite and carbon backfill. It was concluded that properties will highly depend on the quantity of bentonite and carbon-based backfill material.

Keywords: backfill material, bentonite, grounding material, low resistivity

Procedia PDF Downloads 142
1805 Effect of Birks Constant and Defocusing Parameter on Triple-to-Double Coincidence Ratio Parameter in Monte Carlo Simulation-GEANT4

Authors: Farmesk Abubaker, Francesco Tortorici, Marco Capogni, Concetta Sutera, Vincenzo Bellini

Abstract:

This project concerns with the detection efficiency of the portable triple-to-double coincidence ratio (TDCR) at the National Institute of Metrology of Ionizing Radiation (INMRI-ENEA) which allows direct activity measurement and radionuclide standardization for pure-beta emitter or pure electron capture radionuclides. The dependency of the simulated detection efficiency of the TDCR, by using Monte Carlo simulation Geant4 code, on the Birks factor (kB) and defocusing parameter has been examined especially for low energy beta-emitter radionuclides such as 3H and 14C, for which this dependency is relevant. The results achieved in this analysis can be used for selecting the best kB factor and the defocusing parameter for computing theoretical TDCR parameter value. The theoretical results were compared with the available ones, measured by the ENEA TDCR portable detector, for some pure-beta emitter radionuclides. This analysis allowed to improve the knowledge of the characteristics of the ENEA TDCR detector that can be used as a traveling instrument for in-situ measurements with particular benefits in many applications in the field of nuclear medicine and in the nuclear energy industry.

Keywords: Birks constant, defocusing parameter, GEANT4 code, TDCR parameter

Procedia PDF Downloads 144
1804 Enhancing Scalability in Ethereum Network Analysis: Methods and Techniques

Authors: Stefan K. Behfar

Abstract:

The rapid growth of the Ethereum network has brought forth the urgent need for scalable analysis methods to handle the increasing volume of blockchain data. In this research, we propose efficient methodologies for making Ethereum network analysis scalable. Our approach leverages a combination of graph-based data representation, probabilistic sampling, and parallel processing techniques to achieve unprecedented scalability while preserving critical network insights. Data Representation: We develop a graph-based data representation that captures the underlying structure of the Ethereum network. Each block transaction is represented as a node in the graph, while the edges signify temporal relationships. This representation ensures efficient querying and traversal of the blockchain data. Probabilistic Sampling: To cope with the vastness of the Ethereum blockchain, we introduce a probabilistic sampling technique. This method strategically selects a representative subset of transactions and blocks, allowing for concise yet statistically significant analysis. The sampling approach maintains the integrity of the network properties while significantly reducing the computational burden. Graph Convolutional Networks (GCNs): We incorporate GCNs to process the graph-based data representation efficiently. The GCN architecture enables the extraction of complex spatial and temporal patterns from the sampled data. This combination of graph representation and GCNs facilitates parallel processing and scalable analysis. Distributed Computing: To further enhance scalability, we adopt distributed computing frameworks such as Apache Hadoop and Apache Spark. By distributing computation across multiple nodes, we achieve a significant reduction in processing time and enhanced memory utilization. Our methodology harnesses the power of parallelism, making it well-suited for large-scale Ethereum network analysis. Evaluation and Results: We extensively evaluate our methodology on real-world Ethereum datasets covering diverse time periods and transaction volumes. The results demonstrate its superior scalability, outperforming traditional analysis methods. Our approach successfully handles the ever-growing Ethereum data, empowering researchers and developers with actionable insights from the blockchain. Case Studies: We apply our methodology to real-world Ethereum use cases, including detecting transaction patterns, analyzing smart contract interactions, and predicting network congestion. The results showcase the accuracy and efficiency of our approach, emphasizing its practical applicability in real-world scenarios. Security and Robustness: To ensure the reliability of our methodology, we conduct thorough security and robustness evaluations. Our approach demonstrates high resilience against adversarial attacks and perturbations, reaffirming its suitability for security-critical blockchain applications. Conclusion: By integrating graph-based data representation, GCNs, probabilistic sampling, and distributed computing, we achieve network scalability without compromising analytical precision. This approach addresses the pressing challenges posed by the expanding Ethereum network, opening new avenues for research and enabling real-time insights into decentralized ecosystems. Our work contributes to the development of scalable blockchain analytics, laying the foundation for sustainable growth and advancement in the domain of blockchain research and application.

Keywords: Ethereum, scalable network, GCN, probabilistic sampling, distributed computing

Procedia PDF Downloads 70
1803 Concentric Circle Detection based on Edge Pre-Classification and Extended RANSAC

Authors: Zhongjie Yu, Hancheng Yu

Abstract:

In this paper, we propose an effective method to detect concentric circles with imperfect edges. First, the gradient of edge pixel is coded and a 2-D lookup table is built to speed up normal generation. Then we take an accumulator to estimate the rough center and collect plausible edges of concentric circles through gradient and distance. Later, we take the contour-based method, which takes the contour and edge intersection, to pre-classify the edges. Finally, we use the extended RANSAC method to find all the candidate circles. The center of concentric circles is determined by the two circles with the highest concentricity. Experimental results demonstrate that the proposed method has both good performance and accuracy for the detection of concentric circles.

Keywords: concentric circle detection, gradient, contour, edge pre-classification, RANSAC

Procedia PDF Downloads 127
1802 Assessment of Rock Masses Performance as a Support of Lined Rock Cavern for Isothermal Compressed Air Energy Storage

Authors: Vathna Suy, Ki-Il Song

Abstract:

In order to store highly pressurized gas such as an isothermal compressed air energy storage, Lined Rock Caverns (LRC) are constructed underground and supported by layers of concrete, steel and rock masses. This study aims to numerically investigate the performance of rock masses which serve as a support of Lined Rock Cavern subjected to high cyclic pressure loadings. FLAC3D finite different software is used for the simulation since the software can effectively model the behavior of concrete lining and steel plate with its built-in structural elements. Cyclic pressure loadings are applied onto the inner surface of the cavern which then transmitted to concrete, steel and eventually to the surrounding rock masses. Changes of stress and strain are constantly monitored throughout all the process of loading operations. The results at various monitoring locations are then extracted and analyzed to assess the response of the rock masses, specifically on its ability to absorb energy during loadings induced by the changes of cyclic pressure loadings inside the cavern. By analyzing the obtained data of stress-strain relation and taking into account the behavior of materials under the effect of strain-dependency, conclusions on the performance of rock masses subjected to high cyclic loading conditions are drawn.

Keywords: cyclic loading, FLAC3D, lined rock cavern (LRC), strain-dependency

Procedia PDF Downloads 241
1801 Mechanisms Underlying Comprehension of Visualized Personal Health Information: An Eye Tracking Study

Authors: Da Tao, Mingfu Qin, Wenkai Li, Tieyan Wang

Abstract:

While the use of electronic personal health portals has gained increasing popularity in the healthcare industry, users usually experience difficulty in comprehending and correctly responding to personal health information, partly due to inappropriate or poor presentation of the information. The way personal health information is visualized may affect how users perceive and assess their personal health information. This study was conducted to examine the effects of information visualization format and visualization mode on the comprehension and perceptions of personal health information among personal health information users with eye tracking techniques. A two-factor within-subjects experimental design was employed, where participants were instructed to complete a series of personal health information comprehension tasks under varied types of visualization mode (i.e., whether the information visualization is static or dynamic) and three visualization formats (i.e., bar graph, instrument-like graph, and text-only format). Data on a set of measures, including comprehension performance, perceptions, and eye movement indicators, were collected during the task completion in the experiment. Repeated measure analysis of variance analyses (RM-ANOVAs) was used for data analysis. The results showed that while the visualization format yielded no effects on comprehension performance, it significantly affected users’ perceptions (such as perceived ease of use and satisfaction). The two graphic visualizations yielded significantly higher favorable scores on subjective evaluations than that of the text format. While visualization mode showed no effects on users’ perception measures, it significantly affected users' comprehension performance in that dynamic visualization significantly reduced users' information search time. Both visualization format and visualization mode had significant main effects on eye movement behaviors, and their interaction effects were also significant. While the bar graph format and text format had similar time to first fixation across dynamic and static visualizations, instrument-like graph format had a larger time to first fixation for dynamic visualization than for static visualization. The two graphic visualization formats yielded shorter total fixation duration compared with the text-only format, indicating their ability to improve information comprehension efficiency. The results suggest that dynamic visualization can improve efficiency in comprehending important health information, and graphic visualization formats were favored more by users. The findings are helpful in the underlying comprehension mechanism of visualized personal health information and provide important implications for optimal design and visualization of personal health information.

Keywords: eye tracking, information comprehension, personal health information, visualization

Procedia PDF Downloads 102
1800 Stochastic Default Risk Estimation Evidence from the South African Financial Market

Authors: Mesias Alfeus, Kirsty Fitzhenry, Alessia Lederer

Abstract:

The present paper provides empirical studies to estimate defaultable bonds in the South African financial market. The main goal is to estimate the unobservable factors affecting bond yields for South African major banks. The maximum likelihood approach is adopted for the estimation methodology. Extended Kalman filtering techniques are employed in order to tackle the situation that the factors cannot be observed directly. Multi-dimensional Cox-Ingersoll-Ross (CIR)-type factor models are considered. Results show that default risk increased sharply in the South African financial market during COVID-19 and the CIR model with jumps exhibits a better performance.

Keywords: default intensity, unobservable state variables, CIR, α-CIR, extended kalman filtering

Procedia PDF Downloads 107
1799 Quantization of Damped Systems Based on the Doubling of Degrees of Freedom

Authors: Khaled I. Nawafleh

Abstract:

In this paper, it provide the canonical approach for studying dissipated oscillators based on the doubling of degrees of freedom. Clearly, expressions for Lagrangians of the elementary modes of the system are given, which ends with the familiar classical equations of motion for the dissipative oscillator. The equation for one variable is the time reversed of the motion of the second variable. it discuss in detail the extended Bateman Lagrangian specifically for a dual extended damped oscillator time-dependent. A Hamilton-Jacobi analysis showing the equivalence with the Lagrangian approach is also obtained. For that purpose, the techniques of separation of variables were applied, and the quantization process was achieved.

Keywords: doubling of degrees of freedom, dissipated harmonic oscillator, Hamilton-Jacobi, time-dependent lagrangians, quantization

Procedia PDF Downloads 66
1798 Systematic Evaluation of Convolutional Neural Network on Land Cover Classification from Remotely Sensed Images

Authors: Eiman Kattan, Hong Wei

Abstract:

In using Convolutional Neural Network (CNN) for classification, there is a set of hyperparameters available for the configuration purpose. This study aims to evaluate the impact of a range of parameters in CNN architecture i.e. AlexNet on land cover classification based on four remotely sensed datasets. The evaluation tests the influence of a set of hyperparameters on the classification performance. The parameters concerned are epoch values, batch size, and convolutional filter size against input image size. Thus, a set of experiments were conducted to specify the effectiveness of the selected parameters using two implementing approaches, named pertained and fine-tuned. We first explore the number of epochs under several selected batch size values (32, 64, 128 and 200). The impact of kernel size of convolutional filters (1, 3, 5, 7, 10, 15, 20, 25 and 30) was evaluated against the image size under testing (64, 96, 128, 180 and 224), which gave us insight of the relationship between the size of convolutional filters and image size. To generalise the validation, four remote sensing datasets, AID, RSD, UCMerced and RSCCN, which have different land covers and are publicly available, were used in the experiments. These datasets have a wide diversity of input data, such as number of classes, amount of labelled data, and texture patterns. A specifically designed interactive deep learning GPU training platform for image classification (Nvidia Digit) was employed in the experiments. It has shown efficiency in both training and testing. The results have shown that increasing the number of epochs leads to a higher accuracy rate, as expected. However, the convergence state is highly related to datasets. For the batch size evaluation, it has shown that a larger batch size slightly decreases the classification accuracy compared to a small batch size. For example, selecting the value 32 as the batch size on the RSCCN dataset achieves the accuracy rate of 90.34 % at the 11th epoch while decreasing the epoch value to one makes the accuracy rate drop to 74%. On the other extreme, setting an increased value of batch size to 200 decreases the accuracy rate at the 11th epoch is 86.5%, and 63% when using one epoch only. On the other hand, selecting the kernel size is loosely related to data set. From a practical point of view, the filter size 20 produces 70.4286%. The last performed image size experiment shows a dependency in the accuracy improvement. However, an expensive performance gain had been noticed. The represented conclusion opens the opportunities toward a better classification performance in various applications such as planetary remote sensing.

Keywords: CNNs, hyperparamters, remote sensing, land cover, land use

Procedia PDF Downloads 164
1797 Extending Early High Energy Physics Studies with a Tri-Preon Model

Authors: Peter J. Riley

Abstract:

Introductory courses in High Energy Physics (HEP) can be extended with the Tri-Preon (TP) model to both supplements and challenge the Standard Model (SM) theory. TP supplements by simplifying the tracking of Conserved Quantum Numbers at an interaction vertex, e.g., the lepton number can be seen as a di-preon current. TP challenges by proposing extended particle families to three generations of particle triplets for leptons, quarks, and weak bosons. There are extensive examples discussed at an introductory level in six arXiv publications, including supersymmetry, hyper color, and the Higgs. Interesting exercises include pion decay, kaon-antikaon mixing, neutrino oscillations, and K+ decay to muons. It is a revealing exercise for students to weigh the pros and cons of parallel theories at an early stage in their HEP journey.

Keywords: HEP, particle physics, standard model, Tri-Preon model

Procedia PDF Downloads 68
1796 An Optimized Approach to Generate the Possible States of Football Tournaments Final Table

Authors: Mouslem Damkhi

Abstract:

This paper focuses on possible states of a football tournament final table according to the number of participating teams. Each team holds a position in the table with which it is possible to determine the highest and lowest points for that team. This paper proposes an optimized search space based on the minimum and maximum number of points which can be gained by each team to produce and enumerate the possible states for a football tournament final table. The proposed search space minimizes producing the invalid states which cannot occur during a football tournament. The generated states are filtered by a validity checking algorithm which seeks to reach a tournament graph based on a generated state. Thus, the algorithm provides a way to determine which team’s wins, draws and loses values guarantee a particular table position. The paper also presents and discusses the experimental results of the approach on the tournaments with up to eight teams. Comparing with a blind search algorithm, our proposed approach reduces generating the invalid states up to 99.99%, which results in a considerable optimization in term of the execution time.

Keywords: combinatorics, enumeration, graph, tournament

Procedia PDF Downloads 120
1795 Simulation of Nano Drilling Fluid in an Extended Reach Well

Authors: Lina Jassim, Robiah Yunus, , Amran Salleh

Abstract:

Since nano particles have been assessed as thermo stabilizer, rheology enhancer, and ecology safer, nano drilling fluid can be utilized to overcome the complexity of hole cleaning in highly deviated interval of an extended reach wells. The eccentric annular flow is a flow with special considerations; it forms a vital part of drilling fluid flow analysis in an extended reach wells. In this work eccentric, dual phase flow (different types of rock cuttings with different size were blended with nano fluid) through horizontal well (an extended reach well) are simulated with the help of CFD, Fluent package. In horizontal wells flow occurs in an adverse pressure gradient condition, that makes the particle inside it susceptible to reversed flow. Thus the flow has to be analyzed in a three dimensional manner. Moreover the non-Newtonian behavior of the nano fluid makes the problem really challenging in numerical and physical aspects. The primary objective of the work is to establish a relationship between different flow characteristics with the speed of inner wall rotation. The nano fluid flow characteristics include swirl of flow and its effect on wellbore cleaning ability , wall shear stress and its effect on fluid viscosity to suspend and carry the rock cuttings, axial velocity and its effect on transportation of rock cuttings to the wellbore surface, finally pressure drop and its effect on managed of drilling pressure. The importance of eccentricity of the inner cylinder has to be analyzed as a part of it. Practical horizontal well flows contain a good amount of particles (rock cuttings) with moderate axial velocity, which verified nano drilling fluid ability of carrying and transferring cuttings particles in the highly deviated eccentric annular flow is also of utmost importance.

Keywords: Non-Newtonian, dual phase, eccentric annular, CFD

Procedia PDF Downloads 431
1794 A Polynomial Approach for a Graphical-based Integrated Production and Transport Scheduling with Capacity Restrictions

Authors: M. Ndeley

Abstract:

The performance of global manufacturing supply chains depends on the interaction of production and transport processes. Currently, the scheduling of these processes is done separately without considering mutual requirements, which leads to no optimal solutions. An integrated scheduling of both processes enables the improvement of supply chain performance. The integrated production and transport scheduling problem (PTSP) is NP-hard, so that heuristic methods are necessary to efficiently solve large problem instances as in the case of global manufacturing supply chains. This paper presents a heuristic scheduling approach which handles the integration of flexible production processes with intermodal transport, incorporating flexible land transport. The method is based on a graph that allows a reformulation of the PTSP as a shortest path problem for each job, which can be solved in polynomial time. The proposed method is applied to a supply chain scenario with a manufacturing facility in South Africa and shipments of finished product to customers within the Country. The obtained results show that the approach is suitable for the scheduling of large-scale problems and can be flexibly adapted to different scenarios.

Keywords: production and transport scheduling problem, graph based scheduling, integrated scheduling

Procedia PDF Downloads 471
1793 Knowledge Graph Development to Connect Earth Metadata and Standard English Queries

Authors: Gabriel Montague, Max Vilgalys, Catherine H. Crawford, Jorge Ortiz, Dava Newman

Abstract:

There has never been so much publicly accessible atmospheric and environmental data. The possibilities of these data are exciting, but the sheer volume of available datasets represents a new challenge for researchers. The task of identifying and working with a new dataset has become more difficult with the amount and variety of available data. Datasets are often documented in ways that differ substantially from the common English used to describe the same topics. This presents a barrier not only for new scientists, but for researchers looking to find comparisons across multiple datasets or specialists from other disciplines hoping to collaborate. This paper proposes a method for addressing this obstacle: creating a knowledge graph to bridge the gap between everyday English language and the technical language surrounding these datasets. Knowledge graph generation is already a well-established field, although there are some unique challenges posed by working with Earth data. One is the sheer size of the databases – it would be infeasible to replicate or analyze all the data stored by an organization like The National Aeronautics and Space Administration (NASA) or the European Space Agency. Instead, this approach identifies topics from metadata available for datasets in NASA’s Earthdata database, which can then be used to directly request and access the raw data from NASA. By starting with a single metadata standard, this paper establishes an approach that can be generalized to different databases, but leaves the challenge of metadata harmonization for future work. Topics generated from the metadata are then linked to topics from a collection of English queries through a variety of standard and custom natural language processing (NLP) methods. The results from this method are then compared to a baseline of elastic search applied to the metadata. This comparison shows the benefits of the proposed knowledge graph system over existing methods, particularly in interpreting natural language queries and interpreting topics in metadata. For the research community, this work introduces an application of NLP to the ecological and environmental sciences, expanding the possibilities of how machine learning can be applied in this discipline. But perhaps more importantly, it establishes the foundation for a platform that can enable common English to access knowledge that previously required considerable effort and experience. By making this public data accessible to the full public, this work has the potential to transform environmental understanding, engagement, and action.

Keywords: earth metadata, knowledge graphs, natural language processing, question-answer systems

Procedia PDF Downloads 144
1792 Discovery of the Piano Extended Techniques by Focusing on Symbols That George Crumb Used in Makrokosmos Volumes

Authors: Parham Bakhtiari

Abstract:

George Crumb's Makrokosmos Volumes are considered significant pieces in twentieth-century piano music and showcase the extensive use of different tones and extended techniques on the piano. Crumb's works are known for making references, particularly to music from previous eras which the visual, aural, and numerical characteristics are symbolic in nature. Crumb created a list of symbols and shortened letters to clarify his unique directions to those who performed his compositions. The pianists who prepare to play Makrokosmos must dedicate time to study and analyze Crumb's markings diligently to accurately capture the composer's wishes. The aim of this paper is to provide a collection for pianists looking to perform George Crumb's compositions known as Makrokosmos Volumes. The research traits of unconventional playing techniques and discussions on the music explored by the composer are being described.

Keywords: music, piano, Crumb, Makrokosmos, performance

Procedia PDF Downloads 41
1791 Real-Time Network Anomaly Detection Systems Based on Machine-Learning Algorithms

Authors: Zahra Ramezanpanah, Joachim Carvallo, Aurelien Rodriguez

Abstract:

This paper aims to detect anomalies in streaming data using machine learning algorithms. In this regard, we designed two separate pipelines and evaluated the effectiveness of each separately. The first pipeline, based on supervised machine learning methods, consists of two phases. In the first phase, we trained several supervised models using the UNSW-NB15 data-set. We measured the efficiency of each using different performance metrics and selected the best model for the second phase. At the beginning of the second phase, we first, using Argus Server, sniffed a local area network. Several types of attacks were simulated and then sent the sniffed data to a running algorithm at short intervals. This algorithm can display the results of each packet of received data in real-time using the trained model. The second pipeline presented in this paper is based on unsupervised algorithms, in which a Temporal Graph Network (TGN) is used to monitor a local network. The TGN is trained to predict the probability of future states of the network based on its past behavior. Our contribution in this section is introducing an indicator to identify anomalies from these predicted probabilities.

Keywords: temporal graph network, anomaly detection, cyber security, IDS

Procedia PDF Downloads 100
1790 Row Detection and Graph-Based Localization in Tree Nurseries Using a 3D LiDAR

Authors: Ionut Vintu, Stefan Laible, Ruth Schulz

Abstract:

Agricultural robotics has been developing steadily over recent years, with the goal of reducing and even eliminating pesticides used in crops and to increase productivity by taking over human labor. The majority of crops are arranged in rows. The first step towards autonomous robots, capable of driving in fields and performing crop-handling tasks, is for robots to robustly detect the rows of plants. Recent work done towards autonomous driving between plant rows offers big robotic platforms equipped with various expensive sensors as a solution to this problem. These platforms need to be driven over the rows of plants. This approach lacks flexibility and scalability when it comes to the height of plants or distance between rows. This paper proposes instead an algorithm that makes use of cheaper sensors and has a higher variability. The main application is in tree nurseries. Here, plant height can range from a few centimeters to a few meters. Moreover, trees are often removed, leading to gaps within the plant rows. The core idea is to combine row detection algorithms with graph-based localization methods as they are used in SLAM. Nodes in the graph represent the estimated pose of the robot, and the edges embed constraints between these poses or between the robot and certain landmarks. This setup aims to improve individual plant detection and deal with exception handling, like row gaps, which are falsely detected as an end of rows. Four methods were developed for detecting row structures in the fields, all using a point cloud acquired with a 3D LiDAR as an input. Comparing the field coverage and number of damaged plants, the method that uses a local map around the robot proved to perform the best, with 68% covered rows and 25% damaged plants. This method is further used and combined with a graph-based localization algorithm, which uses the local map features to estimate the robot’s position inside the greater field. Testing the upgraded algorithm in a variety of simulated fields shows that the additional information obtained from localization provides a boost in performance over methods that rely purely on perception to navigate. The final algorithm achieved a row coverage of 80% and an accuracy of 27% damaged plants. Future work would focus on achieving a perfect score of 100% covered rows and 0% damaged plants. The main challenges that the algorithm needs to overcome are fields where the height of the plants is too small for the plants to be detected and fields where it is hard to distinguish between individual plants when they are overlapping. The method was also tested on a real robot in a small field with artificial plants. The tests were performed using a small robot platform equipped with wheel encoders, an IMU and an FX10 3D LiDAR. Over ten runs, the system achieved 100% coverage and 0% damaged plants. The framework built within the scope of this work can be further used to integrate data from additional sensors, with the goal of achieving even better results.

Keywords: 3D LiDAR, agricultural robots, graph-based localization, row detection

Procedia PDF Downloads 136