Search results for: circuit models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7432

Search results for: circuit models

4102 Shark Detection and Classification with Deep Learning

Authors: Jeremy Jenrette, Z. Y. C. Liu, Pranav Chimote, Edward Fox, Trevor Hastie, Francesco Ferretti

Abstract:

Suitable shark conservation depends on well-informed population assessments. Direct methods such as scientific surveys and fisheries monitoring are adequate for defining population statuses, but species-specific indices of abundance and distribution coming from these sources are rare for most shark species. We can rapidly fill these information gaps by boosting media-based remote monitoring efforts with machine learning and automation. We created a database of shark images by sourcing 24,546 images covering 219 species of sharks from the web application spark pulse and the social network Instagram. We used object detection to extract shark features and inflate this database to 53,345 images. We packaged object-detection and image classification models into a Shark Detector bundle. We developed the Shark Detector to recognize and classify sharks from videos and images using transfer learning and convolutional neural networks (CNNs). We applied these models to common data-generation approaches of sharks: boosting training datasets, processing baited remote camera footage and online videos, and data-mining Instagram. We examined the accuracy of each model and tested genus and species prediction correctness as a result of training data quantity. The Shark Detector located sharks in baited remote footage and YouTube videos with an average accuracy of 89\%, and classified located subjects to the species level with 69\% accuracy (n =\ eight species). The Shark Detector sorted heterogeneous datasets of images sourced from Instagram with 91\% accuracy and classified species with 70\% accuracy (n =\ 17 species). Data-mining Instagram can inflate training datasets and increase the Shark Detector’s accuracy as well as facilitate archiving of historical and novel shark observations. Base accuracy of genus prediction was 68\% across 25 genera. The average base accuracy of species prediction within each genus class was 85\%. The Shark Detector can classify 45 species. All data-generation methods were processed without manual interaction. As media-based remote monitoring strives to dominate methods for observing sharks in nature, we developed an open-source Shark Detector to facilitate common identification applications. Prediction accuracy of the software pipeline increases as more images are added to the training dataset. We provide public access to the software on our GitHub page.

Keywords: classification, data mining, Instagram, remote monitoring, sharks

Procedia PDF Downloads 121
4101 Application of the Discrete Rationalized Haar Transform to Distributed Parameter System

Authors: Joon-Hoon Park

Abstract:

In this paper the rationalized Haar transform is applied for distributed parameter system identification and estimation. A distributed parameter system is a dynamical and mathematical model described by a partial differential equation. And system identification concerns the problem of determining mathematical models from observed data. The Haar function has some disadvantages of calculation because it contains irrational numbers, for these reasons the rationalized Haar function that has only rational numbers. The algorithm adopted in this paper is based on the transform and operational matrix of the rationalized Haar function. This approach provides more convenient and efficient computational results.

Keywords: distributed parameter system, rationalized Haar transform, operational matrix, system identification

Procedia PDF Downloads 509
4100 Simulating Drilling Using a CAD System

Authors: Panagiotis Kyratsis, Konstantinos Kakoulis

Abstract:

Nowadays, the rapid development of CAD systems’ programming environments results in the creation of multiple downstream applications, which are developed and becoming increasingly available. CAD based manufacturing simulations is gradually following the same trend. Drilling is the most popular hole-making process used in a variety of industries. A specially built piece of software that deals with the drilling kinematics is presented. The cutting forces are calculated based on the tool geometry, the cutting conditions and the tool/work piece materials. The results are verified by experimental work. Finally, the response surface methodology (RSM) is applied and mathematical models of the total thrust force and the thrust force developed because of the main cutting edges are proposed.

Keywords: CAD, application programming interface, response surface methodology, drilling, RSM

Procedia PDF Downloads 470
4099 The Use of Social Networking Sites in eLearning

Authors: Clifford De Raffaele, Luana Bugeja, Serengul Smith

Abstract:

The adaptation of social networking sites within higher education has garnered significant interest in the recent years with numerous researches considering it as a possible shift from the traditional classroom based learning paradigm. Notwithstanding this increase in research and conducted studies however, the adaption of SNS based modules have failed to proliferate within Universities. This paper, commences its contribution by analyzing the various models and theories proposed in literature and amalgamates together various effective aspects for the inclusion of social technology within e-Learning. A three phased framework is further proposed which details the necessary considerations for the successful adaptation of SNS in enhancing the students learning experience. This proposal outlines the theoretical foundations which will be analyzed in practical implementation across international university campuses.

Keywords: eLearning, higher education, social network sites, student learning

Procedia PDF Downloads 340
4098 Experimental Squeeze Flow of Bitumen: Rheological Properties

Authors: A. Kraiem, A. Ayadi

Abstract:

The squeeze flow tests were studied by many authors to measure the rheological properties of fluid. Experimental squeezing flow test with constant area between two parallel disks of bitumen is investigated in the present work. The effect of the temperature, the process of preparing the sample and the gap between the discs were discussed. The obtained results were compared with the theoretical models. The behavior of bitumen depends on the viscosity and the yield stress. Thus, the bitumen was presented as a power law for a small power law exponent and as a biviscous fluid when the viscosity ratio was smaller than one. Also, the influence of the ambient temperature is required for the compression test. Therefore, for a high temperature the yield stress decrease.

Keywords: bitumen, biviscous fluid, squeeze flow, viscosity, yield stress

Procedia PDF Downloads 140
4097 Predictive Modelling of Curcuminoid Bioaccessibility as a Function of Food Formulation and Associated Properties

Authors: Kevin De Castro Cogle, Mirian Kubo, Maria Anastasiadi, Fady Mohareb, Claire Rossi

Abstract:

Background: The bioaccessibility of bioactive compounds is a critical determinant of the nutritional quality of various food products. Despite its importance, there is a limited number of comprehensive studies aimed at assessing how the composition of a food matrix influences the bioaccessibility of a compound of interest. This knowledge gap has prompted a growing need to investigate the intricate relationship between food matrix formulations and the bioaccessibility of bioactive compounds. One such class of bioactive compounds that has attracted considerable attention is curcuminoids. These naturally occurring phytochemicals, extracted from the roots of Curcuma longa, have gained popularity owing to their purported health benefits and also well known for their poor bioaccessibility Project aim: The primary objective of this research project is to systematically assess the influence of matrix composition on the bioaccessibility of curcuminoids. Additionally, this study aimed to develop a series of predictive models for bioaccessibility, providing valuable insights for optimising the formula for functional foods and provide more descriptive nutritional information to potential consumers. Methods: Food formulations enriched with curcuminoids were subjected to in vitro digestion simulation, and their bioaccessibility was characterized with chromatographic and spectrophotometric techniques. The resulting data served as the foundation for the development of predictive models capable of estimating bioaccessibility based on specific physicochemical properties of the food matrices. Results: One striking finding of this study was the strong correlation observed between the concentration of macronutrients within the food formulations and the bioaccessibility of curcuminoids. In fact, macronutrient content emerged as a very informative explanatory variable of bioaccessibility and was used, alongside other variables, as predictors in a Bayesian hierarchical model that predicted curcuminoid bioaccessibility accurately (optimisation performance of 0.97 R2) for the majority of cross-validated test formulations (LOOCV of 0.92 R2). These preliminary results open the door to further exploration, enabling researchers to investigate a broader spectrum of food matrix types and additional properties that may influence bioaccessibility. Conclusions: This research sheds light on the intricate interplay between food matrix composition and the bioaccessibility of curcuminoids. This study lays a foundation for future investigations, offering a promising avenue for advancing our understanding of bioactive compound bioaccessibility and its implications for the food industry and informed consumer choices.

Keywords: bioactive bioaccessibility, food formulation, food matrix, machine learning, probabilistic modelling

Procedia PDF Downloads 68
4096 Fragility Assessment for Torsionally Asymmetric Buildings in Plan

Authors: S. Feli, S. Tavousi Tafreshi, A. Ghasemi

Abstract:

The present paper aims at evaluating the response of three-dimensional buildings with in-plan stiffness irregularities that have been subjected to two-way excitation ground motion records simultaneously. This study is broadly-based fragility assessment with greater emphasis on structural response at in-plan flexible and stiff sides. To this end, three type of three-dimensional 5-story steel building structures with stiffness eccentricities, were subjected to extensive nonlinear incremental dynamic analyses (IDA) utilizing Ibarra-Krawinkler deterioration models. Fragility assessment was implemented for different configurations of braces to investigate the losses in buildings with center of resisting (CR) eccentricities.

Keywords: Ibarra-Krawinkler, fragility assessment, flexible and stiff side, center of resisting

Procedia PDF Downloads 205
4095 Real Activities Manipulation vs. Accrual Earnings Management: The Effect of Political Risk

Authors: Heba Abdelmotaal, Magdy Abdel-Kader

Abstract:

Purpose: This study explores whether a firm’s effective political risk management is preventing real and accrual earnings management . Design/methodology/approach: Based on a sample of 130 firms operating in Egypt during the period 2008-2013, two hypotheses are tested using the panel data regression models. Findings: The empirical findings indicate a significant relation between real and accrual earnings management and political risk. Originality/value: This paper provides a statistically evidence on the effects of the political risk management failure on the mangers’ engagement in the real and accrual earnings management practices, and its impact on the firm’s performance.

Keywords: political risk, risk management failure, real activities manipulation, accrual earnings management

Procedia PDF Downloads 439
4094 A Spin and Valley Modulating Device in Grapheme heterostructure: Controlling Valley and Spin Current

Authors: Adel Belayadi

Abstract:

The investigation of two-dimensional (2D) heterostructures, whether in the presence or the absence of magnetic substrates that sustain several induced spin-orbit couplings, has shown a promising/essential application for advancing the emerging fields of spintronics and valleytronics. In this contribution, we study spin/valley transport in graphene-like substrates in the presence of one or several locally induced spin-orbit coupling (SOC) terms resulting from graphene-based heterostructures. The models we proposed are based on the tight-binding approach, and our findings imply an alternative approach for conducting valley-polarized currents and suggest a corresponding mechanism for valley-dependent electron optics and optoelectronic devices.

Keywords: graphene-heterostructures, tight binding pproch, Spintronics, Valleytronics

Procedia PDF Downloads 25
4093 Studies on the Applicability of Artificial Neural Network (ANN) in Prediction of Thermodynamic Behavior of Sodium Chloride Aqueous System Containing a Non-Electrolytes

Authors: Dariush Jafari, S. Mostafa Nowee

Abstract:

In this study a ternary system containing sodium chloride as solute, water as primary solvent and ethanol as the antisolvent was considered to investigate the application of artificial neural network (ANN) in prediction of sodium solubility in the mixture of water as the solvent and ethanol as the antisolvent. The system was previously studied using by Extended UNIQUAC model by the authors of this study. The comparison between the results of the two models shows an excellent agreement between them (R2=0.99), and also approves the capability of ANN to predict the thermodynamic behavior of ternary electrolyte systems which are difficult to model.

Keywords: thermodynamic modeling, ANN, solubility, ternary electrolyte system

Procedia PDF Downloads 385
4092 Lee-Carter Mortality Forecasting Method with Dynamic Normal Inverse Gaussian Mortality Index

Authors: Funda Kul, İsmail Gür

Abstract:

Pension scheme providers have to price mortality risk by accurate mortality forecasting method. There are many mortality-forecasting methods constructed and used in literature. The Lee-Carter model is the first model to consider stochastic improvement trends in life expectancy. It is still precisely used. Mortality forecasting is done by mortality index in the Lee-Carter model. It is assumed that mortality index fits ARIMA time series model. In this paper, we propose and use dynamic normal inverse gaussian distribution to modeling mortality indes in the Lee-Carter model. Using population mortality data for Italy, France, and Turkey, the model is forecasting capability is investigated, and a comparative analysis with other models is ensured by some well-known benchmarking criterions.

Keywords: mortality, forecasting, lee-carter model, normal inverse gaussian distribution

Procedia PDF Downloads 360
4091 Tumor Detection Using Convolutional Neural Networks (CNN) Based Neural Network

Authors: Vinai K. Singh

Abstract:

In Neural Network-based Learning techniques, there are several models of Convolutional Networks. Whenever the methods are deployed with large datasets, only then can their applicability and appropriateness be determined. Clinical and pathological pictures of lobular carcinoma are thought to exhibit a large number of random formations and textures. Working with such pictures is a difficult problem in machine learning. Focusing on wet laboratories and following the outcomes, numerous studies have been published with fresh commentaries in the investigation. In this research, we provide a framework that can operate effectively on raw photos of various resolutions while easing the issues caused by the existence of patterns and texturing. The suggested approach produces very good findings that may be used to make decisions in the diagnosis of cancer.

Keywords: lobular carcinoma, convolutional neural networks (CNN), deep learning, histopathological imagery scans

Procedia PDF Downloads 136
4090 Artificial Intelligence Based Comparative Analysis for Supplier Selection in Multi-Echelon Automotive Supply Chains via GEP and ANN Models

Authors: Seyed Esmail Seyedi Bariran, Laysheng Ewe, Amy Ling

Abstract:

Since supplier selection appears as a vital decision, selecting supplier based on the best and most accurate ways has a lot of importance for enterprises. In this study, a new Artificial Intelligence approach is exerted to remove weaknesses of supplier selection. The paper has three parts. First part is choosing the appropriate criteria for assessing the suppliers’ performance. Next one is collecting the data set based on experts. Afterwards, the data set is divided into two parts, the training data set and the testing data set. By the training data set the best structure of GEP and ANN are selected and to evaluate the power of the mentioned methods the testing data set is used. The result obtained shows that the accuracy of GEP is more than ANN. Moreover, unlike ANN, a mathematical equation is presented by GEP for the supplier selection.

Keywords: supplier selection, automotive supply chains, ANN, GEP

Procedia PDF Downloads 631
4089 A Genetic-Neural-Network Modeling Approach for Self-Heating in GaN High Electron Mobility Transistors

Authors: Anwar Jarndal

Abstract:

In this paper, a genetic-neural-network (GNN) based large-signal model for GaN HEMTs is presented along with its parameters extraction procedure. The model is easy to construct and implement in CAD software and requires only DC and S-parameter measurements. An improved decomposition technique is used to model self-heating effect. Two GNN models are constructed to simulate isothermal drain current and power dissipation, respectively. The two model are then composed to simulate the drain current. The modeling procedure was applied to a packaged GaN-on-Si HEMT and the developed model is validated by comparing its large-signal simulation with measured data. A very good agreement between the simulation and measurement is obtained.

Keywords: GaN HEMT, computer-aided design and modeling, neural networks, genetic optimization

Procedia PDF Downloads 382
4088 A Blind Three-Dimensional Meshes Watermarking Using the Interquartile Range

Authors: Emad E. Abdallah, Alaa E. Abdallah, Bajes Y. Alskarnah

Abstract:

We introduce a robust three-dimensional watermarking algorithm for copyright protection and indexing. The basic idea behind our technique is to measure the interquartile range or the spread of the 3D model vertices. The algorithm starts by converting all the vertices to spherical coordinate followed by partitioning them into small groups. The proposed algorithm is slightly altering the interquartile range distribution of the small groups based on predefined watermark. The experimental results on several 3D meshes prove perceptual invisibility and the robustness of the proposed technique against the most common attacks including compression, noise, smoothing, scaling, rotation as well as combinations of these attacks.

Keywords: watermarking, three-dimensional models, perceptual invisibility, interquartile range, 3D attacks

Procedia PDF Downloads 474
4087 Model-Driven and Data-Driven Approaches for Crop Yield Prediction: Analysis and Comparison

Authors: Xiangtuo Chen, Paul-Henry Cournéde

Abstract:

Crop yield prediction is a paramount issue in agriculture. The main idea of this paper is to find out efficient way to predict the yield of corn based meteorological records. The prediction models used in this paper can be classified into model-driven approaches and data-driven approaches, according to the different modeling methodologies. The model-driven approaches are based on crop mechanistic modeling. They describe crop growth in interaction with their environment as dynamical systems. But the calibration process of the dynamic system comes up with much difficulty, because it turns out to be a multidimensional non-convex optimization problem. An original contribution of this paper is to propose a statistical methodology, Multi-Scenarios Parameters Estimation (MSPE), for the parametrization of potentially complex mechanistic models from a new type of datasets (climatic data, final yield in many situations). It is tested with CORNFLO, a crop model for maize growth. On the other hand, the data-driven approach for yield prediction is free of the complex biophysical process. But it has some strict requirements about the dataset. A second contribution of the paper is the comparison of these model-driven methods with classical data-driven methods. For this purpose, we consider two classes of regression methods, methods derived from linear regression (Ridge and Lasso Regression, Principal Components Regression or Partial Least Squares Regression) and machine learning methods (Random Forest, k-Nearest Neighbor, Artificial Neural Network and SVM regression). The dataset consists of 720 records of corn yield at county scale provided by the United States Department of Agriculture (USDA) and the associated climatic data. A 5-folds cross-validation process and two accuracy metrics: root mean square error of prediction(RMSEP), mean absolute error of prediction(MAEP) were used to evaluate the crop prediction capacity. The results show that among the data-driven approaches, Random Forest is the most robust and generally achieves the best prediction error (MAEP 4.27%). It also outperforms our model-driven approach (MAEP 6.11%). However, the method to calibrate the mechanistic model from dataset easy to access offers several side-perspectives. The mechanistic model can potentially help to underline the stresses suffered by the crop or to identify the biological parameters of interest for breeding purposes. For this reason, an interesting perspective is to combine these two types of approaches.

Keywords: crop yield prediction, crop model, sensitivity analysis, paramater estimation, particle swarm optimization, random forest

Procedia PDF Downloads 231
4086 Enhancing Large Language Models' Data Analysis Capability with Planning-and-Execution and Code Generation Agents: A Use Case for Southeast Asia Real Estate Market Analytics

Authors: Kien Vu, Jien Min Soh, Mohamed Jahangir Abubacker, Piyawut Pattamanon, Soojin Lee, Suvro Banerjee

Abstract:

Recent advances in Generative Artificial Intelligence (GenAI), in particular Large Language Models (LLMs) have shown promise to disrupt multiple industries at scale. However, LLMs also present unique challenges, notably, these so-called "hallucination" which is the generation of outputs that are not grounded in the input data that hinders its adoption into production. Common practice to mitigate hallucination problem is utilizing Retrieval Agmented Generation (RAG) system to ground LLMs'response to ground truth. RAG converts the grounding documents into embeddings, retrieve the relevant parts with vector similarity between user's query and documents, then generates a response that is not only based on its pre-trained knowledge but also on the specific information from the retrieved documents. However, the RAG system is not suitable for tabular data and subsequent data analysis tasks due to multiple reasons such as information loss, data format, and retrieval mechanism. In this study, we have explored a novel methodology that combines planning-and-execution and code generation agents to enhance LLMs' data analysis capabilities. The approach enables LLMs to autonomously dissect a complex analytical task into simpler sub-tasks and requirements, then convert them into executable segments of code. In the final step, it generates the complete response from output of the executed code. When deployed beta version on DataSense, the property insight tool of PropertyGuru, the approach yielded promising results, as it was able to provide market insights and data visualization needs with high accuracy and extensive coverage by abstracting the complexities for real-estate agents and developers from non-programming background. In essence, the methodology not only refines the analytical process but also serves as a strategic tool for real estate professionals, aiding in market understanding and enhancement without the need for programming skills. The implication extends beyond immediate analytics, paving the way for a new era in the real estate industry characterized by efficiency and advanced data utilization.

Keywords: large language model, reasoning, planning and execution, code generation, natural language processing, prompt engineering, data analysis, real estate, data sense, PropertyGuru

Procedia PDF Downloads 87
4085 Optimization of Electric Vehicle (EV) Charging Station Allocation Based on Multiple Data - Taking Nanjing (China) as an Example

Authors: Yue Huang, Yiheng Feng

Abstract:

Due to the global pressure on climate and energy, many countries are vigorously promoting electric vehicles and building charging (public) charging facilities. Faced with the supply-demand gap of existing electric vehicle charging stations and unreasonable space usage in China, this paper takes the central city of Nanjing as an example, establishes a site selection model through multivariate data integration, conducts multiple linear regression SPSS analysis, gives quantitative site selection results, and provides optimization models and suggestions for charging station layout planning.

Keywords: electric vehicle, charging station, allocation optimization, urban mobility, urban infrastructure, nanjing

Procedia PDF Downloads 92
4084 Geostatistical Analysis of Contamination of Soils in an Urban Area in Ghana

Authors: S. K. Appiah, E. N. Aidoo, D. Asamoah Owusu, M. W. Nuonabuor

Abstract:

Urbanization remains one of the unique predominant factors which is linked to the destruction of urban environment and its associated cases of soil contamination by heavy metals through the natural and anthropogenic activities. These activities are important sources of toxic heavy metals such as arsenic (As), cadmium (Cd), chromium (Cr), copper (Cu), iron (Fe), manganese (Mn), and lead (Pb), nickel (Ni) and zinc (Zn). Often, these heavy metals lead to increased levels in some areas due to the impact of atmospheric deposition caused by their proximity to industrial plants or the indiscriminately burning of substances. Information gathered on potentially hazardous levels of these heavy metals in soils leads to establish serious health and urban agriculture implications. However, characterization of spatial variations of soil contamination by heavy metals in Ghana is limited. Kumasi is a Metropolitan city in Ghana, West Africa and is challenged with the recent spate of deteriorating soil quality due to rapid economic development and other human activities such as “Galamsey”, illegal mining operations within the metropolis. The paper seeks to use both univariate and multivariate geostatistical techniques to assess the spatial distribution of heavy metals in soils and the potential risk associated with ingestion of sources of soil contamination in the Metropolis. Geostatistical tools have the ability to detect changes in correlation structure and how a good knowledge of the study area can help to explain the different scales of variation detected. To achieve this task, point referenced data on heavy metals measured from topsoil samples in a previous study, were collected at various locations. Linear models of regionalisation and coregionalisation were fitted to all experimental semivariograms to describe the spatial dependence between the topsoil heavy metals at different spatial scales, which led to ordinary kriging and cokriging at unsampled locations and production of risk maps of soil contamination by these heavy metals. Results obtained from both the univariate and multivariate semivariogram models showed strong spatial dependence with range of autocorrelations ranging from 100 to 300 meters. The risk maps produced show strong spatial heterogeneity for almost all the soil heavy metals with extremely risk of contamination found close to areas with commercial and industrial activities. Hence, ongoing pollution interventions should be geared towards these highly risk areas for efficient management of soil contamination to avert further pollution in the metropolis.

Keywords: coregionalization, heavy metals, multivariate geostatistical analysis, soil contamination, spatial distribution

Procedia PDF Downloads 300
4083 Machine Learning Prediction of Diabetes Prevalence in the U.S. Using Demographic, Physical, and Lifestyle Indicators: A Study Based on NHANES 2009-2018

Authors: Oluwafunmibi Omotayo Fasanya, Augustine Kena Adjei

Abstract:

To develop a machine learning model to predict diabetes (DM) prevalence in the U.S. population using demographic characteristics, physical indicators, and lifestyle habits, and to analyze how these factors contribute to the likelihood of diabetes. We analyzed data from 23,546 participants aged 20 and older, who were non-pregnant, from the 2009-2018 National Health and Nutrition Examination Survey (NHANES). The dataset included key demographic (age, sex, ethnicity), physical (BMI, leg length, total cholesterol [TCHOL], fasting plasma glucose), and lifestyle indicators (smoking habits). A weighted sample was used to account for NHANES survey design features such as stratification and clustering. A classification machine learning model was trained to predict diabetes status. The target variable was binary (diabetes or non-diabetes) based on fasting plasma glucose measurements. The following models were evaluated: Logistic Regression (baseline), Random Forest Classifier, Gradient Boosting Machine (GBM), Support Vector Machine (SVM). Model performance was assessed using accuracy, F1-score, AUC-ROC, and precision-recall metrics. Feature importance was analyzed using SHAP values to interpret the contributions of variables such as age, BMI, ethnicity, and smoking status. The Gradient Boosting Machine (GBM) model outperformed other classifiers with an AUC-ROC score of 0.85. Feature importance analysis revealed the following key predictors: Age: The most significant predictor, with diabetes prevalence increasing with age, peaking around the 60s for males and 70s for females. BMI: Higher BMI was strongly associated with a higher risk of diabetes. Ethnicity: Black participants had the highest predicted prevalence of diabetes (14.6%), followed by Mexican-Americans (13.5%) and Whites (10.6%). TCHOL: Diabetics had lower total cholesterol levels, particularly among White participants (mean decline of 23.6 mg/dL). Smoking: Smoking showed a slight increase in diabetes risk among Whites (0.2%) but had a limited effect in other ethnic groups. Using machine learning models, we identified key demographic, physical, and lifestyle predictors of diabetes in the U.S. population. The results confirm that diabetes prevalence varies significantly across age, BMI, and ethnic groups, with lifestyle factors such as smoking contributing differently by ethnicity. These findings provide a basis for more targeted public health interventions and resource allocation for diabetes management.

Keywords: diabetes, NHANES, random forest, gradient boosting machine, support vector machine

Procedia PDF Downloads 8
4082 Modelling of Pervaporation Separation of Butanol from Aqueous Solutions Using Polydimethylsiloxane Mixed Matrix Membranes

Authors: Arian Ebneyamini, Hoda Azimi, Jules Thibaults, F. Handan Tezel

Abstract:

In this study, a modification of Hennepe model for pervaporation separation of butanol from aqueous solutions using Polydimethylsiloxane (PDMS) mixed matrix membranes has been introduced and validated by experimental data. The model was compared to the original Hennepe model and few other models which are applicable for membrane gas separation processes such as Maxwell, Lewis Nielson and Pal. Theoretical modifications for non-ideal interface morphology have been offered to predict the permeability in case of interface void, interface rigidification and pore-blockage. The model was in a good agreement with experimental data.

Keywords: butanol, PDMS, modeling, pervaporation, mixed matrix membranes

Procedia PDF Downloads 221
4081 On Disaggregation and Consolidation of Imperfect Quality Shipments in an Extended EPQ Model

Authors: Hung-Chi Chang

Abstract:

For an extended EPQ model with random yield, the existent study revealed that both the disaggregating and consolidating shipment policies for the imperfect quality items are independent of holding cost, and recommended a model with economic benefit by comparing the least total cost for each of the three models investigated. To better capture the real situation, we generalize the existent study to include different holding costs for perfect and imperfect quality items. Through analysis, we show that the above shipment policies are dependent on holding costs. Furthermore, we derive a simple decision rule solely based on the thresholds of problem parameters to select a superior model. The results are illustrated analytically and numerically.

Keywords: consolidating shipments, disaggregating shipments, EPQ, imperfect quality, inventory

Procedia PDF Downloads 376
4080 3D Model Completion Based on Similarity Search with Slim-Tree

Authors: Alexis Aldo Mendoza Villarroel, Ademir Clemente Villena Zevallos, Cristian Jose Lopez Del Alamo

Abstract:

With the advancement of technology it is now possible to scan entire objects and obtain their digital representation by using point clouds or polygon meshes. However, some objects may be broken or have missing parts; thus, several methods focused on this problem have been proposed based on Geometric Deep Learning, such as GCNN, ACNN, PointNet, among others. In this article an approach from a different paradigm is proposed, using metric data structures to index global descriptors in the spectral domain and allow the recovery of a set of similar models in polynomial time; to later use the Iterative Close Point algorithm and recover the parts of the incomplete model using the geometry and topology of the model with less Hausdorff distance.

Keywords: 3D reconstruction method, point cloud completion, shape completion, similarity search

Procedia PDF Downloads 121
4079 A Summary-Based Text Classification Model for Graph Attention Networks

Authors: Shuo Liu

Abstract:

In Chinese text classification tasks, redundant words and phrases can interfere with the formation of extracted and analyzed text information, leading to a decrease in the accuracy of the classification model. To reduce irrelevant elements, extract and utilize text content information more efficiently and improve the accuracy of text classification models. In this paper, the text in the corpus is first extracted using the TextRank algorithm for abstraction, the words in the abstract are used as nodes to construct a text graph, and then the graph attention network (GAT) is used to complete the task of classifying the text. Testing on a Chinese dataset from the network, the classification accuracy was improved over the direct method of generating graph structures using text.

Keywords: Chinese natural language processing, text classification, abstract extraction, graph attention network

Procedia PDF Downloads 100
4078 Application of Neuro-Fuzzy Technique for Optimizing the PVC Membrane Sensor

Authors: Majid Rezayi, Sh. Shahaboddin, HNM E. Mahmud, A. Yadollah, A. Saeid, A. Yatimah

Abstract:

In this study, the adaptive neuro-fuzzy inference system (ANFIS) was applied to obtain the membrane composition model affecting the potential response of our reported polymeric PVC sensor for determining the titanium (III) ions. The performance statistics of the artificial neural network (ANN) and linear regression models for potential slope prediction of membrane composition of titanium (III) ion selective electrode were compared with ANFIS technique. The results show that the ANFIS model can be used as a practical tool for obtaining the Nerntian slope of the proposed sensor in this study.

Keywords: adaptive neuro fuzzy inference, PVC sensor, titanium (III) ions, Nerntian slope

Procedia PDF Downloads 287
4077 An Examination of Economic Evaluation Approaches in Mental Health Promotion Initiatives Targeted at Black and Asian Minority Ethnic Communities in the UK: A Critical Discourse Analysis

Authors: Phillipa Denise Peart

Abstract:

Black Asian and Minority Ethnic (BAME) people are more at risk of developing mental health disorders because they are more exposed to unfavorable social, economic, and environmental circumstances. These include housing, education, employment, community development, stigma, and discrimination. However, the majority of BAME mental health intervention studies focus on treatment with therapeutically effective drugs and use basic economic methods to evaluate their effectiveness; as a result, little is invested in the economic assessment of psychosocial interventions in BAME mental health. The UK government’s austerity programme and reduced funds for mental health services, has increased the need for the evaluation and assessment of initiatives to focus on value for money. The No Health without Mental Health policy (2011) provides practice guidance to practitioners, but there is little or no mention of the need to provide mental health initiatives targeted at BAME communities that are effective in terms of their impact and the cost-effectiveness. This, therefore, appears to contradict with and is at odds with the wider political discourse, which suggests there should be an increasing focus on health economic evaluation. As a consequence, it could be argued that whilst such policies provide direction to organisations to provide mental health services to the BAME community, by not requesting effective governance, assurance, and evaluation processes, they are merely paying lip service to address these problems and not helping advance knowledge and practice through evidence-based approaches. As a result, BAME communities suffer due to lack of efficient resources that can aid in the recovery process. This research study explores the mental health initiatives targeted at BAME communities, and analyses the techniques used when examining the cost effectiveness of mental health initiatives for BAME mental health communities. Using critical discourse analysis as an approach and method, mental health services will be selected as case studies, and their evaluations will be examined, alongside the political drivers that frame, shape, and direct their work. In doing so, it will analyse what the mental health policies initiatives are, how the initiatives are directed and demonstrate how economic models of evaluation are used in mental health programmes and how the value for money impacts and outcomes are articulated by mental health programme staff. It is anticipated that this study will further our understanding in order to provide adequate mental health resources and will deliver creative, supportive research to ensure evaluation is effective for the government to provide and maintain high quality and efficient mental health initiatives targeted at BAME communities.

Keywords: black, Asian and ethnic minority, economic models, mental health, health policy

Procedia PDF Downloads 111
4076 Analog Railway Signal Object Controller Development

Authors: Ercan Kızılay, Mustafa Demi̇rel, Selçuk Coşkun

Abstract:

Railway signaling systems consist of vital products that regulate railway traffic and provide safe route arrangements and maneuvers of trains. SIL 4 signal lamps are produced by many manufacturers today. There is a need for systems that enable these signal lamps to be controlled by commands from the interlocking. These systems should act as fail-safe and give error indications to the interlocking system when an unexpected situation occurs for the safe operation of railway systems from the RAMS perspective. In the past, driving and proving the lamp in relay-based systems was typically done via signaling relays. Today, the proving of lamps is done by comparing the current values read over the return circuit, the lower and upper threshold values. The purpose is an analog electronic object controller with the possibility of easy integration with vital systems and the signal lamp itself. During the study, the EN50126 standard approach was considered, and the concept, definition, risk analysis, requirements, architecture, design, and prototyping were performed throughout this study. FMEA (Failure Modes and Effects Analysis) and FTA (Fault Tree) Analysis) have been used for safety analysis in accordance with EN 50129. Concerning these analyzes, the 1oo2D reactive fail-safe hardware design of a controller has been researched. Electromagnetic compatibility (EMC) effects on the functional safety of equipment, insulation coordination, and over-voltage protection were discussed during hardware design according to EN 50124 and EN 50122 standards. As vital equipment for railway signaling, railway signal object controllers should be developed according to EN 50126 and EN 50129 standards which identify the steps and requirements of the development in accordance with the SIL 4(Safety Integrity Level) target. In conclusion of this study, an analog railway signal object controller, which takes command from the interlocking system, is processed in driver cards. Driver cards arrange the voltage level according to desired visibility by means of semiconductors. Additionally, prover cards evaluate the current upper and lower thresholds. Evaluated values are processed via logic gates which are composed as 1oo2D by means of analog electronic technologies. This logic evaluates the voltage level of the lamp and mitigates the risks of undue dimming.

Keywords: object controller, railway electronic, analog electronic, safety, railway signal

Procedia PDF Downloads 100
4075 An Integrated Intuitionistic Fuzzy Elimination Et Choix Traduisant La REalite (IFELECTRE) Model

Authors: Babak Daneshvar Rouyendegh

Abstract:

The aim of this study is to develop and describe a new methodology for the Multi-Criteria Decision-Making (MCDM) problem using Intuitionistic Fuzzy Elimination Et Choix Traduisant La REalite (IFELECTRE) model. The proposed models enable Decision-Makers (DMs) on the assessment and use Intuitionistic Fuzzy numbers (IFN). A numerical example is provided to demonstrate and clarify the proposed analysis procedure. Also, an empirical experiment is conducted to validation the effectiveness.

Keywords: Decision-Makers (DMs), Multi-Criteria Decision-Making (MCDM), Intuitionistic Fuzzy Elimination Et Choix Traduisant La REalite (IFELECTRE), Intuitionistic Fuzzy Numbers (IFN)

Procedia PDF Downloads 678
4074 Mechanism Design and Dynamic Analysis of Active Independent Front Steering System

Authors: Cheng-Chi Yu, Yu-Shiue Wang, Kei-Lin Kuo

Abstract:

Active Independent Front Steering system is a steering system which can according to vehicle driving situation adjusts the relation of steering angle between inner wheel and outer wheel. In low-speed cornering, AIFS sets the steering angles of inner and outer wheel into Ackerman steering geometry to make vehicle has less cornering radius. Besides, AIFS changes the steering geometry to parallel or even anti-Ackerman steering geometry to keep vehicle stability in high-speed cornering. Therefore, based on the analysis of the vehicle steering behavior from different steering geometries, this study develops a new screw type of active independent front steering system to make vehicles best cornering performance at any speeds. The screw type of active independent front steering system keeps the pinion and separates the rack into main rack and second rack. Two racks connect by a screw. Extra screw rotated motion powered by assistant motor through coupler makes second rack move relative to main rack, which can adjust both steering ratio and steering geometry. First of all, this study distinguishes the steering geometry by using Ackerman percentage and utilizes the software of ADAMS/Car to construct diverse steering geometry models. The different steering geometries are compared at low-speed and high-speed cornering, and then control strategies of the active independent front steering systems could be formulated. Secondly, this study applies closed loop equation to analyze tire steering angles and carries out optimization calculations to make the steering geometry from traditional rack and pinion steering system near to Ackerman steering geometry. Steering characteristics of the optimum steering mechanism and motion characteristics of vehicle installed the steering mechanism are verified by ADAMS/Car models of front suspension and full vehicle respectively. By adding dual auxiliary rack and dual motor to the optimum steering mechanism, the active independent front steering system could be developed to achieve the functions of variable steering ratio and variable steering geometry. At last, this study uses ADAMS/Car and Matlab/Simulink to co-simulate the cornering motion of vehicles confirms the vehicle installed the Active Independent Front Steering (AIFS) system has better handling performance than that with Active Independent Steering (AFS) system or with Electric Power Steering (EPS) system. At low-speed cornering, the vehicles with AIFS system and with AFS system have better maneuverability, less cornering radius, than the traditional vehicle with EPS system because that AIFS and AFS systems both provide function of variable steering ratio. However, there is a slight penalty in the motor(s) power consumption. In addition, because of the capability of variable steering geometry, the vehicle with AIFS system has better high-speed cornering stability, trajectory keeping, and even less motor(s) power consumption than that with EPS system and also with AFS system.

Keywords: active front steering system, active independent front steering system, steering geometry, steering ratio

Procedia PDF Downloads 189
4073 Translating the Australian National Health and Medical Research Council Obesity Guidelines into Practice into a Rural/Regional Setting in Tasmania, Australia

Authors: Giuliana Murfet, Heidi Behrens

Abstract:

Chronic disease is Australia’s biggest health concern and obesity the leading risk factor for many. Obesity and chronic disease have a higher representation in rural Tasmania, where levels of socio-disadvantage are also higher. People living outside major cities have less access to health services and poorer health outcomes. To help primary healthcare professionals manage obesity, the Australian NHMRC evidence-based clinical practice guidelines for management of overweight and obesity in adults were developed. They include recommendations for practice and models for obesity management. To our knowledge there has been no research conducted that investigates translation of these guidelines into practice in rural-regional areas; where implementation can be complicated by limited financial and staffing resources. Also, the systematic review that informed the guidelines revealed a lack of evidence for chronic disease models of obesity care. The aim was to establish and evaluate a multidisciplinary model for obesity management in a group of adult people with type 2 diabetes in a dispersed rural population in Australia. Extensive stakeholder engagement was undertaken to both garner support for an obesity clinic and develop a sustainable model of care. A comprehensive nurse practitioner-led outpatient model for obesity care was designed. Multidisciplinary obesity clinics for adults with type 2 diabetes including a dietitian, psychologist, physiotherapist and nurse practitioner were set up in the north-west of Tasmania at two geographically-rural towns. Implementation was underpinned by the NHMRC guidelines and recommendations focused on: assessment approaches; promotion of health benefits of weight loss; identification of relevant programs for individualising care; medication and bariatric surgery options for obesity management; and, the importance of long-term weight management. A clinical pathway for adult weight management is delivered by the multidisciplinary team with recognition of the impact of and adjustments needed for other comorbidities. The model allowed for intensification of intervention such as bariatric surgery according to recommendations, patient desires and suitability. A randomised controlled trial is ongoing, with the aim to evaluate standard care (diabetes-focused management) compared with an obesity-related approach with additional dietetic, physiotherapy, psychology and lifestyle advice. Key barriers and enablers to guideline implementation were identified that fall under the following themes: 1) health care delivery changes and the project framework development; 2) capacity and team-building; 3) stakeholder engagement; and, 4) the research project and partnerships. Engagement of not only local hospital but also state-wide health executives and surgical services committee were paramount to the success of the project. Staff training and collective development of the framework allowed for shared understanding. Staff capacity was increased with most taking on other activities (e.g., surgery coordination). Barriers were often related to differences of opinions in focus of the project; a desire to remain evidenced based (e.g., exercise prescription) without adjusting the model to allow for consideration of comorbidities. While barriers did exist and challenges overcome; the development of critical partnerships did enable the capacity for a potential model of obesity care for rural regional areas. Importantly, the findings contribute to the evidence base for models of diabetes and obesity care that coordinate limited resources.

Keywords: diabetes, interdisciplinary, model of care, obesity, rural regional

Procedia PDF Downloads 228