Search results for: generalized autoregressive score model
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 18403

Search results for: generalized autoregressive score model

18073 Fusion Models for Cyber Threat Defense: Integrating Clustering, Random Forests, and Support Vector Machines to Against Windows Malware

Authors: Azita Ramezani, Atousa Ramezani

Abstract:

In the ever-escalating landscape of windows malware the necessity for pioneering defense strategies turns into undeniable this study introduces an avant-garde approach fusing the capabilities of clustering random forests and support vector machines SVM to combat the intricate web of cyber threats our fusion model triumphs with a staggering accuracy of 98.67 and an equally formidable f1 score of 98.68 a testament to its effectiveness in the realm of windows malware defense by deciphering the intricate patterns within malicious code our model not only raises the bar for detection precision but also redefines the paradigm of cybersecurity preparedness this breakthrough underscores the potential embedded in the fusion of diverse analytical methodologies and signals a paradigm shift in fortifying against the relentless evolution of windows malicious threats as we traverse through the dynamic cybersecurity terrain this research serves as a beacon illuminating the path toward a resilient future where innovative fusion models stand at the forefront of cyber threat defense.

Keywords: fusion models, cyber threat defense, windows malware, clustering, random forests, support vector machines (SVM), accuracy, f1-score, cybersecurity, malicious code detection

Procedia PDF Downloads 43
18072 Forecasting the Volatility of Geophysical Time Series with Stochastic Volatility Models

Authors: Maria C. Mariani, Md Al Masum Bhuiyan, Osei K. Tweneboah, Hector G. Huizar

Abstract:

This work is devoted to the study of modeling geophysical time series. A stochastic technique with time-varying parameters is used to forecast the volatility of data arising in geophysics. In this study, the volatility is defined as a logarithmic first-order autoregressive process. We observe that the inclusion of log-volatility into the time-varying parameter estimation significantly improves forecasting which is facilitated via maximum likelihood estimation. This allows us to conclude that the estimation algorithm for the corresponding one-step-ahead suggested volatility (with ±2 standard prediction errors) is very feasible since it possesses good convergence properties.

Keywords: Augmented Dickey Fuller Test, geophysical time series, maximum likelihood estimation, stochastic volatility model

Procedia PDF Downloads 291
18071 Invasive Ranges of Gorse (Ulex europaeus) in South Australia and Sri Lanka Using Species Distribution Modelling

Authors: Champika S. Kariyawasam

Abstract:

The distribution of gorse (Ulex europaeus) plants in South Australia has been modelled using 126 presence-only location data as a function of seven climate parameters. The predicted range of U. europaeus is mainly along the Mount Lofty Ranges in the Adelaide Hills and on Kangaroo Island. Annual precipitation and yearly average aridity index appeared to be the highest contributing variables to the final model formulation. The Jackknife procedure was employed to identify the contribution of different variables to gorse model outputs and response curves were used to predict changes with changing environmental variables. Based on this analysis, it was revealed that the combined effect of one or more variables could make a completely different impact to the original variables on their own to the model prediction. This work also demonstrates the need for a careful approach when selecting environmental variables for projecting correlative models to climatically distinct area. Maxent acts as a robust model when projecting the fitted species distribution model to another area with changing climatic conditions, whereas the generalized linear model, bioclim, and domain models to be less robust in this regard. These findings are important not only for predicting and managing invasive alien gorse in South Australia and Sri Lanka but also in other countries of the invasive range.

Keywords: invasive species, Maxent, species distribution modelling, Ulex europaeus

Procedia PDF Downloads 113
18070 E-Hailing Taxi Industry Management Mode Innovation Based on the Credit Evaluation

Authors: Yuan-lin Liu, Ye Li, Tian Xia

Abstract:

There are some shortcomings in Chinese existing taxi management modes. This paper suggests to establish the third-party comprehensive information management platform and put forward an evaluation model based on credit. Four indicators are used to evaluate the drivers’ credit, they are passengers’ evaluation score, driving behavior evaluation, drivers’ average bad record number, and personal credit score. A weighted clustering method is used to achieve credit level evaluation for taxi drivers. The management of taxi industry is based on the credit level, while the grade of the drivers is accorded to their credit rating. Credit rating determines the cost, income levels, the market access, useful period of license and the level of wage and bonus, as well as violation fine. These methods can make the credit evaluation effective. In conclusion, more credit data will help to set up a more accurate and detailed classification standard library.

Keywords: credit, mobile internet, e-hailing taxi, management mode, weighted cluster

Procedia PDF Downloads 295
18069 Nonstationary Modeling of Extreme Precipitation in the Wei River Basin, China

Authors: Yiyuan Tao

Abstract:

Under the impact of global warming together with the intensification of human activities, the hydrological regimes may be altered, and the traditional stationary assumption was no longer satisfied. However, most of the current design standards of water infrastructures were still based on the hypothesis of stationarity, which may inevitably result in severe biases. Many critical impacts of climate on ecosystems, society, and the economy are controlled by extreme events rather than mean values. Therefore, it is of great significance to identify the non-stationarity of precipitation extremes and model the precipitation extremes in a nonstationary framework. The Wei River Basin (WRB), located in a continental monsoon climate zone in China, is selected as a case study in this study. Six extreme precipitation indices were employed to investigate the changing patterns and stationarity of precipitation extremes in the WRB. To identify if precipitation extremes are stationary, the Mann-Kendall trend test and the Pettitt test, which is used to examine the occurrence of abrupt changes are adopted in this study. Extreme precipitation indices series are fitted with non-stationary distributions that selected from six widely used distribution functions: Gumbel, lognormal, Weibull, gamma, generalized gamma and exponential distributions by means of the time-varying moments model generalized additive models for location, scale and shape (GAMLSS), where the distribution parameters are defined as a function of time. The results indicate that: (1) the trends were not significant for the whole WRB, but significant positive/negative trends were still observed in some stations, abrupt changes for consecutive wet days (CWD) mainly occurred in 1985, and the assumption of stationarity is invalid for some stations; (2) for these nonstationary extreme precipitation indices series with significant positive/negative trends, the GAMLSS models are able to capture well the temporal variations of the indices, and perform better than the stationary model. Finally, the differences between the quantiles of nonstationary and stationary models are analyzed, which highlight the importance of nonstationary modeling of precipitation extremes in the WRB.

Keywords: extreme precipitation, GAMLSSS, non-stationary, Wei River Basin

Procedia PDF Downloads 101
18068 An Efficient Discrete Chaos in Generalized Logistic Maps with Applications in Image Encryption

Authors: Ashish Ashish

Abstract:

In the last few decades, the discrete chaos of difference equations has gained a massive attention of academicians and scholars due to its tremendous applications in each and every branch of science, such as cryptography, traffic control models, secure communications, weather forecasting, and engineering. In this article, a generalized logistic discrete map is established and discrete chaos is reported through period doubling bifurcation, period three orbit and Lyapunov exponent. It is interesting to see that the generalized logistic map exhibits superior chaos due to the presence of an extra degree of freedom of an ordered parameter. The period doubling bifurcation and Lyapunov exponent are demonstrated for some particular values of parameter and the discrete chaos is determined in the sense of Devaney's definition of chaos theoretically as well as numerically. Moreover, the study discusses an extended chaos based image encryption and decryption scheme in cryptography using this novel system. Surprisingly, a larger key space for coding and more sensitive dependence on initial conditions are examined for encryption and decryption of text messages, images and videos which secure the system strongly from external cyber attacks, coding attacks, statistic attacks and differential attacks.

Keywords: chaos, period-doubling, logistic map, Lyapunov exponent, image encryption

Procedia PDF Downloads 126
18067 EEG and ABER Abnormalities in Children with Speech and Language Delay

Authors: Bharati Mehta, Manish Parakh, Bharti Bhandari, Sneha Ambwani

Abstract:

Speech and language delay (SLD) is seen commonly as a co-morbidity in children having severe resistant focal and generalized, syndromic and symptomatic epilepsies. It is however not clear whether epilepsy contributes to or is a mere association in the pathogenesis of SLD. Also, it is acknowledged that Auditory Brainstem Evoked Responses (ABER), besides used for evaluating hearing threshold, also aid in prognostication of neurological disorders and abnormalities in the hearing pathway in the brainstem. There is no circumscribed or surrogate neurophysiologic laboratory marker to adjudge the extent of SLD. The current study was designed to evaluate the abnormalities in Electroencephalography (EEG) and ABER in children with SLD who do not have an overt hearing deficit or autism. 94 children of age group 2-8 years with predominant SLD and without any gross motor developmental delay, head injury, gross hearing disorder, cleft lip/palate and autism were selected. Standard video Electroencephalography using the 10:20 international system and ABER after click stimulus with intensities 110 db until 40 db was performed in all children. EEG was abnormal in 47.9% (n= 45; 36 boys and 9 girls) children. In the children with abnormal EEG, 64.5% (n=29) had an abnormal background, 57.8% (n=27) had presence of generalized interictal epileptiform discharges (IEDs), 20% (n=9) had focal epileptiform discharges exclusively from left side and 33.3% (n=15) had multifocal IEDs occurring both in isolation or associated with generalised abnormalities. In ABER, surprisingly, the peak latencies for waves I, III & V, inter-peak latencies I-III & I-V, III-V and wave amplitude ratio V/I, were found within normal limits in both ears of all the children. Thus in the current study it is certain that presence of generalized IEDs in EEG are seen in higher frequency with SLD and focal IEDs are seen exclusively in left hemisphere in these children. It may be possible that even with generalized EEG abnormalities present in these children, left hemispheric abnormalities as a part of this generalized dysfunction may be responsible for the speech and language dysfunction. The current study also emphasizes that ABER may not be routinely recommended as diagnostic or prognostic tool in children with SLD without frank hearing deficit or autism, thus reducing the burden on electro physiologists, laboratories and saving time and financial resources.

Keywords: ABER, EEG, speech, language delay

Procedia PDF Downloads 501
18066 Common Fixed Point Results and Stability of a Modified Jungck Iterative Scheme

Authors: Hudson Akewe

Abstract:

In this study, we introduce a modified Jungck (Dual Jungck) iterative scheme and use the scheme to approximate the unique common fixed point of a pair of generalized contractive-like operators in a Banach space. The iterative scheme is also shown to be stable with respect to the maps (S,T). An example is taken to justify the convergence of the scheme. Our result is a generalization and improvement of several results in the literature on single map T.

Keywords: generalized contractive-like operators, modified Jungck iterative scheme, stability results, weakly compatible maps, unique common fixed point

Procedia PDF Downloads 324
18065 Comparative Study of Estimators of Population Means in Two Phase Sampling in the Presence of Non-Response

Authors: Syed Ali Taqi, Muhammad Ismail

Abstract:

A comparative study of estimators of population means in two phase sampling in the presence of non-response when Unknown population means of the auxiliary variable(s) and incomplete information of study variable y as well as of auxiliary variable(s) is made. Three real data sets of University students, hospital and unemployment are used for comparison of all the available techniques in two phase sampling in the presence of non-response with the newly generalized ratio estimators.

Keywords: two-phase sampling, ratio estimator, product estimator, generalized estimators

Procedia PDF Downloads 216
18064 Developing a Model to Objectively Assess the Culture of Individuals and Teams in Order to Effectively and Efficiently Achieve Sustainability in the Manpower

Authors: Ahmed Mohamed Elnady Mohamed Elsafty

Abstract:

This paper explains a developed applied objective model to measure the culture qualitatively and quantitatively, whether in individuals or in teams, in order to be able to use culture correctly or modify it efficiently. This model provides precise measurements and consistent interpretations by being comprehensive, updateable, and protected from being misled by imitations. Methodically, the provided model divides the culture into seven dimensions (total 43 cultural factors): First dimension is outcome-orientation which consists of five factors and should be highest in leaders. Second dimension is details-orientation which consists of eight factors and should be in highest intelligence members. Third dimension is team-orientation which consists of five factors and should be highest in instructors or coaches. Fourth dimension is change-orientation which consists of five factors and should be highest in soldiers. Fifth dimension is people-orientation which consists of eight factors and should be highest in media members. Sixth dimension is masculinity which consists of seven factors and should be highest in hard workers. Last dimension is stability which consists of seven factors and should be highest in soft workers. In this paper, the details of all cultural factors are explained. Practically, information collection about each cultural factor in the targeted person or team is essential in order to calculate the degrees of all cultural factors using the suggested equation of multiplying 'the score of factor presence' by 'the score of factor strength'. In this paper, the details of how to build each score are explained. Based on the highest degrees - to identify which cultural dimension is the prominent - choosing the tested individual or team in the supposedly right position at the right time will provide a chance to use minimal efforts to make everyone aligned to the organization’s objectives. In other words, making everyone self-motivated by setting him/her at the right source of motivation is the most effective and efficient method to achieve high levels of competency, commitment, and sustainability. Modifying a team culture can be achieved by excluding or including new members with relatively high or low degrees in specific cultural factors. For conclusion, culture is considered as the software of the human beings and it is one of the major compression factors on the managerial discretion. It represents the behaviors, attitudes, and motivations of the human resources which are vital to enhance quality and safety, expanding the market share, and defending against attacks from external environments. Thus, it is tremendously essential and useful to use such a comprehensive model to measure, use, and modify culture.

Keywords: culture dimensions, culture factors, culture measurement, cultural analysis, cultural modification, self-motivation, alignment to objectives, competency, sustainability

Procedia PDF Downloads 151
18063 Regional Hydrological Extremes Frequency Analysis Based on Statistical and Hydrological Models

Authors: Hadush Kidane Meresa

Abstract:

The hydrological extremes frequency analysis is the foundation for the hydraulic engineering design, flood protection, drought management and water resources management and planning to utilize the available water resource to meet the desired objectives of different organizations and sectors in a country. This spatial variation of the statistical characteristics of the extreme flood and drought events are key practice for regional flood and drought analysis and mitigation management. For different hydro-climate of the regions, where the data set is short, scarcity, poor quality and insufficient, the regionalization methods are applied to transfer at-site data to a region. This study aims in regional high and low flow frequency analysis for Poland River Basins. Due to high frequent occurring of hydrological extremes in the region and rapid water resources development in this basin have caused serious concerns over the flood and drought magnitude and frequencies of the river in Poland. The magnitude and frequency result of high and low flows in the basin is needed for flood and drought planning, management and protection at present and future. Hydrological homogeneous high and low flow regions are formed by the cluster analysis of site characteristics, using the hierarchical and C- mean clustering and PCA method. Statistical tests for regional homogeneity are utilized, by Discordancy and Heterogeneity measure tests. In compliance with results of the tests, the region river basin has been divided into ten homogeneous regions. In this study, frequency analysis of high and low flows using AM for high flow and 7-day minimum low flow series is conducted using six statistical distributions. The use of L-moment and LL-moment method showed a homogeneous region over entire province with Generalized logistic (GLOG), Generalized extreme value (GEV), Pearson type III (P-III), Generalized Pareto (GPAR), Weibull (WEI) and Power (PR) distributions as the regional drought and flood frequency distributions. The 95% percentile and Flow duration curves of 1, 7, 10, 30 days have been plotted for 10 stations. However, the cluster analysis performed two regions in west and east of the province where L-moment and LL-moment method demonstrated the homogeneity of the regions and GLOG and Pearson Type III (PIII) distributions as regional frequency distributions for each region, respectively. The spatial variation and regional frequency distribution of flood and drought characteristics for 10 best catchment from the whole region was selected and beside the main variable (streamflow: high and low) we used variables which are more related to physiographic and drainage characteristics for identify and delineate homogeneous pools and to derive best regression models for ungauged sites. Those are mean annual rainfall, seasonal flow, average slope, NDVI, aspect, flow length, flow direction, maximum soil moisture, elevation, and drainage order. The regional high-flow or low-flow relationship among one streamflow characteristics with (AM or 7-day mean annual low flows) some basin characteristics is developed using Generalized Linear Mixed Model (GLMM) and Generalized Least Square (GLS) regression model, providing a simple and effective method for estimation of flood and drought of desired return periods for ungauged catchments.

Keywords: flood , drought, frequency, magnitude, regionalization, stochastic, ungauged, Poland

Procedia PDF Downloads 580
18062 Estimating the Timing Interval for Malarial Indoor Residual Spraying: A Modelling Approach

Authors: Levicatus Mugenyi, Joaniter Nankabirwa, Emmanuel Arinaitwe, John Rek, Niel Hens, Moses Kamya, Grant Dorsey

Abstract:

Background: Indoor residual spraying (IRS) reduces vector densities and malaria transmission, however, the most effective spraying intervals for IRS have not been well established. We aim to estimate the optimal timing interval for IRS using a modeling approach. Methods: We use a generalized additive model to estimate the optimal timing interval for IRS using the predicted malaria incidence. The model is applied to post IRS cohort clinical data from children aged 0.5–10 years in selected households in Tororo, historically a high malaria transmission setting in Uganda. Six rounds of IRS were implemented in Tororo during the study period (3 rounds with bendiocarb: December 2014 to December 2015, and 3 rounds with actellic: June 2016 to July 2018). Results: Monthly incidence of malaria from October 2014 to February 2019 decreased from 3.25 to 0.0 per person-years in the children under 5 years, and 1.57 to 0.0 for 5-10 year-olds. The optimal time interval for IRS differed between bendiocarb and actellic and by IRS round. It was estimated to be 17 and 40 weeks after the first round of bendiocarb and actellic, respectively. After the third round of actellic, 36 weeks was estimated to be optimal. However, we could not estimate from the data the optimal time after the second and third rounds of bendiocarb and after the second round of actellic. Conclusion: We conclude that to sustain the effect of IRS in a high-medium transmission setting, the second rounds of bendiocarb need to be applied roughly 17 weeks and actellic 40 weeks after the first round, and the timing differs for subsequent rounds. The amount of rainfall did not influence the trend in malaria incidence after IRS, as well as the IRS timing intervals. Our results suggest that shorter intervals for the IRS application can be more effective compared to the current practice, which is about 24 weeks for bendiocarb and 48 weeks for actellic. However, when considering our findings, one should account for the cost and drug resistance associated with IRS. We also recommend that the timing and incidence should be monitored in the future to improve these estimates.

Keywords: incidence, indoor residual spraying, generalized additive model, malaria

Procedia PDF Downloads 102
18061 A Generalized Framework for Adaptive Machine Learning Deployments in Algorithmic Trading

Authors: Robert Caulk

Abstract:

A generalized framework for adaptive machine learning deployments in algorithmic trading is introduced, tested, and released as open-source code. The presented software aims to test the hypothesis that recent data contains enough information to form a probabilistically favorable short-term price prediction. Further, the framework contains various adaptive machine learning techniques that are geared toward generating profit during strong trends and minimizing losses during trend changes. Results demonstrate that this adaptive machine learning approach is capable of capturing trends and generating profit. The presentation also discusses the importance of defining the parameter space associated with the dynamic training data-set and using the parameter space to identify and remove outliers from prediction data points. Meanwhile, the generalized architecture enables common users to exploit the powerful machinery while focusing on high-level feature engineering and model testing. The presentation also highlights common strengths and weaknesses associated with the presented technique and presents a broad range of well-tested starting points for feature set construction, target setting, and statistical methods for enforcing risk management and maintaining probabilistically favorable entry and exit points. The presentation also describes the end-to-end data processing tools associated with FreqAI, including automatic data fetching, data aggregation, feature engineering, safe and robust data pre-processing, outlier detection, custom machine learning and statistical tools, data post-processing, and adaptive training backtest emulation, and deployment of adaptive training in live environments. Finally, the generalized user interface is also discussed in the presentation. Feature engineering is simplified so that users can seed their feature sets with common indicator libraries (e.g. TA-lib, pandas-ta). The user also feeds data expansion parameters to fill out a large feature set for the model, which can contain as many as 10,000+ features. The presentation describes the various object-oriented programming techniques employed to make FreqAI agnostic to third-party libraries and external data sources. In other words, the back-end is constructed in such a way that users can leverage a broad range of common regression libraries (Catboost, LightGBM, Sklearn, etc) as well as common Neural Network libraries (TensorFlow, PyTorch) without worrying about the logistical complexities associated with data handling and API interactions. The presentation finishes by drawing conclusions about the most important parameters associated with a live deployment of the adaptive learning framework and provides the road map for future development in FreqAI.

Keywords: machine learning, market trend detection, open-source, adaptive learning, parameter space exploration

Procedia PDF Downloads 71
18060 Rectenna Modeling Based on MoM-GEC Method for RF Energy Harvesting

Authors: Soulayma Smirani, Mourad Aidi, Taoufik Aguili

Abstract:

Energy harvesting has arisen as a prominent research area for low power delivery to RF devices. Rectennas have become a key element in this technology. In this paper, electromagnetic modeling of a rectenna system is presented. In our approach, a hybrid technique was demonstrated to associate both the method of auxiliary sources (MAS) and MoM-GEC (the method of moments combined with the generalized equivalent circuit technique). Auxiliary sources were used in order to substitute specific electronic devices. Therefore, a simple and controllable model is obtained. Also, it can easily be interconnected to form different topologies of rectenna arrays for more energy harvesting. At last, simulation results show the feasibility and simplicity of the proposed rectenna model with high precision and computation efficiency.

Keywords: computational electromagnetics, MoM-GEC method, rectennas, RF energy harvesting

Procedia PDF Downloads 149
18059 Proficient Estimation Procedure for a Rare Sensitive Attribute Using Poisson Distribution

Authors: S. Suman, G. N. Singh

Abstract:

The present manuscript addresses the estimation procedure of population parameter using Poisson probability distribution when characteristic under study possesses a rare sensitive attribute. The generalized form of unrelated randomized response model is suggested in order to acquire the truthful responses from respondents. The resultant estimators have been proposed for two situations when the information on an unrelated rare non-sensitive characteristic is known as well as unknown. The properties of the proposed estimators are derived, and the measure of confidentiality of respondent is also suggested for respondents. Empirical studies are carried out in the support of discussed theory.

Keywords: Poisson distribution, randomized response model, rare sensitive attribute, non-sensitive attribute

Procedia PDF Downloads 242
18058 Coronary Artery Calcium Score and Statin Treatment Effect on Myocardial Infarction and Major Adverse Cardiovascular Event of Atherosclerotic Cardiovascular Disease: A Systematic Review and Meta-Analysis

Authors: Yusra Pintaningrum, Ilma Fahira Basyir, Sony Hilal Wicaksono, Vito A. Damay

Abstract:

Background: Coronary artery calcium (CAC) scores play an important role in improving prognostic accuracy and can be selectively used to guide the allocation of statin therapy for atherosclerotic cardiovascular disease outcomes and potentially associated with the occurrence of MACE (Major Adverse Cardiovascular Event) and MI (Myocardial Infarction). Objective: This systematic review and meta-analysis aim to analyze the findings of a study about CAC Score and statin treatment effect on MI and MACE risk. Methods: Search for published scientific articles using the PRISMA (Preferred Reporting, Items for Systematic Reviews and Meta-Analysis) method conducted on PubMed, Cochrane Library, and Medline databases published in the last 20 years on “coronary artery calcium” AND “statin” AND “cardiovascular disease” Further systematic review and meta-analysis using RevMan version 5.4 were performed based on the included published scientific articles. Results: Based on 11 studies included with a total of 1055 participants, we performed a meta-analysis and found that individuals with CAC score > 0 increased risk ratio of MI 8.48 (RR = 9.48: 95% CI: 6.22 – 14.45) times and MACE 2.48 (RR = 3.48: 95% CI: 2.98 – 4.05) times higher than CAC score 0 individual. Statin compared against non-statin treatment showed a statistically insignificant overall effect on the risk of MI (P = 0.81) and MACE (P = 0.89) in an individual with elevated CAC score 1 – 100 (P = 0.65) and > 100 (P = 0.11). Conclusions: This study found that an elevated CAC scores individual has a higher risk of MI and MACE than a non-elevated CAC score individual. There is no significant effect of statin compared against non-statin treatment to reduce MI and MACE in elevated CAC score individuals of 1 – 100 or > 100.

Keywords: coronary artery calcium, statin, cardiovascular disease, myocardial infarction, MACE

Procedia PDF Downloads 78
18057 Detecting HCC Tumor in Three Phasic CT Liver Images with Optimization of Neural Network

Authors: Mahdieh Khalilinezhad, Silvana Dellepiane, Gianni Vernazza

Abstract:

The aim of the present work is to build a model based on tissue characterization that is able to discriminate pathological and non-pathological regions from three-phasic CT images. Based on feature selection in different phases, in this research, we design a neural network system that has optimal neuron number in a hidden layer. Our approach consists of three steps: feature selection, feature reduction, and classification. For each ROI, 6 distinct set of texture features are extracted such as first order histogram parameters, absolute gradient, run-length matrix, co-occurrence matrix, autoregressive model, and wavelet, for a total of 270 texture features. We show that with the injection of liquid and the analysis of more phases the high relevant features in each region changed. Our results show that for detecting HCC tumor phase3 is the best one in most of the features that we apply to the classification algorithm. The percentage of detection between these two classes according to our method, relates to first order histogram parameters with the accuracy of 85% in phase 1, 95% phase 2, and 95% in phase 3.

Keywords: multi-phasic liver images, texture analysis, neural network, hidden layer

Procedia PDF Downloads 245
18056 Evaluation of the Irritation Potential of Three Topical Formulations of Minoxidil 5% Using Patch Test

Authors: Sule Pallavi, Shah Priyank, Thavkar Amit, Mehta Suyog, Rohira Poonam

Abstract:

Minoxidil is used topically to help hair growth in the treatment of male androgenetic alopecia. The objective of this study is to compare irritation potential of three conventional formulation of minoxidil 5% topical solution of in human patch test. The study was a single centre, double blind, non-randomized controlled study in 56 healthy adult Indian subjects. Occlusive patch test for 24 hours was performed with three formulation of minoxidil 5% topical solution. Products tested included aqueous based minoxidil 5% (AnasureTM 5%, Sun Pharma, India – Brand A), alcohol based minoxidil 5% (Brand B) and aqueous based minoxidil 5% (Brand C). Isotonic saline 0.9% and 1% w/w sodium lauryl sulphate were included as negative control and positive control respectively. Patches were applied and removed after 24hours. The skin reaction was assessed and clinically scored 24 hours after the removal of the patches under constant artificial daylight source using Draize scale (0-4 points scale for erythema/wrinkles/dryness and for oedema). A combined mean score up to 2.0/8.0 indicates a product is “non-irritant” and score between 2.0/8.0 and 4.0/8.0 indicates “mildly irritant” and score above 4.0/8.0 indicates “irritant”. Follow-up was scheduled after one week to confirm recovery for any reaction. The procedure of the patch test followed the principles outlined by Bureau of Indian standards (BIS) (IS 4011:2018; Methods of Test for safety evaluation of Cosmetics-3rd revision). Fifty six subjects with mean age 30.9 years (27 males and 29 females) participated in the study. The combined mean score (± standard deviation) were: 0.13 ± 0.33 (Brand A), 0.39 ± 0.49 (Brand B), 0.22 ± 0.41 (Brand C), 2.91 ± 0.79 (Positive control) and 0.02 ± 0.13 (Negative control). The mean score of Brand A (Sun Pharma product) was significantly lower than Brand B (p=0.001) and was comparable with Brand C (p=0.21). The combined mean erythema score (± standard deviation) were: 0.09 ± 0.29 (Brand A), 0.27 ± 0.5 (Brand B), 0.18 ± 0.39 (Brand C), 2.02 ± 0.49 (Positive control) and 0.0 ± 0.0 (Negative control). The mean erythema score of Brand A was significantly lower than Brand B (p=0.01) and was comparable with Brand C (p=0.16). Any reaction observed at 24hours after patch removal subsided in a week. All the three topical formulation of minoxidil 5% were non-irritant. Brand A of 5% minoxidil (Sun Pharma) was found to be least irritant than Brand B and Brand C based on the combined mean score and mean erythema score in the human patch test as per the BIS, IS 4011;2018.

Keywords: erythema, irritation, minoxidil, patch test

Procedia PDF Downloads 83
18055 Transverse Vibration of Non-Homogeneous Rectangular Plates of Variable Thickness Using GDQ

Authors: R. Saini, R. Lal

Abstract:

The effect of non-homogeneity on the free transverse vibration of thin rectangular plates of bilinearly varying thickness has been analyzed using generalized differential quadrature (GDQ) method. The non-homogeneity of the plate material is assumed to arise due to linear variations in Young’s modulus and density of the plate material with the in-plane coordinates x and y. Numerical results have been computed for fully clamped and fully simply supported boundary conditions. The solution procedure by means of GDQ method has been implemented in a MATLAB code. The effect of various plate parameters has been investigated for the first three modes of vibration. A comparison of results with those available in literature has been presented.

Keywords: rectangular, non-homogeneous, bilinear thickness, generalized differential quadrature (GDQ)

Procedia PDF Downloads 370
18054 Error Amount in Viscoelasticity Analysis Depending on Time Step Size and Method used in ANSYS

Authors: A. Fettahoglu

Abstract:

Theory of viscoelasticity is used by many researchers to represent behavior of many materials such as pavements on roads or bridges. Several researches used analytical methods and rheology to predict the material behaviors of simple models. Today, more complex engineering structures are analyzed using Finite Element Method, in which material behavior is embedded by means of three dimensional viscoelastic material laws. As a result, structures of unordinary geometry and domain like pavements of bridges can be analyzed by means of Finite Element Method and three dimensional viscoelastic equations. In the scope of this study, rheological models embedded in ANSYS, namely, generalized Maxwell elements and Prony series, which are two methods used by ANSYS to represent viscoelastic material behavior, are presented explicitly. Subsequently, a practical problem, which has an analytical solution given in literature, is used to verify the applicability of viscoelasticity tool embedded in ANSYS. Finally, amount of error in the results of ANSYS is compared with the analytical results to indicate the influence of used method and time step size.

Keywords: generalized Maxwell model, finite element method, prony series, time step size, viscoelasticity

Procedia PDF Downloads 346
18053 Reservoir Inflow Prediction for Pump Station Using Upstream Sewer Depth Data

Authors: Osung Im, Neha Yadav, Eui Hoon Lee, Joong Hoon Kim

Abstract:

Artificial Neural Network (ANN) approach is commonly used in lots of fields for forecasting. In water resources engineering, forecast of water level or inflow of reservoir is useful for various kind of purposes. Due to advantages of ANN, many papers were written for inflow prediction in river networks, but in this study, ANN is used in urban sewer networks. The growth of severe rain storm in Korea has increased flood damage severely, and the precipitation distribution is getting more erratic. Therefore, effective pump operation in pump station is an essential task for the reduction in urban area. If real time inflow of pump station reservoir can be predicted, it is possible to operate pump effectively for reducing the flood damage. This study used ANN model for pump station reservoir inflow prediction using upstream sewer depth data. For this study, rainfall events, sewer depth, and inflow into Banpo pump station reservoir between years of 2013-2014 were considered. Feed – Forward Back Propagation (FFBF), Cascade – Forward Back Propagation (CFBP), Elman Back Propagation (EBP) and Nonlinear Autoregressive Exogenous (NARX) were used as ANN model for prediction. A comparison of results with ANN model suggests that ANN is a powerful tool for inflow prediction using the sewer depth data.

Keywords: artificial neural network, forecasting, reservoir inflow, sewer depth

Procedia PDF Downloads 292
18052 An Educational Program Based on Health Belief Model to Prevent of Non-alcoholic Fatty Liver Disease Among Iranian Women

Authors: Arezoo Fallahi

Abstract:

Background and purpose: Non-alcoholic fatty liver is one of the most common liver disorders, which, as the most important cause of death from liver disease, has unpleasant consequences and complications. The aim of this study was to investigate the effect of an educational intervention based on a health belief model to prevent non-alcoholic fatty liver among women. Materials and Methods: This experimental study was performed among 110 women referring to comprehensive health service centers in Malayer City, west of Iran, in 2023. Using the available sampling method, 110 Participants were divided into experimental and control groups. The data collection tool included demographic characteristics and a questionnaire based on the health belief model. In The experimental group, three one-hour training sessions were conducted in the form of pamphlets, lectures and group discussions. Data were analyzed using SPSS software version 21, by correlation tests, paired t-tests independent t-tests. Results: The mean age of participants was 38.07±6.28 years, and Most of the participants were middle-aged, married, housewives with academic education, middle-income and overweight. After the educational intervention, the mean scores of the constructs include perceived sensitivity (p=0.01), perceived severity (p=0.01), perceived benefits (p=0.01), guidance for internal (p=0.01) and external action (p=0.01), and perceived self-efficacy (p=0.01) in the experimental group were significantly higher than the control group. The score of perceived barriers in the experimental group decreased after training. The perceived obstacles score in the test group decreased after the training (15.2 ± 3.9 v.s 11.2 ± 3.3, (p<0.01). Conclusion: The findings of the study showed that the design and implementation of educational programs based on the constructs of the health belief model can be effective in preventing women from developing higher levels of non-alcoholic fatty liver.

Keywords: health, education, believe, behaviour

Procedia PDF Downloads 24
18051 Using Computer Vision to Detect and Localize Fractures in Wrist X-ray Images

Authors: John Paul Q. Tomas, Mark Wilson L. de los Reyes, Kirsten Joyce P. Vasquez

Abstract:

The most frequent type of fracture is a wrist fracture, which often makes it difficult for medical professionals to find and locate. In this study, fractures in wrist x-ray pictures were located and identified using deep learning and computer vision. The researchers used image filtering, masking, morphological operations, and data augmentation for the image preprocessing and trained the RetinaNet and Faster R-CNN models with ResNet50 backbones and Adam optimizers separately for each image filtering technique and projection. The RetinaNet model with Anisotropic Diffusion Smoothing filter trained with 50 epochs has obtained the greatest accuracy of 99.14%, precision of 100%, sensitivity/recall of 98.41%, specificity of 100%, and an IoU score of 56.44% for the Posteroanterior projection utilizing augmented data. For the Lateral projection using augmented data, the RetinaNet model with an Anisotropic Diffusion filter trained with 50 epochs has produced the highest accuracy of 98.40%, precision of 98.36%, sensitivity/recall of 98.36%, specificity of 98.43%, and an IoU score of 58.69%. When comparing the test results of the different individual projections, models, and image filtering techniques, the Anisotropic Diffusion filter trained with 50 epochs has produced the best classification and regression scores for both projections.

Keywords: Artificial Intelligence, Computer Vision, Wrist Fracture, Deep Learning

Procedia PDF Downloads 57
18050 Self-Efficacy Perceptions and the Attitudes of Prospective Teachers towards Assessment and Evaluation

Authors: Münevver Başman, Ezel Tavşancıl

Abstract:

Making the right decisions about students depends on teachers’ use of the assessment and evaluation techniques effectively. In order to do that, teachers should have positive attitudes and adequate self-efficacy perception towards assessment and evaluation. The purpose of this study is to investigate relationship between self-efficacy perception and the attitudes of prospective teachers towards assessment and evaluation and what kind of differences these issues have in terms of a variety of demographic variables. The study group consisted of 277 prospective teachers who have been studying in different departments of Marmara University, Faculty of Education. In this study, ‘Personal Information Form’, ‘A Perceptual Scale for Measurement and Evaluation of Prospective Teachers Self-Efficacy in Education’ and ‘Attitudes toward Educational Measurement Inventory’ are applied. As a result, positive correlation was found between self-efficacy perceptions and the attitudes of prospective teachers towards assessment and evaluation. Considering different departments, there is a significant difference between the mean score of attitudes of prospective teachers and between the mean score of self-efficacy perceptions of them. However, considering variables of attending statistics class and the class types at the graduated high school, there is no significant difference between the mean score of attitudes of prospective teachers and between the mean score of self-efficacy perceptions of them.

Keywords: attitude, perception, prospective teacher, self-efficacy

Procedia PDF Downloads 277
18049 Using a Train-the-Trainer Model to Deliver Post-Partum Haemorrhage Simulation in Rural Uganda

Authors: Michael Campbell, Malaz Elsaddig, Kevin Jones

Abstract:

Background: Despite encouraging progress, global maternal mortality has remained stubbornly high since the declaration of the Millennium development goals. Sub-Saharan Africa accounts for well over half of maternal deaths with Post-Partum Haemorrhage (PPH) being the lead cause. ‘In house’ simulation training delivered by local doctors may be a sustainable approach for improving emergency obstetric care. The aim of this study was to evaluate the use of a Train-the-Trainer (TtT) model in a rural Ugandan hospital to ascertain whether it can feasibly improve practitioners’ management of PPH. Methods: Three Ugandan doctors underwent a training course to enable them to design and deliver simulation training. These doctors used MamaNatalie® models to simulate PPH scenarios for midwives, nurses and medical students. The main outcome was improvement in participants’ knowledge and confidence, assessed using self-reported scores on a 10-point scale. Results: The TtT model produced significant improvements in the confidence and knowledge scores of the ten participants. The mean confidence score rose significantly (p=0.0005) from 6.4 to 8.6 following the simulation training. There was also a significant increase in the mean knowledge score from 7.2 to 9.0 (p=0.04). Medical students demonstrated the greatest overall increase in confidence scores whilst increases in knowledge scores were largest amongst nurses. Conclusions: This study demonstrates that a TtT model can be used in a low resource setting to improve healthcare professionals’ confidence and knowledge in managing obstetric emergencies. This Train-the-Trainer model represents a sustainable approach to addressing skill deficits in low resource settings. We believe that its expansion across healthcare institutions in Sub-Saharan Africa will help to reduce the region’s high maternal mortality rate and step closer to achieving the ambitions of the Millennium development goals.

Keywords: low resource setting, post-partum haemorrhage, simulation training, train the trainer

Procedia PDF Downloads 161
18048 Asymmetrical Informative Estimation for Macroeconomic Model: Special Case in the Tourism Sector of Thailand

Authors: Chukiat Chaiboonsri, Satawat Wannapan

Abstract:

This paper used an asymmetric informative concept to apply in the macroeconomic model estimation of the tourism sector in Thailand. The variables used to statistically analyze are Thailand international and domestic tourism revenues, the expenditures of foreign and domestic tourists, service investments by private sectors, service investments by the government of Thailand, Thailand service imports and exports, and net service income transfers. All of data is a time-series index which was observed between 2002 and 2015. Empirically, the tourism multiplier and accelerator were estimated by two statistical approaches. The first was the result of the Generalized Method of Moments model (GMM) based on the assumption which the tourism market in Thailand had perfect information (Symmetrical data). The second was the result of the Maximum Entropy Bootstrapping approach (MEboot) based on the process that attempted to deal with imperfect information and reduced uncertainty in data observations (Asymmetrical data). In addition, the tourism leakages were investigated by a simple model based on the injections and leakages concept. The empirical findings represented the parameters computed from the MEboot approach which is different from the GMM method. However, both of the MEboot estimation and GMM model suggests that Thailand’s tourism sectors are in a period capable of stimulating the economy.

Keywords: TThailand tourism, Maximum Entropy Bootstrapping approach, macroeconomic model, asymmetric information

Procedia PDF Downloads 275
18047 Continuous Land Cover Change Detection in Subtropical Thicket Ecosystems

Authors: Craig Mahlasi

Abstract:

The Subtropical Thicket Biome has been in peril of transformation. Estimates indicate that as much as 63% of the Subtropical Thicket Biome is severely degraded. Agricultural expansion is the main driver of transformation. While several studies have sought to document and map the long term transformations, there is a lack of information on disturbance events that allow for timely intervention by authorities. Furthermore, tools that seek to perform continuous land cover change detection are often developed for forests and thus tend to perform poorly in thicket ecosystems. This study investigates the utility of Earth Observation data for continuous land cover change detection in Subtropical Thicket ecosystems. Temporal Neural Networks are implemented on a time series of Sentinel-2 observations. The model obtained 0.93 accuracy, a recall score of 0.93, and a precision score of 0.91 in detecting Thicket disturbances. The study demonstrates the potential of continuous land cover change in Subtropical Thicket ecosystems.

Keywords: remote sensing, land cover change detection, subtropical thickets, near-real time

Procedia PDF Downloads 138
18046 Prediction of Survival Rate after Gastrointestinal Surgery Based on The New Japanese Association for Acute Medicine (JAAM Score) With Neural Network Classification Method

Authors: Ayu Nabila Kusuma Pradana, Aprinaldi Jasa Mantau, Tomohiko Akahoshi

Abstract:

The incidence of Disseminated intravascular coagulation (DIC) following gastrointestinal surgery has a poor prognosis. Therefore, it is important to determine the factors that can predict the prognosis of DIC. This study will investigate the factors that may influence the outcome of DIC in patients after gastrointestinal surgery. Eighty-one patients were admitted to the intensive care unit after gastrointestinal surgery in Kyushu University Hospital from 2003 to 2021. Acute DIC scores were estimated using the new Japanese Association for Acute Medicine (JAAM) score from before and after surgery from day 1, day 3, and day 7. Acute DIC scores will be compared with The Sequential Organ Failure Assessment (SOFA) score, platelet count, lactate level, and a variety of biochemical parameters. This study applied machine learning algorithms to predict the prognosis of DIC after gastrointestinal surgery. The results of this study are expected to be used as an indicator for evaluating patient prognosis so that it can increase life expectancy and reduce mortality from cases of DIC patients after gastrointestinal surgery.

Keywords: the survival rate, gastrointestinal surgery, JAAM score, neural network, machine learning, disseminated intravascular coagulation (DIC)

Procedia PDF Downloads 228
18045 Modeling Bessel Beams and Their Discrete Superpositions from the Generalized Lorenz-Mie Theory to Calculate Optical Forces over Spherical Dielectric Particles

Authors: Leonardo A. Ambrosio, Carlos. H. Silva Santos, Ivan E. L. Rodrigues, Ayumi K. de Campos, Leandro A. Machado

Abstract:

In this work, we propose an algorithm developed under Python language for the modeling of ordinary scalar Bessel beams and their discrete superpositions and subsequent calculation of optical forces exerted over dielectric spherical particles. The mathematical formalism, based on the generalized Lorenz-Mie theory, is implemented in Python for its large number of free mathematical (as SciPy and NumPy), data visualization (Matplotlib and PyJamas) and multiprocessing libraries. We also propose an approach, provided by a synchronized Software as Service (SaaS) in cloud computing, to develop a user interface embedded on a mobile application, thus providing users with the necessary means to easily introduce desired unknowns and parameters and see the graphical outcomes of the simulations right at their mobile devices. Initially proposed as a free Android-based application, such an App enables data post-processing in cloud-based architectures and visualization of results, figures and numerical tables.

Keywords: Bessel Beams and Frozen Waves, Generalized Lorenz-Mie Theory, Numerical Methods, optical forces

Procedia PDF Downloads 360
18044 Evaluation of a Personalized Online Decision Aid for Colorectal Cancer Screening: A Randomized Controlled Trial

Authors: Linda P. M. Pluymen, Mariska M. G. Leeflang, I. Stegeman, Henock G. Yebyo, Anne E. M. Brabers, Patrick M. Bossuyt, E. Dekker, Anke J. Woudstra, Mirjam P. Fransen

Abstract:

Weighing the benefits and harms of colorectal cancer screening can be difficult for individuals. An existing online decision aid was expanded with a benefit-harm analysis to help people make an informed decision about participating in colorectal cancer screening. In a randomized controlled trial, we investigated whether those in the intervention group who used the decision aid with benefit-harm analysis were more certain about their decision than those in the control group who used the decision aid without benefit-harm analysis. Participants were 623 (39% of those invited) men and women aged 45 until 75 years old. Analyses were performed in those 386 participants (62%) who reported to have completed the entire decision aid. No statistically significant differences were observed between intervention and control group in decisional conflict score (mean difference 2.4, 95% CI -0.9, 5.6), clarity of values (mean difference 1.0, 95% CI -4.4, 6.6), deliberation score (mean difference 0.5, 95% CI -0.6, 1.7), anxiety score (mean difference 0.0, 95% CI -0.3, 0.3) and risk perception score (mean difference 0.1, -0.1, 0.3). Adding a benefit-harm analysis to an online decision aid did not improve informed decision making about participating in colorectal cancer screening.

Keywords: benefit-harm analysis, decision aid, informed decision making, personalized decision making

Procedia PDF Downloads 151