Search results for: data exploitation
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 24858

Search results for: data exploitation

19638 Marketing Mix, Motivation and the Tendency of Consumer Decision Making in Buying Condominium

Authors: Bundit Pungnirund

Abstract:

This research aimed to study the relationship between marketing mix attitudes, motivation of buying decision and tendency of consumer decision making in buying the condominiums in Thailand. This study employed by survey and quantitative research. The questionnaire was used to collect the data from 400 sampled of customers who interested in buying condominium in Bangkok. The descriptive statistics and Pearson’s correlation coefficient analysis were used to analyze data. The research found that marketing mixed factors in terms of product and price were related to buying decision making tendency in terms of price and room size. Marketing mixed factors in terms of price, place and promotion were related to buying decision making tendency in term of word of mouth. Consumers’ buying motivation in terms of social acceptance, self-esteemed and self-actualization were related to buying decision making tendency in term of room size. In addition, motivation in self-esteemed was related to buying decision making tendency within a year.

Keywords: condominium, marketing mix, motivation, tendency of consumer decision making

Procedia PDF Downloads 294
19637 Analysis of Mutation Associated with Male Infertility in Patients and Healthy Males in the Russian Population

Authors: Svetlana Zhikrivetskaya, Nataliya Shirokova, Roman Bikanov, Elizaveta Musatova, Yana Kovaleva, Nataliya Vetrova, Ekaterina Pomerantseva

Abstract:

Nowadays there is a growing number of couples with conceiving problems due to male or female infertility. Genetic abnormalities are responsible for about 31% of all cases of male infertility. These abnormalities include both chromosomal aberrations or aneuploidies and mutations in certain genes. Chromosomal abnormalities can be easily identified, thus the development of screening panels able to reveal genetic reasons of male infertility on gene level is of current interest. There are approximately 2,000 genes involved in male fertility that is the reason why it is very important to determine the most clinically relevant in certain population and ethnic conditions. An infertility screening panel containing 48 mutations in genes AMHR2, CFTR, DNAI1, HFE, KAL1, TSSK2 and AZF locus which are the most clinically relevant for the European population according to databases NCBI and ClinVar was designed. The aim of this research was to confirm clinic relevance of these mutations in the Russian population. Genotyping was performed in 220 patients with different types of male infertility and in 57 healthy males with normozoospermia. Mutations were identified by end-point PCR with TaqMan probes in microfluidic plates. The frequency of 5 mutations in healthy males and 13 mutations in patients with infertility was revealed and estimated. The frequency of mutation c.187C>G in HFE gene was significantly lower for healthy males (8.8%) compared with patients (17.7%) and the values for the European population according to ExAc database (13.7%) and dbSNP (17.2%). Analysis of c.3454G>C, and c.1545_1546delTA mutations in the CFTR gene revealed increased frequency (0.9 and 0.2%, respectively) in patients with infertility compared with data for the European population (0.04%, respectively (ExAc, European (Non-Finnish) and for the Aggregated Populations (0.002% (ExAc), because there is no data for European population for c.1545_1546delTA mutation. The frequency of del508 mutation (CFTR) in patients (1.59%) were lower comparing with male infertility Europeans (3.34-6.25% depending on nationality) and at the same level with healthy Europeans (1.06%, ExAc, European (Non-Finnish). Analysis of c.845G>A (HFE) mutation resulted in decreased frequency in patients (1.8%) in contrast with the European population data (5.1%, respectively, ExAc, European (Non-Finnish). Moreover, obtained data revealed no statistically significant frequency difference for c.845G>A mutation (HFE) between healthy males in the Russian and the European populations. Allele frequencies of mutations c.350G>A (CFTR), c.193A>T (HFE), c.774C>T, and c.80A>G (gene TSSK2) showed no significantly difference among patients with infertility, healthy males and Europeans. Analysis of AZF locus revealed increased frequency for AZFc microdeletion in patients with male infertility. Thereby, the new data of the allele frequencies in infertility patients in the Russian population was obtained. As well as the frequency differences of mutations associated with male infertility among patients, healthy males in the Russian population and the European one were estimated. The revealed differences showed that for high effectiveness of screening panel detecting genetically caused male infertility it is very important to consider ethnic and population characteristics of patients which will be screened.

Keywords: allele frequency, azoospermia, male infertility, mutation, population

Procedia PDF Downloads 378
19636 Scope of Rainwater Harvesting in Residential Plots of Dhaka City

Authors: Jubaida Gulshan Ara, Zebun Nasreen Ahmed

Abstract:

Urban flood and drought has been a major problem of Dhaka city, particularly in recent years. Continuous increase of the city built up area, and limiting rainwater infiltration zone, are thought to be the main causes of the problem. Proper rainwater management, even at the individual plot level, might bring significant improvement in this regard. As residential use pattern occupies a significant portion of the city surface, the scope of rainwater harvesting (RWH) in residential buildings can be investigated. This paper reports on a research which explored the scope of rainwater harvesting in residential plots, with multifamily apartment buildings, in Dhaka city. The research investigated the basics of RWH, contextual information, i.e., hydro-geological, meteorological data of Dhaka city and the rules and legislations for residential building construction. The study also explored contemporary rainwater harvesting practices in the local and international contexts. On the basis of theoretical understanding, 21 sample case-studies, in different phases of construction, were selected from seven different categories of plot sizes, in different residential areas of Dhaka city. Primary data from the 21 case-study buildings were collected from a physical survey, from design drawings, accompanied by a questionnaire survey. All necessary secondary data were gathered from published and other relevant sources. Collected primary and secondary data were used to calculate and analyze the RWH needs for each case study, based on the theoretical understanding. The main findings have been compiled and compared, to observe residential development trends with regards to building rainwater harvesting system. The study has found that, in ‘Multifamily Apartment Building’ of Dhaka city, storage, and recharge structure size for rainwater harvesting, increases along with occupants’ number, and with the increasing size of the plot. Hence, demand vs. supply ratio remains almost the same for different sizes of plots, and consequently, the size of the storage structure increases significantly, in large-scale plots. It has been found that rainwater can meet only 12%-30% of the total restricted water demand of these residential buildings of Dhaka city. Therefore, artificial groundwater recharge might be the more suitable option for RWH, than storage. The study came up with this conclusion that, in multifamily residential apartments of Dhaka city, artificial groundwater recharge might be the more suitable option for RWH, than storing the rainwater on site.

Keywords: Dhaka city, rainwater harvesting, residential plots, urban flood

Procedia PDF Downloads 174
19635 Security of Database Using Chaotic Systems

Authors: Eman W. Boghdady, A. R. Shehata, M. A. Azem

Abstract:

Database (DB) security demands permitting authorized users and prohibiting non-authorized users and intruders actions on the DB and the objects inside it. Organizations that are running successfully demand the confidentiality of their DBs. They do not allow the unauthorized access to their data/information. They also demand the assurance that their data is protected against any malicious or accidental modification. DB protection and confidentiality are the security concerns. There are four types of controls to obtain the DB protection, those include: access control, information flow control, inference control, and cryptographic. The cryptographic control is considered as the backbone for DB security, it secures the DB by encryption during storage and communications. Current cryptographic techniques are classified into two types: traditional classical cryptography using standard algorithms (DES, AES, IDEA, etc.) and chaos cryptography using continuous (Chau, Rossler, Lorenz, etc.) or discreet (Logistics, Henon, etc.) algorithms. The important characteristics of chaos are its extreme sensitivity to initial conditions of the system. In this paper, DB-security systems based on chaotic algorithms are described. The Pseudo Random Numbers Generators (PRNGs) from the different chaotic algorithms are implemented using Matlab and their statistical properties are evaluated using NIST and other statistical test-suits. Then, these algorithms are used to secure conventional DB (plaintext), where the statistical properties of the ciphertext are also tested. To increase the complexity of the PRNGs and to let pass all the NIST statistical tests, we propose two hybrid PRNGs: one based on two chaotic Logistic maps and another based on two chaotic Henon maps, where each chaotic algorithm is running side-by-side and starting from random independent initial conditions and parameters (encryption keys). The resulted hybrid PRNGs passed the NIST statistical test suit.

Keywords: algorithms and data structure, DB security, encryption, chaotic algorithms, Matlab, NIST

Procedia PDF Downloads 257
19634 Using Speech Emotion Recognition as a Longitudinal Biomarker for Alzheimer’s Diseases

Authors: Yishu Gong, Liangliang Yang, Jianyu Zhang, Zhengyu Chen, Sihong He, Xusheng Zhang, Wei Zhang

Abstract:

Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that affects millions of people worldwide and is characterized by cognitive decline and behavioral changes. People living with Alzheimer’s disease often find it hard to complete routine tasks. However, there are limited objective assessments that aim to quantify the difficulty of certain tasks for AD patients compared to non-AD people. In this study, we propose to use speech emotion recognition (SER), especially the frustration level, as a potential biomarker for quantifying the difficulty patients experience when describing a picture. We build an SER model using data from the IEMOCAP dataset and apply the model to the DementiaBank data to detect the AD/non-AD group difference and perform longitudinal analysis to track the AD disease progression. Our results show that the frustration level detected from the SER model can possibly be used as a cost-effective tool for objective tracking of AD progression in addition to the Mini-Mental State Examination (MMSE) score.

Keywords: Alzheimer’s disease, speech emotion recognition, longitudinal biomarker, machine learning

Procedia PDF Downloads 95
19633 Establishment of a Classifier Model for Early Prediction of Acute Delirium in Adult Intensive Care Unit Using Machine Learning

Authors: Pei Yi Lin

Abstract:

Objective: The objective of this study is to use machine learning methods to build an early prediction classifier model for acute delirium to improve the quality of medical care for intensive care patients. Background: Delirium is a common acute and sudden disturbance of consciousness in critically ill patients. After the occurrence, it is easy to prolong the length of hospital stay and increase medical costs and mortality. In 2021, the incidence of delirium in the intensive care unit of internal medicine was as high as 59.78%, which indirectly prolonged the average length of hospital stay by 8.28 days, and the mortality rate is about 2.22% in the past three years. Therefore, it is expected to build a delirium prediction classifier through big data analysis and machine learning methods to detect delirium early. Method: This study is a retrospective study, using the artificial intelligence big data database to extract the characteristic factors related to delirium in intensive care unit patients and let the machine learn. The study included patients aged over 20 years old who were admitted to the intensive care unit between May 1, 2022, and December 31, 2022, excluding GCS assessment <4 points, admission to ICU for less than 24 hours, and CAM-ICU evaluation. The CAMICU delirium assessment results every 8 hours within 30 days of hospitalization are regarded as an event, and the cumulative data from ICU admission to the prediction time point are extracted to predict the possibility of delirium occurring in the next 8 hours, and collect a total of 63,754 research case data, extract 12 feature selections to train the model, including age, sex, average ICU stay hours, visual and auditory abnormalities, RASS assessment score, APACHE-II Score score, number of invasive catheters indwelling, restraint and sedative and hypnotic drugs. Through feature data cleaning, processing and KNN interpolation method supplementation, a total of 54595 research case events were extracted to provide machine learning model analysis, using the research events from May 01 to November 30, 2022, as the model training data, 80% of which is the training set for model training, and 20% for the internal verification of the verification set, and then from December 01 to December 2022 The CU research event on the 31st is an external verification set data, and finally the model inference and performance evaluation are performed, and then the model has trained again by adjusting the model parameters. Results: In this study, XG Boost, Random Forest, Logistic Regression, and Decision Tree were used to analyze and compare four machine learning models. The average accuracy rate of internal verification was highest in Random Forest (AUC=0.86), and the average accuracy rate of external verification was in Random Forest and XG Boost was the highest, AUC was 0.86, and the average accuracy of cross-validation was the highest in Random Forest (ACC=0.77). Conclusion: Clinically, medical staff usually conduct CAM-ICU assessments at the bedside of critically ill patients in clinical practice, but there is a lack of machine learning classification methods to assist ICU patients in real-time assessment, resulting in the inability to provide more objective and continuous monitoring data to assist Clinical staff can more accurately identify and predict the occurrence of delirium in patients. It is hoped that the development and construction of predictive models through machine learning can predict delirium early and immediately, make clinical decisions at the best time, and cooperate with PADIS delirium care measures to provide individualized non-drug interventional care measures to maintain patient safety, and then Improve the quality of care.

Keywords: critically ill patients, machine learning methods, delirium prediction, classifier model

Procedia PDF Downloads 51
19632 Design and Implementation of Generative Models for Odor Classification Using Electronic Nose

Authors: Kumar Shashvat, Amol P. Bhondekar

Abstract:

In the midst of the five senses, odor is the most reminiscent and least understood. Odor testing has been mysterious and odor data fabled to most practitioners. The delinquent of recognition and classification of odor is important to achieve. The facility to smell and predict whether the artifact is of further use or it has become undesirable for consumption; the imitation of this problem hooked on a model is of consideration. The general industrial standard for this classification is color based anyhow; odor can be improved classifier than color based classification and if incorporated in machine will be awfully constructive. For cataloging of odor for peas, trees and cashews various discriminative approaches have been used Discriminative approaches offer good prognostic performance and have been widely used in many applications but are incapable to make effectual use of the unlabeled information. In such scenarios, generative approaches have better applicability, as they are able to knob glitches, such as in set-ups where variability in the series of possible input vectors is enormous. Generative models are integrated in machine learning for either modeling data directly or as a transitional step to form an indeterminate probability density function. The algorithms or models Linear Discriminant Analysis and Naive Bayes Classifier have been used for classification of the odor of cashews. Linear Discriminant Analysis is a method used in data classification, pattern recognition, and machine learning to discover a linear combination of features that typifies or divides two or more classes of objects or procedures. The Naive Bayes algorithm is a classification approach base on Bayes rule and a set of qualified independence theory. Naive Bayes classifiers are highly scalable, requiring a number of restraints linear in the number of variables (features/predictors) in a learning predicament. The main recompenses of using the generative models are generally a Generative Models make stronger assumptions about the data, specifically, about the distribution of predictors given the response variables. The Electronic instrument which is used for artificial odor sensing and classification is an electronic nose. This device is designed to imitate the anthropological sense of odor by providing an analysis of individual chemicals or chemical mixtures. The experimental results have been evaluated in the form of the performance measures i.e. are accuracy, precision and recall. The investigational results have proven that the overall performance of the Linear Discriminant Analysis was better in assessment to the Naive Bayes Classifier on cashew dataset.

Keywords: odor classification, generative models, naive bayes, linear discriminant analysis

Procedia PDF Downloads 365
19631 Ribotaxa: Combined Approaches for Taxonomic Resolution Down to the Species Level from Metagenomics Data Revealing Novelties

Authors: Oshma Chakoory, Sophie Comtet-Marre, Pierre Peyret

Abstract:

Metagenomic classifiers are widely used for the taxonomic profiling of metagenomic data and estimation of taxa relative abundance. Small subunit rRNA genes are nowadays a gold standard for the phylogenetic resolution of complex microbial communities, although the power of this marker comes down to its use as full-length. We benchmarked the performance and accuracy of rRNA-specialized versus general-purpose read mappers, reference-targeted assemblers and taxonomic classifiers. We then built a pipeline called RiboTaxa to generate a highly sensitive and specific metataxonomic approach. Using metagenomics data, RiboTaxa gave the best results compared to other tools (Kraken2, Centrifuge (1), METAXA2 (2), PhyloFlash (3)) with precise taxonomic identification and relative abundance description, giving no false positive detection. Using real datasets from various environments (ocean, soil, human gut) and from different approaches (metagenomics and gene capture by hybridization), RiboTaxa revealed microbial novelties not seen by current bioinformatics analysis opening new biological perspectives in human and environmental health. In a study focused on corals’ health involving 20 metagenomic samples (4), an affiliation of prokaryotes was limited to the family level with Endozoicomonadaceae characterising healthy octocoral tissue. RiboTaxa highlighted 2 species of uncultured Endozoicomonas which were dominant in the healthy tissue. Both species belonged to a genus not yet described, opening new research perspectives on corals’ health. Applied to metagenomics data from a study on human gut and extreme longevity (5), RiboTaxa detected the presence of an uncultured archaeon in semi-supercentenarians (aged 105 to 109 years) highlighting an archaeal genus, not yet described, and 3 uncultured species belonging to the Enorma genus that could be species of interest participating in the longevity process. RiboTaxa is user-friendly, rapid, allowing microbiota structure description from any environment and the results can be easily interpreted. This software is freely available at https://github.com/oschakoory/RiboTaxa under the GNU Affero General Public License 3.0.

Keywords: metagenomics profiling, microbial diversity, SSU rRNA genes, full-length phylogenetic marker

Procedia PDF Downloads 98
19630 A Comparative Analysis of Geometric and Exponential Laws in Modelling the Distribution of the Duration of Daily Precipitation

Authors: Mounia El Hafyani, Khalid El Himdi

Abstract:

Precipitation is one of the key variables in water resource planning. The importance of modeling wet and dry durations is a crucial pointer in engineering hydrology. The objective of this study is to model and analyze the distribution of wet and dry durations. For this purpose, the daily rainfall data from 1967 to 2017 of the Moroccan city of Kenitra’s station are used. Three models are implemented for the distribution of wet and dry durations, namely the first-order Markov chain, the second-order Markov chain, and the truncated negative binomial law. The adherence of the data to the proposed models is evaluated using Chi-square and Kolmogorov-Smirnov tests. The Akaike information criterion is applied to assess the most effective model distribution. We go further and study the law of the number of wet and dry days among k consecutive days. The calculation of this law is done through an algorithm that we have implemented based on conditional laws. We complete our work by comparing the observed moments of the numbers of wet/dry days among k consecutive days to the calculated moment of the three estimated models. The study shows the effectiveness of our approach in modeling wet and dry durations of daily precipitation.

Keywords: Markov chain, rainfall, truncated negative binomial law, wet and dry durations

Procedia PDF Downloads 111
19629 The Effect of Antibiotic Use on Blood Cultures: Implications for Future Policy

Authors: Avirup Chowdhury, Angus K. McFadyen, Linsey Batchelor

Abstract:

Blood cultures (BCs) are an important aspect of management of the septic patient, identifying the underlying pathogen and its antibiotic sensitivities. However, while the current literature outlines indications for initial BCs to be taken, there is little guidance for repeat sampling in the following 5-day period and little information on how antibiotic use can affect the usefulness of this investigation. A retrospective cohort study was conducted using inpatients who had undergone 2 or more BCs within 5 days between April 2016 and April 2017 at a 400-bed hospital in the west of Scotland and received antibiotic therapy between the first and second BCs. The data for BC sampling was collected from the electronic microbiology database, and cross-referenced with data from the hospital electronic prescribing system. Overall, 283 BCs were included in the study, taken from 92 patients (mean 3.08 cultures per patient, range 2-10). All 92 patients had initial BCs, of which 83 were positive (90%). 65 had a further sample within 24 hours of commencement of antibiotics, with 35 positive (54%). 23 had samples within 24-48 hours, with 4 (17%) positive; 12 patients had sampling at 48-72 hours, 12 at 72-96 hours, and 10 at 96-120 hours, with none positive. McNemar’s Exact Test was used to calculate statistical significance for patients who received blood cultures in multiple time blocks (Initial, < 24h, 24-120h, > 120h). For initial vs. < 24h-post BCs (53 patients tested), the proportion of positives fell from 46/53 to 29/53 (one-tailed P=0.002, OR 3.43, 95% CI 1.48-7.96). For initial vs 24-120h (n=42), the proportions were 38/42 and 4/42 respectively (P < 0.001, OR 35.0, 95% CI 4.79-255.48). For initial vs > 120h (n=36), these were 33/36 and 2/36 (P < 0.001,OR ∞). These were also calculated for a positive in initial or < 24h vs. 24-120h (n=42), with proportions of 41/42 and 4/42 (P < 0.001, OR 38.0, 95% CI 5.22-276.78); and for initial or < 24h vs > 120h (n=36), with proportions of 35/36 and 2/36 respectively (P < 0.001, OR ∞). This data appears to show that taking an initial BC followed by a BC within 24 hours of antibiotic commencement would maximise blood culture yield while minimising the risk of false negative results. This could potentially remove the need for as many as 46% of BC samples without adversely affecting patient care. BC yield decreases sharply after 48 hours of antibiotic use, and may not provide any clinically useful information after this time. Further multi-centre studies would validate these findings, and provide a foundation for future health policy generation.

Keywords: antibiotics, blood culture, efficacy, inpatient

Procedia PDF Downloads 160
19628 Data Clustering Algorithm Based on Multi-Objective Periodic Bacterial Foraging Optimization with Two Learning Archives

Authors: Chen Guo, Heng Tang, Ben Niu

Abstract:

Clustering splits objects into different groups based on similarity, making the objects have higher similarity in the same group and lower similarity in different groups. Thus, clustering can be treated as an optimization problem to maximize the intra-cluster similarity or inter-cluster dissimilarity. In real-world applications, the datasets often have some complex characteristics: sparse, overlap, high dimensionality, etc. When facing these datasets, simultaneously optimizing two or more objectives can obtain better clustering results than optimizing one objective. However, except for the objectives weighting methods, traditional clustering approaches have difficulty in solving multi-objective data clustering problems. Due to this, evolutionary multi-objective optimization algorithms are investigated by researchers to optimize multiple clustering objectives. In this paper, the Data Clustering algorithm based on Multi-objective Periodic Bacterial Foraging Optimization with two Learning Archives (DC-MPBFOLA) is proposed. Specifically, first, to reduce the high computing complexity of the original BFO, periodic BFO is employed as the basic algorithmic framework. Then transfer the periodic BFO into a multi-objective type. Second, two learning strategies are proposed based on the two learning archives to guide the bacterial swarm to move in a better direction. On the one hand, the global best is selected from the global learning archive according to the convergence index and diversity index. On the other hand, the personal best is selected from the personal learning archive according to the sum of weighted objectives. According to the aforementioned learning strategies, a chemotaxis operation is designed. Third, an elite learning strategy is designed to provide fresh power to the objects in two learning archives. When the objects in these two archives do not change for two consecutive times, randomly initializing one dimension of objects can prevent the proposed algorithm from falling into local optima. Fourth, to validate the performance of the proposed algorithm, DC-MPBFOLA is compared with four state-of-art evolutionary multi-objective optimization algorithms and one classical clustering algorithm on evaluation indexes of datasets. To further verify the effectiveness and feasibility of designed strategies in DC-MPBFOLA, variants of DC-MPBFOLA are also proposed. Experimental results demonstrate that DC-MPBFOLA outperforms its competitors regarding all evaluation indexes and clustering partitions. These results also indicate that the designed strategies positively influence the performance improvement of the original BFO.

Keywords: data clustering, multi-objective optimization, bacterial foraging optimization, learning archives

Procedia PDF Downloads 123
19627 The Usage of Bridge Estimator for Hegy Seasonal Unit Root Tests

Authors: Huseyin Guler, Cigdem Kosar

Abstract:

The aim of this study is to propose Bridge estimator for seasonal unit root tests. Seasonality is an important factor for many economic time series. Some variables may contain seasonal patterns and forecasts that ignore important seasonal patterns have a high variance. Therefore, it is very important to eliminate seasonality for seasonal macroeconomic data. There are some methods to eliminate the impacts of seasonality in time series. One of them is filtering the data. However, this method leads to undesired consequences in unit root tests, especially if the data is generated by a stochastic seasonal process. Another method to eliminate seasonality is using seasonal dummy variables. Some seasonal patterns may result from stationary seasonal processes, which are modelled using seasonal dummies but if there is a varying and changing seasonal pattern over time, so the seasonal process is non-stationary, deterministic seasonal dummies are inadequate to capture the seasonal process. It is not suitable to use seasonal dummies for modeling such seasonally nonstationary series. Instead of that, it is necessary to take seasonal difference if there are seasonal unit roots in the series. Different alternative methods are proposed in the literature to test seasonal unit roots, such as Dickey, Hazsa, Fuller (DHF) and Hylleberg, Engle, Granger, Yoo (HEGY) tests. HEGY test can be also used to test the seasonal unit root in different frequencies (monthly, quarterly, and semiannual). Another issue in unit root tests is the lag selection. Lagged dependent variables are added to the model in seasonal unit root tests as in the unit root tests to overcome the autocorrelation problem. In this case, it is necessary to choose the lag length and determine any deterministic components (i.e., a constant and trend) first, and then use the proper model to test for seasonal unit roots. However, this two-step procedure might lead size distortions and lack of power in seasonal unit root tests. Recent studies show that Bridge estimators are good in selecting optimal lag length while differentiating nonstationary versus stationary models for nonseasonal data. The advantage of this estimator is the elimination of the two-step nature of conventional unit root tests and this leads a gain in size and power. In this paper, the Bridge estimator is proposed to test seasonal unit roots in a HEGY model. A Monte-Carlo experiment is done to determine the efficiency of this approach and compare the size and power of this method with HEGY test. Since Bridge estimator performs well in model selection, our approach may lead to some gain in terms of size and power over HEGY test.

Keywords: bridge estimators, HEGY test, model selection, seasonal unit root

Procedia PDF Downloads 318
19626 Determinant Factor of Farm Household Fruit Tree Planting: The Case of Habru Woreda, North Wollo

Authors: Getamesay Kassaye Dimru

Abstract:

The cultivation of fruit tree in degraded areas has two-fold importance. Firstly, it improves food availability and income, and secondly, it promotes the conservation of soil and water improving, in turn, the productivity of the land. The main objectives of this study are to identify the determinant of farmer's fruit trees plantation decision and to major fruit production challenges and opportunities of the study area. The analysis was made using primary data collected from 60 sample household selected randomly from the study area in 2016. The primary data was supplemented by data collected from a key informant. In addition to the descriptive statistics and statistical tests (Chi-square test and t-test), a logit model was employed to identify the determinant of fruit tree plantation decision. Drought, pest incidence, land degradation, lack of input, lack of capital and irrigation schemes maintenance, lack of misuse of irrigation water and limited agricultural personnel are the major production constraints identified. The opportunities that need to further exploited are better access to irrigation, main road access, endowment of preferred guava variety, experience of farmers, and proximity of the study area to research center. The result of logit model shows that from different factors hypothesized to determine fruit tree plantation decision, age of the household head accesses to market and perception of farmers about fruits' disease and pest resistance are found to be significant. The result has revealed important implications for the promotion of fruit production for both land degradation control and rehabilitation and increasing the livelihood of farming households.

Keywords: degradation, fruit, irrigation, pest

Procedia PDF Downloads 207
19625 Calculation of Pressure-Varying Langmuir and Brunauer-Emmett-Teller Isotherm Adsorption Parameters

Authors: Trevor C. Brown, David J. Miron

Abstract:

Gas-solid physical adsorption methods are central to the characterization and optimization of the effective surface area, pore size and porosity for applications such as heterogeneous catalysis, and gas separation and storage. Properties such as adsorption uptake, capacity, equilibrium constants and Gibbs free energy are dependent on the composition and structure of both the gas and the adsorbent. However, challenges remain, in accurately calculating these properties from experimental data. Gas adsorption experiments involve measuring the amounts of gas adsorbed over a range of pressures under isothermal conditions. Various constant-parameter models, such as Langmuir and Brunauer-Emmett-Teller (BET) theories are used to provide information on adsorbate and adsorbent properties from the isotherm data. These models typically do not provide accurate interpretations across the full range of pressures and temperatures. The Langmuir adsorption isotherm is a simple approximation for modelling equilibrium adsorption data and has been effective in estimating surface areas and catalytic rate laws, particularly for high surface area solids. The Langmuir isotherm assumes the systematic filling of identical adsorption sites to a monolayer coverage. The BET model is based on the Langmuir isotherm and allows for the formation of multiple layers. These additional layers do not interact with the first layer and the energetics are equal to the adsorbate as a bulk liquid. This BET method is widely used to measure the specific surface area of materials. Both Langmuir and BET models assume that the affinity of the gas for all adsorption sites are identical and so the calculated adsorbent uptake at the monolayer and equilibrium constant are independent of coverage and pressure. Accurate representations of adsorption data have been achieved by extending the Langmuir and BET models to include pressure-varying uptake capacities and equilibrium constants. These parameters are determined using a novel regression technique called flexible least squares for time-varying linear regression. For isothermal adsorption the adsorption parameters are assumed to vary slowly and smoothly with increasing pressure. The flexible least squares for pressure-varying linear regression (FLS-PVLR) approach assumes two distinct types of discrepancy terms, dynamic and measurement for all parameters in the linear equation used to simulate the data. Dynamic terms account for pressure variation in successive parameter vectors, and measurement terms account for differences between observed and theoretically predicted outcomes via linear regression. The resultant pressure-varying parameters are optimized by minimizing both dynamic and measurement residual squared errors. Validation of this methodology has been achieved by simulating adsorption data for n-butane and isobutane on activated carbon at 298 K, 323 K and 348 K and for nitrogen on mesoporous alumina at 77 K with pressure-varying Langmuir and BET adsorption parameters (equilibrium constants and uptake capacities). This modeling provides information on the adsorbent (accessible surface area and micropore volume), adsorbate (molecular areas and volumes) and thermodynamic (Gibbs free energies) variations of the adsorption sites.

Keywords: Langmuir adsorption isotherm, BET adsorption isotherm, pressure-varying adsorption parameters, adsorbate and adsorbent properties and energetics

Procedia PDF Downloads 217
19624 Citation Analysis of New Zealand Court Decisions

Authors: Tobias Milz, L. Macpherson, Varvara Vetrova

Abstract:

The law is a fundamental pillar of human societies as it shapes, controls and governs how humans conduct business, behave and interact with each other. Recent advances in computer-assisted technologies such as NLP, data science and AI are creating opportunities to support the practice, research and study of this pervasive domain. It is therefore not surprising that there has been an increase in investments into supporting technologies for the legal industry (also known as “legal tech” or “law tech”) over the last decade. A sub-discipline of particular appeal is concerned with assisted legal research. Supporting law researchers and practitioners to retrieve information from the vast amount of ever-growing legal documentation is of natural interest to the legal research community. One tool that has been in use for this purpose since the early nineteenth century is legal citation indexing. Among other use cases, they provided an effective means to discover new precedent cases. Nowadays, computer-assisted network analysis tools can allow for new and more efficient ways to reveal the “hidden” information that is conveyed through citation behavior. Unfortunately, access to openly available legal data is still lacking in New Zealand and access to such networks is only commercially available via providers such as LexisNexis. Consequently, there is a need to create, analyze and provide a legal citation network with sufficient data to support legal research tasks. This paper describes the development and analysis of a legal citation Network for New Zealand containing over 300.000 decisions from 125 different courts of all areas of law and jurisdiction. Using python, the authors assembled web crawlers, scrapers and an OCR pipeline to collect and convert court decisions from openly available sources such as NZLII into uniform and machine-readable text. This facilitated the use of regular expressions to identify references to other court decisions from within the decision text. The data was then imported into a graph-based database (Neo4j) with the courts and their respective cases represented as nodes and the extracted citations as links. Furthermore, additional links between courts of connected cases were added to indicate an indirect citation between the courts. Neo4j, as a graph-based database, allows efficient querying and use of network algorithms such as PageRank to reveal the most influential/most cited courts and court decisions over time. This paper shows that the in-degree distribution of the New Zealand legal citation network resembles a power-law distribution, which indicates a possible scale-free behavior of the network. This is in line with findings of the respective citation networks of the U.S. Supreme Court, Austria and Germany. The authors of this paper provide the database as an openly available data source to support further legal research. The decision texts can be exported from the database to be used for NLP-related legal research, while the network can be used for in-depth analysis. For example, users of the database can specify the network algorithms and metrics to only include specific courts to filter the results to the area of law of interest.

Keywords: case citation network, citation analysis, network analysis, Neo4j

Procedia PDF Downloads 92
19623 'Sea Power: Concept, Influence and Securitization'; the Nigerian Navy's Role in a Developing State like Nigeria

Authors: William Abiodun Duyile

Abstract:

It is common knowledge that marine food has always been found from the sea, energy can also be found underneath and, to a growing extent; other mineral resources have come from the sea spaces. It is the importance of the sea and the sea lines of communication to littoral nations that has made concepts such as sea power, naval power, etc., significant to them. The study relied on documentary data. The documentary data were sourced from government annual departmental reports, newspapers and correspondence. The secondary sources used were subjected to internal and external criticism for authentication, and then to textual and contextual analyses. The study found that the differential level of seamanship amongst states defined their relationship. It was sea power that gave some states an edge over the others. The study proves that over the ages sea power has been core to the development of States or Empires. The study found that the Nigerian Navy was centre to Nigeria’s conquest of the littoral areas of Biafra, like Bonny, Port-Harcourt, and Calabar; it was also an important turning point of the Nigerian civil war since by it Biafra became landlocked. The research was able to identify succinctly the Nigerian Navy’s contribution to the security and development of the Nigerian State.

Keywords: sea power, naval power, land locked states, warship

Procedia PDF Downloads 124
19622 Heterogeneous Reactions to Digital Opportunities: A Field Study

Authors: Bangaly Kaba

Abstract:

In the global information society, the importance of the Internet cannot be overemphasized. Africa needs access to the powerful information and communication tools of the Internet in order to obtain the resources and efficiency essential for sustainable development. Unfortunately, in 2013, the data from Internetworldstats showed only 15% of African populations have access to Internet. This relative low Internet penetration rate signals a problem that may threaten the economic development, governmental efficiency, and ultimately the global competitiveness of African countries. Many initiatives were undertaken to bring the benefits of the global information revolution to the people of Africa, through connection to the Internet and other Global Information Infrastructure technologies. The purpose is to understand differences between socio-economically advantaged and disadvantaged internet users. From that, we will determine what prevents disadvantaged groups from benefiting from Internet usage. Data were collected through a survey from Internet users in Ivory Coast. The results reveal that Personal network exposure, Self-efficacy and Availability are the key drivers of continued use intention for the socio-economically disadvantaged group. The theoretical and practical implications are also described.

Keywords: digital inequality, internet, integrative model, socio-economically advantaged and disadvantaged, use continuance, Africa

Procedia PDF Downloads 460
19621 Local Binary Patterns-Based Statistical Data Analysis for Accurate Soccer Match Prediction

Authors: Mohammad Ghahramani, Fahimeh Saei Manesh

Abstract:

Winning a soccer game is based on thorough and deep analysis of the ongoing match. On the other hand, giant gambling companies are in vital need of such analysis to reduce their loss against their customers. In this research work, we perform deep, real-time analysis on every soccer match around the world that distinguishes our work from others by focusing on particular seasons, teams and partial analytics. Our contributions are presented in the platform called “Analyst Masters.” First, we introduce various sources of information available for soccer analysis for teams around the world that helped us record live statistical data and information from more than 50,000 soccer matches a year. Our second and main contribution is to introduce our proposed in-play performance evaluation. The third contribution is developing new features from stable soccer matches. The statistics of soccer matches and their odds before and in-play are considered in the image format versus time including the halftime. Local Binary patterns, (LBP) is then employed to extract features from the image. Our analyses reveal incredibly interesting features and rules if a soccer match has reached enough stability. For example, our “8-minute rule” implies if 'Team A' scores a goal and can maintain the result for at least 8 minutes then the match would end in their favor in a stable match. We could also make accurate predictions before the match of scoring less/more than 2.5 goals. We benefit from the Gradient Boosting Trees, GBT, to extract highly related features. Once the features are selected from this pool of data, the Decision trees decide if the match is stable. A stable match is then passed to a post-processing stage to check its properties such as betters’ and punters’ behavior and its statistical data to issue the prediction. The proposed method was trained using 140,000 soccer matches and tested on more than 100,000 samples achieving 98% accuracy to select stable matches. Our database from 240,000 matches shows that one can get over 20% betting profit per month using Analyst Masters. Such consistent profit outperforms human experts and shows the inefficiency of the betting market. Top soccer tipsters achieve 50% accuracy and 8% monthly profit in average only on regional matches. Both our collected database of more than 240,000 soccer matches from 2012 and our algorithm would greatly benefit coaches and punters to get accurate analysis.

Keywords: soccer, analytics, machine learning, database

Procedia PDF Downloads 221
19620 Nitrous Oxide Wastage: Putting Strategies “In the Pipeline” to Reduce Carbon Emissions from Nitrous Oxide

Authors: F. Gallop, C. Ward, M. Zaky, M. Vaghela, R. Sabaratnam

Abstract:

Nitrous oxide (N₂O) has been used in anaesthesia for over 150 years owing to advantageous physical and pharmacological properties. However, with a global warming potential of 310, we have an urgent responsibility to reduce its usage and emission. Anecdotal evidence in our hospital trust suggests minimal N₂O usage, yet our theatres receive a staggering supply. This warranted further investigation. We used a data collection tool to prospectively capture quantitative and qualitative data regarding N₂O cases during one week: this recorded demographics, N₂O indications, clinical management, and total N₂O consumption in litres. In addition, N₂O usage in dental sedation suites and paediatric theatres was separately quantified. Pipeline supply data was acquired from British Oxygen Company accounts. We captured 490 cases. 4% (n=19) used N₂O, 63% (n=12) of these in dental theatres. Common N₂0 indications were induction speed (37%) and rapidly increasing anaesthesia depth (32%). In adult cases, N₂O was always used intraoperatively rather than solely at induction. 74% (n=14) of anaesthetists reported environmental concern over using N₂O. The week’s total N₂O usage was 8109 litres, amounting to 421,668 litres annually. However, the annual N₂O pipeline supply is 2,997,000 litres; an enormous 1.8 million Kg of CO₂. Our results supportively demonstrate that the N₂O pipeline supply greatly exceeds its clinical use. Acknowledging clinical areas not audited, the discrepancy between supply and usage suggests approximately 2.5 million litres of yearly wastage. We consequently recommend terminating the N₂O pipeline supply in minimally used areas, eliminating 1.5 million Kg of CO₂ emissions. High usage clinical areas could consider portable N₂O cylinders as an alternative. In Sweden, N₂O destruction technology is routinely used to minimise CO₂ emissions. Our results support National Health System investment in similar infrastructure.

Keywords: anaesthesia, environment, medical gases, nitrous oxide, sustainability

Procedia PDF Downloads 122
19619 Rangeland Monitoring by Computerized Technologies

Authors: H. Arzani, Z. Arzani

Abstract:

Every piece of rangeland has a different set of physical and biological characteristics. This requires the manager to synthesis various information for regular monitoring to define changes trend to get wright decision for sustainable management. So range managers need to use computerized technologies to monitor rangeland, and select. The best management practices. There are four examples of computerized technologies that can benefit sustainable management: (1) Photographic method for cover measurement: The method was tested in different vegetation communities in semi humid and arid regions. Interpretation of pictures of quadrats was done using Arc View software. Data analysis was done by SPSS software using paired t test. Based on the results, generally, photographic method can be used to measure ground cover in most vegetation communities. (2) GPS application for corresponding ground samples and satellite pixels: In two provinces of Tehran and Markazi, six reference points were selected and in each point, eight GPS models were tested. Significant relation among GPS model, time and location with accuracy of estimated coordinates was found. After selection of suitable method, in Markazi province coordinates of plots along four transects in each 6 sites of rangelands was recorded. The best time of GPS application was in the morning hours, Etrex Vista had less error than other models, and a significant relation among GPS model, time and location with accuracy of estimated coordinates was found. (3) Application of satellite data for rangeland monitoring: Focusing on the long term variation of vegetation parameters such as vegetation cover and production is essential. Our study in grass and shrub lands showed that there were significant correlations between quantitative vegetation characteristics and satellite data. So it is possible to monitor rangeland vegetation using digital data for sustainable utilization. (4) Rangeland suitability classification with GIS: Range suitability assessment can facilitate sustainable management planning. Three sub-models of sensitivity to erosion, water suitability and forage production out puts were entered to final range suitability classification model. GIS was facilitate classification of range suitability and produced suitability maps for sheep grazing. Generally digital computers assist range managers to interpret, modify, calibrate or integrating information for correct management.

Keywords: computer, GPS, GIS, remote sensing, photographic method, monitoring, rangeland ecosystem, management, suitability, sheep grazing

Procedia PDF Downloads 347
19618 Impact of Water Storage Structures on Groundwater Recharge in Jeloula Basin, Central Tunisia

Authors: I. Farid, K. Zouari

Abstract:

An attempt has been made to examine the effect of water storage structures on groundwater recharge in a semi-arid agroclimatic setting in Jeloula Basin (Central Tunisia). In this area, surface water in rivers is seasonal, and therefore groundwater is the perennial source of water supply for domestic and agricultural purposes. Three pumped storage water power plants (PSWPP) have been built to increase the overall water availability in the basin and support agricultural livelihoods of rural smallholders. The scale and geographical dispersion of these multiple lakes restrict the understanding of these coupled human-water systems and the identification of adequate strategies to support riparian farmers. In the present review, hydrochemistry and isotopic tools were combined to get an insight into the processes controlling mineralization and recharge conditions in the investigated aquifer system. This study showed a slight increase in the groundwater level, especially after the artificial recharge operations and a decline when the water volume moves down during drought periods. Chemical data indicate that the main sources of salinity in the waters are related to water-rock interactions. Data inferred from stable isotopes in groundwater samples indicated recharge with modern rainfall. The investigated surface water samples collected from the PSWPP are affected by a significant evaporation and reveal large seasonal variations, which could be controlled by the water volume changes in the open surface reservoirs and the meteorological conditions during evaporation, condensation, and precipitation. The geochemical information is comparable to the isotopic results and illustrates that the chemical and isotopic signatures of reservoir waters differ clearly from those of groundwaters. These data confirm that the contribution of the artificial recharge operations from the PSWPP is very limited.

Keywords: Jeloula basin, recharge, hydrochemistry, isotopes

Procedia PDF Downloads 132
19617 Drying Kinetics of Vacuum Dried Beef Meat Slices

Authors: Elif Aykin Dincer, Mustafa Erbas

Abstract:

The vacuum drying behavior of beef slices (10 x 4 x 0.2 cm3) was experimentally investigated at the temperature of 60, 70, and 80°C under 25 mbar ultimate vacuum pressure and the mathematical models (Lewis, Page, Midilli, Two-term, Wangh and Singh and Modified Henderson and Pabis) were used to fit the vacuum drying of beef slices. The increase in drying air temperature resulted in a decrease in drying time. It took approximately 206, 180 and 157 min to dry beef slices from an initial moisture content to a final moisture content of 0.05 kg water/kg dry matter at 60, 70 and 80 °C of vacuum drying, respectively. It is also observed that the drying rate increased with increasing drying temperature. The coefficients (R2), the reduced chi-square (x²) and root mean square error (RMSE) values were obtained by application of six models to the experimental drying data. The best model with the highest R2 and, the lowest x² and RMSE values was selected to describe the drying characteristics of beef slices. The Page model has shown a better fit to the experimental drying data as compared to other models. In addition, the effective moisture diffusivities of beef slices in the vacuum drying at 60 - 80 °C varied in the range of 1.05 – 1.09 x 10-10 m2/s. Consequently, this results can be used to simulate vacuum drying process of beef slices and improve efficiency of the drying process.

Keywords: beef slice, drying models, effective diffusivity, vacuum

Procedia PDF Downloads 276
19616 Re-Evaluation of Field X Located in Northern Lake Albert Basin to Refine the Structural Interpretation

Authors: Calorine Twebaze, Jesca Balinga

Abstract:

Field X is located on the Eastern shores of L. Albert, Uganda, on the rift flank where the gross sedimentary fill is typically less than 2,000m. The field was discovered in 2006 and encountered about 20.4m of net pay across three (3) stratigraphic intervals within the discovery well. The field covers an area of 3 km2, with the structural configuration comprising a 3-way dip-closed hanging wall anticline that seals against the basement to the southeast along the bounding fault. Field X had been mapped on reprocessed 3D seismic data, which was originally acquired in 2007 and reprocessed in 2013. The seismic data quality is good across the field, and reprocessing work reduced the uncertainty in the location of the bounding fault and enhanced the lateral continuity of reservoir reflectors. The current study was a re-evaluation of Field X to refine fault interpretation and understand the structural uncertainties associated with the field. The seismic data, and three (3) wells datasets were used during the study. The evaluation followed standard workflows using Petrel software and structural attribute analysis. The process spanned from seismic- -well tie, structural interpretation, and structural uncertainty analysis. Analysis of three (3) well ties generated for the 3 wells provided a geophysical interpretation that was consistent with geological picks. The generated time-depth curves showed a general increase in velocity with burial depth. However, separation in curve trends observed below 1100m was mainly attributed to minimal lateral variation in velocity between the wells. In addition to Attribute analysis, three velocity modeling approaches were evaluated, including the Time-Depth Curve, Vo+ kZ, and Average Velocity Method. The generated models were calibrated at well locations using well tops to obtain the best velocity model for Field X. The Time-depth method resulted in more reliable depth surfaces with good structural coherence between the TWT and depth maps with minimal error at well locations of 2 to 5m. Both the NNE-SSW rift border fault and minor faults in the existing interpretation were reevaluated. However, the new interpretation delineated an E-W trending fault in the northern part of the field that had not been interpreted before. The fault was interpreted at all stratigraphic levels and thus propagates from the basement to the surface and is an active fault today. It was also noted that the entire field is less faulted with more faults in the deeper part of the field. The major structural uncertainties defined included 1) The time horizons due to reduced data quality, especially in the deeper parts of the structure, an error equal to one-third of the reflection time thickness was assumed, 2) Check shot analysis showed varying velocities within the wells thus varying depth values for each well, and 3) Very few average velocity points due to limited wells produced a pessimistic average Velocity model.

Keywords: 3D seismic data interpretation, structural uncertainties, attribute analysis, velocity modelling approaches

Procedia PDF Downloads 38
19615 Social Norms around Adolescent Girls’ Marriage Practices in Ethiopia: A Qualitative Exploration

Authors: Dagmawit Tewahido

Abstract:

Purpose: This qualitative study was conducted to explore social norms around adolescent girls’ marriage practices in West Hararghe, Ethiopia, where early marriage is prohibited by law. Methods: Twenty Focus Group Discussions were conducted with Married and Unmarried adolescent girls, adolescent boys and parents of girls using locally developed vignettes. A total of 32 in-depth interviews were conducted with married and unmarried adolescent girls, husbands of adolescent girls and mothers-in-law. Key informant interviews were conducted with 36 district officials. Data analysis was assisted by Open Code computer software. The Social Norms Analysis Plot (SNAP) framework developed by CARE guided the development and analysis of vignettes. A thematic data analysis approach was utilized to summarize the data. Results: Early marriage is seen as a positive phenomenon in our study context, and girls who are not married by the perceived ideal age of 15 are socially sanctioned. They are particularly influenced by their peers to marry. Marrying early is considered a chance given by God and a symbol of good luck. The two common types of marriage are decided: 1) by adolescent girl and boy themselves without seeking parental permission (’Jalaa-deemaa’- meaning ‘to go along’), and 2) by just informing girl’s parents (‘Cabsaa’- meaning ‘to break the culture’). Relatives and marriage brokers also arrange early marriages. Girls usually accept the first marriage proposal regardless of their age. Parents generally tend not to oppose marriage arrangements chosen by their daughters. Conclusions: In the study context social norms encourage early marriage despite the existence of a law prohibiting marriage before the age of eighteen years. Early marriage commonly happens through consensual arrangements between adolescent girls and boys. Interventions to reduce early marriage need to consider the influence of Reference Groups on the decision makers for marriages, especially girls’ own peers.

Keywords: adolescent girls, social norms, early marriage, Ethiopia

Procedia PDF Downloads 120
19614 Modelling Operational Risk Using Extreme Value Theory and Skew t-Copulas via Bayesian Inference

Authors: Betty Johanna Garzon Rozo, Jonathan Crook, Fernando Moreira

Abstract:

Operational risk losses are heavy tailed and are likely to be asymmetric and extremely dependent among business lines/event types. We propose a new methodology to assess, in a multivariate way, the asymmetry and extreme dependence between severity distributions, and to calculate the capital for Operational Risk. This methodology simultaneously uses (i) several parametric distributions and an alternative mix distribution (the Lognormal for the body of losses and the Generalized Pareto Distribution for the tail) via extreme value theory using SAS®, (ii) the multivariate skew t-copula applied for the first time for operational losses and (iii) Bayesian theory to estimate new n-dimensional skew t-copula models via Markov chain Monte Carlo (MCMC) simulation. This paper analyses a newly operational loss data set, SAS Global Operational Risk Data [SAS OpRisk], to model operational risk at international financial institutions. All the severity models are constructed in SAS® 9.2. We implement the procedure PROC SEVERITY and PROC NLMIXED. This paper focuses in describing this implementation.

Keywords: operational risk, loss distribution approach, extreme value theory, copulas

Procedia PDF Downloads 582
19613 Innovative Screening Tool Based on Physical Properties of Blood

Authors: Basant Singh Sikarwar, Mukesh Roy, Ayush Goyal, Priya Ranjan

Abstract:

This work combines two bodies of knowledge which includes biomedical basis of blood stain formation and fluid communities’ wisdom that such formation of blood stain depends heavily on physical properties. Moreover biomedical research tells that different patterns in stains of blood are robust indicator of blood donor’s health or lack thereof. Based on these valuable insights an innovative screening tool is proposed which can act as an aide in the diagnosis of diseases such Anemia, Hyperlipidaemia, Tuberculosis, Blood cancer, Leukemia, Malaria etc., with enhanced confidence in the proposed analysis. To realize this powerful technique, simple, robust and low-cost micro-fluidic devices, a micro-capillary viscometer and a pendant drop tensiometer are designed and proposed to be fabricated to measure the viscosity, surface tension and wettability of various blood samples. Once prognosis and diagnosis data has been generated, automated linear and nonlinear classifiers have been applied into the automated reasoning and presentation of results. A support vector machine (SVM) classifies data on a linear fashion. Discriminant analysis and nonlinear embedding’s are coupled with nonlinear manifold detection in data and detected decisions are made accordingly. In this way, physical properties can be used, using linear and non-linear classification techniques, for screening of various diseases in humans and cattle. Experiments are carried out to validate the physical properties measurement devices. This framework can be further developed towards a real life portable disease screening cum diagnostics tool. Small-scale production of screening cum diagnostic devices is proposed to carry out independent test.

Keywords: blood, physical properties, diagnostic, nonlinear, classifier, device, surface tension, viscosity, wettability

Procedia PDF Downloads 365
19612 Ethiopia as a Tourist Destination: An Exploration of Italian Tourists’ Market Demand

Authors: Frezer Okubay Weldegebriel

Abstract:

The tourism sector in Ethiopia plays a significant role in the national economy. The government is granting its pledge and readiness to develop this sector through various initiatives since to eradicate poverty and encourage economic development of the country is one of the Millennium Development plans. The tourism sector has been identified as one of the priority economic sectors by many countries, and the Government of Ethiopia has planned to make Ethiopia among the top five African destinations by 2020. Nevertheless, the international tourism demand for Ethiopia currently lags behind other African countries such as South Africa, Egypt, Morocco, Tanzania, and Kenya. Meanwhile, the number of international tourists’ arrival in Ethiopia is recently increasing even if it cannot be competitive with other African countries. Therefore, to offer demand-driven tourism products, the Ethiopian government, Tourism planners, Tour & Travel operators need to understand the important factors, which affect international tourists’ decision to visit Ethiopian destinations. This study was intended to analyze Italian Tourists Demand towards Ethiopian destination. The researcher aimed to identify the demand for Italian tourists’ preference to Ethiopian destinations comparing to the top East African countries. This study uses both qualitative and quantitative research methodology, and the data is manipulating through primary data collection method using questionnaires, interviews, and secondary data by reviewing books, journals, magazines, past researches, and websites. An active and potential Italian tourist cohort, five well-functioning tour operators based in Ethiopia for Italian tourists and professionals from Ethiopian Ministry of Tourism and Culture participated. Based on the analysis of the data collected through the questionnaire, interviews, and reviews of different materials, the study disclosed that the majority of Italian tourists have a high demand on Ethiopian Tourist destination. Historical and cultural interest, safety and security, the hospitality of the people and affordable accommodation coast are the main reason for them. However, some Italian tourists prefer to visit Kenya, Tanzania, and Uganda due to the fact that they are fascinated by adventure, safari and beaches, while Ethiopia cannot provide these attractions. Most Italian tourists have little information and practical experiences on Ethiopian tourism possibilities via a tour and travel companies. Moreover, the insufficient marketing campaign and promotion by Ethiopian Government and Ministry of Tourism could also contribute to the failure of Ethiopian tourism.

Keywords: The demand of Italian tourists, Ethiopia economy, Ethiopia tourism destination, promoting Ethiopia tourism

Procedia PDF Downloads 183
19611 Health State Utility Values Related to COVID-19 Pandemic Using EQ-5D: A Systematic Review and Meta-Analysis

Authors: Xu Feifei

Abstract:

The prevalence of COVID-19 currently is the biggest challenge to improving people's quality of life. Its impact on the health-related quality of life (HRQoL) is highly uncertain and has not been summarized so far. The aim of the present systematic review was to assess and provide an up-to-date analysis of the impact of the COVID-19 pandemic on the HRQoL of participants who have been infected, have not been infected but isolated, frontline, with different diseases, and the general population. Therefore, an electronic search of the literature in PubMed databases was performed from 2019 to July 2022 (without date restriction). PRISMA guideline methodology was employed, and data regarding the HRQoL were extracted from eligible studies. Articles were included if they met the following inclusion criteria: (a) reports on the data collection of the health state utility values (HSUVs) related to COVID-19 from 2019 to 2021; (b) English language and peer-reviewed journals; and (c) original HSUV data; (d) using EQ-5D tool to quantify the HRQoL. To identify studies that reported the effects on COVID-19, data on the proportion of overall HSUVs of participants who had the outcome were collected and analyzed using a one-group meta-analysis. As a result, thirty-two studies fulfilled the inclusion criteria and, therefore, were included in the systematic review. A total of 45295 participants and provided 219 means of HSUVs during COVID-19 were included in this systematic review. The range of utility is from 0.224 to 1. The study included participants from Europe (n=16), North America (n=4), Asia (n=10), South America (n=1), and Africa (n=1). Twelve articles showed that the HRQoL of the participants who have been infected with COVID-19 (range of overall HSUVs from 0.6125 to 0.863). Two studies reported the population of frontline workers (the range of overall HSUVs from 0.82 to 0.93). Seven of the articles researched the participants who had not been infected with COVID-19 but suffered from morbidities during the pandemic (range of overall HSUVs from 0.5 to 0.96). Thirteen studies showed that the HRQoL of the respondents who have not been infected with COVID-19 and without any morbidities (range of overall HSUVs from 0.64 to 0.964). Moreover, eighteen articles reported the outcomes of overall HSUVs during the COVID-19 pandemic in different population groups. The estimate of overall HSUVs of direct COVID-19 experience population (n=1333) was 0.751 (95% CI 0.670 - 0.832, I2 = 98.64%); the estimate of frontline population (n=610) was 0.906 ((95% CI 0.854 – 0.957, I2 = 98.61%); participants with different disease (n=132) were 0.768 (95% CI 0.515 - 1.021, I2= 99.26%); general population without infection history (n=29,892) was 0.825 (95% CI 0.766 - 0.885, I2 =99.69%). Conclusively, taking into account these results, this systematic review might confirm that COVID-19 has a negative impact on the HRQoL of the infected population and illness population. It provides practical value for cost-effectiveness model analysis of health states related to COVID-19.

Keywords: COVID-19, health-related quality of life, meta-analysis, systematic review, utility value

Procedia PDF Downloads 67
19610 A Bio-Inspired Approach for Self-Managing Wireless Sensor and Actor Networks

Authors: Lyamine Guezouli, Kamel Barka, Zineb Seghir

Abstract:

Wireless sensor and actor networks (WSANs) present a research challenge for different practice areas. Researchers are trying to optimize the use of such networks through their research work. This optimization is done on certain criteria, such as improving energy efficiency, exploiting node heterogeneity, self-adaptability and self-configuration. In this article, we present our proposal for BIFSA (Biologically-Inspired Framework for Wireless Sensor and Actor networks). Indeed, BIFSA is a middleware that addresses the key issues of wireless sensor and actor networks. BIFSA consists of two types of agents: sensor agents (SA) that operate at the sensor level to collect and transport data to actors and actor agents (AA) that operate at the actor level to transport data to base stations. Once the sensor agent arrives at the actor, it becomes an actor agent, which can exploit the resources of the actors and vice versa. BIFSA allows agents to evolve their genetic structures and adapt to the current network conditions. The simulation results show that BIFSA allows the agents to make better use of all the resources available in each type of node, which improves the performance of the network.

Keywords: wireless sensor and actor networks, self-management, genetic algorithm, agent.

Procedia PDF Downloads 69
19609 Techno-Economic Analysis of Solar Energy for Cathodic Protection of Oil and Gas Buried Pipelines in Southwestern of Iran

Authors: M. Goodarzi, M. Mohammadi, A. Gharib

Abstract:

Solar energy is a renewable energy which has attracted special attention in many countries. Solar cathodic protectionsystems harness the sun’senergy to protect underground pipelinesand tanks from galvanic corrosion. The object of this study is to design and the economic analysis a cathodic protection system by impressed current supplied with solar energy panels applied to underground pipelines. In the present study, the technical and economic analysis of using solar energy for cathodic protection system in southwestern of Iran (Khuzestan province) is investigated. For this purpose, the ecological conditions such as the weather data, air clearness and sunshine hours are analyzed. The economic analyses were done using computer code to investigate the feasibility analysis from the using of various energy sources in order to cathodic protection system. The overall research methodology is divided into four components: Data collection, design of elements, techno economical evaluation, and output analysis. According to the results, solar renewable energy systems can supply adequate power for cathodic protection system purposes.

Keywords: renewable energy, solar energy, solar cathodic protection station, lifecycle cost method

Procedia PDF Downloads 521