Search results for: stock market data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 27047

Search results for: stock market data

19127 Security of Database Using Chaotic Systems

Authors: Eman W. Boghdady, A. R. Shehata, M. A. Azem

Abstract:

Database (DB) security demands permitting authorized users and prohibiting non-authorized users and intruders actions on the DB and the objects inside it. Organizations that are running successfully demand the confidentiality of their DBs. They do not allow the unauthorized access to their data/information. They also demand the assurance that their data is protected against any malicious or accidental modification. DB protection and confidentiality are the security concerns. There are four types of controls to obtain the DB protection, those include: access control, information flow control, inference control, and cryptographic. The cryptographic control is considered as the backbone for DB security, it secures the DB by encryption during storage and communications. Current cryptographic techniques are classified into two types: traditional classical cryptography using standard algorithms (DES, AES, IDEA, etc.) and chaos cryptography using continuous (Chau, Rossler, Lorenz, etc.) or discreet (Logistics, Henon, etc.) algorithms. The important characteristics of chaos are its extreme sensitivity to initial conditions of the system. In this paper, DB-security systems based on chaotic algorithms are described. The Pseudo Random Numbers Generators (PRNGs) from the different chaotic algorithms are implemented using Matlab and their statistical properties are evaluated using NIST and other statistical test-suits. Then, these algorithms are used to secure conventional DB (plaintext), where the statistical properties of the ciphertext are also tested. To increase the complexity of the PRNGs and to let pass all the NIST statistical tests, we propose two hybrid PRNGs: one based on two chaotic Logistic maps and another based on two chaotic Henon maps, where each chaotic algorithm is running side-by-side and starting from random independent initial conditions and parameters (encryption keys). The resulted hybrid PRNGs passed the NIST statistical test suit.

Keywords: algorithms and data structure, DB security, encryption, chaotic algorithms, Matlab, NIST

Procedia PDF Downloads 259
19126 Using Speech Emotion Recognition as a Longitudinal Biomarker for Alzheimer’s Diseases

Authors: Yishu Gong, Liangliang Yang, Jianyu Zhang, Zhengyu Chen, Sihong He, Xusheng Zhang, Wei Zhang

Abstract:

Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that affects millions of people worldwide and is characterized by cognitive decline and behavioral changes. People living with Alzheimer’s disease often find it hard to complete routine tasks. However, there are limited objective assessments that aim to quantify the difficulty of certain tasks for AD patients compared to non-AD people. In this study, we propose to use speech emotion recognition (SER), especially the frustration level, as a potential biomarker for quantifying the difficulty patients experience when describing a picture. We build an SER model using data from the IEMOCAP dataset and apply the model to the DementiaBank data to detect the AD/non-AD group difference and perform longitudinal analysis to track the AD disease progression. Our results show that the frustration level detected from the SER model can possibly be used as a cost-effective tool for objective tracking of AD progression in addition to the Mini-Mental State Examination (MMSE) score.

Keywords: Alzheimer’s disease, speech emotion recognition, longitudinal biomarker, machine learning

Procedia PDF Downloads 100
19125 Establishment of a Classifier Model for Early Prediction of Acute Delirium in Adult Intensive Care Unit Using Machine Learning

Authors: Pei Yi Lin

Abstract:

Objective: The objective of this study is to use machine learning methods to build an early prediction classifier model for acute delirium to improve the quality of medical care for intensive care patients. Background: Delirium is a common acute and sudden disturbance of consciousness in critically ill patients. After the occurrence, it is easy to prolong the length of hospital stay and increase medical costs and mortality. In 2021, the incidence of delirium in the intensive care unit of internal medicine was as high as 59.78%, which indirectly prolonged the average length of hospital stay by 8.28 days, and the mortality rate is about 2.22% in the past three years. Therefore, it is expected to build a delirium prediction classifier through big data analysis and machine learning methods to detect delirium early. Method: This study is a retrospective study, using the artificial intelligence big data database to extract the characteristic factors related to delirium in intensive care unit patients and let the machine learn. The study included patients aged over 20 years old who were admitted to the intensive care unit between May 1, 2022, and December 31, 2022, excluding GCS assessment <4 points, admission to ICU for less than 24 hours, and CAM-ICU evaluation. The CAMICU delirium assessment results every 8 hours within 30 days of hospitalization are regarded as an event, and the cumulative data from ICU admission to the prediction time point are extracted to predict the possibility of delirium occurring in the next 8 hours, and collect a total of 63,754 research case data, extract 12 feature selections to train the model, including age, sex, average ICU stay hours, visual and auditory abnormalities, RASS assessment score, APACHE-II Score score, number of invasive catheters indwelling, restraint and sedative and hypnotic drugs. Through feature data cleaning, processing and KNN interpolation method supplementation, a total of 54595 research case events were extracted to provide machine learning model analysis, using the research events from May 01 to November 30, 2022, as the model training data, 80% of which is the training set for model training, and 20% for the internal verification of the verification set, and then from December 01 to December 2022 The CU research event on the 31st is an external verification set data, and finally the model inference and performance evaluation are performed, and then the model has trained again by adjusting the model parameters. Results: In this study, XG Boost, Random Forest, Logistic Regression, and Decision Tree were used to analyze and compare four machine learning models. The average accuracy rate of internal verification was highest in Random Forest (AUC=0.86), and the average accuracy rate of external verification was in Random Forest and XG Boost was the highest, AUC was 0.86, and the average accuracy of cross-validation was the highest in Random Forest (ACC=0.77). Conclusion: Clinically, medical staff usually conduct CAM-ICU assessments at the bedside of critically ill patients in clinical practice, but there is a lack of machine learning classification methods to assist ICU patients in real-time assessment, resulting in the inability to provide more objective and continuous monitoring data to assist Clinical staff can more accurately identify and predict the occurrence of delirium in patients. It is hoped that the development and construction of predictive models through machine learning can predict delirium early and immediately, make clinical decisions at the best time, and cooperate with PADIS delirium care measures to provide individualized non-drug interventional care measures to maintain patient safety, and then Improve the quality of care.

Keywords: critically ill patients, machine learning methods, delirium prediction, classifier model

Procedia PDF Downloads 55
19124 Design and Implementation of Generative Models for Odor Classification Using Electronic Nose

Authors: Kumar Shashvat, Amol P. Bhondekar

Abstract:

In the midst of the five senses, odor is the most reminiscent and least understood. Odor testing has been mysterious and odor data fabled to most practitioners. The delinquent of recognition and classification of odor is important to achieve. The facility to smell and predict whether the artifact is of further use or it has become undesirable for consumption; the imitation of this problem hooked on a model is of consideration. The general industrial standard for this classification is color based anyhow; odor can be improved classifier than color based classification and if incorporated in machine will be awfully constructive. For cataloging of odor for peas, trees and cashews various discriminative approaches have been used Discriminative approaches offer good prognostic performance and have been widely used in many applications but are incapable to make effectual use of the unlabeled information. In such scenarios, generative approaches have better applicability, as they are able to knob glitches, such as in set-ups where variability in the series of possible input vectors is enormous. Generative models are integrated in machine learning for either modeling data directly or as a transitional step to form an indeterminate probability density function. The algorithms or models Linear Discriminant Analysis and Naive Bayes Classifier have been used for classification of the odor of cashews. Linear Discriminant Analysis is a method used in data classification, pattern recognition, and machine learning to discover a linear combination of features that typifies or divides two or more classes of objects or procedures. The Naive Bayes algorithm is a classification approach base on Bayes rule and a set of qualified independence theory. Naive Bayes classifiers are highly scalable, requiring a number of restraints linear in the number of variables (features/predictors) in a learning predicament. The main recompenses of using the generative models are generally a Generative Models make stronger assumptions about the data, specifically, about the distribution of predictors given the response variables. The Electronic instrument which is used for artificial odor sensing and classification is an electronic nose. This device is designed to imitate the anthropological sense of odor by providing an analysis of individual chemicals or chemical mixtures. The experimental results have been evaluated in the form of the performance measures i.e. are accuracy, precision and recall. The investigational results have proven that the overall performance of the Linear Discriminant Analysis was better in assessment to the Naive Bayes Classifier on cashew dataset.

Keywords: odor classification, generative models, naive bayes, linear discriminant analysis

Procedia PDF Downloads 369
19123 Ribotaxa: Combined Approaches for Taxonomic Resolution Down to the Species Level from Metagenomics Data Revealing Novelties

Authors: Oshma Chakoory, Sophie Comtet-Marre, Pierre Peyret

Abstract:

Metagenomic classifiers are widely used for the taxonomic profiling of metagenomic data and estimation of taxa relative abundance. Small subunit rRNA genes are nowadays a gold standard for the phylogenetic resolution of complex microbial communities, although the power of this marker comes down to its use as full-length. We benchmarked the performance and accuracy of rRNA-specialized versus general-purpose read mappers, reference-targeted assemblers and taxonomic classifiers. We then built a pipeline called RiboTaxa to generate a highly sensitive and specific metataxonomic approach. Using metagenomics data, RiboTaxa gave the best results compared to other tools (Kraken2, Centrifuge (1), METAXA2 (2), PhyloFlash (3)) with precise taxonomic identification and relative abundance description, giving no false positive detection. Using real datasets from various environments (ocean, soil, human gut) and from different approaches (metagenomics and gene capture by hybridization), RiboTaxa revealed microbial novelties not seen by current bioinformatics analysis opening new biological perspectives in human and environmental health. In a study focused on corals’ health involving 20 metagenomic samples (4), an affiliation of prokaryotes was limited to the family level with Endozoicomonadaceae characterising healthy octocoral tissue. RiboTaxa highlighted 2 species of uncultured Endozoicomonas which were dominant in the healthy tissue. Both species belonged to a genus not yet described, opening new research perspectives on corals’ health. Applied to metagenomics data from a study on human gut and extreme longevity (5), RiboTaxa detected the presence of an uncultured archaeon in semi-supercentenarians (aged 105 to 109 years) highlighting an archaeal genus, not yet described, and 3 uncultured species belonging to the Enorma genus that could be species of interest participating in the longevity process. RiboTaxa is user-friendly, rapid, allowing microbiota structure description from any environment and the results can be easily interpreted. This software is freely available at https://github.com/oschakoory/RiboTaxa under the GNU Affero General Public License 3.0.

Keywords: metagenomics profiling, microbial diversity, SSU rRNA genes, full-length phylogenetic marker

Procedia PDF Downloads 101
19122 A Comparative Analysis of Geometric and Exponential Laws in Modelling the Distribution of the Duration of Daily Precipitation

Authors: Mounia El Hafyani, Khalid El Himdi

Abstract:

Precipitation is one of the key variables in water resource planning. The importance of modeling wet and dry durations is a crucial pointer in engineering hydrology. The objective of this study is to model and analyze the distribution of wet and dry durations. For this purpose, the daily rainfall data from 1967 to 2017 of the Moroccan city of Kenitra’s station are used. Three models are implemented for the distribution of wet and dry durations, namely the first-order Markov chain, the second-order Markov chain, and the truncated negative binomial law. The adherence of the data to the proposed models is evaluated using Chi-square and Kolmogorov-Smirnov tests. The Akaike information criterion is applied to assess the most effective model distribution. We go further and study the law of the number of wet and dry days among k consecutive days. The calculation of this law is done through an algorithm that we have implemented based on conditional laws. We complete our work by comparing the observed moments of the numbers of wet/dry days among k consecutive days to the calculated moment of the three estimated models. The study shows the effectiveness of our approach in modeling wet and dry durations of daily precipitation.

Keywords: Markov chain, rainfall, truncated negative binomial law, wet and dry durations

Procedia PDF Downloads 112
19121 The Effect of Antibiotic Use on Blood Cultures: Implications for Future Policy

Authors: Avirup Chowdhury, Angus K. McFadyen, Linsey Batchelor

Abstract:

Blood cultures (BCs) are an important aspect of management of the septic patient, identifying the underlying pathogen and its antibiotic sensitivities. However, while the current literature outlines indications for initial BCs to be taken, there is little guidance for repeat sampling in the following 5-day period and little information on how antibiotic use can affect the usefulness of this investigation. A retrospective cohort study was conducted using inpatients who had undergone 2 or more BCs within 5 days between April 2016 and April 2017 at a 400-bed hospital in the west of Scotland and received antibiotic therapy between the first and second BCs. The data for BC sampling was collected from the electronic microbiology database, and cross-referenced with data from the hospital electronic prescribing system. Overall, 283 BCs were included in the study, taken from 92 patients (mean 3.08 cultures per patient, range 2-10). All 92 patients had initial BCs, of which 83 were positive (90%). 65 had a further sample within 24 hours of commencement of antibiotics, with 35 positive (54%). 23 had samples within 24-48 hours, with 4 (17%) positive; 12 patients had sampling at 48-72 hours, 12 at 72-96 hours, and 10 at 96-120 hours, with none positive. McNemar’s Exact Test was used to calculate statistical significance for patients who received blood cultures in multiple time blocks (Initial, < 24h, 24-120h, > 120h). For initial vs. < 24h-post BCs (53 patients tested), the proportion of positives fell from 46/53 to 29/53 (one-tailed P=0.002, OR 3.43, 95% CI 1.48-7.96). For initial vs 24-120h (n=42), the proportions were 38/42 and 4/42 respectively (P < 0.001, OR 35.0, 95% CI 4.79-255.48). For initial vs > 120h (n=36), these were 33/36 and 2/36 (P < 0.001,OR ∞). These were also calculated for a positive in initial or < 24h vs. 24-120h (n=42), with proportions of 41/42 and 4/42 (P < 0.001, OR 38.0, 95% CI 5.22-276.78); and for initial or < 24h vs > 120h (n=36), with proportions of 35/36 and 2/36 respectively (P < 0.001, OR ∞). This data appears to show that taking an initial BC followed by a BC within 24 hours of antibiotic commencement would maximise blood culture yield while minimising the risk of false negative results. This could potentially remove the need for as many as 46% of BC samples without adversely affecting patient care. BC yield decreases sharply after 48 hours of antibiotic use, and may not provide any clinically useful information after this time. Further multi-centre studies would validate these findings, and provide a foundation for future health policy generation.

Keywords: antibiotics, blood culture, efficacy, inpatient

Procedia PDF Downloads 163
19120 Eco-Products in Day-to-Day Life: A Catalyst for Achieving Sustainability

Authors: Rani Fernandez

Abstract:

As global concerns regarding environmental degradation and climate change intensify, the imperative for sustainable living has never been more critical. This research delves into the role of eco-products in everyday life as a pivotal strategy for achieving sustainability. The study investigates the awareness, adoption, and impact of eco-friendly products on individual and community levels. The research employs a mixed-methods approach, combining surveys, interviews, and case studies to explore consumer perceptions, behaviours, and motivations surrounding the use of eco-products. Additionally, life cycle assessments are conducted to evaluate the environmental footprint of selected eco-products, shedding light on their tangible contributions to sustainability. The findings reveal the diverse range of eco-products available in the market, from biodegradable packaging to energy-efficient appliances, and the extent to which consumers integrate these products into their daily routines. Moreover, the research examines the challenges and opportunities associated with widespread adoption, considering factors such as cost, accessibility, and efficacy. In addition to individual consumption patterns, the study investigates the broader societal impact of eco-product integration. It explores the potential for eco-products to drive systemic change by influencing supply chains, corporate practices, and government policies. The research highlights successful case studies of communities or businesses that have effectively incorporated eco-products, providing valuable insights into scalable models for sustainability. Ultimately, this research contributes to the discourse on sustainable living by elucidating the pivotal role of eco-products in shaping environmentally conscious behaviours. By understanding the dynamics of eco-product adoption, policymakers, businesses, and individuals can collaboratively work towards a more sustainable future. The implications of this study extend beyond academia, informing practical strategies for fostering a global shift towards sustainable consumption and production.

Keywords: eco-friendly, sustainablity, environment, climate change

Procedia PDF Downloads 32
19119 Data Clustering Algorithm Based on Multi-Objective Periodic Bacterial Foraging Optimization with Two Learning Archives

Authors: Chen Guo, Heng Tang, Ben Niu

Abstract:

Clustering splits objects into different groups based on similarity, making the objects have higher similarity in the same group and lower similarity in different groups. Thus, clustering can be treated as an optimization problem to maximize the intra-cluster similarity or inter-cluster dissimilarity. In real-world applications, the datasets often have some complex characteristics: sparse, overlap, high dimensionality, etc. When facing these datasets, simultaneously optimizing two or more objectives can obtain better clustering results than optimizing one objective. However, except for the objectives weighting methods, traditional clustering approaches have difficulty in solving multi-objective data clustering problems. Due to this, evolutionary multi-objective optimization algorithms are investigated by researchers to optimize multiple clustering objectives. In this paper, the Data Clustering algorithm based on Multi-objective Periodic Bacterial Foraging Optimization with two Learning Archives (DC-MPBFOLA) is proposed. Specifically, first, to reduce the high computing complexity of the original BFO, periodic BFO is employed as the basic algorithmic framework. Then transfer the periodic BFO into a multi-objective type. Second, two learning strategies are proposed based on the two learning archives to guide the bacterial swarm to move in a better direction. On the one hand, the global best is selected from the global learning archive according to the convergence index and diversity index. On the other hand, the personal best is selected from the personal learning archive according to the sum of weighted objectives. According to the aforementioned learning strategies, a chemotaxis operation is designed. Third, an elite learning strategy is designed to provide fresh power to the objects in two learning archives. When the objects in these two archives do not change for two consecutive times, randomly initializing one dimension of objects can prevent the proposed algorithm from falling into local optima. Fourth, to validate the performance of the proposed algorithm, DC-MPBFOLA is compared with four state-of-art evolutionary multi-objective optimization algorithms and one classical clustering algorithm on evaluation indexes of datasets. To further verify the effectiveness and feasibility of designed strategies in DC-MPBFOLA, variants of DC-MPBFOLA are also proposed. Experimental results demonstrate that DC-MPBFOLA outperforms its competitors regarding all evaluation indexes and clustering partitions. These results also indicate that the designed strategies positively influence the performance improvement of the original BFO.

Keywords: data clustering, multi-objective optimization, bacterial foraging optimization, learning archives

Procedia PDF Downloads 125
19118 The Usage of Bridge Estimator for Hegy Seasonal Unit Root Tests

Authors: Huseyin Guler, Cigdem Kosar

Abstract:

The aim of this study is to propose Bridge estimator for seasonal unit root tests. Seasonality is an important factor for many economic time series. Some variables may contain seasonal patterns and forecasts that ignore important seasonal patterns have a high variance. Therefore, it is very important to eliminate seasonality for seasonal macroeconomic data. There are some methods to eliminate the impacts of seasonality in time series. One of them is filtering the data. However, this method leads to undesired consequences in unit root tests, especially if the data is generated by a stochastic seasonal process. Another method to eliminate seasonality is using seasonal dummy variables. Some seasonal patterns may result from stationary seasonal processes, which are modelled using seasonal dummies but if there is a varying and changing seasonal pattern over time, so the seasonal process is non-stationary, deterministic seasonal dummies are inadequate to capture the seasonal process. It is not suitable to use seasonal dummies for modeling such seasonally nonstationary series. Instead of that, it is necessary to take seasonal difference if there are seasonal unit roots in the series. Different alternative methods are proposed in the literature to test seasonal unit roots, such as Dickey, Hazsa, Fuller (DHF) and Hylleberg, Engle, Granger, Yoo (HEGY) tests. HEGY test can be also used to test the seasonal unit root in different frequencies (monthly, quarterly, and semiannual). Another issue in unit root tests is the lag selection. Lagged dependent variables are added to the model in seasonal unit root tests as in the unit root tests to overcome the autocorrelation problem. In this case, it is necessary to choose the lag length and determine any deterministic components (i.e., a constant and trend) first, and then use the proper model to test for seasonal unit roots. However, this two-step procedure might lead size distortions and lack of power in seasonal unit root tests. Recent studies show that Bridge estimators are good in selecting optimal lag length while differentiating nonstationary versus stationary models for nonseasonal data. The advantage of this estimator is the elimination of the two-step nature of conventional unit root tests and this leads a gain in size and power. In this paper, the Bridge estimator is proposed to test seasonal unit roots in a HEGY model. A Monte-Carlo experiment is done to determine the efficiency of this approach and compare the size and power of this method with HEGY test. Since Bridge estimator performs well in model selection, our approach may lead to some gain in terms of size and power over HEGY test.

Keywords: bridge estimators, HEGY test, model selection, seasonal unit root

Procedia PDF Downloads 323
19117 Calculation of Pressure-Varying Langmuir and Brunauer-Emmett-Teller Isotherm Adsorption Parameters

Authors: Trevor C. Brown, David J. Miron

Abstract:

Gas-solid physical adsorption methods are central to the characterization and optimization of the effective surface area, pore size and porosity for applications such as heterogeneous catalysis, and gas separation and storage. Properties such as adsorption uptake, capacity, equilibrium constants and Gibbs free energy are dependent on the composition and structure of both the gas and the adsorbent. However, challenges remain, in accurately calculating these properties from experimental data. Gas adsorption experiments involve measuring the amounts of gas adsorbed over a range of pressures under isothermal conditions. Various constant-parameter models, such as Langmuir and Brunauer-Emmett-Teller (BET) theories are used to provide information on adsorbate and adsorbent properties from the isotherm data. These models typically do not provide accurate interpretations across the full range of pressures and temperatures. The Langmuir adsorption isotherm is a simple approximation for modelling equilibrium adsorption data and has been effective in estimating surface areas and catalytic rate laws, particularly for high surface area solids. The Langmuir isotherm assumes the systematic filling of identical adsorption sites to a monolayer coverage. The BET model is based on the Langmuir isotherm and allows for the formation of multiple layers. These additional layers do not interact with the first layer and the energetics are equal to the adsorbate as a bulk liquid. This BET method is widely used to measure the specific surface area of materials. Both Langmuir and BET models assume that the affinity of the gas for all adsorption sites are identical and so the calculated adsorbent uptake at the monolayer and equilibrium constant are independent of coverage and pressure. Accurate representations of adsorption data have been achieved by extending the Langmuir and BET models to include pressure-varying uptake capacities and equilibrium constants. These parameters are determined using a novel regression technique called flexible least squares for time-varying linear regression. For isothermal adsorption the adsorption parameters are assumed to vary slowly and smoothly with increasing pressure. The flexible least squares for pressure-varying linear regression (FLS-PVLR) approach assumes two distinct types of discrepancy terms, dynamic and measurement for all parameters in the linear equation used to simulate the data. Dynamic terms account for pressure variation in successive parameter vectors, and measurement terms account for differences between observed and theoretically predicted outcomes via linear regression. The resultant pressure-varying parameters are optimized by minimizing both dynamic and measurement residual squared errors. Validation of this methodology has been achieved by simulating adsorption data for n-butane and isobutane on activated carbon at 298 K, 323 K and 348 K and for nitrogen on mesoporous alumina at 77 K with pressure-varying Langmuir and BET adsorption parameters (equilibrium constants and uptake capacities). This modeling provides information on the adsorbent (accessible surface area and micropore volume), adsorbate (molecular areas and volumes) and thermodynamic (Gibbs free energies) variations of the adsorption sites.

Keywords: Langmuir adsorption isotherm, BET adsorption isotherm, pressure-varying adsorption parameters, adsorbate and adsorbent properties and energetics

Procedia PDF Downloads 219
19116 Citation Analysis of New Zealand Court Decisions

Authors: Tobias Milz, L. Macpherson, Varvara Vetrova

Abstract:

The law is a fundamental pillar of human societies as it shapes, controls and governs how humans conduct business, behave and interact with each other. Recent advances in computer-assisted technologies such as NLP, data science and AI are creating opportunities to support the practice, research and study of this pervasive domain. It is therefore not surprising that there has been an increase in investments into supporting technologies for the legal industry (also known as “legal tech” or “law tech”) over the last decade. A sub-discipline of particular appeal is concerned with assisted legal research. Supporting law researchers and practitioners to retrieve information from the vast amount of ever-growing legal documentation is of natural interest to the legal research community. One tool that has been in use for this purpose since the early nineteenth century is legal citation indexing. Among other use cases, they provided an effective means to discover new precedent cases. Nowadays, computer-assisted network analysis tools can allow for new and more efficient ways to reveal the “hidden” information that is conveyed through citation behavior. Unfortunately, access to openly available legal data is still lacking in New Zealand and access to such networks is only commercially available via providers such as LexisNexis. Consequently, there is a need to create, analyze and provide a legal citation network with sufficient data to support legal research tasks. This paper describes the development and analysis of a legal citation Network for New Zealand containing over 300.000 decisions from 125 different courts of all areas of law and jurisdiction. Using python, the authors assembled web crawlers, scrapers and an OCR pipeline to collect and convert court decisions from openly available sources such as NZLII into uniform and machine-readable text. This facilitated the use of regular expressions to identify references to other court decisions from within the decision text. The data was then imported into a graph-based database (Neo4j) with the courts and their respective cases represented as nodes and the extracted citations as links. Furthermore, additional links between courts of connected cases were added to indicate an indirect citation between the courts. Neo4j, as a graph-based database, allows efficient querying and use of network algorithms such as PageRank to reveal the most influential/most cited courts and court decisions over time. This paper shows that the in-degree distribution of the New Zealand legal citation network resembles a power-law distribution, which indicates a possible scale-free behavior of the network. This is in line with findings of the respective citation networks of the U.S. Supreme Court, Austria and Germany. The authors of this paper provide the database as an openly available data source to support further legal research. The decision texts can be exported from the database to be used for NLP-related legal research, while the network can be used for in-depth analysis. For example, users of the database can specify the network algorithms and metrics to only include specific courts to filter the results to the area of law of interest.

Keywords: case citation network, citation analysis, network analysis, Neo4j

Procedia PDF Downloads 94
19115 'Sea Power: Concept, Influence and Securitization'; the Nigerian Navy's Role in a Developing State like Nigeria

Authors: William Abiodun Duyile

Abstract:

It is common knowledge that marine food has always been found from the sea, energy can also be found underneath and, to a growing extent; other mineral resources have come from the sea spaces. It is the importance of the sea and the sea lines of communication to littoral nations that has made concepts such as sea power, naval power, etc., significant to them. The study relied on documentary data. The documentary data were sourced from government annual departmental reports, newspapers and correspondence. The secondary sources used were subjected to internal and external criticism for authentication, and then to textual and contextual analyses. The study found that the differential level of seamanship amongst states defined their relationship. It was sea power that gave some states an edge over the others. The study proves that over the ages sea power has been core to the development of States or Empires. The study found that the Nigerian Navy was centre to Nigeria’s conquest of the littoral areas of Biafra, like Bonny, Port-Harcourt, and Calabar; it was also an important turning point of the Nigerian civil war since by it Biafra became landlocked. The research was able to identify succinctly the Nigerian Navy’s contribution to the security and development of the Nigerian State.

Keywords: sea power, naval power, land locked states, warship

Procedia PDF Downloads 127
19114 Heterogeneous Reactions to Digital Opportunities: A Field Study

Authors: Bangaly Kaba

Abstract:

In the global information society, the importance of the Internet cannot be overemphasized. Africa needs access to the powerful information and communication tools of the Internet in order to obtain the resources and efficiency essential for sustainable development. Unfortunately, in 2013, the data from Internetworldstats showed only 15% of African populations have access to Internet. This relative low Internet penetration rate signals a problem that may threaten the economic development, governmental efficiency, and ultimately the global competitiveness of African countries. Many initiatives were undertaken to bring the benefits of the global information revolution to the people of Africa, through connection to the Internet and other Global Information Infrastructure technologies. The purpose is to understand differences between socio-economically advantaged and disadvantaged internet users. From that, we will determine what prevents disadvantaged groups from benefiting from Internet usage. Data were collected through a survey from Internet users in Ivory Coast. The results reveal that Personal network exposure, Self-efficacy and Availability are the key drivers of continued use intention for the socio-economically disadvantaged group. The theoretical and practical implications are also described.

Keywords: digital inequality, internet, integrative model, socio-economically advantaged and disadvantaged, use continuance, Africa

Procedia PDF Downloads 463
19113 Linking Access to Land, Tenure Security with Food Sufficiency of Tenants/Landless or Small Holder Farmers of Parsa District

Authors: Subesh Panta

Abstract:

The land is a one of the major boosting factors of production for the agricultural country like Nepal where access to land has been a major source of livelihood of tenants and small farmers. But there is an absence of secure land tenure arrangement which drastically affect the overall production of farmers leading towards food insecurity. Sharecropping is practiced in Nepal especially in tarai region from early period, but there is the gap in the academic study whether the sharecropping has benefitted tenant farmers and make them food sufficient or not. This study attempts to find out the food sufficiency among the tenant households. The research was carried in the three VDCs of Parsa district -Paterwa (Sugauli), Jitpur and Nirchuta. A total of 111 households were determined as the sample size from each of the three VDCs was randomly visited for interview in the study. The size of land rent-in was found to be very small and fragmented. At the same time, the land tenure security was not found to be secured among the tenants. Due to lack of land tenure security, on one hand tenants and small farmers were not found to be motivated to investment in agriculture as they need to share fifty percent of their production with the land owners, and on other hand land owners were also not interested in investing as they have other alternative sources of livelihood rather than agriculture. In conclusion, the study highpoint that the crop production and food sufficiency level of the tenants’ farmers of the Parsa district are decreasing. Many tenants’ farmers are seeking alternative opportunities for livelihood rather than sharecropping due to insecure land tenure, feudalistic practice, lack of storage for agriculture production, lack of proper agro-market. The situation is such that, if no action is taken timely, there may be a situation that we will have to depend on imports for all the food requirements. Thus, the study discloses that the sharecropping could act as catalyst for ensuring food sufficiency for all, if proper land tenure police are promoted to tenants/small farmers with legal titles to their land or promoted with sustainable agriculture methods.

Keywords: agriculture, food sufficiency, land, tenant farmes

Procedia PDF Downloads 229
19112 Nitrous Oxide Wastage: Putting Strategies “In the Pipeline” to Reduce Carbon Emissions from Nitrous Oxide

Authors: F. Gallop, C. Ward, M. Zaky, M. Vaghela, R. Sabaratnam

Abstract:

Nitrous oxide (N₂O) has been used in anaesthesia for over 150 years owing to advantageous physical and pharmacological properties. However, with a global warming potential of 310, we have an urgent responsibility to reduce its usage and emission. Anecdotal evidence in our hospital trust suggests minimal N₂O usage, yet our theatres receive a staggering supply. This warranted further investigation. We used a data collection tool to prospectively capture quantitative and qualitative data regarding N₂O cases during one week: this recorded demographics, N₂O indications, clinical management, and total N₂O consumption in litres. In addition, N₂O usage in dental sedation suites and paediatric theatres was separately quantified. Pipeline supply data was acquired from British Oxygen Company accounts. We captured 490 cases. 4% (n=19) used N₂O, 63% (n=12) of these in dental theatres. Common N₂0 indications were induction speed (37%) and rapidly increasing anaesthesia depth (32%). In adult cases, N₂O was always used intraoperatively rather than solely at induction. 74% (n=14) of anaesthetists reported environmental concern over using N₂O. The week’s total N₂O usage was 8109 litres, amounting to 421,668 litres annually. However, the annual N₂O pipeline supply is 2,997,000 litres; an enormous 1.8 million Kg of CO₂. Our results supportively demonstrate that the N₂O pipeline supply greatly exceeds its clinical use. Acknowledging clinical areas not audited, the discrepancy between supply and usage suggests approximately 2.5 million litres of yearly wastage. We consequently recommend terminating the N₂O pipeline supply in minimally used areas, eliminating 1.5 million Kg of CO₂ emissions. High usage clinical areas could consider portable N₂O cylinders as an alternative. In Sweden, N₂O destruction technology is routinely used to minimise CO₂ emissions. Our results support National Health System investment in similar infrastructure.

Keywords: anaesthesia, environment, medical gases, nitrous oxide, sustainability

Procedia PDF Downloads 125
19111 Rangeland Monitoring by Computerized Technologies

Authors: H. Arzani, Z. Arzani

Abstract:

Every piece of rangeland has a different set of physical and biological characteristics. This requires the manager to synthesis various information for regular monitoring to define changes trend to get wright decision for sustainable management. So range managers need to use computerized technologies to monitor rangeland, and select. The best management practices. There are four examples of computerized technologies that can benefit sustainable management: (1) Photographic method for cover measurement: The method was tested in different vegetation communities in semi humid and arid regions. Interpretation of pictures of quadrats was done using Arc View software. Data analysis was done by SPSS software using paired t test. Based on the results, generally, photographic method can be used to measure ground cover in most vegetation communities. (2) GPS application for corresponding ground samples and satellite pixels: In two provinces of Tehran and Markazi, six reference points were selected and in each point, eight GPS models were tested. Significant relation among GPS model, time and location with accuracy of estimated coordinates was found. After selection of suitable method, in Markazi province coordinates of plots along four transects in each 6 sites of rangelands was recorded. The best time of GPS application was in the morning hours, Etrex Vista had less error than other models, and a significant relation among GPS model, time and location with accuracy of estimated coordinates was found. (3) Application of satellite data for rangeland monitoring: Focusing on the long term variation of vegetation parameters such as vegetation cover and production is essential. Our study in grass and shrub lands showed that there were significant correlations between quantitative vegetation characteristics and satellite data. So it is possible to monitor rangeland vegetation using digital data for sustainable utilization. (4) Rangeland suitability classification with GIS: Range suitability assessment can facilitate sustainable management planning. Three sub-models of sensitivity to erosion, water suitability and forage production out puts were entered to final range suitability classification model. GIS was facilitate classification of range suitability and produced suitability maps for sheep grazing. Generally digital computers assist range managers to interpret, modify, calibrate or integrating information for correct management.

Keywords: computer, GPS, GIS, remote sensing, photographic method, monitoring, rangeland ecosystem, management, suitability, sheep grazing

Procedia PDF Downloads 348
19110 Quantitative Ethno-Botanical Analysis and Conservation Issues of Medicinal Flora from Alpine and Sub-Alpine, Hindukush Region of Pakistan

Authors: Gul Jan

Abstract:

It is the first quantitative ethno-botanical analysis and conservation issues of medicinal flora of Alpine and Sub-alpine, Hindikush region of Pakistan. The objective of the study aims to report, compare the uses and highlight the ethno-Botanical significance of medicinal plants for treatment of various diseases. A total of 250 (242 males and 8 females) local informants including 10 Local Traditional Healers were interviewed. Information was collected through semi-structured interviews, analyzed and compared by quantitative ethno-botanical indices such as Jaccard index (JI), Informant Consensus Factor (ICF), use value (UV) and Relative frequency of citation (RFC).Thorough survey indicated that 57 medicinal plants belongs to 43 families were investigated to treat various illnesses. The highest ICF is recorded for digestive system (0.69%), Circolatory system (0.61%), urinary tract system, (0.53%) and respiratory system (0.52%). Used value indicated that, Achillea mellefolium (UV = 0.68), Aconitum violaceum (UV = 0.69), Valeriana jatamansi (UV = 0.63), Berberis lyceum (UV = 0.65) and are exceedingly medicinal plant species used in the region. In comparison, highest similarity index is recorded in these studies with JI 17.72 followed by 16.41. According to DMR output, Pinus williciana ranked first due to multipurpose uses among all species and was found most threatened with higher market value. Unwise used of natural assets pooled with unsuitable harvesting practices have exaggerated pressure on plant species of the research region. The main issues causative to natural variety loss found were over grazing of animals, forest violation, wild animal hunting, fodder, plant collection as medicine, fuel wood, forest fire, and invasive species negatively affect the natural resources. For viable utilization, in situ and ex situ conservation, skillful collecting, and reforestation project may be the resolution. Further wide field management research is required.

Keywords: quantitative analysis, conservations issues, medicinal flora, alpine and sub-alpine, Hindukush region

Procedia PDF Downloads 292
19109 Impact of Water Storage Structures on Groundwater Recharge in Jeloula Basin, Central Tunisia

Authors: I. Farid, K. Zouari

Abstract:

An attempt has been made to examine the effect of water storage structures on groundwater recharge in a semi-arid agroclimatic setting in Jeloula Basin (Central Tunisia). In this area, surface water in rivers is seasonal, and therefore groundwater is the perennial source of water supply for domestic and agricultural purposes. Three pumped storage water power plants (PSWPP) have been built to increase the overall water availability in the basin and support agricultural livelihoods of rural smallholders. The scale and geographical dispersion of these multiple lakes restrict the understanding of these coupled human-water systems and the identification of adequate strategies to support riparian farmers. In the present review, hydrochemistry and isotopic tools were combined to get an insight into the processes controlling mineralization and recharge conditions in the investigated aquifer system. This study showed a slight increase in the groundwater level, especially after the artificial recharge operations and a decline when the water volume moves down during drought periods. Chemical data indicate that the main sources of salinity in the waters are related to water-rock interactions. Data inferred from stable isotopes in groundwater samples indicated recharge with modern rainfall. The investigated surface water samples collected from the PSWPP are affected by a significant evaporation and reveal large seasonal variations, which could be controlled by the water volume changes in the open surface reservoirs and the meteorological conditions during evaporation, condensation, and precipitation. The geochemical information is comparable to the isotopic results and illustrates that the chemical and isotopic signatures of reservoir waters differ clearly from those of groundwaters. These data confirm that the contribution of the artificial recharge operations from the PSWPP is very limited.

Keywords: Jeloula basin, recharge, hydrochemistry, isotopes

Procedia PDF Downloads 134
19108 Drying Kinetics of Vacuum Dried Beef Meat Slices

Authors: Elif Aykin Dincer, Mustafa Erbas

Abstract:

The vacuum drying behavior of beef slices (10 x 4 x 0.2 cm3) was experimentally investigated at the temperature of 60, 70, and 80°C under 25 mbar ultimate vacuum pressure and the mathematical models (Lewis, Page, Midilli, Two-term, Wangh and Singh and Modified Henderson and Pabis) were used to fit the vacuum drying of beef slices. The increase in drying air temperature resulted in a decrease in drying time. It took approximately 206, 180 and 157 min to dry beef slices from an initial moisture content to a final moisture content of 0.05 kg water/kg dry matter at 60, 70 and 80 °C of vacuum drying, respectively. It is also observed that the drying rate increased with increasing drying temperature. The coefficients (R2), the reduced chi-square (x²) and root mean square error (RMSE) values were obtained by application of six models to the experimental drying data. The best model with the highest R2 and, the lowest x² and RMSE values was selected to describe the drying characteristics of beef slices. The Page model has shown a better fit to the experimental drying data as compared to other models. In addition, the effective moisture diffusivities of beef slices in the vacuum drying at 60 - 80 °C varied in the range of 1.05 – 1.09 x 10-10 m2/s. Consequently, this results can be used to simulate vacuum drying process of beef slices and improve efficiency of the drying process.

Keywords: beef slice, drying models, effective diffusivity, vacuum

Procedia PDF Downloads 277
19107 Re-Evaluation of Field X Located in Northern Lake Albert Basin to Refine the Structural Interpretation

Authors: Calorine Twebaze, Jesca Balinga

Abstract:

Field X is located on the Eastern shores of L. Albert, Uganda, on the rift flank where the gross sedimentary fill is typically less than 2,000m. The field was discovered in 2006 and encountered about 20.4m of net pay across three (3) stratigraphic intervals within the discovery well. The field covers an area of 3 km2, with the structural configuration comprising a 3-way dip-closed hanging wall anticline that seals against the basement to the southeast along the bounding fault. Field X had been mapped on reprocessed 3D seismic data, which was originally acquired in 2007 and reprocessed in 2013. The seismic data quality is good across the field, and reprocessing work reduced the uncertainty in the location of the bounding fault and enhanced the lateral continuity of reservoir reflectors. The current study was a re-evaluation of Field X to refine fault interpretation and understand the structural uncertainties associated with the field. The seismic data, and three (3) wells datasets were used during the study. The evaluation followed standard workflows using Petrel software and structural attribute analysis. The process spanned from seismic- -well tie, structural interpretation, and structural uncertainty analysis. Analysis of three (3) well ties generated for the 3 wells provided a geophysical interpretation that was consistent with geological picks. The generated time-depth curves showed a general increase in velocity with burial depth. However, separation in curve trends observed below 1100m was mainly attributed to minimal lateral variation in velocity between the wells. In addition to Attribute analysis, three velocity modeling approaches were evaluated, including the Time-Depth Curve, Vo+ kZ, and Average Velocity Method. The generated models were calibrated at well locations using well tops to obtain the best velocity model for Field X. The Time-depth method resulted in more reliable depth surfaces with good structural coherence between the TWT and depth maps with minimal error at well locations of 2 to 5m. Both the NNE-SSW rift border fault and minor faults in the existing interpretation were reevaluated. However, the new interpretation delineated an E-W trending fault in the northern part of the field that had not been interpreted before. The fault was interpreted at all stratigraphic levels and thus propagates from the basement to the surface and is an active fault today. It was also noted that the entire field is less faulted with more faults in the deeper part of the field. The major structural uncertainties defined included 1) The time horizons due to reduced data quality, especially in the deeper parts of the structure, an error equal to one-third of the reflection time thickness was assumed, 2) Check shot analysis showed varying velocities within the wells thus varying depth values for each well, and 3) Very few average velocity points due to limited wells produced a pessimistic average Velocity model.

Keywords: 3D seismic data interpretation, structural uncertainties, attribute analysis, velocity modelling approaches

Procedia PDF Downloads 40
19106 Social Norms around Adolescent Girls’ Marriage Practices in Ethiopia: A Qualitative Exploration

Authors: Dagmawit Tewahido

Abstract:

Purpose: This qualitative study was conducted to explore social norms around adolescent girls’ marriage practices in West Hararghe, Ethiopia, where early marriage is prohibited by law. Methods: Twenty Focus Group Discussions were conducted with Married and Unmarried adolescent girls, adolescent boys and parents of girls using locally developed vignettes. A total of 32 in-depth interviews were conducted with married and unmarried adolescent girls, husbands of adolescent girls and mothers-in-law. Key informant interviews were conducted with 36 district officials. Data analysis was assisted by Open Code computer software. The Social Norms Analysis Plot (SNAP) framework developed by CARE guided the development and analysis of vignettes. A thematic data analysis approach was utilized to summarize the data. Results: Early marriage is seen as a positive phenomenon in our study context, and girls who are not married by the perceived ideal age of 15 are socially sanctioned. They are particularly influenced by their peers to marry. Marrying early is considered a chance given by God and a symbol of good luck. The two common types of marriage are decided: 1) by adolescent girl and boy themselves without seeking parental permission (’Jalaa-deemaa’- meaning ‘to go along’), and 2) by just informing girl’s parents (‘Cabsaa’- meaning ‘to break the culture’). Relatives and marriage brokers also arrange early marriages. Girls usually accept the first marriage proposal regardless of their age. Parents generally tend not to oppose marriage arrangements chosen by their daughters. Conclusions: In the study context social norms encourage early marriage despite the existence of a law prohibiting marriage before the age of eighteen years. Early marriage commonly happens through consensual arrangements between adolescent girls and boys. Interventions to reduce early marriage need to consider the influence of Reference Groups on the decision makers for marriages, especially girls’ own peers.

Keywords: adolescent girls, social norms, early marriage, Ethiopia

Procedia PDF Downloads 127
19105 Modelling Operational Risk Using Extreme Value Theory and Skew t-Copulas via Bayesian Inference

Authors: Betty Johanna Garzon Rozo, Jonathan Crook, Fernando Moreira

Abstract:

Operational risk losses are heavy tailed and are likely to be asymmetric and extremely dependent among business lines/event types. We propose a new methodology to assess, in a multivariate way, the asymmetry and extreme dependence between severity distributions, and to calculate the capital for Operational Risk. This methodology simultaneously uses (i) several parametric distributions and an alternative mix distribution (the Lognormal for the body of losses and the Generalized Pareto Distribution for the tail) via extreme value theory using SAS®, (ii) the multivariate skew t-copula applied for the first time for operational losses and (iii) Bayesian theory to estimate new n-dimensional skew t-copula models via Markov chain Monte Carlo (MCMC) simulation. This paper analyses a newly operational loss data set, SAS Global Operational Risk Data [SAS OpRisk], to model operational risk at international financial institutions. All the severity models are constructed in SAS® 9.2. We implement the procedure PROC SEVERITY and PROC NLMIXED. This paper focuses in describing this implementation.

Keywords: operational risk, loss distribution approach, extreme value theory, copulas

Procedia PDF Downloads 585
19104 Innovative Screening Tool Based on Physical Properties of Blood

Authors: Basant Singh Sikarwar, Mukesh Roy, Ayush Goyal, Priya Ranjan

Abstract:

This work combines two bodies of knowledge which includes biomedical basis of blood stain formation and fluid communities’ wisdom that such formation of blood stain depends heavily on physical properties. Moreover biomedical research tells that different patterns in stains of blood are robust indicator of blood donor’s health or lack thereof. Based on these valuable insights an innovative screening tool is proposed which can act as an aide in the diagnosis of diseases such Anemia, Hyperlipidaemia, Tuberculosis, Blood cancer, Leukemia, Malaria etc., with enhanced confidence in the proposed analysis. To realize this powerful technique, simple, robust and low-cost micro-fluidic devices, a micro-capillary viscometer and a pendant drop tensiometer are designed and proposed to be fabricated to measure the viscosity, surface tension and wettability of various blood samples. Once prognosis and diagnosis data has been generated, automated linear and nonlinear classifiers have been applied into the automated reasoning and presentation of results. A support vector machine (SVM) classifies data on a linear fashion. Discriminant analysis and nonlinear embedding’s are coupled with nonlinear manifold detection in data and detected decisions are made accordingly. In this way, physical properties can be used, using linear and non-linear classification techniques, for screening of various diseases in humans and cattle. Experiments are carried out to validate the physical properties measurement devices. This framework can be further developed towards a real life portable disease screening cum diagnostics tool. Small-scale production of screening cum diagnostic devices is proposed to carry out independent test.

Keywords: blood, physical properties, diagnostic, nonlinear, classifier, device, surface tension, viscosity, wettability

Procedia PDF Downloads 368
19103 Health State Utility Values Related to COVID-19 Pandemic Using EQ-5D: A Systematic Review and Meta-Analysis

Authors: Xu Feifei

Abstract:

The prevalence of COVID-19 currently is the biggest challenge to improving people's quality of life. Its impact on the health-related quality of life (HRQoL) is highly uncertain and has not been summarized so far. The aim of the present systematic review was to assess and provide an up-to-date analysis of the impact of the COVID-19 pandemic on the HRQoL of participants who have been infected, have not been infected but isolated, frontline, with different diseases, and the general population. Therefore, an electronic search of the literature in PubMed databases was performed from 2019 to July 2022 (without date restriction). PRISMA guideline methodology was employed, and data regarding the HRQoL were extracted from eligible studies. Articles were included if they met the following inclusion criteria: (a) reports on the data collection of the health state utility values (HSUVs) related to COVID-19 from 2019 to 2021; (b) English language and peer-reviewed journals; and (c) original HSUV data; (d) using EQ-5D tool to quantify the HRQoL. To identify studies that reported the effects on COVID-19, data on the proportion of overall HSUVs of participants who had the outcome were collected and analyzed using a one-group meta-analysis. As a result, thirty-two studies fulfilled the inclusion criteria and, therefore, were included in the systematic review. A total of 45295 participants and provided 219 means of HSUVs during COVID-19 were included in this systematic review. The range of utility is from 0.224 to 1. The study included participants from Europe (n=16), North America (n=4), Asia (n=10), South America (n=1), and Africa (n=1). Twelve articles showed that the HRQoL of the participants who have been infected with COVID-19 (range of overall HSUVs from 0.6125 to 0.863). Two studies reported the population of frontline workers (the range of overall HSUVs from 0.82 to 0.93). Seven of the articles researched the participants who had not been infected with COVID-19 but suffered from morbidities during the pandemic (range of overall HSUVs from 0.5 to 0.96). Thirteen studies showed that the HRQoL of the respondents who have not been infected with COVID-19 and without any morbidities (range of overall HSUVs from 0.64 to 0.964). Moreover, eighteen articles reported the outcomes of overall HSUVs during the COVID-19 pandemic in different population groups. The estimate of overall HSUVs of direct COVID-19 experience population (n=1333) was 0.751 (95% CI 0.670 - 0.832, I2 = 98.64%); the estimate of frontline population (n=610) was 0.906 ((95% CI 0.854 – 0.957, I2 = 98.61%); participants with different disease (n=132) were 0.768 (95% CI 0.515 - 1.021, I2= 99.26%); general population without infection history (n=29,892) was 0.825 (95% CI 0.766 - 0.885, I2 =99.69%). Conclusively, taking into account these results, this systematic review might confirm that COVID-19 has a negative impact on the HRQoL of the infected population and illness population. It provides practical value for cost-effectiveness model analysis of health states related to COVID-19.

Keywords: COVID-19, health-related quality of life, meta-analysis, systematic review, utility value

Procedia PDF Downloads 71
19102 Technology Roadmapping in Defense Industry

Authors: Sevgi Özlem Bulu, Arif Furkan Mendi, Tolga Erol, İzzet Gökhan Özbilgin

Abstract:

The rapid progress of technology in today's competitive conditions has also accelerated companies' technology development activities. As a result, companies are paying more attention to R&D studies and are beginning to allocate a larger share to R&D projects. A more systematic, comprehensive, target-oriented implementation of R&D studies is crucial for the company to achieve successful results. As a consequence, Technology Roadmap (TRM) is gaining importance as a management tool. It has critical prospects for achieving medium and long term success as it contains decisions about past business, future plans, technological infrastructure. When studies on TRM are examined, projects to be placed on the roadmap are selected by many different methods. Generally preferred methods are based on multi-criteria decision making methods. Management of selected projects becomes an important point after the selection phase of the projects. At this stage, TRM are used. TRM can be created in many different ways so that each institution can prepare its own Technology Roadmap according to their strategic plan. Depending on the intended use, there can be TRM with different layers at different sizes. In the evaluation phase of the R&D projects and in the creation of the TRM, HAVELSAN, Turkey's largest defense company in the software field, carries out this process with great care and diligence. At the beginning, suggested R&D projects are evaluated by the Technology Management Board (TMB) of HAVELSAN in accordance with the company's resources, objectives, and targets. These projects are presented to the TMB periodically for evaluation within the framework of certain criteria by board members. After the necessary steps have been passed, the approved projects are added to the time-based TRM, which is composed of four layers as market, product, project and technology. The use of a four-layered roadmap provides a clearer understanding and visualization of company strategy and objectives. This study demonstrates the benefits of using TRM, four-layered Technology Roadmapping and the possibilities for the institutions in the defense industry.

Keywords: technology roadmap, research and development project, project selection, research development in defense industry

Procedia PDF Downloads 168
19101 A Bio-Inspired Approach for Self-Managing Wireless Sensor and Actor Networks

Authors: Lyamine Guezouli, Kamel Barka, Zineb Seghir

Abstract:

Wireless sensor and actor networks (WSANs) present a research challenge for different practice areas. Researchers are trying to optimize the use of such networks through their research work. This optimization is done on certain criteria, such as improving energy efficiency, exploiting node heterogeneity, self-adaptability and self-configuration. In this article, we present our proposal for BIFSA (Biologically-Inspired Framework for Wireless Sensor and Actor networks). Indeed, BIFSA is a middleware that addresses the key issues of wireless sensor and actor networks. BIFSA consists of two types of agents: sensor agents (SA) that operate at the sensor level to collect and transport data to actors and actor agents (AA) that operate at the actor level to transport data to base stations. Once the sensor agent arrives at the actor, it becomes an actor agent, which can exploit the resources of the actors and vice versa. BIFSA allows agents to evolve their genetic structures and adapt to the current network conditions. The simulation results show that BIFSA allows the agents to make better use of all the resources available in each type of node, which improves the performance of the network.

Keywords: wireless sensor and actor networks, self-management, genetic algorithm, agent.

Procedia PDF Downloads 73
19100 Techno-Economic Analysis of Solar Energy for Cathodic Protection of Oil and Gas Buried Pipelines in Southwestern of Iran

Authors: M. Goodarzi, M. Mohammadi, A. Gharib

Abstract:

Solar energy is a renewable energy which has attracted special attention in many countries. Solar cathodic protectionsystems harness the sun’senergy to protect underground pipelinesand tanks from galvanic corrosion. The object of this study is to design and the economic analysis a cathodic protection system by impressed current supplied with solar energy panels applied to underground pipelines. In the present study, the technical and economic analysis of using solar energy for cathodic protection system in southwestern of Iran (Khuzestan province) is investigated. For this purpose, the ecological conditions such as the weather data, air clearness and sunshine hours are analyzed. The economic analyses were done using computer code to investigate the feasibility analysis from the using of various energy sources in order to cathodic protection system. The overall research methodology is divided into four components: Data collection, design of elements, techno economical evaluation, and output analysis. According to the results, solar renewable energy systems can supply adequate power for cathodic protection system purposes.

Keywords: renewable energy, solar energy, solar cathodic protection station, lifecycle cost method

Procedia PDF Downloads 523
19099 Different Perceptions of Distance and Full-time Teaching Depending on Different Cultural Backgrounds: A Comparative Study

Authors: Daniel Ecler

Abstract:

This paper aims to compare the data obtained using semi-structured questionnaires and find some connections between them, which could help to understand what factors affect the perception of the advantages and disadvantages of distance learning compared to conventional education. The data collected came from respondents from Czech and Chinese university students, and expectations were such that the different cultural environments from which the two groups come would have an impact on different experiences of distance education. With the help of variation-finding comparison, it turned out that Chinese students did not have such difficulties with the transition to distance learning as students from the Czech Republic, as most of them came into contact with some form of distance education in the past. In addition, it has also been shown that Chinese students use modern technology to a much greater extent, which has also made it easier for them to become accustomed to another form of teaching. In conclusion, Chinese students have greater preconditions for easier management of distance learning, while Czech students prefer more personal contact, and thus full-time teaching. It is obvious that both approaches have their pros and cons; now, it is necessary to find out how to use them for maximum efficiency of the educational process.

Keywords: Chinese college students, cultural background, Czech college students, distance learning, full-time teaching

Procedia PDF Downloads 139
19098 Neural Network Approach to Classifying Truck Traffic

Authors: Ren Moses

Abstract:

The process of classifying vehicles on a highway is hereby viewed as a pattern recognition problem in which connectionist techniques such as artificial neural networks (ANN) can be used to assign vehicles to their correct classes and hence to establish optimum axle spacing thresholds. In the United States, vehicles are typically classified into 13 classes using a methodology commonly referred to as “Scheme F”. In this research, the ANN model was developed, trained, and applied to field data of vehicles. The data comprised of three vehicular features—axle spacing, number of axles per vehicle, and overall vehicle weight. The ANN reduced the classification error rate from 9.5 percent to 6.2 percent when compared to an existing classification algorithm that is not ANN-based and which uses two vehicular features for classification, that is, axle spacing and number of axles. The inclusion of overall vehicle weight as a third classification variable further reduced the error rate from 6.2 percent to only 3.0 percent. The promising results from the neural networks were used to set up new thresholds that reduce classification error rate.

Keywords: artificial neural networks, vehicle classification, traffic flow, traffic analysis, and highway opera-tions

Procedia PDF Downloads 298