Search results for: large woody debris
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7179

Search results for: large woody debris

5979 Rapid Identification of Thermophilic Campylobacter Species from Retail Poultry Meat Using Matrix-Assisted Laser Desorption Ionization-Time of Flight Mass Spectrometry

Authors: Graziella Ziino, Filippo Giarratana, Stefania Maria Marotta, Alessandro Giuffrida, Antonio Panebianco

Abstract:

In Europe, North America and Japan, campylobacteriosis is one of the leading food-borne bacterial illnesses, often related to the consumption of poultry meats and/or by-products. The aim of this study was the evaluation of Campylobacter contamination of poultry meats marketed in Sicily (Italy) using both traditional methods and Matrix-Assisted Laser Desorption Ionization-Time of Flight Mass Spectrometry (MALDI-TOF MS). MALDI-TOF MS is considered a promising rapid (less than 1 hour) identification method for food borne pathogens bacteria. One hundred chicken and turkey meat preparations (no. 68 hamburgers, no. 21 raw sausages, no. 4 meatballs and no. 7 meat rolls) were taken from different butcher’s shops and large scale retailers and submitted to detection/enumeration of Campylobacter spp. according to EN ISO 10272-1:2006 and EN ISO 10272-2:2006. Campylobacter spp. was detected with general low counts in 44 samples (44%), of which 30 from large scale retailers and 14 from butcher’s shops. Chicken meats were significantly more contaminated than turkey meats. Among the preparations, Campylobacter spp. was found in 85.71% of meat rolls, 50% of meatballs, 44.12% of hamburgers and 28.57% of raw sausages. A total of 100 strains, 2-3 from each positive samples, were isolated for the identification by phenotypic, biomolecular and MALDI-TOF MS methods. C. jejuni was the predominant strains (63%), followed by C. coli (33%) and C. lari (4%). MALDI-TOF MS correctly identified 98% of the strains at the species level, only 1% of the tested strains were not identified. In the last 1%, a mixture of two different species was mixed in the same sample and MALDI-TOF MS correctly identified at least one of the strains. Considering the importance of rapid identification of pathogens in the food matrix, this method is highly recommended for the identification of suspected colonies of Campylobacteria.

Keywords: campylobacter spp., Food Microbiology, matrix-assisted laser desorption ionization-time of flight mass spectrometry, rapid microbial identification

Procedia PDF Downloads 292
5978 Human Behavioral Assessment to Derive Land-Use for Sustenance of River in India

Authors: Juhi Sah

Abstract:

Habitat is characterized by the inter-dependency of environmental elements. Anthropocentric development approach is increasing our vulnerability towards natural hazards. Hence, manmade interventions should have a higher level of sensitivity towards the natural settings. Sensitivity towards the environment can be assessed by the behavior of the stakeholders involved. This led to the establishment of a hypothesis: there exists a legitimate relationship between the behavioral sciences, land use evolution and environment conservation, in the planning process. An attempt has been made to establish this relationship by reviewing the existing set of knowledge and case examples pertaining to the three disciplines under inquiry. Understanding the scarce & deteriorating nature of fresh-water reserves of earth and experimenting the above concept, a case study of a growing urban center's river flood plain is selected, in a developing economy, India. Cases of urban flooding in Chennai, Delhi and other mega cities of India, imposes a high risk on the unauthorized settlement, on the floodplains of the rivers. The issue addressed here is the encroachment of floodplains, through psychological enlightenment and modification through knowledge building. The reaction of an individual or society can be compared to a cognitive process. This study documents all the stakeholders' behavior and perception for their immediate natural environment (water body), and produce various land uses suitable along a river in an urban settlement as per different stakeholder's perceptions. To assess and induce morally responsible behavior in a community (small scale or large scale), tools of psychological inquiry is used for qualitative analysis. The analysis will deal with varied data sets from two sectors namely: River and its geology, Land use planning and regulation. Identification of a distinctive pattern in the built up growth, river ecology degradation, and human behavior, by handling large quantum of data from the diverse sector and comments on the availability of relevant data and its implications, has been done. Along the whole river stretch, condition and usage of its bank vary, hence stakeholder specific survey questionnaires have been prepared to accurately map the responses and habits of the rational inhabitants. A conceptual framework has been designed to move forward with the empirical analysis. The classical principle of virtues says "virtue of a human depends on its character" but another concept defines that the behavior or response is a derivative of situations and to bring about a behavioral change one needs to introduce a disruption in the situation/environment. Owing to the present trends, blindly following the results of data analytics and using it to construct policy, is not proving to be in favor of planned development and natural resource conservation. Thus behavioral assessment of the rational inhabitants of the planet is also required, as their activities and interests have a large impact on the earth's pre-set systems and its sustenance.

Keywords: behavioral assessment, flood plain encroachment, land use planning, river sustenance

Procedia PDF Downloads 117
5977 Analyzing Industry-University Collaboration Using Complex Networks and Game Theory

Authors: Elnaz Kanani-Kuchesfehani, Andrea Schiffauerova

Abstract:

Due to the novelty of the nanotechnology science, its highly knowledge intensive content, and its invaluable application in almost all technological fields, the close interaction between university and industry is essential. A possible gap between academic strengths to generate good nanotechnology ideas and industrial capacity to receive them can thus have far-reaching consequences. In order to be able to enhance the collaboration between the two parties, a better understanding of knowledge transfer within the university-industry relationship is needed. The objective of this research is to investigate the research collaboration between academia and industry in Canadian nanotechnology and to propose the best cooperative strategy to maximize the quality of the produced knowledge. First, a network of all Canadian academic and industrial nanotechnology inventors is constructed using the patent data from the USPTO (United States Patent and Trademark Office), and it is analyzed with social network analysis software. The actual level of university-industry collaboration in Canadian nanotechnology is determined and the significance of each group of actors in the network (academic vs. industrial inventors) is assessed. Second, a novel methodology is proposed, in which the network of nanotechnology inventors is assessed from a game theoretic perspective. It involves studying a cooperative game with n players each having at most n-1 decisions to choose from. The equilibrium leads to a strategy for all the players to choose their co-worker in the next period in order to maximize the correlated payoff of the game. The payoffs of the game represent the quality of the produced knowledge based on the citations of the patents. The best suggestion for the next collaborative relationship is provided for each actor from a game theoretic point of view in order to maximize the quality of the produced knowledge. One of the major contributions of this work is the novel approach which combines game theory and social network analysis for the case of large networks. This approach can serve as a powerful tool in the analysis of the strategic interactions of the network actors within the innovation systems and other large scale networks.

Keywords: cooperative strategy, game theory, industry-university collaboration, knowledge production, social network analysis

Procedia PDF Downloads 258
5976 Indirect Intergranular Slip Transfer Modeling Through Continuum Dislocation Dynamics

Authors: A. Kalaei, A. H. W. Ngan

Abstract:

In this study, a mesoscopic continuum dislocation dynamics (CDD) approach is applied to simulate the intergranular slip transfer. The CDD scheme applies an efficient kinematics equation to model the evolution of the “all-dislocation density,” which is the line-length of dislocations of each character per unit volume. As the consideration of every dislocation line can be a limiter for the simulation of slip transfer in large scales with a large quantity of participating dislocations, a coarse-grained, extensive description of dislocations in terms of their density is utilized to resolve the effect of collective motion of dislocation lines. For dynamics closure, namely, to obtain the dislocation velocity from a velocity law involving the effective glide stress, mutual elastic interaction of dislocations is calculated using Mura’s equation after singularity removal at the core of dislocation lines. The developed scheme for slip transfer can therefore resolve the effects of the elastic interaction and pile-up of dislocations, which are important physics omitted in coarser models like crystal plasticity finite element methods (CPFEMs). Also, the length and timescales of the simulationareconsiderably larger than those in molecular dynamics (MD) and discrete dislocation dynamics (DDD) models. The present work successfully simulates that, as dislocation density piles up in front of a grain boundary, the elastic stress on the other side increases, leading to dislocation nucleation and stress relaxation when the local glide stress exceeds the operation stress of dislocation sources seeded on the other side of the grain boundary. More importantly, the simulation verifiesa phenomenological misorientation factor often used by experimentalists, namely, the ease of slip transfer increases with the product of the cosines of misorientation angles of slip-plane normals and slip directions on either side of the grain boundary. Furthermore, to investigate the effects of the critical stress-intensity factor of the grain boundary, dislocation density sources are seeded at different distances from the grain boundary, and the critical applied stress to make slip transfer happen is studied.

Keywords: grain boundary, dislocation dynamics, slip transfer, elastic stress

Procedia PDF Downloads 123
5975 Wireless Backhauling for 5G Small Cell Networks

Authors: Abdullah A. Al Orainy

Abstract:

Small cell backhaul solutions need to be cost-effective, scalable, and easy to install. This paper presents an overview of small cell backhaul technologies. Wireless solutions including TV white space, satellite, sub-6 GHz radio wave, microwave and mmWave with their backhaul characteristics are discussed. Recent research on issues like beamforming, backhaul architecture, precoding and large antenna arrays, and energy efficiency for dense small cell backhaul with mmWave communications is reviewed. Recent trials of 5G technologies are summarized.

Keywords: backhaul, small cells, wireless, 5G

Procedia PDF Downloads 512
5974 The Role of Muzara’ah Islamic Financing in Supporting Smallholder Farmers among Muslim Communities: An Empirical Experience of Yobe Microfinance Bank

Authors: Sheriff Muhammad Ibrahim

Abstract:

The contemporary world has seen many agents of market liberalization, globalization, and expansion in agribusiness, which pose a big threat to the existence of smallholder farmers in the farming business or, at most, being marginalized against government interventions, investors' partnerships and further stretched by government policies in an effort to promote subsistent farming that can generate profits and speedy growth through attracting foreign businesses. The consequence of these modern shifts ends basically at the expense of smallholder farmers. Many scholars believed that this shift was among the major causes of urban-rural drift facing almost all communities in the World. In an effort to address these glaring economic crises, various governments at different levels and development agencies have created different programs trying to identify other sources of income generation for rural farmers. However, despite the different approaches adopted by many communities and states, the mass rural exodus continues to increase as the rural farmers continue to lose due to a lack of reliable sources for cost-efficient inputs such as agricultural extension services, mechanization supports, quality, and improved seeds, soil matching fertilizers and access to credit facilities and profitable markets for rural farmers output. Unfortunately for them, they see these agricultural requirements provided by large-scale farmers making their farming activities cheaper and yields higher. These have further created other social problems between the smallholder farmers and the large-scale farmers in many areas. This study aims to suggest the Islamic mode of agricultural financing named Muzara’ah for smallholder farmers as a microfinance banking product adopted and practiced by Yobe Microfinance Bank as a model to promote agricultural financing to be adopted in other communities. The study adopts a comparative research method to conclude that the Muzara’ah model of financing can be adopted as a valid means of financing smallholder farmers and reducing food insecurity.

Keywords: Muzara'ah, Islamic finance, agricultural financing, microfinance, smallholder farmers

Procedia PDF Downloads 62
5973 Influence of Model Hydrometeor Form on Probability of Discharge Initiation from Artificial Charged Water Aerosol Cloud

Authors: A. G. Temnikov, O. S. Belova, L. L. Chernensky, T. K. Gerastenok, N. Y. Lysov, A. V. Orlov, D. S. Zhuravkova

Abstract:

Hypothesis of the lightning initiation on the arrays of large hydrometeors are in the consideration. There is no agreement about the form the hydrometeors that could be the best for the lightning initiation from the thundercloud. Artificial charged water aerosol clouds of the positive or negative polarity could help investigate the possible influence of the hydrometeor form on the peculiarities and the probability of the lightning discharge initiation between the thundercloud and the ground. Artificial charged aerosol clouds that could create the electric field strength in the range of 5-6 kV/cm to 16-18 kV/cm have been used in experiments. The array of the model hydrometeors of the volume and plate form has been disposed near the bottom cloud boundary. It was established that the different kinds of the discharge could be initiated in the presence of the model hydrometeors array – from the cloud discharges up to the diffuse and channel discharges between the charged cloud and the ground. It was found that the form of the model hydrometeors could significantly influence the channel discharge initiation from the artificial charged aerosol cloud of the negative or positive polarity correspondingly. Analysis and generalization of the experimental results have shown that the maximal probability of the channel discharge initiation and propagation stimulation has been observed for the artificial charged cloud of the positive polarity when the arrays of the model hydrometeors of the cylinder revolution form have been used. At the same time, for the artificial charged clouds of the negative polarity, application of the model hydrometeor array of the plate rhombus form has provided the maximal probability of the channel discharge formation between the charged cloud and the ground. The established influence of the form of the model hydrometeors on the channel discharge initiation and from the artificial charged water aerosol cloud and its following successful propagation has been related with the different character of the positive and negative streamer and volume leader development on the model hydrometeors array being near the bottom boundary of the charged cloud. The received experimental results have shown the possibly important role of the form of the large hail particles precipitated in thundercloud on the discharge initiation.

Keywords: cloud and channel discharges, hydrometeor form, lightning initiation, negative and positive artificial charged aerosol cloud

Procedia PDF Downloads 316
5972 Ergonomics Aspects of Work with Computers

Authors: Leena Korpinen, Rauno Pääkkönen, Fabriziomaria Gobba

Abstract:

This paper is based on a large questionnaire study. The paper presents how all participants and subgroups (upper- and lower-level white-collar workers) answered the question, “Have you had an ache, pain, or numbness, which you associate with desktop computer use, in the different body parts during the last 12 months?’ 14.6% of participants (19.4% of women and 8.2% of men) reported that they had often or very often physical symptoms in the neck. Even if our results cannot prove a causal relation of symptoms with computer use, show that workers believe that computer use can influence their wellbeing: This is important when devising treatment modalities to decrease these physical symptoms.

Keywords: ergonomics, work, computer, symptoms

Procedia PDF Downloads 402
5971 Oncolytic Efficacy of Thymidine Kinase-Deleted Vaccinia Virus Strain Tiantan (oncoVV-TT) in Glioma

Authors: Seyedeh Nasim Mirbahari, Taha Azad, Mehdi Totonchi

Abstract:

Oncolytic viruses, which only replicate in tumor cells, are being extensively studied for their use in cancer therapy. A particular virus known as the vaccinia virus, a member of the poxvirus family, has demonstrated oncolytic abilities glioma. Treating Glioma with traditional methods such as chemotherapy and radiotherapy is quite challenging. Even though oncolytic viruses have shown immense potential in cancer treatment, their effectiveness in glioblastoma treatment is still low. Therefore, there is a need to improve and optimize immunotherapies for better results. In this study, we have designed oncoVV-TT, which can more effectively target tumor cells while minimizing replication in normal cells by replacing the thymidine kinase gene with a luc-p2a-GFP gene expression cassette. Human glioblastoma cell line U251 MG, rat glioblastoma cell line C6, and non-tumor cell line HFF were plated at 105 cells in a 12-well plates in 2 mL of DMEM-F2 medium with 10% FBS added to each well. Then incubated at 37°C. After 16 hours, the cells were treated with oncoVV-TT at an MOI of 0.01, 0.1 and left in the incubator for a further 24, 48, 72 and 96 hours. Viral replication assay, fluorescence imaging and viability tests, including trypan blue and crystal violet, were conducted to evaluate the cytotoxic effect of oncoVV-TT. The finding shows that oncoVV-TT had significantly higher cytotoxic activity and proliferation rates in tumor cells in a dose and time-dependent manner, with the strongest effect observed in U251 MG. To conclude, oncoVV-TT has the potential to be a promising oncolytic virus for cancer treatment, with a more cytotoxic effect in human glioblastoma cells versus rat glioma cells. To assess the effectiveness of vaccinia virus-mediated viral therapy, we have tested U251mg and C6 tumor cell lines taken from human and rat gliomas, respectively. The study evaluated oncoVV-TT's ability to replicate and lyse cells and analyzed the survival rates of the tested cell lines when treated with different doses of oncoVV-TT. Additionally, we compared the sensitivity of human and mouse glioma cell lines to the oncolytic vaccinia virus. All experiments regarding viruses were conducted under biosafety level 2. We engineered a Vaccinia-based oncolytic virus called oncoVV-TT to replicate specifically in tumor cells. To propagate the oncoVV-TT virus, HeLa cells (5 × 104/well) were plated in 24-well plates and incubated overnight to attach to the bottom of the wells. Subsequently, 10 MOI virus was added. After 48 h, cells were harvested by scraping, and viruses were collected by 3 sequential freezing and thawing cycles followed by removal of cell debris by centrifugation (1500 rpm, 5 min). The supernatant was stored at −80 ◦C for the following experiments. To measure the replication of the virus in Hela, cells (5 × 104/well) were plated in 24-well plates and incubated overnight to attach to the bottom of the wells. Subsequently, 5 MOI virus or equal dilution of PBS was added. At the treatment time of 0 h, 24 h, 48 h, 72 h and 96 h, the viral titers were determined under the fluorescence microscope (BZ-X700; Keyence, Osaka, Japan). Fluorescence intensity was quantified using the imagej software according to the manufacturer’s protocol. For the isolation of single-virus clones, HeLa cells seeded in six-well plates (5×105 cells/well). After 24 h (100% confluent), the cells were infected with a 10-fold dilution series of TianTan green fluorescent protein (GFP)virus and incubated for 4 h. To examine the cytotoxic effect of oncoVV-TT virus ofn U251mg and C6 cell, trypan blue and crystal violet assay was used.

Keywords: oncolytic virus, immune therapy, glioma, vaccinia virus

Procedia PDF Downloads 79
5970 Examining the Role of Corporate Culture in Driving Firm Performance

Authors: Lovorka Galetić, Ivana Načinović Braje, Nevenka Čavlek

Abstract:

The purpose of this paper is to analyze the relationship between corporate culture and firm performance. Extensive theoretical and empirical evidence on this issue is provided. A quantitative methodology was used to explore relationship between corporate culture and performance among large Croatian companies. Corporate culture was explored by using Denison framework. The research revealed a positive, statistically significant relationship between mission and performance. Other dimensions of corporate culture (involvement, consistency and adaptability) show only partial relationship with performance.

Keywords: corporate culture, Croatia, Denison culture model, performance

Procedia PDF Downloads 528
5969 Multiscale Analysis of Shale Heterogeneity in Silurian Longmaxi Formation from South China

Authors: Xianglu Tang, Zhenxue Jiang, Zhuo Li

Abstract:

Characterization of shale multi scale heterogeneity is an important part to evaluate size and space distribution of shale gas reservoirs in sedimentary basins. The origin of shale heterogeneity has always been a hot research topic for it determines shale micro characteristics description and macro quality reservoir prediction. Shale multi scale heterogeneity was discussed based on thin section observation, FIB-SEM, QEMSCAN, TOC, XRD, mercury intrusion porosimetry (MIP), and nitrogen adsorption analysis from 30 core samples in Silurian Longmaxi formation. Results show that shale heterogeneity can be characterized by pore structure and mineral composition. The heterogeneity of shale pore is showed by different size pores at nm-μm scale. Macropores (pore diameter > 50 nm) have a large percentage of pore volume than mesopores (pore diameter between 2~ 50 nm) and micropores (pore diameter < 2nm). However, they have a low specific surface area than mesopores and micropores. Fractal dimensions of the pores from nitrogen adsorption data are higher than 2.7, what are higher than 2.8 from MIP data, showing extremely complex pore structure. This complexity in pore structure is mainly due to the organic matter and clay minerals with complex pore network structures, and diagenesis makes it more complicated. The heterogeneity of shale minerals is showed by mineral grains, lamina, and different lithology at nm-km scale under the continuous changing horizon. Through analyzing the change of mineral composition at each scale, random arrangement of mineral equal proportion, seasonal climate changes, large changes of sedimentary environment, and provenance supply are considered to be the main reasons that cause shale minerals heterogeneity from microcosmic to macroscopic. Due to scale effect, the change of shale multi scale heterogeneity is a discontinuous process, and there is a transformation boundary between homogeneous and in homogeneous. Therefore, a shale multi scale heterogeneity changing model is established by defining four types of homogeneous unit at different scales, which can be used to guide the prediction of shale gas distribution from micro scale to macro scale.

Keywords: heterogeneity, homogeneous unit, multiscale, shale

Procedia PDF Downloads 452
5968 A New Modification of Nonlinear Conjugate Gradient Coefficients with Global Convergence Properties

Authors: Ahmad Alhawarat, Mustafa Mamat, Mohd Rivaie, Ismail Mohd

Abstract:

Conjugate gradient method has been enormously used to solve large scale unconstrained optimization problems due to the number of iteration, memory, CPU time, and convergence property, in this paper we find a new class of nonlinear conjugate gradient coefficient with global convergence properties proved by exact line search. The numerical results for our new βK give a good result when it compared with well-known formulas.

Keywords: conjugate gradient method, conjugate gradient coefficient, global convergence

Procedia PDF Downloads 463
5967 High-Fidelity Materials Screening with a Multi-Fidelity Graph Neural Network and Semi-Supervised Learning

Authors: Akeel A. Shah, Tong Zhang

Abstract:

Computational approaches to learning the properties of materials are commonplace, motivated by the need to screen or design materials for a given application, e.g., semiconductors and energy storage. Experimental approaches can be both time consuming and costly. Unfortunately, computational approaches such as ab-initio electronic structure calculations and classical or ab-initio molecular dynamics are themselves can be too slow for the rapid evaluation of materials, often involving thousands to hundreds of thousands of candidates. Machine learning assisted approaches have been developed to overcome the time limitations of purely physics-based approaches. These approaches, on the other hand, require large volumes of data for training (hundreds of thousands on many standard data sets such as QM7b). This means that they are limited by how quickly such a large data set of physics-based simulations can be established. At high fidelity, such as configuration interaction, composite methods such as G4, and coupled cluster theory, gathering such a large data set can become infeasible, which can compromise the accuracy of the predictions - many applications require high accuracy, for example band structures and energy levels in semiconductor materials and the energetics of charge transfer in energy storage materials. In order to circumvent this problem, multi-fidelity approaches can be adopted, for example the Δ-ML method, which learns a high-fidelity output from a low-fidelity result such as Hartree-Fock or density functional theory (DFT). The general strategy is to learn a map between the low and high fidelity outputs, so that the high-fidelity output is obtained a simple sum of the physics-based low-fidelity and correction, Although this requires a low-fidelity calculation, it typically requires far fewer high-fidelity results to learn the correction map, and furthermore, the low-fidelity result, such as Hartree-Fock or semi-empirical ZINDO, is typically quick to obtain, For high-fidelity outputs the result can be an order of magnitude or more in speed up. In this work, a new multi-fidelity approach is developed, based on a graph convolutional network (GCN) combined with semi-supervised learning. The GCN allows for the material or molecule to be represented as a graph, which is known to improve accuracy, for example SchNet and MEGNET. The graph incorporates information regarding the numbers of, types and properties of atoms; the types of bonds; and bond angles. They key to the accuracy in multi-fidelity methods, however, is the incorporation of low-fidelity output to learn the high-fidelity equivalent, in this case by learning their difference. Semi-supervised learning is employed to allow for different numbers of low and high-fidelity training points, by using an additional GCN-based low-fidelity map to predict high fidelity outputs. It is shown on 4 different data sets that a significant (at least one order of magnitude) increase in accuracy is obtained, using one to two orders of magnitude fewer low and high fidelity training points. One of the data sets is developed in this work, pertaining to 1000 simulations of quinone molecules (up to 24 atoms) at 5 different levels of fidelity, furnishing the energy, dipole moment and HOMO/LUMO.

Keywords: .materials screening, computational materials, machine learning, multi-fidelity, graph convolutional network, semi-supervised learning

Procedia PDF Downloads 41
5966 Unsupervised Part-of-Speech Tagging for Amharic Using K-Means Clustering

Authors: Zelalem Fantahun

Abstract:

Part-of-speech tagging is the process of assigning a part-of-speech or other lexical class marker to each word into naturally occurring text. Part-of-speech tagging is the most fundamental and basic task almost in all natural language processing. In natural language processing, the problem of providing large amount of manually annotated data is a knowledge acquisition bottleneck. Since, Amharic is one of under-resourced language, the availability of tagged corpus is the bottleneck problem for natural language processing especially for POS tagging. A promising direction to tackle this problem is to provide a system that does not require manually tagged data. In unsupervised learning, the learner is not provided with classifications. Unsupervised algorithms seek out similarity between pieces of data in order to determine whether they can be characterized as forming a group. This paper explicates the development of unsupervised part-of-speech tagger using K-Means clustering for Amharic language since large amount of data is produced in day-to-day activities. In the development of the tagger, the following procedures are followed. First, the unlabeled data (raw text) is divided into 10 folds and tokenization phase takes place; at this level, the raw text is chunked at sentence level and then into words. The second phase is feature extraction which includes word frequency, syntactic and morphological features of a word. The third phase is clustering. Among different clustering algorithms, K-means is selected and implemented in this study that brings group of similar words together. The fourth phase is mapping, which deals with looking at each cluster carefully and the most common tag is assigned to a group. This study finds out two features that are capable of distinguishing one part-of-speech from others these are morphological feature and positional information and show that it is possible to use unsupervised learning for Amharic POS tagging. In order to increase performance of the unsupervised part-of-speech tagger, there is a need to incorporate other features that are not included in this study, such as semantic related information. Finally, based on experimental result, the performance of the system achieves a maximum of 81% accuracy.

Keywords: POS tagging, Amharic, unsupervised learning, k-means

Procedia PDF Downloads 451
5965 Morphometric and Radiographic Studies on the Tarsal Bones of Adult Chinkara (Gazella bennettii)

Authors: Salahud Din, Saima Masood, Hafsa Zaneb, Habib-Ur Rehman, Imad Khan, Muqader Shah

Abstract:

The present study was carried out on the gross anatomy, biometery and radiographic analysis of tarsal bones in twenty specimens of adult chinkara (Gazella bennettii). The desired bones were collected from the graveyards present in the locality of the different safari parks and zoos in Pakistan. To observe the edges and articulations between the bones, the radiographic images were acquired in craniocaudals and mediolateral views of the intact limbs. The gross and radiographic studies of the tarsus of adult Chinkara were carried out in University of Veterinary and Animal Sciences, Lahore, Pakistan. The tarsus of chinkara comprised of five bones both grossly and radiographically, settled in three transverse rows: tibial and fibular tarsal in the proximal, central and fourth fused tarsal in the middle row, the first, second and third fused tarsal in the distal row. The fibular tarsal was the largest and longest bone of the hock, situated on the lateral side and had a bulbous tuber calcis 'point of the hock' at the proximal extremity which projects upward and backward. The average maximum height and breadth for fibular tarsal was 5.61 ± 0.23 cm and 2.06 ± 0.13 cm, respectively. The tibial tarsal bones were the 2nd largest bone of the proximal row and lie on the medial side of the tarsus bears trochlea at either end. The average maximum height and breadth for tibial tarsal was 2.79 ± 0.05 cm and 1.74 ± 0.01 cm, respectively. The central and the fourth tarsals were fused to form a large bone which extends across the entire width of the tarsus and articulates with all bones of the tarsus. A nutrient foramen was present in the center of the non auricular area, more prominent on the ventral surface. The average maximum height and breadth for central and fourth fused tarsal was 1.51 ± 0.13 cm and 2.08 ± 0.07 cm, respectively. The first tarsal was a quadrilateral piece of bone placed on the poteriomedial surface of the hock. The greatest length and maximum breadth of the first tarsal was 0.94 ± 0.01 cm and 1.01 ± 0.01 cm, respectively. The second and third fused tarsal bone resembles the central but was smaller and triangular in outline. It was situated between the central above and the large metatarsal bone below. The greatest length and maximum breadth of second and third fused tarsal was 0.98 ± 0.01 cm and 1.49 ± 0.01 cm.

Keywords: chinkara, morphometry, radiography, tarsal bone

Procedia PDF Downloads 171
5964 Evaluation of a Piecewise Linear Mixed-Effects Model in the Analysis of Randomized Cross-over Trial

Authors: Moses Mwangi, Geert Verbeke, Geert Molenberghs

Abstract:

Cross-over designs are commonly used in randomized clinical trials to estimate efficacy of a new treatment with respect to a reference treatment (placebo or standard). The main advantage of using cross-over design over conventional parallel design is its flexibility, where every subject become its own control, thereby reducing confounding effect. Jones & Kenward, discuss in detail more recent developments in the analysis of cross-over trials. We revisit the simple piecewise linear mixed-effects model, proposed by Mwangi et. al, (in press) for its first application in the analysis of cross-over trials. We compared performance of the proposed piecewise linear mixed-effects model with two commonly cited statistical models namely, (1) Grizzle model; and (2) Jones & Kenward model, used in estimation of the treatment effect, in the analysis of randomized cross-over trial. We estimate two performance measurements (mean square error (MSE) and coverage probability) for the three methods, using data simulated from the proposed piecewise linear mixed-effects model. Piecewise linear mixed-effects model yielded lowest MSE estimates compared to Grizzle and Jones & Kenward models for both small (Nobs=20) and large (Nobs=600) sample sizes. It’s coverage probability were highest compared to Grizzle and Jones & Kenward models for both small and large sample sizes. A piecewise linear mixed-effects model is a better estimator of treatment effect than its two competing estimators (Grizzle and Jones & Kenward models) in the analysis of cross-over trials. The data generating mechanism used in this paper captures two time periods for a simple 2-Treatments x 2-Periods cross-over design. Its application is extendible to more complex cross-over designs with multiple treatments and periods. In addition, it is important to note that, even for single response models, adding more random effects increases the complexity of the model and thus may be difficult or impossible to fit in some cases.

Keywords: Evaluation, Grizzle model, Jones & Kenward model, Performance measures, Simulation

Procedia PDF Downloads 122
5963 Top-Down Approach for Fabricating Hematite Nanowire Arrays

Authors: Seungmin Shin, Jin-Baek Kim

Abstract:

Hematite (α-Fe2O3) has very good semiconducting properties with a band gap of 2.1 eV and is antiferromagnetic. Due to its electrochemical stability, low toxicity, wide abundance, and low-cost, hematite, it is a particularly attractive material for photoelectrochemical cells. Additionally, hematite has also found applications in gas sensing, field emission, heterogeneous catalysis, and lithium-ion battery electrodes. Here, we discovered a new universal top-down method for the synthesis of one-dimensional hematite nanowire arrays. Various shapes and lengths of hematite nanowire have been easily fabricated over large areas by sequential processes. The obtained hematite nanowire arrays are promising candidates as photoanodes in photoelectrochemical solar cells.

Keywords: hematite, lithography, nanowire, top-down process

Procedia PDF Downloads 249
5962 Monitoring the Production of Large Composite Structures Using Dielectric Tool Embedded Capacitors

Authors: Galatee Levadoux, Trevor Benson, Chris Worrall

Abstract:

With the rise of public awareness on climate change comes an increasing demand for renewable sources of energy. As a result, the wind power sector is striving to manufacture longer, more efficient and reliable wind turbine blades. Currently, one of the leading causes of blade failure in service is improper cure of the resin during manufacture. The infusion process creating the main part of the composite blade structure remains a critical step that is yet to be monitored in real time. This stage consists of a viscous resin being drawn into a mould under vacuum, then undergoing a curing reaction until solidification. Successful infusion assumes the resin fills all the voids and cures completely. Given that the electrical properties of the resin change significantly during its solidification, both the filling of the mould and the curing reaction are susceptible to be followed using dieletrometry. However, industrially available dielectrics sensors are currently too small to monitor the entire surface of a wind turbine blade. The aim of the present research project is to scale up the dielectric sensor technology and develop a device able to monitor the manufacturing process of large composite structures, assessing the conformity of the blade before it even comes out of the mould. An array of flat copper wires acting as electrodes are embedded in a polymer matrix fixed in an infusion mould. A multi-frequency analysis from 1 Hz to 10 kHz is performed during the filling of the mould with an epoxy resin and the hardening of the said resin. By following the variations of the complex admittance Y*, the filling of the mould and curing process are monitored. Results are compared to numerical simulations of the sensor in order to validate a virtual cure-monitoring system. The results obtained by drawing glycerol on top of the copper sensor displayed a linear relation between the wetted length of the sensor and the complex admittance measured. Drawing epoxy resin on top of the sensor and letting it cure at room temperature for 24 hours has provided characteristic curves obtained when conventional interdigitated sensor are used to follow the same reaction. The response from the developed sensor has shown the different stages of the polymerization of the resin, validating the geometry of the prototype. The model created and analysed using COMSOL has shown that the dielectric cure process can be simulated, so long as a sufficient time and temperature dependent material properties can be determined. The model can be used to help design larger sensors suitable for use with full-sized blades. The preliminary results obtained with the sensor prototype indicate that the infusion and curing process of an epoxy resin can be followed with the chosen configuration on a scale of several decimeters. Further work is to be devoted to studying the influence of the sensor geometry and the infusion parameters on the results obtained. Ultimately, the aim is to develop a larger scale sensor able to monitor the flow and cure of large composite panels industrially.

Keywords: composite manufacture, dieletrometry, epoxy, resin infusion, wind turbine blades

Procedia PDF Downloads 166
5961 Agricultural Cooperative Model: A Panacea for Economic Development of Small Scale Business Famers in Ilesha, Osun State, Nigeria

Authors: Folasade Adegbaju, Olusola Arowolo, Olufisayo Onawumi

Abstract:

Owolowo ile – ege garri processing industry which is a small scale cassava processing industry, located in Ilesha, Osun State was purposively selected as a case study because it is a cooperative business. This industry was established in 1991 by eight men (8) who were mostly retirees. A researcher made questionnaire was used to collect information from thirty (30) respondents: the manager, four official staffs and 25 randomly selected processors in the industry. The study found that within twelve years of the utilization of their self raised initial capital of N240, 000 naira (Two hundred and forty thousand naira) this cassava – based industry had impacted on and attracted the involvement of many more people because within the period of the study (i.e. 2007-2011) the processors had quadrupled in number (e.g. 8 to 30), the facilities (equipment) in use had increased from one machine and a frying pot to many, this translated into being able to produce large quantities of fried garri, fufu and also starch for marketing to the people in Ilesha and neighbouring cities like Ibadan, Lagos, etc. This is indicative of economic growth. The industry also became a source of employment for community members in the sense that, as at the time of study four staffs were employed to work and coordinate the industry. It was observed that despite all odds of small-scale industry and the problem of people migrating from rural to urban area, this agro-based industry still existed successfully in the community, and many of such industry can be replicated by such agricultural cooperative groups nationwide so as to further boost the productivity as well as the economy of the area and nation at large. However, government and individual still have major roles to play in ensuring the growth and development of the nation in this respect.The local agricultural cooperative groups should form regional cooperative consortium with more networking for the farmers, in order to create more jobs for the young ones and to increase agricultural productivity in the country thus resulting in a better and more sustainable economy.

Keywords: agricultural cooperative, cassava processing industry, model, small scale enterprise

Procedia PDF Downloads 290
5960 Saudi Twitter Corpus for Sentiment Analysis

Authors: Adel Assiri, Ahmed Emam, Hmood Al-Dossari

Abstract:

Sentiment analysis (SA) has received growing attention in Arabic language research. However, few studies have yet to directly apply SA to Arabic due to lack of a publicly available dataset for this language. This paper partially bridges this gap due to its focus on one of the Arabic dialects which is the Saudi dialect. This paper presents annotated data set of 4700 for Saudi dialect sentiment analysis with (K= 0.807). Our next work is to extend this corpus and creation a large-scale lexicon for Saudi dialect from the corpus.

Keywords: Arabic, sentiment analysis, Twitter, annotation

Procedia PDF Downloads 630
5959 Beneath the Leisurely Surface: An Analysis of the Piano Lesson Frenzy among Chinese Middle-Class Parents

Authors: Yijie Wang, Tianyue Wang

Abstract:

In the past two decades, there has been a great ‘piano lesson frenzy’ among Chinese middle-class families, with a large number of parents adding piano training to children’s extra-curriculum lists. Superficially, the frenzy reflects a rather ‘leisurely’ attitude: parents typically claim that pianos lessons are ‘just for fun’ and will hopefully render children’s life more exciting. However, a closer scrutiny reveals that there is great social-status anxiety hidden beneath this ‘leisurely’ surface. Based on pre-interviews of six Chinese middle-class parents who have enthusiastically signed their children up for piano lessons, several tentative analysis are made: 1. Owing to a series of historical and social factors, the Chinese middle-class have yet to establish their cultural norms in the past few decades, resulting in great confusion concerning how to cultivate cultural tastes in their offspring. And partly due to the fact that the middle-class status of the past Chinese generation is mostly self-acquired rather than inherited, parents are much less confident about their cultural resources—which require long-time accumulation—than material ones. Both factors combine to lead to a sort of blind, overcompensating enthusiasm in culture-related education, and the piano frenzy is but a demonstration. 2. The piano has been chosen to be the object of the frenzy partly because of its inherent characteristics as well as socially-constructed ones. Costly, large in size, imported from another culture and so forth, the piano has acquired the meaning of being exclusive, high-end and exotic, which renders it a token of top-tier status among Chinese people, and piano lessons for offspring have therefore become parents’ paths towards a kind of ‘symbolic elevation’. A child playing piano is an exhibition as well as psychological assurance of the families’ middle-class status. 3. A closer look at children’s piano training process reveals that there is much more anxiety than leisurely elements involved. Despite parents’ claim that ‘piano is mainly for kids to have fun,’ the whole process is evidently of a rather ‘ascetic’ nature, with the demands of diligence and senses of time urgency throughout, and techniques rather than flair or styles are emphasized. This either means that the apparent ‘piano-for-fun’ stance is unauthentic and is only other motives in disguise, or that the Chinese middle-class parents are not yet capable of shaking off the sense of anxiety even if they sincerely intend to. 4. When viewed in relation to Chinese formal school system as well as the job market at large, it can be said that by signing children up for piano lessons, parents are consciously or unconsciously seeking to prepare for, or reduce the risks of, their children’s future social mobility. In face of possible failures in the highly-crucial, highly-competitive formal school system, piano-playing as an extra-curriculum activity may be conveniently transferred into an alternative career path. Besides, in contemporary China, as the occupational structure goes through change, and the school-related certificates decline in value, aspects such as a person’s overall deportment, which can be gained or proved by piano-learning, have been gaining in significance.

Keywords: extra-curriculum activities, middle class, piano lesson frenzy, status anxiety

Procedia PDF Downloads 245
5958 Experimental Studies on the Effect of Premixing Methods in Anaerobic Digestor with Corn Stover

Authors: M. Sagarika, M. Chandra Sekhar

Abstract:

Agricultural residues are producing in large quantities in India and account for abundant but underutilized source of renewable biomass in agriculture. In India, the amount of crop residues available is estimated to be approximately 686 million tons. Anaerobic digestion is a promising option to utilize the surplus agricultural residues and can produce biogas and digestate. Biogas is mainly methane (CH4), which can be utilized as an energy source in replacement for fossil fuels such as natural gas, oil, in other hand, digestate contains high amounts of nutrients, can be employed as fertilizer. Solid state anaerobic digestion (total solids ≥ 15%) is suitable for agricultural residues, as it reduces the problems like stratification and floating issues that occur in liquid anaerobic digestion (total solids < 15%). The major concern in solid-state anaerobic digestion is the low mass transfer of feedstock and inoculum that resulting in low performance. To resolve this low mass transfer issue, effective mixing of feedstock and inoculum is required. Mechanical mixing using stirrer at the time of digestion process can be done, but it is difficult to operate the stirring of feedstock with high solids percentage and high viscosity. Complete premixing of feedstock and inoculum is an alternative method, which is usual in lab scale studies but may not be affordable due to high energy demand in large-scale digesters. Developing partial premixing methods may reduce this problem. Current study is to improve the performance of solid-state anaerobic digestion of corn stover at feedstock to inoculum ratios 3 and 5, by applying partial premixing methods and to compare the complete premixing method with two partial premixing methods which are two alternative layers of feedstock and inoculum and three alternative layers of feedstock and inoculum where higher inoculum ratios in the top layers. From experimental studies it is observed that, partial premixing method with three alternative layers of feedstock and inoculum yielded good methane.

Keywords: anaerobic digestion, premixing methods, methane yield, corn stover, volatile solids

Procedia PDF Downloads 234
5957 The Impact of Water Reservoirs on Biodiversity and Food Security and the Creation of Adaptation Mechanisms

Authors: Inom S. Normatov, Abulqosim Muminov, Parviz I. Normatov

Abstract:

Problems of food security and the preservation of reserved zones in the region of Central Asia under the conditions of the climate change induced by the placement and construction of large reservoirs are considered. The criteria for the optimum placement and construction of reservoirs that entail the minimum impact on the environment are established. The need for the accounting of climatic parameters is shown by the calculation of the water quantity required for the irrigation of agricultural lands.

Keywords: adaptation, biodiversity, food security, water reservoir, risk

Procedia PDF Downloads 256
5956 Mitigation of Lithium-ion Battery Thermal Runaway Propagation Through the Use of Phase Change Materials Containing Expanded Graphite

Authors: Jayson Cheyne, David Butler, Iain Bomphray

Abstract:

In recent years, lithium-ion batteries have been used increasingly for electric vehicles and large energy storage systems due to their high-power density and long lifespan. Despite this, thermal runaway remains a significant safety problem because of its uncontrollable and irreversible nature - which can lead to fires and explosions. In large-scale lithium-ion packs and modules, thermal runaway propagation between cells can escalate fire hazards and cause significant damage. Thus, safety measures are required to mitigate thermal runaway propagation. The current research explores composite phase change materials (PCM) containing expanded graphite (EG) for thermal runaway mitigation. PCMs are an area of significant interest for battery thermal management due to their ability to absorb substantial quantities of heat during phase change. Moreover, the introduction of EG can support heat transfer from the cells to the PCM (owing to its high thermal conductivity) and provide shape stability to the PCM during phase change. During the research, a thermal model was established for an array of 16 cylindrical cells to simulate heat dissipation with and without the composite PCM. Two conditions were modeled, including the behavior during charge/discharge cycles (i.e., throughout regular operation) and thermal runaway. Furthermore, parameters including cell spacing, composite PCM thickness, and EG weight percentage (WT%) were varied to establish the optimal material parameters for enabling thermal runaway mitigation and effective thermal management. Although numerical modeling is still ongoing, initial findings suggest that a 3mm PCM containing 15WT% EG can effectively suppress thermal runaway propagation while maintaining shape stability. The next step in the research is to validate the model through controlled experimental tests. Additionally, with the perceived fire safety concerns relating to PCM materials, fire safety tests, including UL-94 and Limiting Oxygen Index (LOI), shall be conducted to explore the flammability risk.

Keywords: battery safety, electric vehicles, phase change materials, thermal management, thermal runaway

Procedia PDF Downloads 144
5955 Smartphone-Based Human Activity Recognition by Machine Learning Methods

Authors: Yanting Cao, Kazumitsu Nawata

Abstract:

As smartphones upgrading, their software and hardware are getting smarter, so the smartphone-based human activity recognition will be described as more refined, complex, and detailed. In this context, we analyzed a set of experimental data obtained by observing and measuring 30 volunteers with six activities of daily living (ADL). Due to the large sample size, especially a 561-feature vector with time and frequency domain variables, cleaning these intractable features and training a proper model becomes extremely challenging. After a series of feature selection and parameters adjustment, a well-performed SVM classifier has been trained.

Keywords: smart sensors, human activity recognition, artificial intelligence, SVM

Procedia PDF Downloads 144
5954 Adsorption and Desorption Behavior of Ionic and Nonionic Surfactants on Polymer Surfaces

Authors: Giulia Magi Meconi, Nicholas Ballard, José M. Asua, Ronen Zangi

Abstract:

Experimental and computational studies are combined to elucidate the adsorption proprieties of ionic and nonionic surfactants on hydrophobic polymer surface such us poly(styrene). To present these two types of surfactants, sodium dodecyl sulfate and poly(ethylene glycol)-block-poly(ethylene), commonly utilized in emulsion polymerization, are chosen. By applying quartz crystal microbalance with dissipation monitoring it is found that, at low surfactant concentrations, it is easier to desorb (as measured by rate) ionic surfactants than nonionic surfactants. From molecular dynamics simulations, the effective, attractive force of these nonionic surfactants to the surface increases with the decrease of their concentration, whereas, the ionic surfactant exhibits mildly the opposite trend. The contrasting behavior of ionic and nonionic surfactants critically relies on two observations obtained from the simulations. The first is that there is a large degree of interweavement between head and tails groups in the adsorbed layer formed by the nonionic surfactant (PEO/PE systems). The second is that water molecules penetrate this layer. In the disordered layer, these nonionic surfactants generate at the surface, only oxygens of the head groups present at the interface with the water phase or oxygens next to the penetrating waters can form hydrogen bonds. Oxygens inside this layer lose this favorable energy, with a magnitude that increases with the surfactants density at the interface. This reduced stability of the surfactants diminishes their driving force for adsorption. All that is shown to be in accordance with experimental results on the dynamics of surfactants desorption. Ionic surfactants assemble into an ordered structure and the attraction to the surface was even slightly augmented at higher surfactant concentration, in agreement with the experimentally determined adsorption isotherm. The reason these two types of surfactants behave differently is because the ionic surfactant has a small head group that is strongly hydrophilic, whereas the head groups of the nonionic surfactants are large and only weakly attracted to water.

Keywords: emulsion polymerization process, molecular dynamics simulations, polymer surface, surfactants adsorption

Procedia PDF Downloads 343
5953 The Formation of the Diminutive in Colloquial Jordanian Arabic

Authors: Yousef Barahmeh

Abstract:

This paper is a linguistic and pragmatic analysis of the use of the diminutive in Colloquial Jordanian Arabic (CJA). It demonstrates a peculiar form of the diminutive in CJA inflected by means of feminine plural ends with -aat suffix. The analysis shows that the pragmatic function(s) of the diminutive in CJA refers primarily to ‘littleness’ while the morphological inflection conveys the message of ‘the plethora’. Examples of this linguistic phenomenon are intelligible and often include a large number of words that are culture-specific to the rural dialect in the north of Jordan. In both cases, the diminutive in CJA is an adaptive strategy relative to its pragmatic and social contexts.

Keywords: Colloquial Jordanian Arabic, diminutive, morphology, pragmatics

Procedia PDF Downloads 266
5952 Digital Twin for University Campus: Workflow, Applications and Benefits

Authors: Frederico Fialho Teixeira, Islam Mashaly, Maryam Shafiei, Jurij Karlovsek

Abstract:

The ubiquity of data gathering and smart technologies, advancements in virtual technologies, and the development of the internet of things (IoT) have created urgent demands for the development of frameworks and efficient workflows for data collection, visualisation, and analysis. Digital twin, in different scales of the city into the building, allows for bringing together data from different sources to generate fundamental and illuminating insights for the management of current facilities and the lifecycle of amenities as well as improvement of the performance of current and future designs. Over the past two decades, there has been growing interest in the topic of digital twin and their applications in city and building scales. Most such studies look at the urban environment through a homogeneous or generalist lens and lack specificity in particular characteristics or identities, which define an urban university campus. Bridging this knowledge gap, this paper offers a framework for developing a digital twin for a university campus that, with some modifications, could provide insights for any large-scale digital twin settings like towns and cities. It showcases how currently unused data could be purposefully combined, interpolated and visualised for producing analysis-ready data (such as flood or energy simulations or functional and occupancy maps), highlighting the potential applications of such a framework for campus planning and policymaking. The research integrates campus-level data layers into one spatial information repository and casts light on critical data clusters for the digital twin at the campus level. The paper also seeks to raise insightful and directive questions on how digital twin for campus can be extrapolated to city-scale digital twin. The outcomes of the paper, thus, inform future projects for the development of large-scale digital twin as well as urban and architectural researchers on potential applications of digital twin in future design, management, and sustainable planning, to predict problems, calculate risks, decrease management costs, and improve performance.

Keywords: digital twin, smart campus, framework, data collection, point cloud

Procedia PDF Downloads 68
5951 Big Data and Health: An Australian Perspective Which Highlights the Importance of Data Linkage to Support Health Research at a National Level

Authors: James Semmens, James Boyd, Anna Ferrante, Katrina Spilsbury, Sean Randall, Adrian Brown

Abstract:

‘Big data’ is a relatively new concept that describes data so large and complex that it exceeds the storage or computing capacity of most systems to perform timely and accurate analyses. Health services generate large amounts of data from a wide variety of sources such as administrative records, electronic health records, health insurance claims, and even smart phone health applications. Health data is viewed in Australia and internationally as highly sensitive. Strict ethical requirements must be met for the use of health data to support health research. These requirements differ markedly from those imposed on data use from industry or other government sectors and may have the impact of reducing the capacity of health data to be incorporated into the real time demands of the Big Data environment. This ‘big data revolution’ is increasingly supported by national governments, who have invested significant funds into initiatives designed to develop and capitalize on big data and methods for data integration using record linkage. The benefits to health following research using linked administrative data are recognised internationally and by the Australian Government through the National Collaborative Research Infrastructure Strategy Roadmap, which outlined a multi-million dollar investment strategy to develop national record linkage capabilities. This led to the establishment of the Population Health Research Network (PHRN) to coordinate and champion this initiative. The purpose of the PHRN was to establish record linkage units in all Australian states, to support the implementation of secure data delivery and remote access laboratories for researchers, and to develop the Centre for Data Linkage for the linkage of national and cross-jurisdictional data. The Centre for Data Linkage has been established within Curtin University in Western Australia; it provides essential record linkage infrastructure necessary for large-scale, cross-jurisdictional linkage of health related data in Australia and uses a best practice ‘separation principle’ to support data privacy and security. Privacy preserving record linkage technology is also being developed to link records without the use of names to overcome important legal and privacy constraint. This paper will present the findings of the first ‘Proof of Concept’ project selected to demonstrate the effectiveness of increased record linkage capacity in supporting nationally significant health research. This project explored how cross-jurisdictional linkage can inform the nature and extent of cross-border hospital use and hospital-related deaths. The technical challenges associated with national record linkage, and the extent of cross-border population movements, were explored as part of this pioneering research project. Access to person-level data linked across jurisdictions identified geographical hot spots of cross border hospital use and hospital-related deaths in Australia. This has implications for planning of health service delivery and for longitudinal follow-up studies, particularly those involving mobile populations.

Keywords: data integration, data linkage, health planning, health services research

Procedia PDF Downloads 216
5950 Simulation of the Large Hadrons Collisions Using Monte Carlo Tools

Authors: E. Al Daoud

Abstract:

In many cases, theoretical treatments are available for models for which there is no perfect physical realization. In this situation, the only possible test for an approximate theoretical solution is to compare with data generated from a computer simulation. In this paper, Monte Carlo tools are used to study and compare the elementary particles models. All the experiments are implemented using 10000 events, and the simulated energy is 13 TeV. The mean and the curves of several variables are calculated for each model using MadAnalysis 5. Anomalies in the results can be seen in the muons masses of the minimal supersymmetric standard model and the two Higgs doublet model.

Keywords: Feynman rules, hadrons, Lagrangian, Monte Carlo, simulation

Procedia PDF Downloads 319