Search results for: data interpolating empirical orthogonal function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 29463

Search results for: data interpolating empirical orthogonal function

24903 Experimental Investigation of Mechanical Friction Influence in Semi-Hydraulic Clutch Actuation System Over Mileage

Authors: Abdul Azarrudin M. A., Pothiraj K., Kandasamy Satish

Abstract:

In the current automobile scenario, there comes a demand on more sophistication and comfort drive feel on passenger segments. The clutch pedal effort is one such customer touch feels in manual transmission vehicles, where the driver continuous to operate the clutch pedal in his entire the driving maneuvers. Hence optimum pedal efforts at green condition and over mileage to be ensured for fatigue free the driving. As friction is one the predominant factor and its tendency to challenge the technicality by causing the function degradation. One such semi-hydraulic systems shows load efficiency of about 70-75% over lifetime only due to the increase in friction which leads to the increase in pedal effort and cause fatigue to the vehicle driver. This work deals with the study of friction with different interfaces and its influence in the fulcrum points over mileage, with the objective of understanding the trend over mileage and determining the alternative ways of resolving it. In that one way of methodology is the reduction of friction by experimental investigation of various friction reduction interfaces like metal-to-metal interface and it has been tried out and is detailed further. Also, the specific attention has been put up considering the fulcrum load and its contact interfaces to move on with this study. The main results of the experimental data with the influence of three different contact interfaces are being presented with an ultimate intention of ending up into less fatigue with longer consistent pedal effort, thus smoothens the operation of the end user. The Experimental validation also has been done through rig-level test setup to depict the performance at static condition and in-parallel vehicle level test has also been performed to record the additional influences if any.

Keywords: automobile, clutch, friction, fork

Procedia PDF Downloads 107
24902 Reflecting on Deafblindness: Recommendations for Implementing Effective Strategies

Authors: V. Argyropoulos, M. Nikolaraizi, K. Tanou

Abstract:

There is little available information concerning the cognitive and communicative abilities of the people who are deaf-blind. This mainly stems from the general inadequacy of existing assessment instruments employed with deafblind individuals. Although considerable variability exists with regard to cognitive capacities of the deaf-blind, careful examination of the literature reveals that the majority of these persons suffer from significant deficits in cognitive and adaptive functioning. The few reports available primarily are case studies, narrative program descriptions, or position papers by workers in the field. Without the objective verification afforded by controlled research, specialists in psychology, education, and other rehabilitation services must rely on personal speculations or biases to guide their decisions in the planning, implementation, and evaluation of services to deaf-blind children and adults. This paper highlights the framework and discusses the results of an action research network. The aim of this study was twofold: a) to describe and analyse the different ways in which a student with deafblindness approached a number of developmental issues such as novel tasks, exploration and manipulation of objects, reactions to social stimuli, motor coordination, and quality of play and b) to map the appropriate functional approach for the specific student that could be used to develop strategies for classroom participation and socialization. The persons involved in this collaborative action research scheme were general teachers, a school counsellor, academic staff and student teachers. Rating scales and checklists were used to gather information in natural activities and settings, and additional data were also obtained through interviews with the educators of the student. The findings of this case study indicated that there is a great need to focus on the development of effective intervention strategies. The results showed that the identification of positive reinforcers for this population might represent an important and challenging aspect of behaviour programmes. Finally, the findings suggest that additional empirical work is needed to increase attention to methodological and social validity issues.

Keywords: action research, cognitive and communicative abilities, deafblindness, effective strategies

Procedia PDF Downloads 173
24901 Biophysical Characterization of Archaeal Cyclophilin Like Chaperone Protein

Authors: Vineeta Kaushik, Manisha Goel

Abstract:

Chaperones are proteins that help other proteins fold correctly, and are found in all domains of life i.e., prokaryotes, eukaryotes and archaea. Various comparative genomic studies have suggested that the archaeal protein folding machinery appears to be highly similar to that found in eukaryotes. In case of protein folding; slow rotation of peptide prolyl-imide bond is often the rate limiting step. Formation of the prolyl-imide bond during the folding of a protein requires the assistance of other proteins, termed as peptide prolyl cis-trans isomerases (PPIases). Cyclophilins constitute the class of peptide prolyl isomerases with a wide range of biological function like protein folding, signaling and chaperoning. Most of the cyclophilins exhibit PPIase enzymatic activity and play active role in substrate protein folding which classifies them as a category of molecular chaperones. Till date, there is not very much data available in the literature on archaeal cyclophilins. We aim to compare the structural and biochemical features of the cyclophilin protein from within the three domains to elucidate the features affecting their stability and enzyme activity. In the present study, we carry out in-silico analysis of the cyclophilin proteins to predict their conserved residues, sites under positive selection and compare these proteins to their bacterial and eukaryotic counterparts to predict functional divergence. We also aim to clone and express these proteins in heterologous system and study their biophysical characteristics in detail using techniques like CD and fluorescence spectroscopy. Overall we aim to understand the features contributing to the folding, stability and dynamics of the archaeal cyclophilin proteins.

Keywords: biophysical characterization, x-ray crystallography, chaperone-like activity, cyclophilin, PPIase activity

Procedia PDF Downloads 201
24900 Evaluation of Classification Algorithms for Diagnosis of Asthma in Iranian Patients

Authors: Taha SamadSoltani, Peyman Rezaei Hachesu, Marjan GhaziSaeedi, Maryam Zolnoori

Abstract:

Introduction: Data mining defined as a process to find patterns and relationships along data in the database to build predictive models. Application of data mining extended in vast sectors such as the healthcare services. Medical data mining aims to solve real-world problems in the diagnosis and treatment of diseases. This method applies various techniques and algorithms which have different accuracy and precision. The purpose of this study was to apply knowledge discovery and data mining techniques for the diagnosis of asthma based on patient symptoms and history. Method: Data mining includes several steps and decisions should be made by the user which starts by creation of an understanding of the scope and application of previous knowledge in this area and identifying KD process from the point of view of the stakeholders and finished by acting on discovered knowledge using knowledge conducting, integrating knowledge with other systems and knowledge documenting and reporting.in this study a stepwise methodology followed to achieve a logical outcome. Results: Sensitivity, Specifity and Accuracy of KNN, SVM, Naïve bayes, NN, Classification tree and CN2 algorithms and related similar studies was evaluated and ROC curves were plotted to show the performance of the system. Conclusion: The results show that we can accurately diagnose asthma, approximately ninety percent, based on the demographical and clinical data. The study also showed that the methods based on pattern discovery and data mining have a higher sensitivity compared to expert and knowledge-based systems. On the other hand, medical guidelines and evidence-based medicine should be base of diagnostics methods, therefore recommended to machine learning algorithms used in combination with knowledge-based algorithms.

Keywords: asthma, datamining, classification, machine learning

Procedia PDF Downloads 434
24899 Application of GPRS in Water Quality Monitoring System

Authors: V. Ayishwarya Bharathi, S. M. Hasker, J. Indhu, M. Mohamed Azarudeen, G. Gowthami, R. Vinoth Rajan, N. Vijayarangan

Abstract:

Identification of water quality conditions in a river system based on limited observations is an essential task for meeting the goals of environmental management. The traditional method of water quality testing is to collect samples manually and then send to laboratory for analysis. However, it has been unable to meet the demands of water quality monitoring today. So a set of automatic measurement and reporting system of water quality has been developed. In this project specifies Water quality parameters collected by multi-parameter water quality probe are transmitted to data processing and monitoring center through GPRS wireless communication network of mobile. The multi parameter sensor is directly placed above the water level. The monitoring center consists of GPRS and micro-controller which monitor the data. The collected data can be monitor at any instant of time. In the pollution control board they will monitor the water quality sensor data in computer using Visual Basic Software. The system collects, transmits and processes water quality parameters automatically, so production efficiency and economy benefit are improved greatly. GPRS technology can achieve well within the complex environment of poor water quality non-monitored, and more specifically applicable to the collection point, data transmission automatically generate the field of water analysis equipment data transmission and monitoring.

Keywords: multiparameter sensor, GPRS, visual basic software, RS232

Procedia PDF Downloads 392
24898 Decision Support System in Air Pollution Using Data Mining

Authors: E. Fathallahi Aghdam, V. Hosseini

Abstract:

Environmental pollution is not limited to a specific region or country; that is why sustainable development, as a necessary process for improvement, pays attention to issues such as destruction of natural resources, degradation of biological system, global pollution, and climate change in the world, especially in the developing countries. According to the World Health Organization, as a developing city, Tehran (capital of Iran) is one of the most polluted cities in the world in terms of air pollution. In this study, three pollutants including particulate matter less than 10 microns, nitrogen oxides, and sulfur dioxide were evaluated in Tehran using data mining techniques and through Crisp approach. The data from 21 air pollution measuring stations in different areas of Tehran were collected from 1999 to 2013. Commercial softwares Clementine was selected for this study. Tehran was divided into distinct clusters in terms of the mentioned pollutants using the software. As a data mining technique, clustering is usually used as a prologue for other analyses, therefore, the similarity of clusters was evaluated in this study through analyzing local conditions, traffic behavior, and industrial activities. In fact, the results of this research can support decision-making system, help managers improve the performance and decision making, and assist in urban studies.

Keywords: data mining, clustering, air pollution, crisp approach

Procedia PDF Downloads 417
24897 Quality of Life in People with Hearing Loss: A Study of Patients Referred to an Audiological Service

Authors: Peder O. Laugen Heggdal, Oyvind Nordvik, Jonas Brannstrom, Flemming Vassbotn, Anne Kari Aarstad, Hans Jorgen Aarstad

Abstract:

Background: Hearing loss (HL) affect people of all ages and stages in life. To author's best knowledge, if patients with an HL have reduced Generic Quality of life (QoL), has yet not been answered. Aim: The aim of the present study was to investigate the relationship between HL and generic and disease-specific Health Related Quality of Life (HRQoL) in adult patients (aged 18–78 years) with an HL, seeking Hearing Aid (HA). Material and Methods: 158 adult (aged 18-78 years) patients with HL, referred for HA fitting at Haukeland University Hospital in western Norway, participated in the study. Both first-time users, as well as patients referred for HA renewals, were included. First-time users had been pre-examined by an Ear Nose and Throat specialist. The questionnaires were answered before the actual HA fitting procedure. The pure-tone average (PTA; frequencies 0.5, 1, 2 and 4 kHz) was determined for each ear. The generic European Organization for Research and Treatment of Cancer (EORTC) Quality of Life Questionnaire general part and a shortened version of the Abbreviated Profile of Hearing Aid Benefit (APHAB) were answered. In addition, EORTC HRQoL answers from a general population and patients with former head and neck cancer served as comparison. Results: In general, no lowered HRQoL scores were determined among HL patients compared to the general population. Patients with unilateral HL to some extent showed lower HRQoL than those with bilateral HL (social function and fatigue). The APHAB scores correlated significantly with the EORTC HRQoL scores. By stepwise linear regression analysis, the APHAB scores were scored secondary to PTA (best ear), cognitive and physical function. Conclusion: HRQoL scores in HL patients, in general, seems to be at the population level, but the unilateral HL patients scored to some extent lower than the bilateral HI patients. APHAB and generic QoL scores levels are associated. Both HRQoL and APHAB scores are generated more complexly than anticipated.

Keywords: quality of life, hearing loss, hearing impairment, distress, depression, anxiety, hearing aid

Procedia PDF Downloads 280
24896 Test Suite Optimization Using an Effective Meta-Heuristic BAT Algorithm

Authors: Anuradha Chug, Sunali Gandhi

Abstract:

Regression Testing is a very expensive and time-consuming process carried out to ensure the validity of modified software. Due to the availability of insufficient resources to re-execute all the test cases in time constrained environment, efforts are going on to generate test data automatically without human efforts. Many search based techniques have been proposed to generate efficient, effective as well as optimized test data, so that the overall cost of the software testing can be minimized. The generated test data should be able to uncover all potential lapses that exist in the software or product. Inspired from the natural behavior of bat for searching her food sources, current study employed a meta-heuristic, search-based bat algorithm for optimizing the test data on the basis certain parameters without compromising their effectiveness. Mathematical functions are also applied that can effectively filter out the redundant test data. As many as 50 Java programs are used to check the effectiveness of proposed test data generation and it has been found that 86% saving in testing efforts can be achieved using bat algorithm while covering 100% of the software code for testing. Bat algorithm was found to be more efficient in terms of simplicity and flexibility when the results were compared with another nature inspired algorithms such as Firefly Algorithm (FA), Hill Climbing Algorithm (HC) and Ant Colony Optimization (ACO). The output of this study would be useful to testers as they can achieve 100% path coverage for testing with minimum number of test cases.

Keywords: regression testing, test case selection, test case prioritization, genetic algorithm, bat algorithm

Procedia PDF Downloads 359
24895 The Grand Unified Theory of Everything as a Generalization to the Standard Model Called as the General Standard Model

Authors: Amir Deljoo

Abstract:

The endeavor to comprehend the existence have been the center of thought for human in form of different disciplines and now basically in physics as the theory of everything. Here, after a brief review of the basic frameworks of thought, and a history of thought since ancient up to present, a logical methodology is presented based on a core axiom after which a function, a proto-field and then a coordinates are explained. Afterwards a generalization to Standard Model is proposed as General Standard Model which is believed to be the base of the Unified Theory of Everything.

Keywords: general relativity, grand unified theory, quantum mechanics, standard model, theory of everything

Procedia PDF Downloads 88
24894 Toward Discovering an Architectural Typology Based on the Theory of Affordance

Authors: Falntina Ahmad Alata, Natheer Abu Obeid

Abstract:

This paper revolves around the concept of affordance. It aims to discover and develop an architectural typology based on the ecological concept of affordance. In order to achieve this aim, an analytical study is conducted and two sources were taken into account: 1- Gibson's definition of the concept of affordance and 2- The researches that are concerned on the affordance categorisation. As a result, this paper concluded 16 typologies of affordances, including the possibilities of mixing them based on both sources. To clarify these typologies and provide further understanding, a wide range of architectural examples are presented and proposed in the paper. To prove this vocabulary’s capability to diagnose and evaluate the affordance of different environments, an experimental study with two processes have been adapted: 1. Diagnostic process: the interpretation of the environments with regards to its affordance by using the new vocabulary (the developed typologies). 2. Evaluating process: the evaluation of the environments that have been interpreted and classified with regards to their affordances. By using the measures of emotional experience (the positive affect ‘PA’ and the negative affect ‘NA’) and the architectural evaluation criteria (beauty, economy and function). The experimental study proves that the typologies are capable of reading the affordance within different environments. Additionally, it explains how these different typologies reflect different interactions based on the previous processes. The data which are concluded from the evaluation of measures explain how different typologies of affordance that have already reflected different environments had different evaluations. In fact, some of them are recommended while the others are not. In other words, the paper draws a roadmap for designers to diagnose, evaluate and analyse the affordance into different architectural environments. After that, it guides them through adapting the best interaction (affordance category), which they intend to adapt into their proposed designs.

Keywords: Affordance theory, Affordance categories, Architectural environments, Architectural Evaluation Criteria (AEC), emotional experience (PA, NA)

Procedia PDF Downloads 158
24893 Modified InVEST for Whatsapp Messages Forensic Triage and Search through Visualization

Authors: Agria Rhamdhan

Abstract:

WhatsApp as the most popular mobile messaging app has been used as evidence in many criminal cases. As the use of mobile messages generates large amounts of data, forensic investigation faces the challenge of large data problems. The hardest part of finding this important evidence is because current practice utilizes tools and technique that require manual analysis to check all messages. That way, analyze large sets of mobile messaging data will take a lot of time and effort. Our work offers methodologies based on forensic triage to reduce large data to manageable sets resulting easier to do detailed reviews, then show the results through interactive visualization to show important term, entities and relationship through intelligent ranking using Term Frequency-Inverse Document Frequency (TF-IDF) and Latent Dirichlet Allocation (LDA) Model. By implementing this methodology, investigators can improve investigation processing time and result's accuracy.

Keywords: forensics, triage, visualization, WhatsApp

Procedia PDF Downloads 161
24892 Low Cost Webcam Camera and GNSS Integration for Updating Home Data Using AI Principles

Authors: Mohkammad Nur Cahyadi, Hepi Hapsari Handayani, Agus Budi Raharjo, Ronny Mardianto, Daud Wahyu Imani, Arizal Bawazir, Luki Adi Triawan

Abstract:

PDAM (local water company) determines customer charges by considering the customer's building or house. Charges determination significantly affects PDAM income and customer costs because the PDAM applies a subsidy policy for customers classified as small households. Periodic updates are needed so that pricing is in line with the target. A thorough customer survey in Surabaya is needed to update customer building data. However, the survey that has been carried out so far has been by deploying officers to conduct one-by-one surveys for each PDAM customer. Surveys with this method require a lot of effort and cost. For this reason, this research offers a technology called moblie mapping, a mapping method that is more efficient in terms of time and cost. The use of this tool is also quite simple, where the device will be installed in the car so that it can record the surrounding buildings while the car is running. Mobile mapping technology generally uses lidar sensors equipped with GNSS, but this technology requires high costs. In overcoming this problem, this research develops low-cost mobile mapping technology using a webcam camera sensor added to the GNSS and IMU sensors. The camera used has specifications of 3MP with a resolution of 720 and a diagonal field of view of 78⁰. The principle of this invention is to integrate four camera sensors, a GNSS webcam, and GPS to acquire photo data, which is equipped with location data (latitude, longitude) and IMU (roll, pitch, yaw). This device is also equipped with a tripod and a vacuum cleaner to attach to the car's roof so it doesn't fall off while running. The output data from this technology will be analyzed with artificial intelligence to reduce similar data (Cosine Similarity) and then classify building types. Data reduction is used to eliminate similar data and maintain the image that displays the complete house so that it can be processed for later classification of buildings. The AI method used is transfer learning by utilizing a trained model named VGG-16. From the analysis of similarity data, it was found that the data reduction reached 50%. Then georeferencing is done using the Google Maps API to get address information according to the coordinates in the data. After that, geographic join is done to link survey data with customer data already owned by PDAM Surya Sembada Surabaya.

Keywords: mobile mapping, GNSS, IMU, similarity, classification

Procedia PDF Downloads 69
24891 Electromagnetic Tuned Mass Damper Approach for Regenerative Suspension

Authors: S. Kopylov, C. Z. Bo

Abstract:

This study is aimed at exploring the possibility of energy recovery through the suppression of vibrations. The article describes design of electromagnetic dynamic damper. The magnetic part of the device performs the function of a tuned mass damper, thereby providing both energy regeneration and damping properties to the protected mass. According to the theory of tuned mass damper, equations of mathematical models were obtained. Then, under given properties of current system, amplitude frequency response was investigated. Therefore, main ideas and methods for further research were defined.

Keywords: electromagnetic damper, oscillations with two degrees of freedom, regeneration systems, tuned mass damper

Procedia PDF Downloads 201
24890 An Investigation into the Views of Distant Science Education Students Regarding Teaching Laboratory Work Online

Authors: Abraham Motlhabane

Abstract:

This research analysed the written views of science education students regarding the teaching of laboratory work using the online mode. The research adopted the qualitative methodology. The qualitative research was aimed at investigating small and distinct groups normally regarded as a single-site study. Qualitative research was used to describe and analyze the phenomena from the student’s perspective. This means the research began with assumptions of the world view that use theoretical lenses of research problems inquiring into the meaning of individual students. The research was conducted with three groups of students studying for Postgraduate Certificate in Education, Bachelor of Education and honors Bachelor of Education respectively. In each of the study programmes, the science education module is compulsory. Five science education students from each study programme were purposively selected to participate in this research. Therefore, 15 students participated in the research. In order to analysis the data, the data were first printed and hard copies were used in the analysis. The data was read several times and key concepts and ideas were highlighted. Themes and patterns were identified to describe the data. Coding as a process of organising and sorting data was used. The findings of the study are very diverse; some students are in favour of online laboratory whereas other students argue that science can only be learnt through hands-on experimentation.

Keywords: online learning, laboratory work, views, perceptions

Procedia PDF Downloads 127
24889 Nature of Science in Physics Textbooks – Example of Quebec Province

Authors: Brahim El Fadil

Abstract:

The nature of science as a solution (NOS) to life problems is well established in school activities the world over. However, this study reveals the lack of representation of the NOS in science textbooks used in Quebec Province. A content analysis method was adopted to analyze the NOS in relation to optics knowledge and teaching-learning activities in Grade 9 science and technology textbooks and Grade 11 physics textbooks. The selected textbooks were approved and authorized by the Provincial Ministry of Education. Our analysis points out that most of these editions provided a poor representation of NOS. None of them indicates that scientific knowledge is subject to change, even though the history of optics reveals evolutionary and revolutionary changes. Moreover, the analysis shows that textbooks place little emphasis on the discussion of scientific laws and theories. Few of them argue that scientific inquiries are required to gain a deep understanding of scientific concepts. Moreover, they rarely present empirical evidence to support their arguments.

Keywords: nature of science, history of optics, geometrical theory of optics, wave theory of optics

Procedia PDF Downloads 67
24888 Improvement of Cardiometabolic after 8 Weeks of Weight Loss Intervention

Authors: Boris Bajer, Andrea Havranova, Miroslav Vlcek, Richard Imrich, Adela Penesova

Abstract:

Lifestyle interventions can prevent the deterioration of impaired glucose tolerance to manifest type 2 diabetes, and also prevent cardiovascular diseases, as it showed many studies (the Finnish Diabetes Prevention Study, Diabetes Prevention Program (DPP), . the China Da Qing Diabetes Prevention Study, etc.) Therefore the aim of our study was to compare the effect of intensified lifestyle intervention on cardiometabolic parameters. Methods: It is an ongoing randomized interventional clinical study (NCT02325804) focused on the reduction of body weight/fat. Intervention: hypocaloric diet (30% restriction of calories) and physical activity 150 minutes/week. Before and after 8 weeks of intervention all patients underwent complete medical examination (measurement of physical fitness, resting metabolic rate (RMR), body composition analysis, oral glucose tolerance test, parameters of lipid metabolism, and other cardiometabolic risk factors. Results: So far 39 patients finished the intervention. The average reduction of body weight was 6,8 + 4,9 kg (0-15 kg; p=0,0006), accompanied with significant reduction of body fat percentage (p ≤ 0,0001), amount of fat mass (p=0,03), waist circumference (p=0.02). Amount of lean mass and RMR remained unchanged. Heart rate (p=0,02), systolic and diastolic blood pressure was reduced (p=0,01 p=0,02 resp.) as well as insulin sensitivity was improved. Lipid parameters also changed - cholesterol, LDL decreased (p=0,05, p=0,04 resp.), while triglycerides showed tendency to decrease (p=0,055). Liver function improved, alanine aminotrasnferase (ALT) were reduced (p=0,01). Physical fitness significantly improved (as measure VO2 max (p=0,02). Conclusion: Results of our study are in line with previous results about the beneficial effect of intensive lifestyle changes on the reduction of cardiometabolic risk factors and improvement of liver function. Supported by grants APVV 15-0228; VEGA 2/0161/16

Keywords: obesity, weight loss, diet lipids, blood pressure, liver enzymes

Procedia PDF Downloads 154
24887 The Communication Library DIALOG for iFDAQ of the COMPASS Experiment

Authors: Y. Bai, M. Bodlak, V. Frolov, S. Huber, V. Jary, I. Konorov, D. Levit, J. Novy, D. Steffen, O. Subrt, M. Virius

Abstract:

Modern experiments in high energy physics impose great demands on the reliability, the efficiency, and the data rate of Data Acquisition Systems (DAQ). This contribution focuses on the development and deployment of the new communication library DIALOG for the intelligent, FPGA-based Data Acquisition System (iFDAQ) of the COMPASS experiment at CERN. The iFDAQ utilizing a hardware event builder is designed to be able to readout data at the maximum rate of the experiment. The DIALOG library is a communication system both for distributed and mixed environments, it provides a network transparent inter-process communication layer. Using the high-performance and modern C++ framework Qt and its Qt Network API, the DIALOG library presents an alternative to the previously used DIM library. The DIALOG library was fully incorporated to all processes in the iFDAQ during the run 2016. From the software point of view, it might be considered as a significant improvement of iFDAQ in comparison with the previous run. To extend the possibilities of debugging, the online monitoring of communication among processes via DIALOG GUI is a desirable feature. In the paper, we present the DIALOG library from several insights and discuss it in a detailed way. Moreover, the efficiency measurement and comparison with the DIM library with respect to the iFDAQ requirements is provided.

Keywords: data acquisition system, DIALOG library, DIM library, FPGA, Qt framework, TCP/IP

Procedia PDF Downloads 309
24886 Mining Scientific Literature to Discover Potential Research Data Sources: An Exploratory Study in the Field of Haemato-Oncology

Authors: A. Anastasiou, K. S. Tingay

Abstract:

Background: Discovering suitable datasets is an important part of health research, particularly for projects working with clinical data from patients organized in cohorts (cohort data), but with the proliferation of so many national and international initiatives, it is becoming increasingly difficult for research teams to locate real world datasets that are most relevant to their project objectives. We present a method for identifying healthcare institutes in the European Union (EU) which may hold haemato-oncology (HO) data. A key enabler of this research was the bibInsight platform, a scientometric data management and analysis system developed by the authors at Swansea University. Method: A PubMed search was conducted using HO clinical terms taken from previous work. The resulting XML file was processed using the bibInsight platform, linking affiliations to the Global Research Identifier Database (GRID). GRID is an international, standardized list of institutions, including the city and country in which the institution exists, as well as a category of the main business type, e.g., Academic, Healthcare, Government, Company. Countries were limited to the 28 current EU members, and institute type to 'Healthcare'. An article was considered valid if at least one author was affiliated with an EU-based healthcare institute. Results: The PubMed search produced 21,310 articles, consisting of 9,885 distinct affiliations with correspondence in GRID. Of these articles, 760 were from EU countries, and 390 of these were healthcare institutes. One affiliation was excluded as being a veterinary hospital. Two EU countries did not have any publications in our analysis dataset. The results were analysed by country and by individual healthcare institute. Networks both within the EU and internationally show institutional collaborations, which may suggest a willingness to share data for research purposes. Geographical mapping can ensure that data has broad population coverage. Collaborations with industry or government may exclude healthcare institutes that may have embargos or additional costs associated with data access. Conclusions: Data reuse is becoming increasingly important both for ensuring the validity of results, and economy of available resources. The ability to identify potential, specific data sources from over twenty thousand articles in less than an hour could assist in improving knowledge of, and access to, data sources. As our method has not yet specified if these healthcare institutes are holding data, or merely publishing on that topic, future work will involve text mining of data-specific concordant terms to identify numbers of participants, demographics, study methodologies, and sub-topics of interest.

Keywords: data reuse, data discovery, data linkage, journal articles, text mining

Procedia PDF Downloads 104
24885 Household Earthquake Absorptive Capacity Impact on Food Security: A Case Study in Rural Costa Rica

Authors: Laura Rodríguez Amaya

Abstract:

The impact of natural disasters on food security can be devastating, especially in rural settings where livelihoods are closely tied to their productive assets. In hazards studies, absorptive capacity is seen as a threshold that impacts the degree of people’s recovery after a natural disaster. Increasing our understanding of households’ capacity to absorb natural disaster shocks can provide the international community with viable measurements for assessing at-risk communities’ resilience to food insecurities. The purpose of this study is to identify the most important factors in determining a household’s capacity to absorb the impact of a natural disaster. This is an empirical study conducted in six communities in Costa Rica affected by earthquakes. The Earthquake Impact Index was developed for the selection of the communities in this study. The households coded as total loss in the selected communities constituted the sampling frame from which the sample population was drawn. Because of the study area geographically dispersion over a large surface, the stratified clustered sampling hybrid technique was selected. Of the 302 households identified as total loss in the six communities, a total of 126 households were surveyed, constituting 42 percent of the sampling frame. A list of indicators compiled based on theoretical and exploratory grounds for the absorptive capacity construct served to guide the survey development. These indicators were included in the following variables: (1) use of informal safety nets, (2) Coping Strategy, (3) Physical Connectivity, and (4) Infrastructure Damage. A multivariate data analysis was conducted using Statistical Package for Social Sciences (SPSS). The results show that informal safety nets such as family and friends assistance exerted the greatest influence on the ability of households to absorb the impact of earthquakes. In conclusion, communities that experienced the highest environmental impact and human loss got disconnected from the social networks needed to absorb the shock’s impact. This resulted in higher levels of household food insecurity.

Keywords: absorptive capacity, earthquake, food security, rural

Procedia PDF Downloads 241
24884 Using Data Mining Technique for Scholarship Disbursement

Authors: J. K. Alhassan, S. A. Lawal

Abstract:

This work is on decision tree-based classification for the disbursement of scholarship. Tree-based data mining classification technique is used in other to determine the generic rule to be used to disburse the scholarship. The system based on the defined rules from the tree is able to determine the class (status) to which an applicant shall belong whether Granted or Not Granted. The applicants that fall to the class of granted denote a successful acquirement of scholarship while those in not granted class are unsuccessful in the scheme. An algorithm that can be used to classify the applicants based on the rules from tree-based classification was also developed. The tree-based classification is adopted because of its efficiency, effectiveness, and easy to comprehend features. The system was tested with the data of National Information Technology Development Agency (NITDA) Abuja, a Parastatal of Federal Ministry of Communication Technology that is mandated to develop and regulate information technology in Nigeria. The system was found working according to the specification. It is therefore recommended for all scholarship disbursement organizations.

Keywords: classification, data mining, decision tree, scholarship

Procedia PDF Downloads 355
24883 Public Debt Shocks and Public Goods Provisioning in Nigeria: Implication for National Development

Authors: Amenawo I. Offiong, Hodo B. Riman

Abstract:

Public debt profile of Nigeria has continuously been on the increase over the years. The drop in international crude oil prices has further worsened revenue position of the country, thus, necessitating further acquisition of public debt to bridge the gap in revenue deficit. Yet, when we look back at the increasing public sector spending, there are concerns that the government spending do not amount to increase in public goods provided for the country. Using data from 1980 to 2014 the study therefore seeks to investigate the factors responsible for the poor provision of public goods in the face of increasing public debt profile. Using the unrestricted VAR model Governance and Tax revenue were introduced into the model as structural variables. The result suggested that governance and tax revenue were structural determinants of the effectiveness of public goods provisioning in Nigeria. The study therefore identified weak governance as the major reason for the non-provision of public goods in Nigeria. While tax revenue exerted positive influence on the provisions of public goods, weak/poor governance was observed to crowd the benefits from increase tax revenue. The study therefore recommends reappraisal of the governance system in Nigeria. Elected officers in governance should be more transparent and accountable to the electorates they represent. Furthermore, the study advocates for an annual auditing of all government MDAs accounts by external auditors to ensure (a) accountability of public debts utilization, (b) transparent in implementation of program support funds, (c) integrity of agencies responsible for program management, and (d) measuring program effectiveness with amount of funds expended.

Keywords: impulse response function, public debt shocks, governance, public goods, tax revenue, vector auto-regression

Procedia PDF Downloads 246
24882 [Keynote Speech]: Feature Selection and Predictive Modeling of Housing Data Using Random Forest

Authors: Bharatendra Rai

Abstract:

Predictive data analysis and modeling involving machine learning techniques become challenging in presence of too many explanatory variables or features. Presence of too many features in machine learning is known to not only cause algorithms to slow down, but they can also lead to decrease in model prediction accuracy. This study involves housing dataset with 79 quantitative and qualitative features that describe various aspects people consider while buying a new house. Boruta algorithm that supports feature selection using a wrapper approach build around random forest is used in this study. This feature selection process leads to 49 confirmed features which are then used for developing predictive random forest models. The study also explores five different data partitioning ratios and their impact on model accuracy are captured using coefficient of determination (r-square) and root mean square error (rsme).

Keywords: housing data, feature selection, random forest, Boruta algorithm, root mean square error

Procedia PDF Downloads 307
24881 Towards Accurate Velocity Profile Models in Turbulent Open-Channel Flows: Improved Eddy Viscosity Formulation

Authors: W. Meron Mebrahtu, R. Absi

Abstract:

Velocity distribution in turbulent open-channel flows is organized in a complex manner. This is due to the large spatial and temporal variability of fluid motion resulting from the free-surface turbulent flow condition. This phenomenon is complicated further due to the complex geometry of channels and the presence of solids transported. Thus, several efforts were made to understand the phenomenon and obtain accurate mathematical models that are suitable for engineering applications. However, predictions are inaccurate because oversimplified assumptions are involved in modeling this complex phenomenon. Therefore, the aim of this work is to study velocity distribution profiles and obtain simple, more accurate, and predictive mathematical models. Particular focus will be made on the acceptable simplification of the general transport equations and an accurate representation of eddy viscosity. Wide rectangular open-channel seems suitable to begin the study; other assumptions are smooth-wall, and sediment-free flow under steady and uniform flow conditions. These assumptions will allow examining the effect of the bottom wall and the free surface only, which is a necessary step before dealing with more complex flow scenarios. For this flow condition, two ordinary differential equations are obtained for velocity profiles; from the Reynolds-averaged Navier-Stokes (RANS) equation and equilibrium consideration between turbulent kinetic energy (TKE) production and dissipation. Then different analytic models for eddy viscosity, TKE, and mixing length were assessed. Computation results for velocity profiles were compared to experimental data for different flow conditions and the well-known linear, log, and log-wake laws. Results show that the model based on the RANS equation provides more accurate velocity profiles. In the viscous sublayer and buffer layer, the method based on Prandtl’s eddy viscosity model and Van Driest mixing length give a more precise result. For the log layer and outer region, a mixing length equation derived from Von Karman’s similarity hypothesis provides the best agreement with measured data except near the free surface where an additional correction based on a damping function for eddy viscosity is used. This method allows more accurate velocity profiles with the same value of the damping coefficient that is valid under different flow conditions. This work continues with investigating narrow channels, complex geometries, and the effect of solids transported in sewers.

Keywords: accuracy, eddy viscosity, sewers, velocity profile

Procedia PDF Downloads 100
24880 Image-Based (RBG) Technique for Estimating Phosphorus Levels of Different Crops

Authors: M. M. Ali, Ahmed Al- Ani, Derek Eamus, Daniel K. Y. Tan

Abstract:

In this glasshouse study, we developed the new image-based non-destructive technique for detecting leaf P status of different crops such as cotton, tomato and lettuce. Plants were allowed to grow on nutrient media containing different P concentrations, i.e. 0%, 50% and 100% of recommended P concentration (P0 = no P, L; P1 = 2.5 mL 10 L-1 of P and P2 = 5 mL 10 L-1 of P as NaH2PO4). After 10 weeks of growth, plants were harvested and data on leaf P contents were collected using the standard destructive laboratory method and at the same time leaf images were collected by a handheld crop image sensor. We calculated leaf area, leaf perimeter and RGB (red, green and blue) values of these images. This data was further used in the linear discriminant analysis (LDA) to estimate leaf P contents, which successfully classified these plants on the basis of leaf P contents. The data indicated that P deficiency in crop plants can be predicted using the image and morphological data. Our proposed non-destructive imaging method is precise in estimating P requirements of different crop species.

Keywords: image-based techniques, leaf area, leaf P contents, linear discriminant analysis

Procedia PDF Downloads 362
24879 Correlation between Adherence to Islamic Principles of Success and Academic Achievement

Authors: Zuwaira Abubakar

Abstract:

Islam is the Divine religion which guides Man ways of leading a prosperous life in this life and the hereafter. This study was conducted in order to investigate the possible relationship between adherence to Islamic principles of success and academic performance of university students. Accordingly, a questionnaire based on Islamized principles of success (referred to as 'Islamic character quotient inventory (ICQi)') was correlated with CGPA (Cumulative Grade Point Averages) of 343 students of Usmanu Danfodiyo University Sokoto. The empirical testing indicates that the total score on ICQi correlated positively and significantly with academic performance of the respondent. Students with either high or medium adherence have a significantly (P<0.01) higher CGPA than their counterparts with the low-adherence level. However, the result did not show a significant relationship between the CGPA of highly adherent individuals and that of those with medium adherence level. This may suggests that Islam is not for spiritual life only but also relevant and useful for our practical life.

Keywords: academic, Islam, principles, success

Procedia PDF Downloads 222
24878 Study of the Use of Artificial Neural Networks in Islamic Finance

Authors: Kaoutar Abbahaddou, Mohammed Salah Chiadmi

Abstract:

The need to find a relevant way to predict the next-day price of a stock index is a real concern for many financial stakeholders and researchers. We have known across years the proliferation of several methods. Nevertheless, among all these methods, the most controversial one is a machine learning algorithm that claims to be reliable, namely neural networks. Thus, the purpose of this article is to study the prediction power of neural networks in the particular case of Islamic finance as it is an under-looked area. In this article, we will first briefly present a review of the literature regarding neural networks and Islamic finance. Next, we present the architecture and principles of artificial neural networks most commonly used in finance. Then, we will show its empirical application on two Islamic stock indexes. The accuracy rate would be used to measure the performance of the algorithm in predicting the right price the next day. As a result, we can conclude that artificial neural networks are a reliable method to predict the next-day price for Islamic indices as it is claimed for conventional ones.

Keywords: Islamic finance, stock price prediction, artificial neural networks, machine learning

Procedia PDF Downloads 214
24877 Design of Visual Repository, Constraint and Process Modeling Tool Based on Eclipse Plug-Ins

Authors: Rushiraj Heshi, Smriti Bhandari

Abstract:

Master Data Management requires creation of Central repository, applying constraints on Repository and designing processes to manage data. Designing of Repository, constraints on repository and business processes is very tedious and time consuming task for large Enterprise. Hence Visual Repository, constraints and Process (Workflow) modeling is the most critical step in Master Data Management.In this paper, we realize a Visual Modeling tool for implementing Repositories, Constraints and Processes based on Eclipse Plugin using GMF/EMF which follows principles of Model Driven Engineering (MDE).

Keywords: EMF, GMF, GEF, repository, constraint, process

Procedia PDF Downloads 481
24876 3D CFD Modelling of the Airflow and Heat Transfer in Cold Room Filled with Dates

Authors: Zina Ghiloufi, Tahar Khir

Abstract:

A transient three-dimensional computational fluid dynamics (CFD) model is developed to determine the velocity and temperature distribution in different positions cold room during pre-cooling of dates. The turbulence model used is the k-ω Shear Stress Transport (SST) with the standard wall function, the air. The numerical results obtained show that cooling rate is not uniform inside the room; the product at the medium of room has a slower cooling rate. This cooling heterogeneity has a large effect on the energy consumption during cold storage.

Keywords: CFD, cold room, cooling rate, dDates, numerical simulation, k-ω (SST)

Procedia PDF Downloads 224
24875 Automated Fact-Checking by Incorporating Contextual Knowledge and Multi-Faceted Search

Authors: Wenbo Wang, Yi-Fang Brook Wu

Abstract:

The spread of misinformation and disinformation has become a major concern, particularly with the rise of social media as a primary source of information for many people. As a means to address this phenomenon, automated fact-checking has emerged as a safeguard against the spread of misinformation and disinformation. Existing fact-checking approaches aim to determine whether a news claim is true or false, and they have achieved decent veracity prediction accuracy. However, the state-of-the-art methods rely on manually verified external information to assist the checking model in making judgments, which requires significant human resources. This study introduces a framework, SAC, which focuses on 1) augmenting the representation of a claim by incorporating additional context using general-purpose, comprehensive, and authoritative data; 2) developing a search function to automatically select relevant, new, and credible references; 3) focusing on the important parts of the representations of a claim and its reference that are most relevant to the fact-checking task. The experimental results demonstrate that 1) Augmenting the representations of claims and references through the use of a knowledge base, combined with the multi-head attention technique, contributes to improved performance of fact-checking. 2) SAC with auto-selected references outperforms existing fact-checking approaches with manual selected references. Future directions of this study include I) exploring knowledge graphs in Wikidata to dynamically augment the representations of claims and references without introducing too much noise, II) exploring semantic relations in claims and references to further enhance fact-checking.

Keywords: fact checking, claim verification, deep learning, natural language processing

Procedia PDF Downloads 47
24874 Design and Development of Data Visualization in 2D and 3D Space Using Front-End Technologies

Authors: Sourabh Yaduvanshi, Varsha Namdeo, Namrata Yaduvanshi

Abstract:

This study delves into the design and development intricacies of crafting detailed 2D bar charts via d3.js, recognizing its limitations in generating 3D visuals within the DOM. The study combines three.js with d3.js, facilitating a smooth evolution from 2D to immersive 3D representations. This fusion epitomizes the synergy between front-end technologies, expanding horizons in data visualization. Beyond technical expertise, it symbolizes a creative convergence, pushing boundaries in visual representation. The abstract illuminates methodologies, unraveling the intricate integration of this fusion and guiding enthusiasts. It narrates a compelling story of transcending 2D constraints, propelling data visualization into captivating three-dimensional realms, and igniting creativity in front-end visualization endeavors.

Keywords: design, development, front-end technologies, visualization

Procedia PDF Downloads 61