Search results for: LCA tools and data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 27284

Search results for: LCA tools and data

26144 Ethics Can Enable Open Source Data Research

Authors: Dragana Calic

Abstract:

The openness, availability and the sheer volume of big data have provided, what some regard as, an invaluable and rich dataset. Researchers, businesses, advertising agencies, medical institutions, to name only a few, collect, share, and analyze this data to enable their processes and decision making. However, there are important ethical considerations associated with the use of big data. The rapidly evolving nature of online technologies has overtaken the many legislative, privacy, and ethical frameworks and principles that exist. For example, should we obtain consent to use people’s online data, and under what circumstances can privacy considerations be overridden? Current guidance on how to appropriately and ethically handle big data is inconsistent. Consequently, this paper focuses on two quite distinct but related ethical considerations that are at the core of the use of big data for research purposes. They include empowering the producers of data and empowering researchers who want to study big data. The first consideration focuses on informed consent which is at the core of empowering producers of data. In this paper, we discuss some of the complexities associated with informed consent and consider studies of producers’ perceptions to inform research ethics guidelines and practice. The second consideration focuses on the researcher. Similarly, we explore studies that focus on researchers’ perceptions and experiences.

Keywords: big data, ethics, producers’ perceptions, researchers’ perceptions

Procedia PDF Downloads 282
26143 The Effects of Adlerian Supervision on Enhancing Career Consultants’ Case Conceptualization

Authors: Lin Shang Neng

Abstract:

Due to rapid changes in the societal environment, career development and planning have become increasingly crucial, leading more individuals to seek the assistance of career consultations. However, the training process for career consultants often emphasizes the application of assessment tools and guidance in job-seeking behavior. The abilities of case conceptualization and consulting skills require further in-service supervision. This study aims to inquire about the supervised experiences of employment specialists at the Employment Service Center of the Taiwan Ministry of Labor or career consultants who held private clinics for at least three years. The research participants were continuously supervised by the Adlerian approach twice a month for at least one year, helping them integrate the whole picture of the client through Lifestyle Assessment (the qualitative way, specific diagnosis) and other Adlerian assessment tools (the quantitative way, general diagnosis.) The supervisor was familiar with Adlerian Psychology and certified by the North American Society of Adlerian Psychology. The research method involves semi-structured interviews and qualitative analysis. For the ethical considerations, the participants were invited to interview after the supervision sessions finished. The findings of this research were discussed with possible implications, like how they applied Adlerian Psychology to their career consultations, especially to case conceptualizations and consulting skills. Recommendations for further research and training for career consultants are also discussed.

Keywords: supervision, Adlerian psychology, case conceptualization, career consultant

Procedia PDF Downloads 75
26142 Measuring the Resilience of e-Governments Using an Ontology

Authors: Onyekachi Onwudike, Russell Lock, Iain Phillips

Abstract:

The variability that exists across governments, her departments and the provisioning of services has been areas of concern in the E-Government domain. There is a need for reuse and integration across government departments which are accompanied by varying degrees of risks and threats. There is also the need for assessment, prevention, preparation, response and recovery when dealing with these risks or threats. The ability of a government to cope with the emerging changes that occur within it is known as resilience. In order to forge ahead with concerted efforts to manage reuse and integration induced risks or threats to governments, the ambiguities contained within resilience must be addressed. Enhancing resilience in the E-Government domain is synonymous with reducing risks governments face with provisioning of services as well as reuse of components across departments. Therefore, it can be said that resilience is responsible for the reduction in government’s vulnerability to changes. In this paper, we present the use of the ontology to measure the resilience of governments. This ontology is made up of a well-defined construct for the taxonomy of resilience. A specific class known as ‘Resilience Requirements’ is added to the ontology. This class embraces the concept of resilience into the E-Government domain ontology. Considering that the E-Government domain is a highly complex one made up of different departments offering different services, the reliability and resilience of the E-Government domain have become more complex and critical to understand. We present questions that can help a government access how prepared they are in the face of risks and what steps can be taken to recover from them. These questions can be asked with the use of queries. The ontology focuses on developing a case study section that is used to explore ways in which government departments can become resilient to the different kinds of risks and threats they may face. A collection of resilience tools and resources have been developed in our ontology to encourage governments to take steps to prepare for emergencies and risks that a government may face with the integration of departments and reuse of components across government departments. To achieve this, the ontology has been extended by rules. We present two tools for understanding resilience in the E-Government domain as a risk analysis target and the output of these tools when applied to resilience in the E-Government domain. We introduce the classification of resilience using the defined taxonomy and modelling of existent relationships based on the defined taxonomy. The ontology is constructed on formal theory and it provides a semantic reference framework for the concept of resilience. Key terms which fall under the purview of resilience with respect to E-Governments are defined. Terms are made explicit and the relationships that exist between risks and resilience are made explicit. The overall aim of the ontology is to use it within standards that would be followed by all governments for government-based resilience measures.

Keywords: E-Government, Ontology, Relationships, Resilience, Risks, Threats

Procedia PDF Downloads 336
26141 Screening of Wheat Wild Relatives as a Gene Pool for Improved Photosynthesis in Wheat Breeding

Authors: Amanda J. Burridge, Keith J. Edwards, Paul A. Wilkinson, Tom Batstone, Erik H. Murchie, Lorna McAusland, Ana Elizabete Carmo-Silva, Ivan Jauregui, Tracy Lawson, Silvere R. M. Vialet-Chabrand

Abstract:

The rate of genetic progress in wheat production must be improved to meet global food security targets. However, past selection for domestication traits has reduced the genetic variation in modern wheat cultivars, a fact that could severely limit the future rate of genetic gain. The genetic variation in agronomically important traits for the wild relatives and progenitors of wheat is far greater than that of the current domesticated cultivars, but transferring these traits into modern cultivars is not straightforward. Between the elite cultivars of wheat, photosynthetic capacity is a key trait for which there is limited variation. Early screening of wheat wild relative and progenitors has shown differences in photosynthetic capacity and efficiency not only between wild relative species but marked differences between the accessions of each species. By identifying wild relative accessions with improved photosynthetic traits and characterising the genetic variation responsible, it is possible to incorporate these traits into advanced breeding programmes by wide crossing and introgression programmes. To identify the potential variety of photosynthetic capacity and efficiency available in the secondary and tertiary genepool, a wide scale survey was carried out for over 600 accessions from 80 species including those from the genus Aegilops, Triticum, Thinopyrum, Elymus, and Secale. Genotype data were generated for each accession using a ‘Wheat Wild Relative’ Single Nucleotide Polymorphism (SNP) genotyping array composed of 35,000 SNP markers polymorphic between wild relatives and elite hexaploid wheat. This genotype data was combined with phenotypic measurements such as gas exchange (CO₂, H₂O), chlorophyll fluorescence, growth, morphology, and RuBisCO activity to identify potential breeding material with enhanced photosynthetic capacity and efficiency. The data and associated analysis tools presented here will prove useful to anyone interested in increasing the genetic diversity in hexaploid wheat or the application of complex genotyping data to plant breeding.

Keywords: wheat, wild relatives, pre-breeding, genomics, photosynthesis

Procedia PDF Downloads 220
26140 Hybrid Reliability-Similarity-Based Approach for Supervised Machine Learning

Authors: Walid Cherif

Abstract:

Data mining has, over recent years, seen big advances because of the spread of internet, which generates everyday a tremendous volume of data, and also the immense advances in technologies which facilitate the analysis of these data. In particular, classification techniques are a subdomain of Data Mining which determines in which group each data instance is related within a given dataset. It is used to classify data into different classes according to desired criteria. Generally, a classification technique is either statistical or machine learning. Each type of these techniques has its own limits. Nowadays, current data are becoming increasingly heterogeneous; consequently, current classification techniques are encountering many difficulties. This paper defines new measure functions to quantify the resemblance between instances and then combines them in a new approach which is different from actual algorithms by its reliability computations. Results of the proposed approach exceeded most common classification techniques with an f-measure exceeding 97% on the IRIS Dataset.

Keywords: data mining, knowledge discovery, machine learning, similarity measurement, supervised classification

Procedia PDF Downloads 461
26139 The Power of Geography in the Multipolar World Order

Authors: Norbert Csizmadia

Abstract:

The paper is based on a thorough investigation regarding the recent global, social and geographical processes. The ‘Geofusion’ book series by the author guides the readers with the help of newly illustrated “associative” geographic maps of the global world in the 21st century through the quest for the winning nations, communities, leaders and powers of this age. Hence, the above mentioned represent the research objectives, the preliminary findings of which are presented in this paper. The most significant recognition is that scientists who are recognized as explorers, geostrategists of this century, in this case, are expected to present guidelines for our new world full of global social and economic challenges. To do so, new maps are needed which do not miss the wisdom and tools of the old but complement them with the new structure of knowledge. Using the lately discovered geographic and economic interrelations, the study behind this presentation tries to give a prognosis of the global processes. The methodology applied contains the survey and analysis of many recent publications worldwide regarding geostrategic, cultural, geographical, social, and economic surveys structured into global networks. In conclusion, the author presents the result of the study, which is a collage of the global map of the 21st century as mentioned above, and it can be considered as a potential contribution to the recent scientific literature on the topic. In summary, this paper displays the results of several-year-long research giving the audience an image of how economic navigation tools can help investors, politicians and travelers to get along in the changing new world.

Keywords: geography, economic geography, geo-fusion, geostrategy

Procedia PDF Downloads 129
26138 Seismic Data Scaling: Uncertainties, Potential and Applications in Workstation Interpretation

Authors: Ankur Mundhra, Shubhadeep Chakraborty, Y. R. Singh, Vishal Das

Abstract:

Seismic data scaling affects the dynamic range of a data and with present day lower costs of storage and higher reliability of Hard Disk data, scaling is not suggested. However, in dealing with data of different vintages, which perhaps were processed in 16 bits or even 8 bits and are need to be processed with 32 bit available data, scaling is performed. Also, scaling amplifies low amplitude events in deeper region which disappear due to high amplitude shallow events that saturate amplitude scale. We have focused on significance of scaling data to aid interpretation. This study elucidates a proper seismic loading procedure in workstations without using default preset parameters as available in most software suites. Differences and distribution of amplitude values at different depth for seismic data are probed in this exercise. Proper loading parameters are identified and associated steps are explained that needs to be taken care of while loading data. Finally, the exercise interprets the un-certainties which might arise when correlating scaled and unscaled versions of seismic data with synthetics. As, seismic well tie correlates the seismic reflection events with well markers, for our study it is used to identify regions which are enhanced and/or affected by scaling parameter(s).

Keywords: clipping, compression, resolution, seismic scaling

Procedia PDF Downloads 467
26137 Strategies to Combat the Covid-19 Epidemic

Authors: Marziye Hadian, Alireza Jabbari

Abstract:

Background: The World Health Organization has identified COVID-19 as a public health emergency and is urging governments to stop the virus transmission by adopting appropriate policies. In this regard, the countries have taken different approaches to cutting the chain or controlling the spread of the disease. Methods: The present study was a systematize review of publications relating to prevention strategies for covid-19 disease. The study was carried out based on the PRISMA guidelines and CASP for articles and AACODS for grey literature. Finding: The study findings showed that in order to confront the COVID-19 epidemic, in general, there are three approaches of "mitigation", "active control" and "suppression" and four strategies of "quarantine", "isolation", "social distance" as well as "lockdown" in both individual and social dimensions to deal with epidemics that the choice of each approach requires specific strategies and has different effects when it comes to controlling and inhibiting the disease. Conclusion: The only way to control the disease is to change your behavior and lifestyle. In addition to prevention strategies, use of masks, observance of personal hygiene principles such as regular hand washing and non-contact of contaminated hands with the face, as well as observance of public health principles such as control of sneezing and coughing, safe extermination of personal protective equipment, etc. have not been included in the category of prevention tools. However, it has a great impact on controlling the epidemic, especially the new coronavirus epidemic.

Keywords: novel corona virus, COVID-19, prevention tools, prevention strategies

Procedia PDF Downloads 135
26136 Towards Modern Approaches of Intelligence Measurement for Clinical and Educational Practices

Authors: Alena Kulikova, Tatjana Kanonire

Abstract:

Intelligence research is one of the oldest fields of psychology. Many factors have made a research on intelligence, defined as reasoning and problem solving [1, 2], a very acute and urgent problem. Thus, it has been repeatedly shown that intelligence is a predictor of academic, professional, and social achievement in adulthood (for example, [3]); Moreover, intelligence predicts these achievements better than any other trait or ability [4]. The individual level, a comprehensive assessment of intelligence is a necessary criterion for the diagnosis of various mental conditions. For example, it is a necessary condition for psychological, medical and pedagogical commissions when deciding on educational needs and the most appropriate educational programs for school children. Assessment of intelligence is crucial in clinical psychodiagnostic and needs high-quality intelligence measurement tools. Therefore, it is not surprising that the development of intelligence tests is an essential part of psychological science and practice. Many modern intelligence tests have a long history and have been used for decades, for example, the Stanford-Binet test or the Wechsler test. However, the vast majority of these tests are based on the classic linear test structure, in which all respondents receive all tasks (see, for example, a critical review by [5]). This understanding of the testing procedure is a legacy of the pre-computer era, in which blank testing was the only diagnostic procedure available [6] and has some significant limitations that affect the reliability of the data obtained [7] and increased time costs. Another problem with measuring IQ is that classical line-structured tests do not fully allow to measure respondent's intellectual progress [8], which is undoubtedly a critical limitation. Advances in modern psychometrics allow for avoiding the limitations of existing tools. However, as in any rapidly developing industry, at the moment, psychometrics does not offer ready-made and straightforward solutions and requires additional research. In our presentation we would like to discuss the strengths and weaknesses of the current approaches to intelligence measurement and highlight “points of growth” for creating a test in accordance with modern psychometrics. Whether it is possible to create the instrument that will use all achievements of modern psychometric and remain valid and practically oriented. What would be the possible limitations for such an instrument? The theoretical framework and study design to create and validate the original Russian comprehensive computer test for measuring the intellectual development in school-age children will be presented.

Keywords: Intelligence, psychometrics, psychological measurement, computerized adaptive testing, multistage testing

Procedia PDF Downloads 77
26135 Resale Housing Development Board Price Prediction Considering Covid-19 through Sentiment Analysis

Authors: Srinaath Anbu Durai, Wang Zhaoxia

Abstract:

Twitter sentiment has been used as a predictor to predict price values or trends in both the stock market and housing market. The pioneering works in this stream of research drew upon works in behavioural economics to show that sentiment or emotions impact economic decisions. Latest works in this stream focus on the algorithm used as opposed to the data used. A literature review of works in this stream through the lens of data used shows that there is a paucity of work that considers the impact of sentiments caused due to an external factor on either the stock or the housing market. This is despite an abundance of works in behavioural economics that show that sentiment or emotions caused due to an external factor impact economic decisions. To address this gap, this research studies the impact of Twitter sentiment pertaining to the Covid-19 pandemic on resale Housing Development Board (HDB) apartment prices in Singapore. It leverages SNSCRAPE to collect tweets pertaining to Covid-19 for sentiment analysis, lexicon based tools VADER and TextBlob are used for sentiment analysis, Granger Causality is used to examine the relationship between Covid-19 cases and the sentiment score, and neural networks are leveraged as prediction models. Twitter sentiment pertaining to Covid-19 as a predictor of HDB price in Singapore is studied in comparison with the traditional predictors of housing prices i.e., the structural and neighbourhood characteristics. The results indicate that using Twitter sentiment pertaining to Covid19 leads to better prediction than using only the traditional predictors and performs better as a predictor compared to two of the traditional predictors. Hence, Twitter sentiment pertaining to an external factor should be considered as important as traditional predictors. This paper demonstrates the real world economic applications of sentiment analysis of Twitter data.

Keywords: sentiment analysis, Covid-19, housing price prediction, tweets, social media, Singapore HDB, behavioral economics, neural networks

Procedia PDF Downloads 109
26134 Association of Social Data as a Tool to Support Government Decision Making

Authors: Diego Rodrigues, Marcelo Lisboa, Elismar Batista, Marcos Dias

Abstract:

Based on data on child labor, this work arises questions about how to understand and locate the factors that make up the child labor rates, and which properties are important to analyze these cases. Using data mining techniques to discover valid patterns on Brazilian social databases were evaluated data of child labor in the State of Tocantins (located north of Brazil with a territory of 277000 km2 and comprises 139 counties). This work aims to detect factors that are deterministic for the practice of child labor and their relationships with financial indicators, educational, regional and social, generating information that is not explicit in the government database, thus enabling better monitoring and updating policies for this purpose.

Keywords: social data, government decision making, association of social data, data mining

Procedia PDF Downloads 367
26133 A Particle Filter-Based Data Assimilation Method for Discrete Event Simulation

Authors: Zhi Zhu, Boquan Zhang, Tian Jing, Jingjing Li, Tao Wang

Abstract:

Data assimilation is a model and data hybrid-driven method that dynamically fuses new observation data with a numerical model to iteratively approach the real system state. It is widely used in state prediction and parameter inference of continuous systems. Because of the discrete event system’s non-linearity and non-Gaussianity, traditional Kalman Filter based on linear and Gaussian assumptions cannot perform data assimilation for such systems, so particle filter has gradually become a technical approach for discrete event simulation data assimilation. Hence, we proposed a particle filter-based discrete event simulation data assimilation method and took the unmanned aerial vehicle (UAV) maintenance service system as a proof of concept to conduct simulation experiments. The experimental results showed that the filtered state data is closer to the real state of the system, which verifies the effectiveness of the proposed method. This research can provide a reference framework for the data assimilation process of other complex nonlinear systems, such as discrete-time and agent simulation.

Keywords: discrete event simulation, data assimilation, particle filter, model and data-driven

Procedia PDF Downloads 4
26132 Role of Financial Institutions in Promoting Micro Service Enterprises with Special Reference to Hairdressing Salons

Authors: Gururaj Bhajantri

Abstract:

Financial sector is the backbone of any economy and it plays a crucial role in the mobilisation and allocation of resources. One of the main objectives of financial sector is inclusive growth. The constituents of the financial sector are banks, and financial Institutions, which mobilise the resources from the surplus sector and channelize the same to the different needful sectors in the economy. Micro Small and the Medium Enterprises sector in India cover a wide range of economic activities. These enterprises are divided on the basis of investment on equipment. The micro enterprises are divided into manufacturing and services sector. Micro Service enterprises have investment limit up to ten lakhs on equipment. Hairdresser is one who not only cuts and shaves but also provides different types of hair cut, hairstyles, trimming, hair-dye, massage, manicure, pedicure, nail services, colouring, facial, makeup application, waxing, tanning and other beauty treatments etc., hairdressing salons provide these services with the help of equipment. They need investment on equipment not more than ten lakhs. Hence, they can be considered as Micro service enterprises. Hairdressing salons require more than Rs 2.50,000 to start a moderate salon. Moreover, hairdressers are unable to access the organised finance. Still these individuals access finance from money lenders with high rate of interest to lead life. The socio economic conditions of hairdressers are not known properly. Hence, the present study brings a light on the role of financial institutions in promoting hairdressing salons. The study also focuses the socio-economic background of individuals in hairdressings salons, problems faced by them. The present study is based on primary and secondary data. Primary data collected among hairdressing salons in Davangere city. Samples selected with the help of simple random sampling techniques. Collected data analysed and interpreted with the help of simple statistical tools.

Keywords: micro service enterprises, financial institutions, hairdressing salons, financial sector

Procedia PDF Downloads 202
26131 Effect of Asymmetric Amphiphilic Dicationic Ionic Liquids as Oil Spill Dispersants in Red Sea

Authors: Raghda El-Nagara, Maher I. Nessim, Carmen E. Elshafee, Renee I. Abdallah, Yasser M. Moustafa

Abstract:

Three asymmetric dicationic ionic liquids (ADILs), 1-(2-(1-dodecyl-2-methyl-1H-imidazolium-3-yl)ethyl)-3-methyl pyridinium bromide (IL₁), 1-(6-(1-dodecyl-2-methyl-1H-imidazolium-3-yl)hexyl)-3-methyl pyridinium bromide (IL₂) and 1-(10-(1-dodecyl-2-methyl-1H-imidazolium-3-yl)decyl)-3-methyl pyridinium bromide (IL₃) were synthesized with yield of 83.54, 84.12 & 83.05% respectively. They were elucidated via conventional tools of analysis (elemental analysis, FT-IR, and 1H-NMR). The thermogravimetric analysis confirmed that the three ADILs possessed high thermal stability (up to 500ᵒC). Their critical micelle concentration (CMC) was investigated and exhibited values of 5.5-1*10⁻³ Mol./L. They were evaluated as oil spill dispersants were at different temperatures (10, 30 & 50ᵒC) with different concentrations (750, 1500, 2000, 3000 ppm). Data reveals that the efficiency is ranked as follows: IL₂ > IL₁ > IL₃, which showed high dispersion efficiency reached to 63% with the concentration of 1500 ppm.

Keywords: ionic liquids, amphiphilic, oil spill dispersants, dicationic, efficiency test

Procedia PDF Downloads 148
26130 Outlier Detection in Stock Market Data using Tukey Method and Wavelet Transform

Authors: Sadam Alwadi

Abstract:

Outlier values become a problem that frequently occurs in the data observation or recording process. Thus, the need for data imputation has become an essential matter. In this work, it will make use of the methods described in the prior work to detect the outlier values based on a collection of stock market data. In order to implement the detection and find some solutions that maybe helpful for investors, real closed price data were obtained from the Amman Stock Exchange (ASE). Tukey and Maximum Overlapping Discrete Wavelet Transform (MODWT) methods will be used to impute the detect the outlier values.

Keywords: outlier values, imputation, stock market data, detecting, estimation

Procedia PDF Downloads 80
26129 Pragmatic Competence in Pakistani English Language Learners

Authors: Ghazala Kausar

Abstract:

This study investigates Pakistani first year university students’ perception of the role of pragmatics in their general approach to learning English. The research is triggered by National Curriculum’s initiative to provide holistic opportunities to the students for language development and to equip them with competencies to use English language in academic and social contexts (New English National Curriculum for I-XII). The traditional grammar translation and examination oriented method is believed to reduce learners to silent listener (Zhang, 2008: Zhao 2009). This lead to the inability of the students to interpret discourse by relating utterances to their meaning, understanding the intentions of the users and how language is used in specific setting (Bachman & Palmer, 1996, 2010). Pragmatic competence is a neglected area as far as teaching and learning English in Pakistan is concerned. This study focuses on the different types of pragmatic knowledge, learners perception of such knowledge and learning strategies employed by different learners to process the learning in general and pragmatic in particular. This study employed three data collecting tools; a questionnaire, discourse completion task and interviews to elicit data from first year university students regarding their perception of pragmatic competence. Results showed that Pakistani first year university learners have limited pragmatic knowledge. Although they acknowledged the importance of linguistic knowledge for linguistic competence in the students but argued that insufficient English proficiency, limited knowledge of pragmatics, insufficient language material and tasks were major reasons of pragmatic failure.

Keywords: pragmatic competence, Pakistani college learners, linguistic competence

Procedia PDF Downloads 733
26128 PEINS: A Generic Compression Scheme Using Probabilistic Encoding and Irrational Number Storage

Authors: P. Jayashree, S. Rajkumar

Abstract:

With social networks and smart devices generating a multitude of data, effective data management is the need of the hour for networks and cloud applications. Some applications need effective storage while some other applications need effective communication over networks and data reduction comes as a handy solution to meet out both requirements. Most of the data compression techniques are based on data statistics and may result in either lossy or lossless data reductions. Though lossy reductions produce better compression ratios compared to lossless methods, many applications require data accuracy and miniature details to be preserved. A variety of data compression algorithms does exist in the literature for different forms of data like text, image, and multimedia data. In the proposed work, a generic progressive compression algorithm, based on probabilistic encoding, called PEINS is projected as an enhancement over irrational number stored coding technique to cater to storage issues of increasing data volumes as a cost effective solution, which also offers data security as a secondary outcome to some extent. The proposed work reveals cost effectiveness in terms of better compression ratio with no deterioration in compression time.

Keywords: compression ratio, generic compression, irrational number storage, probabilistic encoding

Procedia PDF Downloads 291
26127 Iot Device Cost Effective Storage Architecture and Real-Time Data Analysis/Data Privacy Framework

Authors: Femi Elegbeleye, Omobayo Esan, Muienge Mbodila, Patrick Bowe

Abstract:

This paper focused on cost effective storage architecture using fog and cloud data storage gateway and presented the design of the framework for the data privacy model and data analytics framework on a real-time analysis when using machine learning method. The paper began with the system analysis, system architecture and its component design, as well as the overall system operations. The several results obtained from this study on data privacy model shows that when two or more data privacy model is combined we tend to have a more stronger privacy to our data, and when fog storage gateway have several advantages over using the traditional cloud storage, from our result shows fog has reduced latency/delay, low bandwidth consumption, and energy usage when been compare with cloud storage, therefore, fog storage will help to lessen excessive cost. This paper dwelt more on the system descriptions, the researchers focused on the research design and framework design for the data privacy model, data storage, and real-time analytics. This paper also shows the major system components and their framework specification. And lastly, the overall research system architecture was shown, its structure, and its interrelationships.

Keywords: IoT, fog, cloud, data analysis, data privacy

Procedia PDF Downloads 95
26126 Comparison of Selected Pier-Scour Equations for Wide Piers Using Field Data

Authors: Nordila Ahmad, Thamer Mohammad, Bruce W. Melville, Zuliziana Suif

Abstract:

Current methods for predicting local scour at wide bridge piers, were developed on the basis of laboratory studies and very limited scour prediction were tested with field data. Laboratory wide pier scour equation from previous findings with field data were presented. A wide range of field data were used and it consists of both live-bed and clear-water scour. A method for assessing the quality of the data was developed and applied to the data set. Three other wide pier-scour equations from the literature were used to compare the performance of each predictive method. The best-performing scour equation were analyzed using statistical analysis. Comparisons of computed and observed scour depths indicate that the equation from the previous publication produced the smallest discrepancy ratio and RMSE value when compared with the large amount of laboratory and field data.

Keywords: field data, local scour, scour equation, wide piers

Procedia PDF Downloads 404
26125 The Maximum Throughput Analysis of UAV Datalink 802.11b Protocol

Authors: Inkyu Kim, SangMan Moon

Abstract:

This IEEE 802.11b protocol provides up to 11Mbps data rate, whereas aerospace industry wants to seek higher data rate COTS data link system in the UAV. The Total Maximum Throughput (TMT) and delay time are studied on many researchers in the past years This paper provides theoretical data throughput performance of UAV formation flight data link using the existing 802.11b performance theory. We operate the UAV formation flight with more than 30 quad copters with 802.11b protocol. We may be predicting that UAV formation flight numbers have to bound data link protocol performance limitations.

Keywords: UAV datalink, UAV formation flight datalink, UAV WLAN datalink application, UAV IEEE 802.11b datalink application

Procedia PDF Downloads 389
26124 Methods for Distinction of Cattle Using Supervised Learning

Authors: Radoslav Židek, Veronika Šidlová, Radovan Kasarda, Birgit Fuerst-Waltl

Abstract:

Machine learning represents a set of topics dealing with the creation and evaluation of algorithms that facilitate pattern recognition, classification, and prediction, based on models derived from existing data. The data can present identification patterns which are used to classify into groups. The result of the analysis is the pattern which can be used for identification of data set without the need to obtain input data used for creation of this pattern. An important requirement in this process is careful data preparation validation of model used and its suitable interpretation. For breeders, it is important to know the origin of animals from the point of the genetic diversity. In case of missing pedigree information, other methods can be used for traceability of animal´s origin. Genetic diversity written in genetic data is holding relatively useful information to identify animals originated from individual countries. We can conclude that the application of data mining for molecular genetic data using supervised learning is an appropriate tool for hypothesis testing and identifying an individual.

Keywords: genetic data, Pinzgau cattle, supervised learning, machine learning

Procedia PDF Downloads 544
26123 Router 1X3 - RTL Design and Verification

Authors: Nidhi Gopal

Abstract:

Routing is the process of moving a packet of data from source to destination and enables messages to pass from one computer to another and eventually reach the target machine. A router is a networking device that forwards data packets between computer networks. It is connected to two or more data lines from different networks (as opposed to a network switch, which connects data lines from one single network). This paper mainly emphasizes upon the study of router device, its top level architecture, and how various sub-modules of router i.e. Register, FIFO, FSM and Synchronizer are synthesized, and simulated and finally connected to its top module.

Keywords: data packets, networking, router, routing

Procedia PDF Downloads 806
26122 Focus-Latent Dirichlet Allocation for Aspect-Level Opinion Mining

Authors: Mohsen Farhadloo, Majid Farhadloo

Abstract:

Aspect-level opinion mining that aims at discovering aspects (aspect identification) and their corresponding ratings (sentiment identification) from customer reviews have increasingly attracted attention of researchers and practitioners as it provides valuable insights about products/services from customer's points of view. Instead of addressing aspect identification and sentiment identification in two separate steps, it is possible to simultaneously identify both aspects and sentiments. In recent years many graphical models based on Latent Dirichlet Allocation (LDA) have been proposed to solve both aspect and sentiment identifications in a single step. Although LDA models have been effective tools for the statistical analysis of document collections, they also have shortcomings in addressing some unique characteristics of opinion mining. Our goal in this paper is to address one of the limitations of topic models to date; that is, they fail to directly model the associations among topics. Indeed in many text corpora, it is natural to expect that subsets of the latent topics have higher probabilities. We propose a probabilistic graphical model called focus-LDA, to better capture the associations among topics when applied to aspect-level opinion mining. Our experiments on real-life data sets demonstrate the improved effectiveness of the focus-LDA model in terms of the accuracy of the predictive distributions over held out documents. Furthermore, we demonstrate qualitatively that the focus-LDA topic model provides a natural way of visualizing and exploring unstructured collection of textual data.

Keywords: aspect-level opinion mining, document modeling, Latent Dirichlet Allocation, LDA, sentiment analysis

Procedia PDF Downloads 92
26121 Project Work with Design Thinking and Blended Learning: A Practical Report from Teaching in Higher Education

Authors: C. Vogeler

Abstract:

Change processes such as individualization and digitalization have an impact on higher education. Graduates are expected to cooperate in creative work processes in their professional life. During their studies, they need to be prepared accordingly. This includes modern learning scenarios that integrate the benefits of digital media. Therefore, design thinking and blended learning have been combined in the project-based seminar conception introduced here. The presented seminar conception has been realized and evaluated with students of information sciences since September 2017. Within the seminar, the students learn to work on a project. They apply the methods in a problem-based learning scenario. Task of the case study is to arrange a conference on the topic gaming in libraries. In order to collaborative develop creative possibilities of realization within the group of students the design thinking method has been chosen. Design thinking is a method, used to create user-centric, problem-solving and need-driven innovation through creative collaboration in multidisciplinary teams. Central characteristics are the openness of this approach to work results and the visualization of ideas. This approach is now also accepted in the field of higher education. Especially in problem-based learning scenarios, the method offers clearly defined process steps for creative ideas and their realization. The creative process can be supported by digital media, such as search engines and tools for the documentation of brainstorming, creation of mind maps, project management etc. Because the students have to do two-thirds of the workload in their private study, design thinking has been combined with a blended learning approach. This supports students’ preparation and follow-up of the joint work in workshops (flipped classroom scenario) as well as the communication and collaboration during the entire project work phase. For this purpose, learning materials are provided on a Moodle-based learning platform as well as various tools that supported the design thinking process as described above. In this paper, the seminar conception with a combination of design thinking and blended learning is described and the potentials and limitations of the chosen strategy for the development of a course with a multimedia approach in higher education are reflected.

Keywords: blended learning, design thinking, digital media tools and methods, flipped classroom

Procedia PDF Downloads 193
26120 Developing Elusive Frame for Creativity, ICT and Teacher Education

Authors: Rahul Malhotra, Anu Malhotra, Veena Bana

Abstract:

Information and Communication Technology (ICT) plays an imperative part in enhancing the quality of life, together with education. This research work is an important consequence to endow with substantiation for the effective use of Information and Communication Technology (ICT) tools for educational rationale. Perspective teachers and students of technical education from various regions of Rajasthan participated in the survey based research work. Condensed from the analysis and interpretations of the data collected from Perspective teachers and students of technical education from various regions of Rajasthan, it is inevitable that use of Information and Communication Technology (ICT) for educational purpose can augment student’s creativity and achievement ability. The Chi-Square statistics produce the evidence of the effective use of Information and Communication Technology (ICT) in enhancing the creativity and achievement ability of the perspective teachers and students of technical education.

Keywords: ICT, education, creativity, elusive frame

Procedia PDF Downloads 459
26119 Teachers’ Conception of and Perception towards the New Curriculum of Ethiopian Higher Education: A Case of Debre Birhan University

Authors: Kassahun Tilahun Dessie

Abstract:

The purpose of this study was to explore the awareness of teachers and the attitude they have to the curriculum they implement as well as to assess the actual and desired magnitude of teachers' participation in curriculum development process. It also aimed at investigating the factors that affect teachers' level of conception and perception towards the new higher education curriculum. The study was carried out in Debre Birhan University. Teachers, course coordinators, team leaders and presidents were included in the study as research subjects. Teachers were proportionally selected from each department (of the six faculties) based on available sampling technique. Accordingly, a total of 103 teachers were chosen as a subject of the study. In order to collect first hand data from the teachers, a questionnaire with four parts was developed by the researcher. To this end, scales were designed for measuring the extent of teachers' awareness and attitude. Each of the scales encompasses 11 and 17 items respectively. An open ended questionnaire was also attached for the purpose of obtaining elaborated data on the issue. Information was also obtained from interviews with presidents, team leaders and course coordinators. The data obtained were analyzed qualitatively using descriptive statistical tools. The overall results of the analysis revealed that the awareness of teachers on the curriculum was low. The meager participation of teachers in the process of curriculum development and the deficiency of trainings on the concern were major factors. Teachers' perception towards the existence and implementation of the new curriculum was also inclined to the negative, though difficult to generalize. Lack of awareness, administrators poor approach and lack of facilitating appropriate incentives as well as absence of room for evaluating the curriculum etc plays big role in endangering teachers attitude while the up to datedness of the new curriculum, involvement of teachers in the curriculum development process, the wide ranging quality of the new curriculum etc laid a better ground to boost teachers attitude towards the curriculum. This may have implication to the university in that there is a need to facilitate workshops or awareness creation trainings, to have positive and cooperative administrators, and embracing committed teachers to implement the curriculum efficiently.

Keywords: conception, perception, curriculum, higher education, Ethiopia

Procedia PDF Downloads 502
26118 Communication Tools Used in Teaching and Their Effects: An Empirical Study on the T. C. Selcuk University Samples

Authors: Sedat Simsek, Tugay Arat

Abstract:

Today's communication concept, which has a great revolution with the printing press which has been found by Gutenberg, has no boundary thanks to advanced communication devices and the internet. It is possible to take advantage in many areas, such as from medicine to social sciences or from mathematics to education, from the computers that was first produced for the purpose of military services. The use of these developing technologies in the field of education has created a great vision changes in both training and having education. Materials, which can be considered as basic communication resources and used in traditional education has begun to lose its significance, and some technologies have begun to replace them such as internet, computers, smart boards, projection devices and mobile phone. On the other hand, the programs and applications used in these technologies have also been developed. University students use virtual books instead of the traditional printed book, use cell phones instead of note books, use the internet and virtual databases instead of the library to research. They even submit their homework with interactive methods rather than printed materials. The traditional education system, these technologies, which increase productivity, have brought a new dimension to education. The aim of this study is to determine the influence of technologies in the learning process of students and to find whether is there any similarities and differences that arise from the their faculty that they have been educated and and their learning process. In addition to this, it is aimed to determine the level of ICT usage of students studying at the university level. In this context, the advantages and conveniences of the technology used by students are also scrutinized. In this study, we used surveys to collect data. The data were analyzed by using SPSS 16 statistical program with the appropriate testing.

Keywords: education, communication technologies, role of technology, teaching

Procedia PDF Downloads 299
26117 Harnessing the Power of Artificial Intelligence: Advancements and Ethical Considerations in Psychological and Behavioral Sciences

Authors: Nayer Mofidtabatabaei

Abstract:

Advancements in artificial intelligence (AI) have transformed various fields, including psychology and behavioral sciences. This paper explores the diverse ways in which AI is applied to enhance research, diagnosis, therapy, and understanding of human behavior and mental health. We discuss the potential benefits and challenges associated with AI in these fields, emphasizing the ethical considerations and the need for collaboration between AI researchers and psychological and behavioral science experts. Artificial Intelligence (AI) has gained prominence in recent years, revolutionizing multiple industries, including healthcare, finance, and entertainment. One area where AI holds significant promise is the field of psychology and behavioral sciences. AI applications in this domain range from improving the accuracy of diagnosis and treatment to understanding complex human behavior patterns. This paper aims to provide an overview of the various AI applications in psychological and behavioral sciences, highlighting their potential impact, challenges, and ethical considerations. Mental Health Diagnosis AI-driven tools, such as natural language processing and sentiment analysis, can analyze large datasets of text and speech to detect signs of mental health issues. For example, chatbots and virtual therapists can provide initial assessments and support to individuals suffering from anxiety or depression. Autism Spectrum Disorder (ASD) Diagnosis AI algorithms can assist in early ASD diagnosis by analyzing video and audio recordings of children's behavior. These tools help identify subtle behavioral markers, enabling earlier intervention and treatment. Personalized Therapy AI-based therapy platforms use personalized algorithms to adapt therapeutic interventions based on an individual's progress and needs. These platforms can provide continuous support and resources for patients, making therapy more accessible and effective. Virtual Reality Therapy Virtual reality (VR) combined with AI can create immersive therapeutic environments for treating phobias, PTSD, and social anxiety. AI algorithms can adapt VR scenarios in real-time to suit the patient's progress and comfort level. Data Analysis AI aids researchers in processing vast amounts of data, including survey responses, brain imaging, and genetic information. Privacy Concerns Collecting and analyzing personal data for AI applications in psychology and behavioral sciences raise significant privacy concerns. Researchers must ensure the ethical use and protection of sensitive information. Bias and Fairness AI algorithms can inherit biases present in training data, potentially leading to biased assessments or recommendations. Efforts to mitigate bias and ensure fairness in AI applications are crucial. Transparency and Accountability AI-driven decisions in psychology and behavioral sciences should be transparent and subject to accountability. Patients and practitioners should understand how AI algorithms operate and make decisions. AI applications in psychological and behavioral sciences have the potential to transform the field by enhancing diagnosis, therapy, and research. However, these advancements come with ethical challenges that require careful consideration. Collaboration between AI researchers and psychological and behavioral science experts is essential to harness AI's full potential while upholding ethical standards and privacy protections. The future of AI in psychology and behavioral sciences holds great promise, but it must be navigated with caution and responsibility.

Keywords: artificial intelligence, psychological sciences, behavioral sciences, diagnosis and therapy, ethical considerations

Procedia PDF Downloads 67
26116 Data Mining and Knowledge Management Application to Enhance Business Operations: An Exploratory Study

Authors: Zeba Mahmood

Abstract:

The modern business organizations are adopting technological advancement to achieve competitive edge and satisfy their consumer. The development in the field of Information technology systems has changed the way of conducting business today. Business operations today rely more on the data they obtained and this data is continuously increasing in volume. The data stored in different locations is difficult to find and use without the effective implementation of Data mining and Knowledge management techniques. Organizations who smartly identify, obtain and then convert data in useful formats for their decision making and operational improvements create additional value for their customers and enhance their operational capabilities. Marketers and Customer relationship departments of firm use Data mining techniques to make relevant decisions, this paper emphasizes on the identification of different data mining and Knowledge management techniques that are applied to different business industries. The challenges and issues of execution of these techniques are also discussed and critically analyzed in this paper.

Keywords: knowledge, knowledge management, knowledge discovery in databases, business, operational, information, data mining

Procedia PDF Downloads 535
26115 Analysis of Energy Efficiency Behavior with the Use of Train Dynamics Simulator and Statistical Tools: Case Study of Vitoria Minas Railway, Brazil

Authors: Eric Wilson Santos Cabral, Marta Monteiro Da Costa Cruz, Fabio Luis Maciel Machado, Henrique Andrade, Rodrigo Pirola Pestana, Vivian Andrea Parreira

Abstract:

The large variation in the price of diesel in Brazil directly affects the variable cost of companies operating in the transportation sector. In rail transport, the great challenge is to overcome the annual budget, cargo and ore transported with cost reduction in relation to previous years, becoming more efficient every year. Some effective measures are necessary to achieve the reduction of the liter ratio consumed by KTKB (Gross Ton per Kilometer multiplied by thousand). This acronym represents the indicator of energy efficiency of some railroads in the world. This study is divided into two parts: the first, to identify using statistical tools, part of the controlled variables in the railways, which have a correlation with the energy efficiency indicator, seeking to aid decision-making. The second, with the use of the train dynamics simulator, within scenarios defined in the operational reality of a railroad, seeks to optimize the train formations and the train stop model for the change of train drivers. With the completion of the study, companies in the rail sector are expected to be able to reduce some of their transportation costs.

Keywords: railway transport, railway simulation, energy efficiency, fuel consumption

Procedia PDF Downloads 333