Search results for: data reliability
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 25627

Search results for: data reliability

24547 Factors Affecting Customer Loyalty in the Independent Surveyor Service Industry in Indonesia

Authors: Sufrin Hannan, Budi Suharjo, Rita Nurmalina, Kirbrandoko

Abstract:

The challenge for independent surveyor service companies now is growing with increasing uncertainty in business. Protection from the government for domestic independent surveyor industry from competitor attack, such as entering the global surveyors to Indonesia also no longer exists. Therefore, building customer loyalty becomes very important to create a long-term relationship between an independent surveyor with its customers. This study aims to develop a model that can be used to build customer loyalty by looking at various factors that determine customer loyalty, especially on independent surveyors for coal inspection in Indonesia. The development of this model uses the relationship marketing approach. Testing of the hypothesis is done by testing the variables that determine customer loyalty, either directly or indirectly, which amounted to 10 variables. The data were collected from 200 questionnaires filled by independent surveyor company decision makers from 51 exporting companies and coal trading companies in Indonesia and analyzed using Structural Equation Model (SEM). The results show that customer loyalty of independent surveyors is influenced by customer satisfaction, trust, switching-barrier, and relationship-bond. Research on customer satisfaction shows that customer satisfaction is influenced by the perceived quality and perceived value, while perceived quality is influenced by reliability, assurance, responsiveness, and empathy.

Keywords: relationship marketing, customer loyalty, customer satisfaction, switching barriers, relationship bonds

Procedia PDF Downloads 159
24546 Security in Resource Constraints Network Light Weight Encryption for Z-MAC

Authors: Mona Almansoori, Ahmed Mustafa, Ahmad Elshamy

Abstract:

Wireless sensor network was formed by a combination of nodes, systematically it transmitting the data to their base stations, this transmission data can be easily compromised if the limited processing power and the data consistency from these nodes are kept in mind; there is always a discussion to address the secure data transfer or transmission in actual time. This will present a mechanism to securely transmit the data over a chain of sensor nodes without compromising the throughput of the network by utilizing available battery resources available in the sensor node. Our methodology takes many different advantages of Z-MAC protocol for its efficiency, and it provides a unique key by sharing the mechanism using neighbor node MAC address. We present a light weighted data integrity layer which is embedded in the Z-MAC protocol to prove that our protocol performs well than Z-MAC when we introduce the different attack scenarios.

Keywords: hybrid MAC protocol, data integrity, lightweight encryption, neighbor based key sharing, sensor node dataprocessing, Z-MAC

Procedia PDF Downloads 133
24545 Using Unilateral Diplomatic Assurances to Evade Provisional Measures' Orders

Authors: William Thomas Worster

Abstract:

This paper will highlight the failure of international adjudication to prevent a state from evading an order of provisional measures by simply issuing a diplomatic assurance to the court. This practice changes the positions of the litigants as equals before a court, prevents the court from inquiring into the reliability of the political pledge as it would with assurances from a state to an individual, and diminishes the court’s ability to control its own proceedings in the face of concerns over sovereignty. Both the European Court of Human Rights (ECtHR) and International Court of Justice (ICJ) will entertain these kinds of unilateral pledges, but they consider them differently when the declaration is made between states or between a state and an individual, and when made directly to the court. In short, diplomatic assurances issued between states or to individuals are usually considered not to be legally binding and are essentially questions of fact, but unilateral assurances issued directly to an international court are questions of law, and usually legally binding. At the same time, orders for provisional measures are now understood also to be legally binding, yet international courts will sometimes permit a state to substitute an assurance in place of an order for provisional measures. This emerging practice has brought the nature of a state as a sovereign capable of creating legal obligations into the forum of adjudication where the parties should have equality of arms and permitted states to create legal obligations that escape inquiry into the reliability of the outcome. While most recent practice has occurred at the ICJ in state-to-state litigation, there is some practice potentially extending the practice to human rights courts. Especially where the litigants are factually unequal – a state and an individual – this practice is problematic since states could more easily overcome factual failings in their pledges and evade the control of the court. Consider, for example, the potential for evading non-refoulement obligations by extending the current diplomatic assurances practice from the state-to-state context to the state-to-court context. The dual nature of assurances, as both legal and factual instruments, should be considered as addressed to distinct questions, each with its own considerations, and that we need to be more demanding about their precise legal and factual effects.

Keywords: unilateral, diplomacy, assurances, undertakings, provisional measures, interim measures

Procedia PDF Downloads 153
24544 Survival Data with Incomplete Missing Categorical Covariates

Authors: Madaki Umar Yusuf, Mohd Rizam B. Abubakar

Abstract:

The survival censored data with incomplete covariate data is a common occurrence in many studies in which the outcome is survival time. With model when the missing covariates are categorical, a useful technique for obtaining parameter estimates is the EM by the method of weights. The survival outcome for the class of generalized linear model is applied and this method requires the estimation of the parameters of the distribution of the covariates. In this paper, we propose some clinical trials with ve covariates, four of which have some missing values which clearly show that they were fully censored data.

Keywords: EM algorithm, incomplete categorical covariates, ignorable missing data, missing at random (MAR), Weibull Distribution

Procedia PDF Downloads 393
24543 Hindrances to Effective Delivery of Infrastructural Development Projects in Nigeria’s Built Environment

Authors: Salisu Gidado Dalibi, Sadiq Gumi Abubakar, JingChun Feng

Abstract:

Nigeria’s population is about 190 million and is on the increase annually making it the seventh most populated nation in the world and first in Africa. This population growth comes with its prospects, needs, and challenges especially on the existing and future infrastructure. Infrastructure refers to structures, systems, and facilities serving the economy of a country, city, town, businesses, industries, etc. These include roads, railways lines, bridges, tunnels, ports, stadiums, dams and water projects, power generation plants and distribution grids, information, and communication technology (ICT), etc. The Nigerian government embarked on several infrastructural development projects (IDPs) to address the deficit as the present infrastructure cannot cater to the needs nor sustain the country. However, delivering such IDPs have not been smooth; comes with challenges from within and outside the project; frequent delays and abandonment. Thus, affecting all the stakeholders involved. Hence, the aim of this paper is to identify and assess the factors that are hindering the effective delivery of IDPs in Nigeria’s built environment with the view to offer more insight into such factors, and ways to address them. The methodology adopted in this study involves the use of secondary sources of data from several materials (official publications, journals, newspapers, internet, etc.) were reviewed within the IDPs field by laying more emphasis on Nigeria’s cases. The hindrance factors in this regard were identified which forms the backbone of the questionnaire. A pilot survey was used to test its suitability; after which it was randomly administered to various project professionals in Nigeria’s construction industry using a 5-point Likert scale format to ascertain the impact of these hindrances. Cronbach’s Alpha reliability test, mean item score computations, relative importance indices, T-test, Chi-Square statistics were used for data analyses. The results outline the impact of various internal, external and project related factors that are hindering IDPs within Nigeria’s built environment.

Keywords: built environment, development, factors, hindrances, infrastructure, Nigeria, project

Procedia PDF Downloads 159
24542 A Study of Blockchain Oracles

Authors: Abdeljalil Beniiche

Abstract:

The limitation with smart contracts is that they cannot access external data that might be required to control the execution of business logic. Oracles can be used to provide external data to smart contracts. An oracle is an interface that delivers data from external data outside the blockchain to a smart contract to consume. Oracle can deliver different types of data depending on the industry and requirements. In this paper, we study and describe the widely used blockchain oracles. Then, we elaborate on his potential role, technical architecture, and design patterns. Finally, we discuss the human oracle and its key role in solving the truth problem by reaching a consensus about a certain inquiry and tasks.

Keywords: blockchain, oracles, oracles design, human oracles

Procedia PDF Downloads 113
24541 Physics-Informed Neural Network for Predicting Strain Demand in Inelastic Pipes under Ground Movement with Geometric and Soil Resistance Nonlinearities

Authors: Pouya Taraghi, Yong Li, Nader Yoosef-Ghodsi, Muntaseer Kainat, Samer Adeeb

Abstract:

Buried pipelines play a crucial role in the transportation of energy products such as oil, gas, and various chemical fluids, ensuring their efficient and safe distribution. However, these pipelines are often susceptible to ground movements caused by geohazards like landslides, fault movements, lateral spreading, and more. Such ground movements can lead to strain-induced failures in pipes, resulting in leaks or explosions, leading to fires, financial losses, environmental contamination, and even loss of human life. Therefore, it is essential to study how buried pipelines respond when traversing geohazard-prone areas to assess the potential impact of ground movement on pipeline design. As such, this study introduces an approach called the Physics-Informed Neural Network (PINN) to predict the strain demand in inelastic pipes subjected to permanent ground displacement (PGD). This method uses a deep learning framework that does not require training data and makes it feasible to consider more realistic assumptions regarding existing nonlinearities. It leverages the underlying physics described by differential equations to approximate the solution. The study analyzes various scenarios involving different geohazard types, PGD values, and crossing angles, comparing the predictions with results obtained from finite element methods. The findings demonstrate a good agreement between the results of the proposed method and the finite element method, highlighting its potential as a simulation-free, data-free, and meshless alternative. This study paves the way for further advancements, such as the simulation-free reliability assessment of pipes subjected to PGD, as part of ongoing research that leverages the proposed method.

Keywords: strain demand, inelastic pipe, permanent ground displacement, machine learning, physics-informed neural network

Procedia PDF Downloads 52
24540 Investigation of the Effect of Teaching Thinking and Research Lesson by Cooperative and Traditional Methods on Creativity of Sixth Grade Students

Authors: Faroogh Khakzad, Marzieh Dehghani, Elahe Hejazi

Abstract:

The present study investigates the effect of teaching a Thinking and Research lesson by cooperative and traditional methods on the creativity of sixth-grade students in Piranshahr province. The statistical society includes all the sixth-grade students of Piranshahr province. The sample of this studytable was selected by available sampling from among male elementary schools of Piranshahr. They were randomly assigned into two groups of cooperative teaching method and traditional teaching method. The design of the study is quasi-experimental with a control group. In this study, to assess students’ creativity, Abedi’s creativity questionnaire was used. Based on Cronbach’s alpha coefficient, the reliability of the factor flow was 0.74, innovation was 0.61, flexibility was 0.63, and expansion was 0.68. To analyze the data, t-test, univariate and multivariate covariance analysis were used for evaluation of the difference of means and the pretest and posttest scores. The findings of the research showed that cooperative teaching method does not significantly increase creativity (p > 0.05). Moreover, cooperative teaching method was found to have significant effect on flow factor (p < 0.05), but in innovation and expansion factors no significant effect was observed (p < 0.05).

Keywords: cooperative teaching method, traditional teaching method, creativity, flow, innovation, flexibility, expansion, thinking and research lesson

Procedia PDF Downloads 311
24539 A New Criterion Using Pose and Shape of Objects for Collision Risk Estimation

Authors: DoHyeung Kim, DaeHee Seo, ByungDoo Kim, ByungGil Lee

Abstract:

As many recent researches being implemented in aviation and maritime aspects, strong doubts have been raised concerning the reliability of the estimation of collision risk. It is shown that using position and velocity of objects can lead to imprecise results. In this paper, therefore, a new approach to the estimation of collision risks using pose and shape of objects is proposed. Simulation results are presented validating the accuracy of the new criterion to adapt to collision risk algorithm based on fuzzy logic.

Keywords: collision risk, pose, shape, fuzzy logic

Procedia PDF Downloads 516
24538 Finding Bicluster on Gene Expression Data of Lymphoma Based on Singular Value Decomposition and Hierarchical Clustering

Authors: Alhadi Bustaman, Soeganda Formalidin, Titin Siswantining

Abstract:

DNA microarray technology is used to analyze thousand gene expression data simultaneously and a very important task for drug development and test, function annotation, and cancer diagnosis. Various clustering methods have been used for analyzing gene expression data. However, when analyzing very large and heterogeneous collections of gene expression data, conventional clustering methods often cannot produce a satisfactory solution. Biclustering algorithm has been used as an alternative approach to identifying structures from gene expression data. In this paper, we introduce a transform technique based on singular value decomposition to identify normalized matrix of gene expression data followed by Mixed-Clustering algorithm and the Lift algorithm, inspired in the node-deletion and node-addition phases proposed by Cheng and Church based on Agglomerative Hierarchical Clustering (AHC). Experimental study on standard datasets demonstrated the effectiveness of the algorithm in gene expression data.

Keywords: agglomerative hierarchical clustering (AHC), biclustering, gene expression data, lymphoma, singular value decomposition (SVD)

Procedia PDF Downloads 267
24537 An Efficient Traceability Mechanism in the Audited Cloud Data Storage

Authors: Ramya P, Lino Abraham Varghese, S. Bose

Abstract:

By cloud storage services, the data can be stored in the cloud, and can be shared across multiple users. Due to the unexpected hardware/software failures and human errors, which make the data stored in the cloud be lost or corrupted easily it affected the integrity of data in cloud. Some mechanisms have been designed to allow both data owners and public verifiers to efficiently audit cloud data integrity without retrieving the entire data from the cloud server. But public auditing on the integrity of shared data with the existing mechanisms will unavoidably reveal confidential information such as identity of the person, to public verifiers. Here a privacy-preserving mechanism is proposed to support public auditing on shared data stored in the cloud. It uses group signatures to compute verification metadata needed to audit the correctness of shared data. The identity of the signer on each block in shared data is kept confidential from public verifiers, who are easily verifying shared data integrity without retrieving the entire file. But on demand, the signer of the each block is reveal to the owner alone. Group private key is generated once by the owner in the static group, where as in the dynamic group, the group private key is change when the users revoke from the group. When the users leave from the group the already signed blocks are resigned by cloud service provider instead of owner is efficiently handled by efficient proxy re-signature scheme.

Keywords: data integrity, dynamic group, group signature, public auditing

Procedia PDF Downloads 379
24536 Critical Thinking in the Moroccan Textbooks of English: Ticket to English as a Case Study

Authors: Mohsine Jebbour

Abstract:

The ultimate aim of this study was to analyze a second-year baccalaureate textbook of English to see to what extent it includes elements of critical thinking. A further purpose was to assess the extent to which the teachers’ teaching practices help students develop some degree of critical thinking. The literature on critical thinking indicated that all the writers agree that critical thinking is skilled and dispositional oriented, and most of the definitions highlight the skill and disposition to select, collect, analyze and evaluate information effectively. In this study, two instruments were used, namely content analysis and questionnaire to ensure validity and reliability. The sample of this study, on the one hand, was a second year textbook of English, namely Ticket to English. The process of collecting data was carried out through designing a checklist to analyze the textbook of English. On the other hand, high school students (second baccalaureate grade) and teachers of English constituted the second sample. Two questionnaires were administered—One was completed by 28 high school teachers (18 males and10 females), and the other was completed by 51 students (26 males and 25 females) from Fez, Morocco. The items of the questionnaire tended to elicit both qualitative and quantitative data. An attempt was made to answer two research questions. One pertained to the extent to which the textbooks of English contain critical thinking elements (Critical thinking skills and dispositions, types of questions, language learning strategies, classroom activities); the second was concerned with whether the teaching practices of teachers of English help improve students’ critical thinking. The results demonstrated that the textbooks of English include elements of critical thinking, and the teachers’ teaching practices help the students develop some degree of critical thinking. Yet, the textbooks do not include problem-solving activities and media analysis and 86% of the teacher-respondents tended to skip activities in the textbooks, mainly the units dealing with Project Work and Study Skills which are necessary for enhancing critical thinking among the students. Therefore, the textbooks need to be designed around additional activities and the teachers are required to cover the units skipped so as to make the teaching of critical thinking effective.

Keywords: critical thinking, language learning strategies, language proficiency, teaching practices

Procedia PDF Downloads 593
24535 Rodriguez Diego, Del Valle Martin, Hargreaves Matias, Riveros Jose Luis

Authors: Nathainail Bashir, Neil Anderson

Abstract:

The objective of this study site was to investigate the current state of the practice with regards to karst detection methods and recommend the best method and pattern of arrays to acquire the desire results. Proper site investigation in karst prone regions is extremely valuable in determining the location of possible voids. Two geophysical techniques were employed: multichannel analysis of surface waves (MASW) and electric resistivity tomography (ERT).The MASW data was acquired at each test location using different array lengths and different array orientations (to increase the probability of getting interpretable data in karst terrain). The ERT data were acquired using a dipole-dipole array consisting of 168 electrodes. The MASW data was interpreted (re: estimated depth to physical top of rock) and used to constrain and verify the interpretation of the ERT data. The ERT data indicates poorer quality MASW data were acquired in areas where there was significant local variation in the depth to top of rock.

Keywords: dipole-dipole, ERT, Karst terrains, MASW

Procedia PDF Downloads 304
24534 Data Science in Military Decision-Making: A Semi-Systematic Literature Review

Authors: H. W. Meerveld, R. H. A. Lindelauf

Abstract:

In contemporary warfare, data science is crucial for the military in achieving information superiority. Yet, to the authors’ knowledge, no extensive literature survey on data science in military decision-making has been conducted so far. In this study, 156 peer-reviewed articles were analysed through an integrative, semi-systematic literature review to gain an overview of the topic. The study examined to what extent literature is focussed on the opportunities or risks of data science in military decision-making, differentiated per level of war (i.e. strategic, operational, and tactical level). A relatively large focus on the risks of data science was observed in social science literature, implying that political and military policymakers are disproportionally influenced by a pessimistic view on the application of data science in the military domain. The perceived risks of data science are, however, hardly addressed in formal science literature. This means that the concerns on the military application of data science are not addressed to the audience that can actually develop and enhance data science models and algorithms. Cross-disciplinary research on both the opportunities and risks of military data science can address the observed research gaps. Considering the levels of war, relatively low attention for the operational level compared to the other two levels was observed, suggesting a research gap with reference to military operational data science. Opportunities for military data science mostly arise at the tactical level. On the contrary, studies examining strategic issues mostly emphasise the risks of military data science. Consequently, domain-specific requirements for military strategic data science applications are hardly expressed. Lacking such applications may ultimately lead to a suboptimal strategic decision in today’s warfare.

Keywords: data science, decision-making, information superiority, literature review, military

Procedia PDF Downloads 148
24533 The Examination of Prospective ICT Teachers’ Attitudes towards Application of Computer Assisted Instruction

Authors: Agâh Tuğrul Korucu, Ismail Fatih Yavuzaslan, Lale Toraman

Abstract:

Nowadays, thanks to development of technology, integration of technology into teaching and learning activities is spreading. Increasing technological literacy which is one of the expected competencies for individuals of 21st century is associated with the effective use of technology in education. The most important factor in effective use of technology in education institutions is ICT teachers. The concept of computer assisted instruction (CAI) refers to the utilization of information and communication technology as a tool aided teachers in order to make education more efficient and improve its quality in the process of educational. Teachers can use computers in different places and times according to owned hardware and software facilities and characteristics of the subject and student in CAI. Analyzing teachers’ use of computers in education is significant because teachers are the ones who manage the course and they are the most important element in comprehending the topic by students. To accomplish computer-assisted instruction efficiently is possible through having positive attitude of teachers. Determination the level of knowledge, attitude and behavior of teachers who get the professional knowledge from educational faculties and elimination of deficiencies if any are crucial when teachers are at the faculty. Therefore, the aim of this paper is to identify ICT teachers' attitudes toward computer-assisted instruction in terms of different variables. Research group consists of 200 prospective ICT teachers studying at Necmettin Erbakan University Ahmet Keleşoğlu Faculty of Education CEIT department. As data collection tool of the study; “personal information form” developed by the researchers and used to collect demographic data and "the attitude scale related to computer-assisted instruction" are used. The scale consists of 20 items. 10 of these items show positive feature, while 10 of them show negative feature. The Kaiser-Meyer-Olkin (KMO) coefficient of the scale is found 0.88 and Barlett test significance value is found 0.000. The Cronbach’s alpha reliability coefficient of the scale is found 0.93. In order to analyze the data collected by data collection tools computer-based statistical software package used; statistical techniques such as descriptive statistics, t-test, and analysis of variance are utilized. It is determined that the attitudes of prospective instructors towards computers do not differ according to their educational branches. On the other hand, the attitudes of prospective instructors who own computers towards computer-supported education are determined higher than those of the prospective instructors who do not own computers. It is established that the departments of students who previously received computer lessons do not affect this situation so much. The result is that; the computer experience affects the attitude point regarding the computer-supported education positively.

Keywords: computer based instruction, teacher candidate, attitude, technology based instruction, information and communication technologies

Procedia PDF Downloads 285
24532 Legal Regulation of Personal Information Data Transmission Risk Assessment: A Case Study of the EU’s DPIA

Authors: Cai Qianyi

Abstract:

In the midst of global digital revolution, the flow of data poses security threats that call China's existing legislative framework for protecting personal information into question. As a preliminary procedure for risk analysis and prevention, the risk assessment of personal data transmission lacks detailed guidelines for support. Existing provisions reveal unclear responsibilities for network operators and weakened rights for data subjects. Furthermore, the regulatory system's weak operability and a lack of industry self-regulation heighten data transmission hazards. This paper aims to compare the regulatory pathways for data information transmission risks between China and Europe from a legal framework and content perspective. It draws on the “Data Protection Impact Assessment Guidelines” to empower multiple stakeholders, including data processors, controllers, and subjects, while also defining obligations. In conclusion, this paper intends to solve China's digital security shortcomings by developing a more mature regulatory framework and industry self-regulation mechanisms, resulting in a win-win situation for personal data protection and the development of the digital economy.

Keywords: personal information data transmission, risk assessment, DPIA, internet service provider, personal information data transimission, risk assessment

Procedia PDF Downloads 45
24531 Wavelets Contribution on Textual Data Analysis

Authors: Habiba Ben Abdessalem

Abstract:

The emergence of giant set of textual data was the push that has encouraged researchers to invest in this field. The purpose of textual data analysis methods is to facilitate access to such type of data by providing various graphic visualizations. Applying these methods requires a corpus pretreatment step, whose standards are set according to the objective of the problem studied. This step determines the forms list contained in contingency table by keeping only those information carriers. This step may, however, lead to noisy contingency tables, so the use of wavelet denoising function. The validity of the proposed approach is tested on a text database that offers economic and political events in Tunisia for a well definite period.

Keywords: textual data, wavelet, denoising, contingency table

Procedia PDF Downloads 271
24530 Customer Churn Analysis in Telecommunication Industry Using Data Mining Approach

Authors: Burcu Oralhan, Zeki Oralhan, Nilsun Sariyer, Kumru Uyar

Abstract:

Data mining has been becoming more and more important and a wide range of applications in recent years. Data mining is the process of find hidden and unknown patterns in big data. One of the applied fields of data mining is Customer Relationship Management. Understanding the relationships between products and customers is crucial for every business. Customer Relationship Management is an approach to focus on customer relationship development, retention and increase on customer satisfaction. In this study, we made an application of a data mining methods in telecommunication customer relationship management side. This study aims to determine the customers profile who likely to leave the system, develop marketing strategies, and customized campaigns for customers. Data are clustered by applying classification techniques for used to determine the churners. As a result of this study, we will obtain knowledge from international telecommunication industry. We will contribute to the understanding and development of this subject in Customer Relationship Management.

Keywords: customer churn analysis, customer relationship management, data mining, telecommunication industry

Procedia PDF Downloads 301
24529 On Pooling Different Levels of Data in Estimating Parameters of Continuous Meta-Analysis

Authors: N. R. N. Idris, S. Baharom

Abstract:

A meta-analysis may be performed using aggregate data (AD) or an individual patient data (IPD). In practice, studies may be available at both IPD and AD level. In this situation, both the IPD and AD should be utilised in order to maximize the available information. Statistical advantages of combining the studies from different level have not been fully explored. This study aims to quantify the statistical benefits of including available IPD when conducting a conventional summary-level meta-analysis. Simulated meta-analysis were used to assess the influence of the levels of data on overall meta-analysis estimates based on IPD-only, AD-only and the combination of IPD and AD (mixed data, MD), under different study scenario. The percentage relative bias (PRB), root mean-square-error (RMSE) and coverage probability were used to assess the efficiency of the overall estimates. The results demonstrate that available IPD should always be included in a conventional meta-analysis using summary level data as they would significantly increased the accuracy of the estimates. On the other hand, if more than 80% of the available data are at IPD level, including the AD does not provide significant differences in terms of accuracy of the estimates. Additionally, combining the IPD and AD has moderating effects on the biasness of the estimates of the treatment effects as the IPD tends to overestimate the treatment effects, while the AD has the tendency to produce underestimated effect estimates. These results may provide some guide in deciding if significant benefit is gained by pooling the two levels of data when conducting meta-analysis.

Keywords: aggregate data, combined-level data, individual patient data, meta-analysis

Procedia PDF Downloads 362
24528 Analyzing On-Line Process Data for Industrial Production Quality Control

Authors: Hyun-Woo Cho

Abstract:

The monitoring of industrial production quality has to be implemented to alarm early warning for unusual operating conditions. Furthermore, identification of their assignable causes is necessary for a quality control purpose. For such tasks many multivariate statistical techniques have been applied and shown to be quite effective tools. This work presents a process data-based monitoring scheme for production processes. For more reliable results some additional steps of noise filtering and preprocessing are considered. It may lead to enhanced performance by eliminating unwanted variation of the data. The performance evaluation is executed using data sets from test processes. The proposed method is shown to provide reliable quality control results, and thus is more effective in quality monitoring in the example. For practical implementation of the method, an on-line data system must be available to gather historical and on-line data. Recently large amounts of data are collected on-line in most processes and implementation of the current scheme is feasible and does not give additional burdens to users.

Keywords: detection, filtering, monitoring, process data

Procedia PDF Downloads 546
24527 A Critical Study of the Performance of Self Compacting Concrete (SCC) Using Locally Supplied Materials in Bahrain

Authors: A. Umar, A. Tamimi

Abstract:

Development of new types of concrete with improved performance is a very important issue for the whole building industry. The development is based on the optimization of the concrete mix design, with an emphasis not only on the workability and mechanical properties but also to the durability and the reliability of the concrete structure in general. Self-compacting concrete (SCC) is a high-performance material designed to flow into formwork under its own weight and without the aid of mechanical vibration. At the same time it is cohesive enough to fill spaces of almost any size and shape without segregation or bleeding. Construction time is shorter and production of SCC is environmentally friendly (no noise, no vibration). Furthermore, SCC produces a good surface finish. Despite these advantages, SCC has not gained much local acceptance though it has been promoted in the Middle East for the last ten to twelve years. The reluctance in utilizing the advantages of SCC, in Bahrain, may be due to lack of research or published data pertaining to locally produced SCC. Therefore, there is a need to conduct studies on SCC using locally available material supplies. From the literature, it has been observed that the use of viscosity modifying admixtures (VMA), micro silica and glass fibers have proved to be very effective in stabilizing the rheological properties and the strength of fresh and hardened properties of self-compacting concrete (SCC). Therefore, in the present study, it is proposed to carry out investigations of SCC with combinations of various dosages of VMAs with and without micro silica and glass fibers and to study their influence on the properties of fresh and hardened concrete.

Keywords: self-compacting concrete, viscosity modifying admixture, micro silica, glass fibers

Procedia PDF Downloads 643
24526 A Review of Travel Data Collection Methods

Authors: Muhammad Awais Shafique, Eiji Hato

Abstract:

Household trip data is of crucial importance for managing present transportation infrastructure as well as to plan and design future facilities. It also provides basis for new policies implemented under Transportation Demand Management. The methods used for household trip data collection have changed with passage of time, starting with the conventional face-to-face interviews or paper-and-pencil interviews and reaching to the recent approach of employing smartphones. This study summarizes the step-wise evolution in the travel data collection methods. It provides a comprehensive review of the topic, for readers interested to know the changing trends in the data collection field.

Keywords: computer, smartphone, telephone, travel survey

Procedia PDF Downloads 302
24525 A Business-to-Business Collaboration System That Promotes Data Utilization While Encrypting Information on the Blockchain

Authors: Hiroaki Nasu, Ryota Miyamoto, Yuta Kodera, Yasuyuki Nogami

Abstract:

To promote Industry 4.0 and Society 5.0 and so on, it is important to connect and share data so that every member can trust it. Blockchain (BC) technology is currently attracting attention as the most advanced tool and has been used in the financial field and so on. However, the data collaboration using BC has not progressed sufficiently among companies on the supply chain of manufacturing industry that handle sensitive data such as product quality, manufacturing conditions, etc. There are two main reasons why data utilization is not sufficiently advanced in the industrial supply chain. The first reason is that manufacturing information is top secret and a source for companies to generate profits. It is difficult to disclose data even between companies with transactions in the supply chain. In the blockchain mechanism such as Bitcoin using PKI (Public Key Infrastructure), in order to confirm the identity of the company that has sent the data, the plaintext must be shared between the companies. Another reason is that the merits (scenarios) of collaboration data between companies are not specifically specified in the industrial supply chain. For these problems this paper proposes a Business to Business (B2B) collaboration system using homomorphic encryption and BC technique. Using the proposed system, each company on the supply chain can exchange confidential information on encrypted data and utilize the data for their own business. In addition, this paper considers a scenario focusing on quality data, which was difficult to collaborate because it is a top secret. In this scenario, we show a implementation scheme and a benefit of concrete data collaboration by proposing a comparison protocol that can grasp the change in quality while hiding the numerical value of quality data.

Keywords: business to business data collaboration, industrial supply chain, blockchain, homomorphic encryption

Procedia PDF Downloads 119
24524 Multivariate Assessment of Mathematics Test Scores of Students in Qatar

Authors: Ali Rashash Alzahrani, Elizabeth Stojanovski

Abstract:

Data on various aspects of education are collected at the institutional and government level regularly. In Australia, for example, students at various levels of schooling undertake examinations in numeracy and literacy as part of NAPLAN testing, enabling longitudinal assessment of such data as well as comparisons between schools and states within Australia. Another source of educational data collected internationally is via the PISA study which collects data from several countries when students are approximately 15 years of age and enables comparisons in the performance of science, mathematics and English between countries as well as ranking of countries based on performance in these standardised tests. As well as student and school outcomes based on the tests taken as part of the PISA study, there is a wealth of other data collected in the study including parental demographics data and data related to teaching strategies used by educators. Overall, an abundance of educational data is available which has the potential to be used to help improve educational attainment and teaching of content in order to improve learning outcomes. A multivariate assessment of such data enables multiple variables to be considered simultaneously and will be used in the present study to help develop profiles of students based on performance in mathematics using data obtained from the PISA study.

Keywords: cluster analysis, education, mathematics, profiles

Procedia PDF Downloads 113
24523 Dataset Quality Index:Development of Composite Indicator Based on Standard Data Quality Indicators

Authors: Sakda Loetpiparwanich, Preecha Vichitthamaros

Abstract:

Nowadays, poor data quality is considered one of the majority costs for a data project. The data project with data quality awareness almost as much time to data quality processes while data project without data quality awareness negatively impacts financial resources, efficiency, productivity, and credibility. One of the processes that take a long time is defining the expectations and measurements of data quality because the expectation is different up to the purpose of each data project. Especially, big data project that maybe involves with many datasets and stakeholders, that take a long time to discuss and define quality expectations and measurements. Therefore, this study aimed at developing meaningful indicators to describe overall data quality for each dataset to quick comparison and priority. The objectives of this study were to: (1) Develop a practical data quality indicators and measurements, (2) Develop data quality dimensions based on statistical characteristics and (3) Develop Composite Indicator that can describe overall data quality for each dataset. The sample consisted of more than 500 datasets from public sources obtained by random sampling. After datasets were collected, there are five steps to develop the Dataset Quality Index (SDQI). First, we define standard data quality expectations. Second, we find any indicators that can measure directly to data within datasets. Thirdly, each indicator aggregates to dimension using factor analysis. Next, the indicators and dimensions were weighted by an effort for data preparing process and usability. Finally, the dimensions aggregate to Composite Indicator. The results of these analyses showed that: (1) The developed useful indicators and measurements contained ten indicators. (2) the developed data quality dimension based on statistical characteristics, we found that ten indicators can be reduced to 4 dimensions. (3) The developed Composite Indicator, we found that the SDQI can describe overall datasets quality of each dataset and can separate into 3 Level as Good Quality, Acceptable Quality, and Poor Quality. The conclusion, the SDQI provide an overall description of data quality within datasets and meaningful composition. We can use SQDI to assess for all data in the data project, effort estimation, and priority. The SDQI also work well with Agile Method by using SDQI to assessment in the first sprint. After passing the initial evaluation, we can add more specific data quality indicators into the next sprint.

Keywords: data quality, dataset quality, data quality management, composite indicator, factor analysis, principal component analysis

Procedia PDF Downloads 125
24522 Predictive Analysis for Big Data: Extension of Classification and Regression Trees Algorithm

Authors: Ameur Abdelkader, Abed Bouarfa Hafida

Abstract:

Since its inception, predictive analysis has revolutionized the IT industry through its robustness and decision-making facilities. It involves the application of a set of data processing techniques and algorithms in order to create predictive models. Its principle is based on finding relationships between explanatory variables and the predicted variables. Past occurrences are exploited to predict and to derive the unknown outcome. With the advent of big data, many studies have suggested the use of predictive analytics in order to process and analyze big data. Nevertheless, they have been curbed by the limits of classical methods of predictive analysis in case of a large amount of data. In fact, because of their volumes, their nature (semi or unstructured) and their variety, it is impossible to analyze efficiently big data via classical methods of predictive analysis. The authors attribute this weakness to the fact that predictive analysis algorithms do not allow the parallelization and distribution of calculation. In this paper, we propose to extend the predictive analysis algorithm, Classification And Regression Trees (CART), in order to adapt it for big data analysis. The major changes of this algorithm are presented and then a version of the extended algorithm is defined in order to make it applicable for a huge quantity of data.

Keywords: predictive analysis, big data, predictive analysis algorithms, CART algorithm

Procedia PDF Downloads 132
24521 Canopy Temperature Acquired from Daytime and Nighttime Aerial Data as an Indicator of Trees’ Health Status

Authors: Agata Zakrzewska, Dominik Kopeć, Adrian Ochtyra

Abstract:

The growing number of new cameras, sensors, and research methods allow for a broader application of thermal data in remote sensing vegetation studies. The aim of this research was to check whether it is possible to use thermal infrared data with a spectral range (3.6-4.9 μm) obtained during the day and the night to assess the health condition of selected species of deciduous trees in an urban environment. For this purpose, research was carried out in the city center of Warsaw (Poland) in 2020. During the airborne data acquisition, thermal data, laser scanning, and orthophoto map images were collected. Synchronously with airborne data, ground reference data were obtained for 617 studied species (Acer platanoides, Acer pseudoplatanus, Aesculus hippocastanum, Tilia cordata, and Tilia × euchlora) in different health condition states. The results were as follows: (i) healthy trees are cooler than trees in poor condition and dying both in the daytime and nighttime data; (ii) the difference in the canopy temperatures between healthy and dying trees was 1.06oC of mean value on the nighttime data and 3.28oC of mean value on the daytime data; (iii) condition classes significantly differentiate on both daytime and nighttime thermal data, but only on daytime data all condition classes differed statistically significantly from each other. In conclusion, the aerial thermal data can be considered as an alternative to hyperspectral data, a method of assessing the health condition of trees in an urban environment. Especially data obtained during the day, which can differentiate condition classes better than data obtained at night. The method based on thermal infrared and laser scanning data fusion could be a quick and efficient solution for identifying trees in poor health that should be visually checked in the field.

Keywords: middle wave infrared, thermal imagery, tree discoloration, urban trees

Procedia PDF Downloads 104
24520 Statistical Analysis Approach for the e-Glassy Mortar And Radiation Shielding Behaviors Using Anova

Authors: Abadou Yacine, Faid Hayette

Abstract:

Significant investigations were performed on the use and impact on physical properties along with the mechanical strength of the recycled and reused E-glass waste powder. However, it has been modelled how recycled display e-waste glass may affect the characteristics and qualities of dune sand mortar. To be involved in this field, an investigation has been done with the substitution of dune sand for recycled E-glass waste and constant water-cement ratios. The linear relationship between the dune sand mortar and E-glass mortar mix % contributes to the model's reliability. The experimental data was exposed to regression analysis using JMP Statistics software. The regression model with one predictor presented the general form of the equation for the prediction of the five properties' characteristics of dune sand mortar from the substitution ratio of E-waste glass and curing age. The results illustrate that curing a long-term process produced an E-glass waste mortar specimen with the highest compressive strength of 68 MPa in the laboratory environment. Anova analysis indicated that the curing at long-term has the utmost importance on the sorptivity level and ultrasonic pulse velocity loss. Furthermore, the E-glass waste powder percentage has the utmost importance on the compressive strength and improvement in dynamic elasticity modulus. Besides, a significant enhancement of radiation-shielding applications.

Keywords: ANOVA analysis, E-glass waste, durability and sustainability, radiation-shielding

Procedia PDF Downloads 52
24519 Hierarchical Clustering Algorithms in Data Mining

Authors: Z. Abdullah, A. R. Hamdan

Abstract:

Clustering is a process of grouping objects and data into groups of clusters to ensure that data objects from the same cluster are identical to each other. Clustering algorithms in one of the areas in data mining and it can be classified into partition, hierarchical, density based, and grid-based. Therefore, in this paper, we do a survey and review for four major hierarchical clustering algorithms called CURE, ROCK, CHAMELEON, and BIRCH. The obtained state of the art of these algorithms will help in eliminating the current problems, as well as deriving more robust and scalable algorithms for clustering.

Keywords: clustering, unsupervised learning, algorithms, hierarchical

Procedia PDF Downloads 874
24518 A Collaborative Action Research on the Teaching of Music Learning Center in Taiwan's Preschool

Authors: Mei-Ying Liao, Lee-Ching Wei, Jung-Hsiang Tseng

Abstract:

The main purpose of this study was to explore the process of planning and execution of the music learning center in preschool. This study was conducted through a collaborative action research method. The research members included a university music professor, a teaching guide, a preschool director, and a preschool teacher, leading a class of 5-6-year-old children to participate in this study. Five teaching cycles were performed with a subject of bird. In the whole process that lasted three months, the research members would maintain the conversation, reflection, and revision repeatedly. A triangular validated method was used to collect data, including archives, interviews, seminars, observations, journals, and learning evaluations to improve research on the validity and reliability. It was found that a successful music learning center required comprehensive planning and execution. It is also important to develop good listening, singing, respect, and homing habits at the beginning of running the music learning center. By timely providing diverse musical instruments, learning materials, and activities according to the teaching goals, children’s desire to learning was highly stimulated. Besides, peer interactions improved their ensemble and problem-solving abilities. The collaborative action research enhanced the preschool teacher’s confidence and promoted professional growth of the research members.

Keywords: collaborative action research, case study, music learning center, music development

Procedia PDF Downloads 362