Search results for: LCA tools and data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 26421

Search results for: LCA tools and data

26301 The Introduction of Modern Diagnostic Techniques and It Impact on Local Garages

Authors: Mustapha Majid

Abstract:

Gone were the days when technicians/mechanics will have to spend too much time trying to identify a mechanical fault and rectify the problem. Now the emphasis is on the use of Automobile diagnosing Equipment through the use of computers and special software. An investigation conducted at Tamale Metropolis and Accra in the Northern and Greater Accra regions of Ghana, respectively. Methodology for data gathering were; questionnaires, physical observation, interviews, and newspaper. The study revealed that majority of mechanics lack computer skills which can enable them use diagnosis tools such as Exhaust Gas Analyzer, Scan Tools, Electronic Wheel Balancing machine, etc.

Keywords: diagnosing, local garages and modern garages, lack of knowledge of diagnosing posing an existential threat, training of local mechanics

Procedia PDF Downloads 121
26300 Methodology for Various Sand Cone Testing

Authors: Abel S. Huaynacho, Yoni D. Huaynacho

Abstract:

The improvement of procedure test ASTM D1556, plays an important role in the developing of testing in field to obtain a higher quality of data QA/QC. The traditional process takes a considerable amount of time for only one test. Even making various testing are tasks repeating and it takes a long time to obtain better results. Moreover, if the adequate tools the help these testing are not properly managed, the improvement in the development for various testing could be stooped. This paper presents an optimized process for various testing ASTM D1556 which uses an initial standard process to another one the uses a simpler and improved management tools.

Keywords: cone sand test, density bulk, ASTM D1556, QA/QC

Procedia PDF Downloads 105
26299 Application of Social Media for Promoting Library and Information Services: A Case Study of Library Science Professionals of India

Authors: Payel Saha

Abstract:

Social media is playing an important role for dissemination of information in society. In 21st century most people have a smart phone and used different social media tools like Facebook, Twitter, Instagram, WhatsApp, Skype etc. in day to day life. It is rapidly growing web-based tool for everyone to share thoughts, ideas and knowledge globally using internet. The study highlights the current use of social media tools for promoting library and information services of Library and Information Professionals of India, which are working in Library. The study was conducted during November, 2017. A structured questionnaire was prepared using google docs and shared using different mailing list, sent to individual email IDs and sharing with other social media tools. Only 90 responses received from the different states of India and analyzed via MS-Excel. The data receive from 17 states and 3 union territories of India; however most of the respondents has come from the states Odisha 23, Himachal Pradesh 14 and Assam 10. The results revealed that out 90 respondents 37 Female and 53 male categories and also majority of respondents 71 have come from academic library followed by special library 15, Public library 3 and corporate library 1 respondent. The study indicates that, out of 90 respondent’s majority of 53 of respondents said that their Library have a social media account while 39 of respondents have not their Library social media account. The study also inform that Facebook, YouTube, Google+, LinkedIn, Twitter and Instagram are using by the LIS professional of India and Facebook 86 was popular social media tool among the other social media tools. Furthermore, respondent reported that they are using social media tools for sharing photos of events and programs of library 72, followed by tips for using different services 64, posting of new arrivals 56, tutorials of database 35 and send brief updates to patrons 32, announcement of library holidays 22. It was also reported by respondents that they are sharing information about scholarships training programs and marketing of library events etc. The study furthermore identify that lack of time is the major problem while using social media with 53 of respondents followed by low speed of internet 35, too many social media tools to learn 17 and some 3 respondents reported that there is no problem while using social media tools. The results also revealed that, majority of the respondents reported that they are using social media tools in daily basis 71 followed by weekly basis 16. It was followed by monthly 1 respondent and other 2 of the respondents. In summary, this study is expected to be useful in further promoting the social media for dissemination of library and information services to the general public.

Keywords: application of social media, India, promoting library services, library professionals

Procedia PDF Downloads 126
26298 Generic Data Warehousing for Consumer Electronics Retail Industry

Authors: S. Habte, K. Ouazzane, P. Patel, S. Patel

Abstract:

The dynamic and highly competitive nature of the consumer electronics retail industry means that businesses in this industry are experiencing different decision making challenges in relation to pricing, inventory control, consumer satisfaction and product offerings. To overcome the challenges facing retailers and create opportunities, we propose a generic data warehousing solution which can be applied to a wide range of consumer electronics retailers with a minimum configuration. The solution includes a dimensional data model, a template SQL script, a high level architectural descriptions, ETL tool developed using C#, a set of APIs, and data access tools. It has been successfully applied by ASK Outlets Ltd UK resulting in improved productivity and enhanced sales growth.

Keywords: consumer electronics, data warehousing, dimensional data model, generic, retail industry

Procedia PDF Downloads 385
26297 Normalized P-Laplacian: From Stochastic Game to Image Processing

Authors: Abderrahim Elmoataz

Abstract:

More and more contemporary applications involve data in the form of functions defined on irregular and topologically complicated domains (images, meshs, points clouds, networks, etc). Such data are not organized as familiar digital signals and images sampled on regular lattices. However, they can be conveniently represented as graphs where each vertex represents measured data and each edge represents a relationship (connectivity or certain affinities or interaction) between two vertices. Processing and analyzing these types of data is a major challenge for both image and machine learning communities. Hence, it is very important to transfer to graphs and networks many of the mathematical tools which were initially developed on usual Euclidean spaces and proven to be efficient for many inverse problems and applications dealing with usual image and signal domains. Historically, the main tools for the study of graphs or networks come from combinatorial and graph theory. In recent years there has been an increasing interest in the investigation of one of the major mathematical tools for signal and image analysis, which are Partial Differential Equations (PDEs) variational methods on graphs. The normalized p-laplacian operator has been recently introduced to model a stochastic game called tug-of-war-game with noise. Part interest of this class of operators arises from the fact that it includes, as particular case, the infinity Laplacian, the mean curvature operator and the traditionnal Laplacian operators which was extensiveley used to models and to solve problems in image processing. The purpose of this paper is to introduce and to study a new class of normalized p-Laplacian on graphs. The introduction is based on the extension of p-harmonious function introduced in as discrete approximation for both infinity Laplacian and p-Laplacian equations. Finally, we propose to use these operators as a framework for solving many inverse problems in image processing.

Keywords: normalized p-laplacian, image processing, stochastic game, inverse problems

Procedia PDF Downloads 485
26296 Faculty Use of Geospatial Tools for Deep Learning in Science and Engineering Courses

Authors: Laura Rodriguez Amaya

Abstract:

Advances in science, technology, engineering, and mathematics (STEM) are viewed as important to countries’ national economies and their capacities to be competitive in the global economy. However, many countries experience low numbers of students entering these disciplines. To strengthen the professional STEM pipelines, it is important that students are retained in these disciplines at universities. Scholars agree that to retain students in universities’ STEM degrees, it is necessary that STEM course content shows the relevance of these academic fields to their daily lives. By increasing students’ understanding on the importance of these degrees and careers, students’ motivation to remain in these academic programs can also increase. An effective way to make STEM content relevant to students’ lives is the use of geospatial technologies and geovisualization in the classroom. The Geospatial Revolution, and the science and technology associated with it, has provided scientists and engineers with an incredible amount of data about Earth and Earth systems. This data can be used in the classroom to support instruction and make content relevant to all students. The purpose of this study was to find out the prevalence use of geospatial technologies and geovisualization as teaching practices in a USA university. The Teaching Practices Inventory survey, which is a modified version of the Carl Wieman Science Education Initiative Teaching Practices Inventory, was selected for the study. Faculty in the STEM disciplines that participated in a summer learning institute at a 4-year university in the USA constituted the population selected for the study. One of the summer learning institute’s main purpose was to have an impact on the teaching of STEM courses, particularly the teaching of gateway courses taken by many STEM majors. The sample population for the study is 97.5 of the total number of summer learning institute participants. Basic descriptive statistics through the Statistical Package for the Social Sciences (SPSS) were performed to find out: 1) The percentage of faculty using geospatial technologies and geovisualization; 2) Did the faculty associated department impact their use of geospatial tools?; and 3) Did the number of years in a teaching capacity impact their use of geospatial tools? Findings indicate that only 10 percent of respondents had used geospatial technologies, and 18 percent had used geospatial visualization. In addition, the use of geovisualization among faculty of different disciplines was broader than the use of geospatial technologies. The use of geospatial technologies concentrated in the engineering departments. Data seems to indicate the lack of incorporation of geospatial tools in STEM education. The use of geospatial tools is an effective way to engage students in deep STEM learning. Future research should look at the effect on student learning and retention in science and engineering programs when geospatial tools are used.

Keywords: engineering education, geospatial technology, geovisualization, STEM

Procedia PDF Downloads 225
26295 Indexing and Incremental Approach Using Map Reduce Bipartite Graph (MRBG) for Mining Evolving Big Data

Authors: Adarsh Shroff

Abstract:

Big data is a collection of dataset so large and complex that it becomes difficult to process using data base management tools. To perform operations like search, analysis, visualization on big data by using data mining; which is the process of extraction of patterns or knowledge from large data set. In recent years, the data mining applications become stale and obsolete over time. Incremental processing is a promising approach to refreshing mining results. It utilizes previously saved states to avoid the expense of re-computation from scratch. This project uses i2MapReduce, an incremental processing extension to Map Reduce, the most widely used framework for mining big data. I2MapReduce performs key-value pair level incremental processing rather than task level re-computation, supports not only one-step computation but also more sophisticated iterative computation, which is widely used in data mining applications, and incorporates a set of novel techniques to reduce I/O overhead for accessing preserved fine-grain computation states. To optimize the mining results, evaluate i2MapReduce using a one-step algorithm and three iterative algorithms with diverse computation characteristics for efficient mining.

Keywords: big data, map reduce, incremental processing, iterative computation

Procedia PDF Downloads 315
26294 Analysis of Factors Used by Farmers to Manage Risk: A Case Study on Italian Farms

Authors: A. Pontrandolfi, G. Enjolras, F. Capitanio

Abstract:

The study analyses the strategies Italian farmers use to cope with the risks that face their production. We specifically explore the potential and the limitations of the economic tools for climatic risk management in agriculture of the Common Agricultural Policy 2014-2020, that foresees contributions for economic tools for risk management, in relation to farms’ needs, exposure and vulnerability of agricultural areas to climatic risk. We consider at the farm level approaches to hedge risks in terms of the use of technical tools (agricultural practices, pesticides, fertilizers, irrigation) and economic/financial instruments (insurances, etc.). We develop cross-sectional and longitudinal analyses as well as analyses of correlation that underline the main differences between the way farms adapt their structure and management towards risk. The results show a preference for technical tools, despite the presence of important public aids on economic tools such as insurances. Therefore, there is a strong need for a more effective and integrated risk management policy scheme. Synergies between economic tools and risk reduction actions of a more technical, structural and management nature (production diversification, irrigation infrastructures, technological and management innovations and formation-information-consultancy, etc.) are emphasized.

Keywords: agriculture and climate change, climatic risk management, insurance schemes, farmers' approaches to risk management

Procedia PDF Downloads 307
26293 Studies on the Characterization and Machinability of Duplex Stainless Steel 2205 during Dry Turning

Authors: Gaurav D. Sonawane, Vikas G. Sargade

Abstract:

The present investigation is a study of the effect of advanced Physical Vapor Deposition (PVD) coatings on cutting temperature residual stresses and surface roughness during Duplex Stainless Steel (DSS) 2205 turning. Austenite stabilizers like nickel, manganese, and molybdenum reduced the cost of DSS. Surface Integrity (SI) plays an important role in determining corrosion resistance and fatigue life. Resistance to various types of corrosion makes DSS suitable for applications with critical environments like Heat exchangers, Desalination plants, Seawater pipes and Marine components. However, lower thermal conductivity, poor chip control and non-uniform tool wear make DSS very difficult to machine. Cemented carbide tools (M grade) were used to turn DSS in a dry environment. AlTiN and AlTiCrN coatings were deposited using advanced PVD High Pulse Impulse Magnetron Sputtering (HiPIMS) technique. Experiments were conducted with cutting speed of 100 m/min, 140 m/min and 180 m/min. A constant feed and depth of cut of 0.18 mm/rev and 0.8 mm were used, respectively. AlTiCrN coated tools followed by AlTiN coated tools outperformed uncoated tools due to properties like lower thermal conductivity, higher adhesion strength and hardness. Residual stresses were found to be compressive for all the tools used for dry turning, increasing the fatigue life of the machined component. Higher cutting temperatures were observed for coated tools due to its lower thermal conductivity, which results in very less tool wear than uncoated tools. Surface roughness with uncoated tools was found to be three times higher than coated tools due to lower coefficient of friction of coating used.

Keywords: cutting temperature, DSS2205, dry turning, HiPIMS, surface integrity

Procedia PDF Downloads 103
26292 Uncertainty Quantification of Corrosion Anomaly Length of Oil and Gas Steel Pipelines Based on Inline Inspection and Field Data

Authors: Tammeen Siraj, Wenxing Zhou, Terry Huang, Mohammad Al-Amin

Abstract:

The high resolution inline inspection (ILI) tool is used extensively in the pipeline industry to identify, locate, and measure metal-loss corrosion anomalies on buried oil and gas steel pipelines. Corrosion anomalies may occur singly (i.e. individual anomalies) or as clusters (i.e. a colony of corrosion anomalies). Although the ILI technology has advanced immensely, there are measurement errors associated with the sizes of corrosion anomalies reported by ILI tools due limitations of the tools and associated sizing algorithms, and detection threshold of the tools (i.e. the minimum detectable feature dimension). Quantifying the measurement error in the ILI data is crucial for corrosion management and developing maintenance strategies that satisfy the safety and economic constraints. Studies on the measurement error associated with the length of the corrosion anomalies (in the longitudinal direction of the pipeline) has been scarcely reported in the literature and will be investigated in the present study. Limitations in the ILI tool and clustering process can sometimes cause clustering error, which is defined as the error introduced during the clustering process by including or excluding a single or group of anomalies in or from a cluster. Clustering error has been found to be one of the biggest contributory factors for relatively high uncertainties associated with ILI reported anomaly length. As such, this study focuses on developing a consistent and comprehensive framework to quantify the measurement errors in the ILI-reported anomaly length by comparing the ILI data and corresponding field measurements for individual and clustered corrosion anomalies. The analysis carried out in this study is based on the ILI and field measurement data for a set of anomalies collected from two segments of a buried natural gas pipeline currently in service in Alberta, Canada. Data analyses showed that the measurement error associated with the ILI-reported length of the anomalies without clustering error, denoted as Type I anomalies is markedly less than that for anomalies with clustering error, denoted as Type II anomalies. A methodology employing data mining techniques is further proposed to classify the Type I and Type II anomalies based on the ILI-reported corrosion anomaly information.

Keywords: clustered corrosion anomaly, corrosion anomaly assessment, corrosion anomaly length, individual corrosion anomaly, metal-loss corrosion, oil and gas steel pipeline

Procedia PDF Downloads 283
26291 Data Privacy: Stakeholders’ Conflicts in Medical Internet of Things

Authors: Benny Sand, Yotam Lurie, Shlomo Mark

Abstract:

Medical Internet of Things (MIoT), AI, and data privacy are linked forever in a gordian knot. This paper explores the conflicts of interests between the stakeholders regarding data privacy in the MIoT arena. While patients are at home during healthcare hospitalization, MIoT can play a significant role in improving the health of large parts of the population by providing medical teams with tools for collecting data, monitoring patients’ health parameters, and even enabling remote treatment. While the amount of data handled by MIoT devices grows exponentially, different stakeholders have conflicting understandings and concerns regarding this data. The findings of the research indicate that medical teams are not concerned by the violation of data privacy rights of the patients' in-home healthcare, while patients are more troubled and, in many cases, are unaware that their data is being used without their consent. MIoT technology is in its early phases, and hence a mixed qualitative and quantitative research approach will be used, which will include case studies and questionnaires in order to explore this issue and provide alternative solutions.

Keywords: MIoT, data privacy, stakeholders, home healthcare, information privacy, AI

Procedia PDF Downloads 75
26290 Survey on Arabic Sentiment Analysis in Twitter

Authors: Sarah O. Alhumoud, Mawaheb I. Altuwaijri, Tarfa M. Albuhairi, Wejdan M. Alohaideb

Abstract:

Large-scale data stream analysis has become one of the important business and research priorities lately. Social networks like Twitter and other micro-blogging platforms hold an enormous amount of data that is large in volume, velocity and variety. Extracting valuable information and trends out of these data would aid in a better understanding and decision-making. Multiple analysis techniques are deployed for English content. Moreover, one of the languages that produce a large amount of data over social networks and is least analyzed is the Arabic language. The proposed paper is a survey on the research efforts to analyze the Arabic content in Twitter focusing on the tools and methods used to extract the sentiments for the Arabic content on Twitter.

Keywords: big data, social networks, sentiment analysis, twitter

Procedia PDF Downloads 537
26289 Optimization of Manufacturing Process Parameters: An Empirical Study from Taiwan's Tech Companies

Authors: Chao-Ton Su, Li-Fei Chen

Abstract:

The parameter design is crucial to improving the uniformity of a product or process. In the product design stage, parameter design aims to determine the optimal settings for the parameters of each element in the system, thereby minimizing the functional deviations of the product. In the process design stage, parameter design aims to determine the operating settings of the manufacturing processes so that non-uniformity in manufacturing processes can be minimized. The parameter design, trying to minimize the influence of noise on the manufacturing system, plays an important role in the high-tech companies. Taiwan has many well-known high-tech companies, which show key roles in the global economy. Quality remains the most important factor that enables these companies to sustain their competitive advantage. In Taiwan however, many high-tech companies face various quality problems. A common challenge is related to root causes and defect patterns. In the R&D stage, root causes are often unknown, and defect patterns are difficult to classify. Additionally, data collection is not easy. Even when high-volume data can be collected, data interpretation is difficult. To overcome these challenges, high-tech companies in Taiwan use more advanced quality improvement tools. In addition to traditional statistical methods and quality tools, the new trend is the application of powerful tools, such as neural network, fuzzy theory, data mining, industrial engineering, operations research, and innovation skills. In this study, several examples of optimizing the parameter settings for the manufacturing process in Taiwan’s tech companies will be presented to illustrate proposed approach’s effectiveness. Finally, a discussion of using traditional experimental design versus the proposed approach for process optimization will be made.

Keywords: quality engineering, parameter design, neural network, genetic algorithm, experimental design

Procedia PDF Downloads 121
26288 Urban Gamification: Analyzing the Effects of UFLab’s Tangible Gamified Tools in Four Hungarian Urban Public Participation Processes

Authors: Olivia Kurucz

Abstract:

Gamification is one of the outstanding new methodological possibilities of urban public participation processes to make the most informed decision possible for the future steps of urban development. This paper examines four Hungarian experimental projects in which gamified tools were applied during the public participation progresses by the Urban Future Laboratory (UFLab) research workshop of Budapest University of Technology and Economics (BUTE). The recently implemented future planning projects (in the cities of Pécel, Kistarcsa, Budapest, and Salgótarján) were initiated by various motives, but the multi-stakeholder dialogues were facilitated through physical gamified tools in all cases. Based on the urban gamification hypothesis, the use of gamified tools supported certain steps of participatory processes in several aspects: it helped to increase the attractiveness of public events, to create a more informal atmosphere, to ensure equal conditions for actors, to recall a design mindset, to bridge contrasting social or cultural differences, to fix opinions and to assist dialogue between city actors, designers, and residents. This statement is confirmed by assessing the applied tools, analyzing the case studies, and comparing them to perceive their effects and interrelations.

Keywords: experimental projects, future planning, gamification, gamified tools, Hungary, public participation, UFLab, urban gamification

Procedia PDF Downloads 109
26287 Accurate HLA Typing at High-Digit Resolution from NGS Data

Authors: Yazhi Huang, Jing Yang, Dingge Ying, Yan Zhang, Vorasuk Shotelersuk, Nattiya Hirankarn, Pak Chung Sham, Yu Lung Lau, Wanling Yang

Abstract:

Human leukocyte antigen (HLA) typing from next generation sequencing (NGS) data has the potential for applications in clinical laboratories and population genetic studies. Here we introduce a novel technique for HLA typing from NGS data based on read-mapping using a comprehensive reference panel containing all known HLA alleles and de novo assembly of the gene-specific short reads. An accurate HLA typing at high-digit resolution was achieved when it was tested on publicly available NGS data, outperforming other newly-developed tools such as HLAminer and PHLAT.

Keywords: human leukocyte antigens, next generation sequencing, whole exome sequencing, HLA typing

Procedia PDF Downloads 627
26286 A Web Service-Based Framework for Mining E-Learning Data

Authors: Felermino D. M. A. Ali, S. C. Ng

Abstract:

E-learning is an evolutionary form of distance learning and has become better over time as new technologies emerged. Today, efforts are still being made to embrace E-learning systems with emerging technologies in order to make them better. Among these advancements, Educational Data Mining (EDM) is one that is gaining a huge and increasing popularity due to its wide application for improving the teaching-learning process in online practices. However, even though EDM promises to bring many benefits to educational industry in general and E-learning environments in particular, its principal drawback is the lack of easy to use tools. The current EDM tools usually require users to have some additional technical expertise to effectively perform EDM tasks. Thus, in response to these limitations, this study intends to design and implement an EDM application framework which aims at automating and simplify the development of EDM in E-learning environment. The application framework introduces a Service-Oriented Architecture (SOA) that hides the complexity of technical details and enables users to perform EDM in an automated fashion. The framework was designed based on abstraction, extensibility, and interoperability principles. The framework implementation was made up of three major modules. The first module provides an abstraction for data gathering, which was done by extending Moodle LMS (Learning Management System) source code. The second module provides data mining methods and techniques as services; it was done by converting Weka API into a set of Web services. The third module acts as an intermediary between the first two modules, it contains a user-friendly interface that allows dynamically locating data provider services, and running knowledge discovery tasks on data mining services. An experiment was conducted to evaluate the overhead of the proposed framework through a combination of simulation and implementation. The experiments have shown that the overhead introduced by the SOA mechanism is relatively small, therefore, it has been concluded that a service-oriented architecture can be effectively used to facilitate educational data mining in E-learning environments.

Keywords: educational data mining, e-learning, distributed data mining, moodle, service-oriented architecture, Weka

Procedia PDF Downloads 218
26285 Multidirectional Product Support System for Decision Making in Textile Industry Using Collaborative Filtering Methods

Authors: A. Senthil Kumar, V. Murali Bhaskaran

Abstract:

In the information technology ground, people are using various tools and software for their official use and personal reasons. Nowadays, people are worrying to choose data accessing and extraction tools at the time of buying and selling their products. In addition, worry about various quality factors such as price, durability, color, size, and availability of the product. The main purpose of the research study is to find solutions to these unsolved existing problems. The proposed algorithm is a Multidirectional Rank Prediction (MDRP) decision making algorithm in order to take an effective strategic decision at all the levels of data extraction, uses a real time textile dataset and analyzes the results. Finally, the results are obtained and compared with the existing measurement methods such as PCC, SLCF, and VSS. The result accuracy is higher than the existing rank prediction methods.

Keywords: Knowledge Discovery in Database (KDD), Multidirectional Rank Prediction (MDRP), Pearson’s Correlation Coefficient (PCC), VSS (Vector Space Similarity)

Procedia PDF Downloads 254
26284 Readiness of Thai Restaurant in Bangkok in Applying for Certification of Halal Food Services Standard for Tourism

Authors: Pongsiri Kingkan

Abstract:

This research aims to study the Readiness of Thai Restaurant in Bangkok in Applying for Certification of Halal Food Services Standard for Tourism. This research was conduct by using mix methodology; both quantitative and qualitative data were used. 420 questionnaires were used as tools to collected data from the samples, the restaurant employees. The results were divided into two parts, the demographic data and the Readiness of Thai Restaurant in Bangkok in Applying for Certification of Halal Food Services Standard for Tourism. The majority of samples are single female age between 18–30 years old, who earn about 282.40 US dollars a month. The result of Thai restaurant readiness study demonstrated that readiness in foods and restaurant operating processes were scored at the lowest level. Readiness in social responsibility, food contact persons and food materials were rated at the low level. The readiness of utensils and kitchen tools, waste management, environmental management, and the availability of space to implement the establishment of halal food were scored at the average level. Location readiness, foods service safety and the relationship with the local community were rated at high level. But interestingly there is none of them rated at the highest level.

Keywords: availability, Bangkok, halal, Thai restaurant, readiness

Procedia PDF Downloads 289
26283 Systematic Review of Misconceptions: Tools for Diagnostics and Remediation Models for Misconceptions in Physics

Authors: Muhammad Iqbal, Edi Istiyono

Abstract:

Misconceptions are one of the problems in physics learning where students' understanding is not in line with scientific theory. The aim of this research is to find diagnostic tools to identify misconceptions and how to remediate physics misconceptions. In this research, the articles that will be reviewed come from the Scopus database related to physics misconceptions from 2013-2023. The articles obtained from the Scopus database were then selected according to the Prisma model, so 29 articles were obtained that focused on discussing physics misconceptions, especially regarding diagnostic tools and remediation methods. Currently, the most widely used diagnostic tool is the four-tier test, which is able to measure students' misconceptions in depth by knowing whether students are guessing or not and from then on, there is also a trend toward five-tier diagnostic tests with additional sources of information obtained. So that the origin of students' misconceptions is known. There are several ways to remediate student misconceptions, namely 11 ways and one of the methods used is digital practicum so that abstract things can be visualized into real ones. This research is limited to knowing what tools are used to diagnose and remediate misconceptions, so it is not yet known how big the effect of remediation methods is on misconceptions. The researcher recommends that in the future further research can be carried out to find out the most appropriate remediation method for remediating student misconceptions.

Keywords: misconception, remediation, systematic review, tools

Procedia PDF Downloads 5
26282 Spatial Information and Urbanizing Futures

Authors: Mohammad Talei, Neda Ranjbar Nosheri, Reza Kazemi Gorzadini

Abstract:

Today municipalities are searching for the new tools for increasing the public participation in different levels of urban planning. This approach of urban planning involves the community in planning process using participatory approaches instead of the long traditional top-down planning methods. These tools can be used to obtain the particular problems of urban furniture form the residents’ point of view. One of the tools that is designed with this goal is public participation GIS (PPGIS) that enables citizen to record and following up their feeling and spatial knowledge regarding main problems of the city, specifically urban furniture, in the form of maps. However, despite the good intentions of PPGIS, its practical implementation in developing countries faces many problems including the lack of basic supporting infrastructure and services and unavailability of sophisticated public participatory models. In this research we develop a PPGIS using of Web 2 to collect voluntary geodataand to perform spatial analysis based on Spatial OnLine Analytical Processing (SOLAP) and Spatial Data Mining (SDM). These tools provide urban planners with proper informationregarding the type, spatial distribution and the clusters of reported problems. This system is implemented in a case study area in Tehran, Iran and the challenges to make it applicable and its potential for real urban planning have been evaluated. It helps decision makers to better understand, plan and allocate scarce resources for providing most requested urban furniture.

Keywords: PPGIS, spatial information, urbanizing futures, urban planning

Procedia PDF Downloads 694
26281 Assessment of Procurement-Demand of Milk Plant Using Quality Control Tools: A Case Study

Authors: Jagdeep Singh, Prem Singh

Abstract:

Milk is considered as an essential and complete food. The present study was conducted at Milk Plant Mohali especially in reference to the procurement section where the cash inflow was maximum, with the objective to achieve higher productivity and reduce wastage of milk. In milk plant it was observed that during the month of Jan-2014 to March-2014 the average procurement of milk was Rs. 4, 19, 361 liter per month and cost of procurement of milk is Rs 35/- per liter. The total cost of procurement thereby equal to Rs. 1crore 46 lakh per month, but there was mismatch in procurement-production of milk, which leads to an average loss of Rs. 12, 94, 405 per month. To solve the procurement-production problem Quality Control Tools like brainstorming, Flow Chart, Cause effect diagram and Pareto analysis are applied wherever applicable. With the successful implementation of Quality Control tools an average saving of Rs. 4, 59, 445 per month is done.

Keywords: milk, procurement-demand, quality control tools,

Procedia PDF Downloads 490
26280 Analyzing On-Line Process Data for Industrial Production Quality Control

Authors: Hyun-Woo Cho

Abstract:

The monitoring of industrial production quality has to be implemented to alarm early warning for unusual operating conditions. Furthermore, identification of their assignable causes is necessary for a quality control purpose. For such tasks many multivariate statistical techniques have been applied and shown to be quite effective tools. This work presents a process data-based monitoring scheme for production processes. For more reliable results some additional steps of noise filtering and preprocessing are considered. It may lead to enhanced performance by eliminating unwanted variation of the data. The performance evaluation is executed using data sets from test processes. The proposed method is shown to provide reliable quality control results, and thus is more effective in quality monitoring in the example. For practical implementation of the method, an on-line data system must be available to gather historical and on-line data. Recently large amounts of data are collected on-line in most processes and implementation of the current scheme is feasible and does not give additional burdens to users.

Keywords: detection, filtering, monitoring, process data

Procedia PDF Downloads 518
26279 Automating Test Activities: Test Cases Creation, Test Execution, and Test Reporting with Multiple Test Automation Tools

Authors: Loke Mun Sei

Abstract:

Software testing has become a mandatory process in assuring the software product quality. Hence, test management is needed in order to manage the test activities conducted in the software test life cycle. This paper discusses on the challenges faced in the software test life cycle, and how the test processes and test activities, mainly on test cases creation, test execution, and test reporting is being managed and automated using several test automation tools, i.e. Jira, Robot Framework, and Jenkins.

Keywords: test automation tools, test case, test execution, test reporting

Procedia PDF Downloads 545
26278 Timing and Noise Data Mining Algorithm and Software Tool in Very Large Scale Integration (VLSI) Design

Authors: Qing K. Zhu

Abstract:

Very Large Scale Integration (VLSI) design becomes very complex due to the continuous integration of millions of gates in one chip based on Moore’s law. Designers have encountered numerous report files during design iterations using timing and noise analysis tools. This paper presented our work using data mining techniques combined with HTML tables to extract and represent critical timing/noise data. When we apply this data-mining tool in real applications, the running speed is important. The software employs table look-up techniques in the programming for the reasonable running speed based on performance testing results. We added several advanced features for the application in one industry chip design.

Keywords: VLSI design, data mining, big data, HTML forms, web, VLSI, EDA, timing, noise

Procedia PDF Downloads 225
26277 Data Mining As A Tool For Knowledge Management: A Review

Authors: Maram Saleh

Abstract:

Knowledge has become an essential resource in today’s economy and become the most important asset of maintaining competition advantage in organizations. The importance of knowledge has made organizations to manage their knowledge assets and resources through all multiple knowledge management stages such as: Knowledge Creation, knowledge storage, knowledge sharing and knowledge use. Researches on data mining are continues growing over recent years on both business and educational fields. Data mining is one of the most important steps of the knowledge discovery in databases process aiming to extract implicit, unknown but useful knowledge and it is considered as significant subfield in knowledge management. Data miming have the great potential to help organizations to focus on extracting the most important information on their data warehouses. Data mining tools and techniques can predict future trends and behaviors, allowing businesses to make proactive, knowledge-driven decisions. This review paper explores the applications of data mining techniques in supporting knowledge management process as an effective knowledge discovery technique. In this paper, we identify the relationship between data mining and knowledge management, and then focus on introducing some application of date mining techniques in knowledge management for some real life domains.

Keywords: Data Mining, Knowledge management, Knowledge discovery, Knowledge creation.

Procedia PDF Downloads 179
26276 The Impact of Interrelationship between Business Intelligence and Knowledge Management on Decision Making Process: An Empirical Investigation of Banking Sector in Jordan

Authors: Issa M. Shehabat, Huda F. Y. Nimri

Abstract:

This paper aims to study the relationship between knowledge management in its processes, including knowledge creation, knowledge sharing, knowledge organization, and knowledge application, and business intelligence tools, including OLAP, data mining, and data warehouse, and their impact on the decision-making process in the banking sector in Jordan. A total of 200 questionnaires were distributed to the sample of the study. The study hypotheses were tested using the statistical package SPSS. Study findings suggest that decision-making processes were positively related to knowledge management processes. Additionally, the components of business intelligence had a positive impact on decision-making. The study recommended conducting studies similar to this study in other sectors such as the industrial, telecommunications, and service sectors to contribute to enhancing understanding of the role of the knowledge management processes and business intelligence tools.

Keywords: business intelligence, knowledge management, decision making, Jordan, banking sector

Procedia PDF Downloads 107
26275 The Impact of Using Technology Tools on Preparing English Language Learners for the 21st Century

Authors: Ozlem Kaya

Abstract:

21st-century learners are energetic and tech-savvy, and the skills and the knowledge required in this century are complex and challenging. Therefore, teachers need to find new ways to appeal to the needs and interests of their students and meet the demands of the 21st century at the same time. One way to do so in English language learning has been to incorporate various technology tools into classroom practices. Although teachers think these practices are effective and their students enjoy them, students may have different perceptions. To find out what students think about the use of technology tools in terms of developing 21st-century skills and knowledge, this study was conducted at Anadolu University School of Foreign Languages. A questionnaire was administered to 40 students at elementary level. Afterward, semi-structured interviews were held with 8 students to provide deeper insight into their perceptions. The details of the findings of the study will be presented and discussed during the presentation.

Keywords: 21st century skills, technology tools, perception, English Language Learning

Procedia PDF Downloads 261
26274 Drawing, Design and Building Information Modelling (BIM): Embedding Advanced Digital Tools in the Academy Programs for Building Engineers and Architects

Authors: Vittorio Caffi, Maria Pignataro, Antonio Cosimo Devito, Marco Pesenti

Abstract:

This paper deals with the integration of advanced digital design and modelling tools and methodologies, known as Building Information Modelling, into the traditional Academy educational programs for building engineers and architects. Nowadays, the challenge the Academy has to face is to present the new tools and their features to the pupils, making sure they acquire the proper skills in order to leverage the potential they offer also for the other courses embedded in the educational curriculum. The syllabus here presented refers to the “Drawing for building engineering”, “2D and 3D laboratory” and “3D modelling” curricula of the MSc in Building Engineering of the Politecnico di Milano. Such topics, included since the first year in the MSc program, are fundamental to give the students the instruments to master the complexity of an architectural or building engineering project with digital tools, so as to represent it in its various forms.

Keywords: BIM, BIM curricula, computational design, digital modelling

Procedia PDF Downloads 638
26273 Advances in Design Decision Support Tools for Early-stage Energy-Efficient Architectural Design: A Review

Authors: Maryam Mohammadi, Mohammadjavad Mahdavinejad, Mojtaba Ansari

Abstract:

The main driving force for increasing movement towards the design of High-Performance Buildings (HPB) are building codes and rating systems that address the various components of the building and their impact on the environment and energy conservation through various methods like prescriptive methods or simulation-based approaches. The methods and tools developed to meet these needs, which are often based on building performance simulation tools (BPST), have limitations in terms of compatibility with the integrated design process (IDP) and HPB design, as well as use by architects in the early stages of design (when the most important decisions are made). To overcome these limitations in recent years, efforts have been made to develop Design Decision Support Systems, which are often based on artificial intelligence. Numerous needs and steps for designing and developing a Decision Support System (DSS), which complies with the early stages of energy-efficient architecture design -consisting of combinations of different methods in an integrated package- have been listed in the literature. While various review studies have been conducted in connection with each of these techniques (such as optimizations, sensitivity and uncertainty analysis, etc.) and their integration of them with specific targets; this article is a critical and holistic review of the researches which leads to the development of applicable systems or introduction of a comprehensive framework for developing models complies with the IDP. Information resources such as Science Direct and Google Scholar are searched using specific keywords and the results are divided into two main categories: Simulation-based DSSs and Meta-simulation-based DSSs. The strengths and limitations of different models are highlighted, two general conceptual models are introduced for each category and the degree of compliance of these models with the IDP Framework is discussed. The research shows movement towards Multi-Level of Development (MOD) models, well combined with early stages of integrated design (schematic design stage and design development stage), which are heuristic, hybrid and Meta-simulation-based, relies on Big-real Data (like Building Energy Management Systems Data or Web data). Obtaining, using and combining of these data with simulation data to create models with higher uncertainty, more dynamic and more sensitive to context and culture models, as well as models that can generate economy-energy-efficient design scenarios using local data (to be more harmonized with circular economy principles), are important research areas in this field. The results of this study are a roadmap for researchers and developers of these tools.

Keywords: integrated design process, design decision support system, meta-simulation based, early stage, big data, energy efficiency

Procedia PDF Downloads 139
26272 Analysis of Tools for Revitalization and Rehabilitation of Brownfields

Authors: Jiří Kugl

Abstract:

Typology and specific opportunities of brownfield revitalization are already largely described. Challenges and opportunities that brownfields represent have been adequately studied and presented, as well as specific ways in which these areas can be used or how they are used abroad. In other words, the questions why (revitalize brownfields) and what (we should do with them) are satisfactorily answered, but the question how (we can work with them) is not. This work will focus on answering this question, which will deal with tools that enable the revitalization and rehabilitation projects in the area. Tools can be divided, for example in terms of spatial planning and urban design, from an environmental perspective, from the perspective of cultural heritage protection and from the perspective of investment opportunities. The result is that the issue of brownfields is handled by numerous institutions and instruments. The aim of this paper is to identify, classify and analyze these instruments. Paper will study instruments from other countries with long-term experience with this issue (eg. France, Great Britain, USA, Germany, Denmark, Czech Republic) and analyse their contribution and the feasibility of their implementation in other countries.

Keywords: brownfields, revitalization, rehabilitation, tools, urban planning

Procedia PDF Downloads 208