Search results for: LCA tools and data
27340 Indexing and Incremental Approach Using Map Reduce Bipartite Graph (MRBG) for Mining Evolving Big Data
Authors: Adarsh Shroff
Abstract:
Big data is a collection of dataset so large and complex that it becomes difficult to process using data base management tools. To perform operations like search, analysis, visualization on big data by using data mining; which is the process of extraction of patterns or knowledge from large data set. In recent years, the data mining applications become stale and obsolete over time. Incremental processing is a promising approach to refreshing mining results. It utilizes previously saved states to avoid the expense of re-computation from scratch. This project uses i2MapReduce, an incremental processing extension to Map Reduce, the most widely used framework for mining big data. I2MapReduce performs key-value pair level incremental processing rather than task level re-computation, supports not only one-step computation but also more sophisticated iterative computation, which is widely used in data mining applications, and incorporates a set of novel techniques to reduce I/O overhead for accessing preserved fine-grain computation states. To optimize the mining results, evaluate i2MapReduce using a one-step algorithm and three iterative algorithms with diverse computation characteristics for efficient mining.Keywords: big data, map reduce, incremental processing, iterative computation
Procedia PDF Downloads 35027339 Readiness of Thai Restaurant in Bangkok in Applying for Certification of Halal Food Services Standard for Tourism
Authors: Pongsiri Kingkan
Abstract:
This research aims to study the Readiness of Thai Restaurant in Bangkok in Applying for Certification of Halal Food Services Standard for Tourism. This research was conduct by using mix methodology; both quantitative and qualitative data were used. 420 questionnaires were used as tools to collected data from the samples, the restaurant employees. The results were divided into two parts, the demographic data and the Readiness of Thai Restaurant in Bangkok in Applying for Certification of Halal Food Services Standard for Tourism. The majority of samples are single female age between 18–30 years old, who earn about 282.40 US dollars a month. The result of Thai restaurant readiness study demonstrated that readiness in foods and restaurant operating processes were scored at the lowest level. Readiness in social responsibility, food contact persons and food materials were rated at the low level. The readiness of utensils and kitchen tools, waste management, environmental management, and the availability of space to implement the establishment of halal food were scored at the average level. Location readiness, foods service safety and the relationship with the local community were rated at high level. But interestingly there is none of them rated at the highest level.Keywords: availability, Bangkok, halal, Thai restaurant, readiness
Procedia PDF Downloads 31627338 Survey on Arabic Sentiment Analysis in Twitter
Authors: Sarah O. Alhumoud, Mawaheb I. Altuwaijri, Tarfa M. Albuhairi, Wejdan M. Alohaideb
Abstract:
Large-scale data stream analysis has become one of the important business and research priorities lately. Social networks like Twitter and other micro-blogging platforms hold an enormous amount of data that is large in volume, velocity and variety. Extracting valuable information and trends out of these data would aid in a better understanding and decision-making. Multiple analysis techniques are deployed for English content. Moreover, one of the languages that produce a large amount of data over social networks and is least analyzed is the Arabic language. The proposed paper is a survey on the research efforts to analyze the Arabic content in Twitter focusing on the tools and methods used to extract the sentiments for the Arabic content on Twitter.Keywords: big data, social networks, sentiment analysis, twitter
Procedia PDF Downloads 57627337 Project Management Tools within SAP S/4 Hana Program Environment
Authors: Jagoda Bruni, Jan Müller-Lucanus, Gernot Stöger-Knes
Abstract:
The purpose of this article is to demonstrate modern project management approaches in the SAP S/R Hana surrounding a programming environment composed of multiple focus-diversified projects. We would like to propose innovative and goal-oriented management standards based on the specificity of the SAP transformations and customer-driven expectations. Due to the regular sprint-based controlling and management tools' application, it has been data-proven that extensive analysis of productive hours of the employees as much as a thorough review of the project progress (per GAP, per business process, and per Lot) within the whole program, can have a positive impact on customer satisfaction and improvement for projects' budget. This has been a collaborative study based on real-life experience and measurements in collaboration with our customers.Keywords: project management, program management, SAP, controlling
Procedia PDF Downloads 9127336 The Effectiveness of Exchange of Tacit and Explicit Knowledge Using Digital and Face to Face Sharing
Authors: Delio I. Castaneda, Paul Toulson
Abstract:
The purpose of this study was to investigate the knowledge sharing effectiveness of two types of knowledge, tacit and explicit, depending on two channels: face to face or digital. Participants were 217 knowledge workers in New Zealand and researchers who attended a knowledge management conference in the United Kingdom. In the study, it was found that digital tools are effective to share explicit knowledge. In addition, digital tools that facilitated dialogue were effective to share tacit knowledge. It was also found that face to face communication was an effective way to share tacit and explicit knowledge. Results of this study contribute to clarify in what cases digital tools are effective to share tacit knowledge. Additionally, even though explicit knowledge can be easily shared using digital tools, this type of knowledge is also possible to be shared through dialogue. Result of this study may support practitioners to redesign programs and activities based on knowledge sharing to make strategies more effective.Keywords: digital knowledge, explicit knowledge, knowledge sharing, tacit knowledge
Procedia PDF Downloads 25527335 Evaluation of Free Technologies as Tools for Business Process Management
Authors: Julio Sotomayor, Daniel Yucra, Jorge Mayhuasca
Abstract:
The article presents an evaluation of free technologies for business process automation, with emphasis only on tools compatible with the general public license (GPL). The compendium of technologies was based on promoting a service-oriented enterprise architecture (SOA) and the establishment of a business process management system (BPMS). The methodology for the selection of tools was Agile UP. This proposal allows businesses to achieve technological sovereignty and independence, in addition to the promotion of service orientation and the development of free software based on components.Keywords: BPM, BPMS suite, open-source software, SOA, enterprise architecture, business process management
Procedia PDF Downloads 28827334 Study on Evaluating the Utilization of Social Media Tools (SMT) in Collaborative Learning Case Study: Faculty of Medicine, King Khalid University
Authors: Vasanthi Muniasamy, Intisar Magboul Ejalani, M.Anandhavalli, K. Gauthaman
Abstract:
Social Media (SM) are websites increasingly popular and built to allow people to express themselves and to interact socially with others. Most SMT are dominated by youth particularly college students. The proliferation of popular social media tools, which can accessed from any communication devices has become pervasive in the lives of today’s student life. Connecting traditional education to social media tools are a relatively new era and any collaborative tool could be used for learning activities. This study focuses (i) how the social media tools are useful for the learning activities of the students of faculty of medicine in King Khalid University (ii) whether the social media affects the collaborative learning with interaction among students, among course instructor, their engagement, perceived ease of use and perceived ease of usefulness (TAM) (iii) overall, the students satisfy with this collaborative learning through Social media.Keywords: social media, Web 2.0, perceived ease of use, perceived usefulness, collaborative Learning
Procedia PDF Downloads 50827333 Data Privacy: Stakeholders’ Conflicts in Medical Internet of Things
Authors: Benny Sand, Yotam Lurie, Shlomo Mark
Abstract:
Medical Internet of Things (MIoT), AI, and data privacy are linked forever in a gordian knot. This paper explores the conflicts of interests between the stakeholders regarding data privacy in the MIoT arena. While patients are at home during healthcare hospitalization, MIoT can play a significant role in improving the health of large parts of the population by providing medical teams with tools for collecting data, monitoring patients’ health parameters, and even enabling remote treatment. While the amount of data handled by MIoT devices grows exponentially, different stakeholders have conflicting understandings and concerns regarding this data. The findings of the research indicate that medical teams are not concerned by the violation of data privacy rights of the patients' in-home healthcare, while patients are more troubled and, in many cases, are unaware that their data is being used without their consent. MIoT technology is in its early phases, and hence a mixed qualitative and quantitative research approach will be used, which will include case studies and questionnaires in order to explore this issue and provide alternative solutions.Keywords: MIoT, data privacy, stakeholders, home healthcare, information privacy, AI
Procedia PDF Downloads 10227332 The Impact of Interrelationship between Business Intelligence and Knowledge Management on Decision Making Process: An Empirical Investigation of Banking Sector in Jordan
Authors: Issa M. Shehabat, Huda F. Y. Nimri
Abstract:
This paper aims to study the relationship between knowledge management in its processes, including knowledge creation, knowledge sharing, knowledge organization, and knowledge application, and business intelligence tools, including OLAP, data mining, and data warehouse, and their impact on the decision-making process in the banking sector in Jordan. A total of 200 questionnaires were distributed to the sample of the study. The study hypotheses were tested using the statistical package SPSS. Study findings suggest that decision-making processes were positively related to knowledge management processes. Additionally, the components of business intelligence had a positive impact on decision-making. The study recommended conducting studies similar to this study in other sectors such as the industrial, telecommunications, and service sectors to contribute to enhancing understanding of the role of the knowledge management processes and business intelligence tools.Keywords: business intelligence, knowledge management, decision making, Jordan, banking sector
Procedia PDF Downloads 14427331 Measurement Tools of the Maturity Model for IT Service Outsourcing in Higher Education Institutions
Authors: Victoriano Valencia García, Luis Usero Aragonés, Eugenio J. Fernández Vicente
Abstract:
Nowadays, the successful implementation of ICTs is vital for almost any kind of organization. Good governance and ICT management are essential for delivering value, managing technological risks, managing resources and performance measurement. In addition, outsourcing is a strategic IT service solution which complements IT services provided internally in organizations. This paper proposes the measurement tools of a new holistic maturity model based on standards ISO/IEC 20000 and ISO/IEC 38500, and the frameworks and best practices of ITIL and COBIT, with a specific focus on IT outsourcing. These measurement tools allow independent validation and practical application in the field of higher education, using a questionnaire, metrics tables, and continuous improvement plan tables as part of the measurement process. Guidelines and standards are proposed in the model for facilitating adaptation to universities and achieving excellence in the outsourcing of IT services.Keywords: IT governance, IT management, IT services, outsourcing, maturity model, measurement tools
Procedia PDF Downloads 59127330 A Web Service-Based Framework for Mining E-Learning Data
Authors: Felermino D. M. A. Ali, S. C. Ng
Abstract:
E-learning is an evolutionary form of distance learning and has become better over time as new technologies emerged. Today, efforts are still being made to embrace E-learning systems with emerging technologies in order to make them better. Among these advancements, Educational Data Mining (EDM) is one that is gaining a huge and increasing popularity due to its wide application for improving the teaching-learning process in online practices. However, even though EDM promises to bring many benefits to educational industry in general and E-learning environments in particular, its principal drawback is the lack of easy to use tools. The current EDM tools usually require users to have some additional technical expertise to effectively perform EDM tasks. Thus, in response to these limitations, this study intends to design and implement an EDM application framework which aims at automating and simplify the development of EDM in E-learning environment. The application framework introduces a Service-Oriented Architecture (SOA) that hides the complexity of technical details and enables users to perform EDM in an automated fashion. The framework was designed based on abstraction, extensibility, and interoperability principles. The framework implementation was made up of three major modules. The first module provides an abstraction for data gathering, which was done by extending Moodle LMS (Learning Management System) source code. The second module provides data mining methods and techniques as services; it was done by converting Weka API into a set of Web services. The third module acts as an intermediary between the first two modules, it contains a user-friendly interface that allows dynamically locating data provider services, and running knowledge discovery tasks on data mining services. An experiment was conducted to evaluate the overhead of the proposed framework through a combination of simulation and implementation. The experiments have shown that the overhead introduced by the SOA mechanism is relatively small, therefore, it has been concluded that a service-oriented architecture can be effectively used to facilitate educational data mining in E-learning environments.Keywords: educational data mining, e-learning, distributed data mining, moodle, service-oriented architecture, Weka
Procedia PDF Downloads 23627329 Accurate HLA Typing at High-Digit Resolution from NGS Data
Authors: Yazhi Huang, Jing Yang, Dingge Ying, Yan Zhang, Vorasuk Shotelersuk, Nattiya Hirankarn, Pak Chung Sham, Yu Lung Lau, Wanling Yang
Abstract:
Human leukocyte antigen (HLA) typing from next generation sequencing (NGS) data has the potential for applications in clinical laboratories and population genetic studies. Here we introduce a novel technique for HLA typing from NGS data based on read-mapping using a comprehensive reference panel containing all known HLA alleles and de novo assembly of the gene-specific short reads. An accurate HLA typing at high-digit resolution was achieved when it was tested on publicly available NGS data, outperforming other newly-developed tools such as HLAminer and PHLAT.Keywords: human leukocyte antigens, next generation sequencing, whole exome sequencing, HLA typing
Procedia PDF Downloads 66327328 Organisational Blogging: Reviewing Its Effectiveness as an Organisational Learning Tool
Authors: Gavin J. Baxter, Mark H. Stansfield
Abstract:
This paper reviews the internal use of blogs and their potential effectiveness as organisational learning tools. Prior to and since the emergence of the concept of ‘Enterprise 2.0’ there still remains a lack of empirical evidence associated with how organisations are applying social media tools and whether they are effective towards supporting organisational learning. Surprisingly, blogs, one of the more traditional social media tools, still remains under-researched in the context of ‘Enterprise 2.0’ and organisational learning. The aim of this paper is to identify the theoretical linkage between blogs and organisational learning in addition to reviewing prior research on organisational blogging with a view towards exploring why this area remains under-researched and identifying what needs to be done to try and move the area forward. Through a review of the literature, one of the principal findings of this paper is that organisational blogs, dependent on their use, do have a mutual compatibility with the interpretivist aspect of organisational learning. This paper further advocates that further empirical work in this subject area is required to substantiate this theoretical assumption.Keywords: Enterprise 2.0, blogs, organisational learning, social media tools
Procedia PDF Downloads 28627327 Advances in Design Decision Support Tools for Early-stage Energy-Efficient Architectural Design: A Review
Authors: Maryam Mohammadi, Mohammadjavad Mahdavinejad, Mojtaba Ansari
Abstract:
The main driving force for increasing movement towards the design of High-Performance Buildings (HPB) are building codes and rating systems that address the various components of the building and their impact on the environment and energy conservation through various methods like prescriptive methods or simulation-based approaches. The methods and tools developed to meet these needs, which are often based on building performance simulation tools (BPST), have limitations in terms of compatibility with the integrated design process (IDP) and HPB design, as well as use by architects in the early stages of design (when the most important decisions are made). To overcome these limitations in recent years, efforts have been made to develop Design Decision Support Systems, which are often based on artificial intelligence. Numerous needs and steps for designing and developing a Decision Support System (DSS), which complies with the early stages of energy-efficient architecture design -consisting of combinations of different methods in an integrated package- have been listed in the literature. While various review studies have been conducted in connection with each of these techniques (such as optimizations, sensitivity and uncertainty analysis, etc.) and their integration of them with specific targets; this article is a critical and holistic review of the researches which leads to the development of applicable systems or introduction of a comprehensive framework for developing models complies with the IDP. Information resources such as Science Direct and Google Scholar are searched using specific keywords and the results are divided into two main categories: Simulation-based DSSs and Meta-simulation-based DSSs. The strengths and limitations of different models are highlighted, two general conceptual models are introduced for each category and the degree of compliance of these models with the IDP Framework is discussed. The research shows movement towards Multi-Level of Development (MOD) models, well combined with early stages of integrated design (schematic design stage and design development stage), which are heuristic, hybrid and Meta-simulation-based, relies on Big-real Data (like Building Energy Management Systems Data or Web data). Obtaining, using and combining of these data with simulation data to create models with higher uncertainty, more dynamic and more sensitive to context and culture models, as well as models that can generate economy-energy-efficient design scenarios using local data (to be more harmonized with circular economy principles), are important research areas in this field. The results of this study are a roadmap for researchers and developers of these tools.Keywords: integrated design process, design decision support system, meta-simulation based, early stage, big data, energy efficiency
Procedia PDF Downloads 16227326 Waste Prevention and Economic Policy: Policy Tools for Increasing Resource Efficiency and Savings
Authors: Sylvia Graczka
Abstract:
Waste related environmental problems are not only exploding but are also spotlighted for capacity shortages in recycling, as China announced its ban on waste imports. According to the waste hierarchy, prevention is the primary solution for waste, and also the cheapest. Waste related environmental pollution as externality puts an ever-growing burden on communities bearing the social costs. Economic policies often claim to be pro-environment, this often appears only theoretically, or at the level of principles. There are few concrete occurrences of tools in economic policies, such as green taxes, that are truly effective in stimulating the shift towards waste reduction. The paper presents theoretical economic policy tools based on literature review, and case studies on applied economic policy tools by analyzing policy papers, strategies in force, in line with ‘polluter pays’ and ‘extended producer responsibility’ principles. The study also emphasizes the differences between the broader notion of waste reduction and that of waste minimization, parallel to the difference between resource efficiency and resource savings. It also puts the issue in the context of neoclassical environmental economics and ecological economics, to present alternatives in approach. The research concludes in identifying effective economic policy tools that support the reduction of material use, and the prevention of waste. Consumer and producer awareness of waste problems and consciousness related to their choices are inevitable to make economic policy tools work effectively.Keywords: economic policy, producer responsibility, resource efficiency, waste prevention
Procedia PDF Downloads 14927325 Using Assessment Criteria as a Pedagogic Tool to Develop Argumentative Essay Writing
Authors: Sruti Akula
Abstract:
Assessment criteria are mostly used for assessing skills like writing and speaking. However, they could be used as a pedagogic tool to develop writing skills. A study was conducted with higher secondary learners (Class XII Kendriya Vidyalaya) to investigate the effectiveness of assessment criteria to develop argumentative essay writing. In order to raise awareness about the features of argumentative essay, assessment criteria were shared with the learners. Along with that, self-evaluation checklists were given to the learners to guide them through the writing process. During the study learners wrote multiple drafts with the help of assessment criteria, self-evaluation checklists and teacher feedback at different stages of their writing. It was observed that learners became more aware of the features of argumentative essay which in turn improved their argumentative essay writing. In addition the self evaluation checklists imporved their ability to reflect on their work there by increasing learner autonomy in the class. Hence, it can be claimed that both assessment criteria and self evaluation checklists are effective pedagogic tools to develop argumentative essay writing. Thus, teachers can be trained to create and use tools like assessment criteria and self-evaluation checklists to develop learners’ writing skills in an effective way. The presentation would discuss the approach adopted in the study to teach argumentative essay writing along with the rationale. The tools used in the study would be shared and the data collected in the form of written scripts, self-evaluation checklists and student interviews will be analyzed to validate the claims. Finally, the practical implication of the study like the ways of using assessment criteria and checklists to raise learner awareness and autonomy, using such tools to keep the learners informed about the task requirements and genre features, and the like will be put forward.Keywords: argumentative essay writing, assessment criteria, self evaluation checklists, pedagogic
Procedia PDF Downloads 51027324 Applications of Digital Tools, Satellite Images and Geographic Information Systems in Data Collection of Greenhouses in Guatemala
Authors: Maria A. Castillo H., Andres R. Leandro, Jose F. Bienvenido B.
Abstract:
During the last 20 years, the globalization of economies, population growth, and the increase in the consumption of fresh agricultural products have generated greater demand for ornamentals, flowers, fresh fruits, and vegetables, mainly from tropical areas. This market situation has demanded greater competitiveness and control over production, with more efficient protected agriculture technologies, which provide greater productivity and allow us to guarantee the quality and quantity that is required in a constant and sustainable way. Guatemala, located in the north of Central America, is one of the largest exporters of agricultural products in the region and exports fresh vegetables, flowers, fruits, ornamental plants, and foliage, most of which were grown in greenhouses. Although there are no official agricultural statistics on greenhouse production, several thesis works, and congress reports have presented consistent estimates. A wide range of protection structures and roofing materials are used, from the most basic and simple ones for rain control to highly technical and automated structures connected with remote sensors for monitoring and control of crops. With this breadth of technological models, it is necessary to analyze georeferenced data related to the cultivated area, to the different existing models, and to the covering materials, integrated with altitude, climate, and soil data. The georeferenced registration of the production units, the data collection with digital tools, the use of satellite images, and geographic information systems (GIS) provide reliable tools to elaborate more complete, agile, and dynamic information maps. This study details a methodology proposed for gathering georeferenced data of high protection structures (greenhouses) in Guatemala, structured in four phases: diagnosis of available information, the definition of the geographic frame, selection of satellite images, and integration with an information system geographic (GIS). It especially takes account of the actual lack of complete data in order to obtain a reliable decision-making system; this gap is solved through the proposed methodology. A summary of the results is presented in each phase, and finally, an evaluation with some improvements and tentative recommendations for further research is added. The main contribution of this study is to propose a methodology that allows to reduce the gap of georeferenced data in protected agriculture in this specific area where data is not generally available and to provide data of better quality, traceability, accuracy, and certainty for the strategic agricultural decision öaking, applicable to other crops, production models and similar/neighboring geographic areas.Keywords: greenhouses, protected agriculture, GIS, Guatemala, satellite image, digital tools, precision agriculture
Procedia PDF Downloads 19427323 Designing the Procedures of Building and Environment Management for Basic Education Schools by Using Quality Management
Authors: Suppara Charoenpoom
Abstract:
This study focuses on 1) a good-quality management procedures of buildings and environment in schools 2) designing the management procedures and 3) creating an operation manual for the procedures. This study is the combination of qualitative and quantitative research method. Populations in the research were 83 deans and directors of primary and secondary schools from the 10th educational district in Samut Songkram. Sample group was selected from the voluntary deans and directors. There were 14 participants in sample group. Research tools in this study were divided into 2 categories. The first one was data-collecting tools, which were in-depth interview and questionnaires. The second one was the designing tools to help creating management procedures: quality business, quality work procedure and key quality indicator of each activity in schools. All data were analyzed by mean and standard deviation. The result from this study has found out 1 effective process of building and environment management for basic education schools which is called Quality Business Process (QBP) and 7 Quality Work Procedures (QWP). In terms of academic feasibility checkup by experts, the research had shown that new design of building and environment management was approved unanimously. It means that new process of building and environment management in schools works very well and can be adapted. After examining the possibility of management process being used in schools by calculating the mean value among sample group (14 school deans and directors), the mean value was between 0.64-1.00. It means that the new design of building and environment management can be operated effectively in schools. For the satisfaction part, deans and school directors gave the satisfaction score in the highest level (Mean = 4.7372, S.D. = 0.4385).Keywords: buildings, environment, procedures, quality management
Procedia PDF Downloads 23327322 Comparison of Existing Predictor and Development of Computational Method for S- Palmitoylation Site Identification in Arabidopsis Thaliana
Authors: Ayesha Sanjana Kawser Parsha
Abstract:
S-acylation is an irreversible bond in which cysteine residues are linked to fatty acids palmitate (74%) or stearate (22%), either at the COOH or NH2 terminal, via a thioester linkage. There are several experimental methods that can be used to identify the S-palmitoylation site; however, since they require a lot of time, computational methods are becoming increasingly necessary. There aren't many predictors, however, that can locate S- palmitoylation sites in Arabidopsis Thaliana with sufficient accuracy. This research is based on the importance of building a better prediction tool. To identify the type of machine learning algorithm that predicts this site more accurately for the experimental dataset, several prediction tools were examined in this research, including the GPS PALM 6.0, pCysMod, GPS LIPID 1.0, CSS PALM 4.0, and NBA PALM. These analyses were conducted by constructing the receiver operating characteristics plot and the area under the curve score. An AI-driven deep learning-based prediction tool has been developed utilizing the analysis and three sequence-based input data, such as the amino acid composition, binary encoding profile, and autocorrelation features. The model was developed using five layers, two activation functions, associated parameters, and hyperparameters. The model was built using various combinations of features, and after training and validation, it performed better when all the features were present while using the experimental dataset for 8 and 10-fold cross-validations. While testing the model with unseen and new data, such as the GPS PALM 6.0 plant and pCysMod mouse, the model performed better, and the area under the curve score was near 1. It can be demonstrated that this model outperforms the prior tools in predicting the S- palmitoylation site in the experimental data set by comparing the area under curve score of 10-fold cross-validation of the new model with the established tools' area under curve score with their respective training sets. The objective of this study is to develop a prediction tool for Arabidopsis Thaliana that is more accurate than current tools, as measured by the area under the curve score. Plant food production and immunological treatment targets can both be managed by utilizing this method to forecast S- palmitoylation sites.Keywords: S- palmitoylation, ROC PLOT, area under the curve, cross- validation score
Procedia PDF Downloads 7627321 Understanding Consumer Behaviors by Using Neuromarketing Tools and Methods
Authors: Tabrej Khan
Abstract:
Neuromarketing can refer to the commercial application of neuroscience technologies and insights to drive business further. On the other side, consumer neuroscience can be seen as the academic use of neuroscience to better understand marketing effects on consumer behavior. Consumer Neuroscience and Neuromarketing is a multidisciplinary effort between economics, psychology, and neuroscience and information technology. Traditional methods are using survey, interviews, focus group people are overtly and consciously reporting on their experience and thoughts. The unconscious side of customer behavior is largely unmeasured in the traditional methods. Neuroscience has a potential to understand the unconscious part. Through this paper, we are going to present specific results of selected tools and methods that are used to understand consumer behaviors.Keywords: neuromarketing, neuroscience, consumer behaviors, tools
Procedia PDF Downloads 40227320 Organizational Innovations of the 20th Century as High Tech of the 21st: Evidence from Patent Data
Authors: Valery Yakubovich, Shuping wu
Abstract:
Organization theorists have long claimed that organizational innovations are nontechnological, in part because they are unpatentable. The claim rests on the assumption that organizational innovations are abstract ideas embodied in persons and contexts rather than in context-free practical tools. However, over the last three decades, organizational knowledge has been increasingly embodied in digital tools which, in principle, can be patented. To provide the first empirical evidence regarding the patentability of organizational innovations, we trained two machine learning algorithms to identify a population of 205,434 patent applications for organizational technologies (OrgTech) and, among them, 141,285 applications that use organizational innovations accumulated over the 20th century. Our event history analysis of the probability of patenting an OrgTech invention shows that ideas from organizational innovations decrease the probability of patent allowance unless they describe a practical tool. We conclude that the present-day digital transformation places organizational innovations in the realm of high tech and turns the debate about organizational technologies into the challenge of designing practical organizational tools that embody big ideas about organizing. We outline an agenda for patent-based research on OrgTech as an emerging phenomenon.Keywords: organizational innovation, organizational technology, high tech, patents, machine learning
Procedia PDF Downloads 12227319 Experimental Investigation and Numerical Simulations of the Cylindrical Machining of a Ti-6Al-4V Tree
Authors: Mohamed Sahli, David Bassir, Thierry Barriere, Xavier Roizard
Abstract:
Predicting the behaviour of the Ti-6Al-4V alloy during the turning operation was very important in the choice of suitable cutting tools and also in the machining strategies. In this study, a 3D model with thermo-mechanical coupling has been proposed to study the influence of cutting parameters and also lubrication on the performance of cutting tools. The constants of the constitutive Johnson-Cook model of Ti-6Al-4V alloy were identified using inverse analysis based on the parameters of the orthogonal cutting process. Then, numerical simulations of the finishing machining operation were developed and experimentally validated for the cylindrical stock removal stage with the finishing cutting tool.Keywords: titanium turning, cutting tools, FE simulation, chip
Procedia PDF Downloads 17327318 Timing and Noise Data Mining Algorithm and Software Tool in Very Large Scale Integration (VLSI) Design
Authors: Qing K. Zhu
Abstract:
Very Large Scale Integration (VLSI) design becomes very complex due to the continuous integration of millions of gates in one chip based on Moore’s law. Designers have encountered numerous report files during design iterations using timing and noise analysis tools. This paper presented our work using data mining techniques combined with HTML tables to extract and represent critical timing/noise data. When we apply this data-mining tool in real applications, the running speed is important. The software employs table look-up techniques in the programming for the reasonable running speed based on performance testing results. We added several advanced features for the application in one industry chip design.Keywords: VLSI design, data mining, big data, HTML forms, web, VLSI, EDA, timing, noise
Procedia PDF Downloads 25427317 Analyzing On-Line Process Data for Industrial Production Quality Control
Authors: Hyun-Woo Cho
Abstract:
The monitoring of industrial production quality has to be implemented to alarm early warning for unusual operating conditions. Furthermore, identification of their assignable causes is necessary for a quality control purpose. For such tasks many multivariate statistical techniques have been applied and shown to be quite effective tools. This work presents a process data-based monitoring scheme for production processes. For more reliable results some additional steps of noise filtering and preprocessing are considered. It may lead to enhanced performance by eliminating unwanted variation of the data. The performance evaluation is executed using data sets from test processes. The proposed method is shown to provide reliable quality control results, and thus is more effective in quality monitoring in the example. For practical implementation of the method, an on-line data system must be available to gather historical and on-line data. Recently large amounts of data are collected on-line in most processes and implementation of the current scheme is feasible and does not give additional burdens to users.Keywords: detection, filtering, monitoring, process data
Procedia PDF Downloads 55927316 Data Mining As A Tool For Knowledge Management: A Review
Authors: Maram Saleh
Abstract:
Knowledge has become an essential resource in today’s economy and become the most important asset of maintaining competition advantage in organizations. The importance of knowledge has made organizations to manage their knowledge assets and resources through all multiple knowledge management stages such as: Knowledge Creation, knowledge storage, knowledge sharing and knowledge use. Researches on data mining are continues growing over recent years on both business and educational fields. Data mining is one of the most important steps of the knowledge discovery in databases process aiming to extract implicit, unknown but useful knowledge and it is considered as significant subfield in knowledge management. Data miming have the great potential to help organizations to focus on extracting the most important information on their data warehouses. Data mining tools and techniques can predict future trends and behaviors, allowing businesses to make proactive, knowledge-driven decisions. This review paper explores the applications of data mining techniques in supporting knowledge management process as an effective knowledge discovery technique. In this paper, we identify the relationship between data mining and knowledge management, and then focus on introducing some application of date mining techniques in knowledge management for some real life domains.Keywords: Data Mining, Knowledge management, Knowledge discovery, Knowledge creation.
Procedia PDF Downloads 20827315 Using Risk Management Indicators in Decision Tree Analysis
Authors: Adel Ali Elshaibani
Abstract:
Risk management indicators augment the reporting infrastructure, particularly for the board and senior management, to identify, monitor, and manage risks. This enhancement facilitates improved decision-making throughout the banking organization. Decision tree analysis is a tool that visually outlines potential outcomes, costs, and consequences of complex decisions. It is particularly beneficial for analyzing quantitative data and making decisions based on numerical values. By calculating the expected value of each outcome, decision tree analysis can help assess the best course of action. In the context of banking, decision tree analysis can assist lenders in evaluating a customer’s creditworthiness, thereby preventing losses. However, applying these tools in developing countries may face several limitations, such as data availability, lack of technological infrastructure and resources, lack of skilled professionals, cultural factors, and cost. Moreover, decision trees can create overly complex models that do not generalize well to new data, known as overfitting. They can also be sensitive to small changes in the data, which can result in different tree structures and can become computationally expensive when dealing with large datasets. In conclusion, while risk management indicators and decision tree analysis are beneficial for decision-making in banks, their effectiveness is contingent upon how they are implemented and utilized by the board of directors, especially in the context of developing countries. It’s important to consider these limitations when planning to implement these tools in developing countries.Keywords: risk management indicators, decision tree analysis, developing countries, board of directors, bank performance, risk management strategy, banking institutions
Procedia PDF Downloads 6027314 Prediction of Saturated Hydraulic Conductivity Dynamics in an Iowan Agriculture Watershed
Authors: Mohamed Elhakeem, A. N. Thanos Papanicolaou, Christopher Wilson, Yi-Jia Chang
Abstract:
In this study, a physically-based, modelling framework was developed to predict saturated hydraulic conductivity (KSAT) dynamics in the Clear Creek Watershed (CCW), Iowa. The modelling framework integrated selected pedotransfer functions and watershed models with geospatial tools. A number of pedotransfer functions and agricultural watershed models were examined to select the appropriate models that represent the study site conditions. Models selection was based on statistical measures of the models’ errors compared to the KSAT field measurements conducted in the CCW under different soil, climate and land use conditions. The study has shown that the predictions of the combined pedotransfer function of Rosetta and the Water Erosion Prediction Project (WEPP) provided the best agreement to the measured KSAT values in the CCW compared to the other tested models. Therefore, Rosetta and WEPP were integrated with the Geographic Information System (GIS) tools for visualization of the data in forms of geospatial maps and prediction of KSAT variability in CCW due to the seasonal changes in climate and land use activities.Keywords: saturated hydraulic conductivity, pedotransfer functions, watershed models, geospatial tools
Procedia PDF Downloads 26027313 Adoption of Noise and Vibration Management Tools for Major Infrastructure Projects in Sydney, Australia
Authors: Adrian Morris, Rodney Phillips, Mattia Tabacchi
Abstract:
Minimizing construction noise and vibration impacts is a key challenge for major infrastructure projects in urban environments. Before commencing construction works, Construction Noise and Vibration Management Plan (CNVMP) and Construction Noise and Vibration Impact Statements (CNVIS) are required to be prepared and submitted to the relevant government authorities for review and approval. However, the assessment of potential impacts from work activities at pre-approval stage may be inaccurate as works methodology and scheduling are yet to be determined. In response, noise and vibration management tools have been developed to refine and supplement the CNVIS as works progress. These tools have been successfully implemented in major infrastructure projects allowing contractors to plan and assess construction works in a cost effective and timely manner. As a result, noise and vibration management tools have been incorporated into management plans and are increasingly required by regulators.Keywords: noise management, environmental noise, infrastructure projects, construction, vibration, cost effective
Procedia PDF Downloads 13127312 BFDD-S: Big Data Framework to Detect and Mitigate DDoS Attack in SDN Network
Authors: Amirreza Fazely Hamedani, Muzzamil Aziz, Philipp Wieder, Ramin Yahyapour
Abstract:
Software-defined networking in recent years came into the sight of so many network designers as a successor to the traditional networking. Unlike traditional networks where control and data planes engage together within a single device in the network infrastructure such as switches and routers, the two planes are kept separated in software-defined networks (SDNs). All critical decisions about packet routing are made on the network controller, and the data level devices forward the packets based on these decisions. This type of network is vulnerable to DDoS attacks, degrading the overall functioning and performance of the network by continuously injecting the fake flows into it. This increases substantial burden on the controller side, and the result ultimately leads to the inaccessibility of the controller and the lack of network service to the legitimate users. Thus, the protection of this novel network architecture against denial of service attacks is essential. In the world of cybersecurity, attacks and new threats emerge every day. It is essential to have tools capable of managing and analyzing all this new information to detect possible attacks in real-time. These tools should provide a comprehensive solution to automatically detect, predict and prevent abnormalities in the network. Big data encompasses a wide range of studies, but it mainly refers to the massive amounts of structured and unstructured data that organizations deal with on a regular basis. On the other hand, it regards not only the volume of the data; but also that how data-driven information can be used to enhance decision-making processes, security, and the overall efficiency of a business. This paper presents an intelligent big data framework as a solution to handle illegitimate traffic burden on the SDN network created by the numerous DDoS attacks. The framework entails an efficient defence and monitoring mechanism against DDoS attacks by employing the state of the art machine learning techniques.Keywords: apache spark, apache kafka, big data, DDoS attack, machine learning, SDN network
Procedia PDF Downloads 16927311 Reliability Enhancement by Parameter Design in Ferrite Magnet Process
Abstract:
Ferrite magnet is widely used in many automotive components such as motors and alternators. Magnets used inside the components must be in good quality to ensure the high level of performance. The purpose of this study is to design input parameters that optimize the ferrite magnet production process to ensure the quality and reliability of manufactured products. Design of Experiments (DOE) and Statistical Process Control (SPC) are used as mutual supplementations to optimize the process. DOE and SPC are quality tools being used in the industry to monitor and improve the manufacturing process condition. These tools are practically used to maintain the process on target and within the limits of natural variation. A mixed Taguchi method is utilized for optimization purpose as a part of DOE analysis. SPC with proportion data is applied to assess the output parameters to determine the optimal operating conditions. An example of case involving the monitoring and optimization of ferrite magnet process was presented to demonstrate the effectiveness of this approach. Through the utilization of these tools, reliable magnets can be produced by following the step by step procedures of proposed framework. One of the main contributions of this study was producing the crack free magnets by applying the proposed parameter design.Keywords: ferrite magnet, crack, reliability, process optimization, Taguchi method
Procedia PDF Downloads 517