Search results for: process data
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 34841

Search results for: process data

34631 Quantifying Automation in the Architectural Design Process via a Framework Based on Task Breakdown Systems and Recursive Analysis: An Exploratory Study

Authors: D. M. Samartsev, A. G. Copping

Abstract:

As with all industries, architects are using increasing amounts of automation within practice, with approaches such as generative design and use of AI becoming more commonplace. However, the discourse on the rate at which the architectural design process is being automated is often personal and lacking in objective figures and measurements. This results in confusion between people and barriers to effective discourse on the subject, in turn limiting the ability of architects, policy makers, and members of the public in making informed decisions in the area of design automation. This paper proposes the use of a framework to quantify the progress of automation within the design process. The use of a reductionist analysis of the design process allows it to be quantified in a manner that enables direct comparison across different times, as well as locations and projects. The methodology is informed by the design of this framework – taking on the aspects of a systematic review but compressed in time to allow for an initial set of data to verify the validity of the framework. The use of such a framework of quantification enables various practical uses such as predicting the future of the architectural industry with regards to which tasks will be automated, as well as making more informed decisions on the subject of automation on multiple levels ranging from individual decisions to policy making from governing bodies such as the RIBA. This is achieved by analyzing the design process as a generic task that needs to be performed, then using principles of work breakdown systems to split the task of designing an entire building into smaller tasks, which can then be recursively split further as required. Each task is then assigned a series of milestones that allow for the objective analysis of its automation progress. By combining these two approaches it is possible to create a data structure that describes how much various parts of the architectural design process are automated. The data gathered in the paper serves the dual purposes of providing the framework with validation, as well as giving insights into the current situation of automation within the architectural design process. The framework can be interrogated in many ways and preliminary analysis shows that almost 40% of the architectural design process has been automated in some practical fashion at the time of writing, with the rate at which progress is made slowly increasing over the years, with the majority of tasks in the design process reaching a new milestone in automation in less than 6 years. Additionally, a further 15% of the design process is currently being automated in some way, with various products in development but not yet released to the industry. Lastly, various limitations of the framework are examined in this paper as well as further areas of study.

Keywords: analysis, architecture, automation, design process, technology

Procedia PDF Downloads 76
34630 Transitivity Analysis in Reading Passage of English Text Book for Senior High School

Authors: Elitaria Bestri Agustina Siregar, Boni Fasius Siregar

Abstract:

The paper concerned with the transitivity in the reading passage of English textbook for Senior High School. The six types of process were occurred in the passages with percentage as follows: Material Process is 166 (42%), Relational Process is 155 (39%), Mental Process is 39 (10%), Verbal Process is 21 (5%), Existential Process is 13 (3), and Behavioral Process is 5 (1%). The material processes were found to be the most frequently used process type in the samples in our corpus (41,60 %). This indicates that the twenty reading passages are centrally concerned with action and events. Related to developmental psychology theory, this book fits the needs of students of this age.

Keywords: transitivity, types of processes, reading passages, developmental psycholoy

Procedia PDF Downloads 374
34629 A Schema of Building an Efficient Quality Gate throughout the Software Development with Tools

Authors: Le Chen

Abstract:

This paper presents an efficient tool platform scheme to ensure quality protection throughout the software development process. The main principle is to manage the information of requirements, design, development, testing, operation and maintenance process with proper tools, and to set up the quality standards of each process. Through the tools’ display and summary of quality standards, the quality standards can be visualizad and ready for policy decision, which is called Quality Gate in this paper. In addition, the tools are also integrated to achieve the exchange and relation of information which highly improving operational efficiency. In this paper, the feasibility of the scheme is verified by practical application of development projects, and the overall information display and data mining are proposed to be further improved.

Keywords: efficiency, quality gate, software process, tools

Procedia PDF Downloads 334
34628 Enhancing Predictive Accuracy in Pharmaceutical Sales through an Ensemble Kernel Gaussian Process Regression Approach

Authors: Shahin Mirshekari, Mohammadreza Moradi, Hossein Jafari, Mehdi Jafari, Mohammad Ensaf

Abstract:

This research employs Gaussian Process Regression (GPR) with an ensemble kernel, integrating Exponential Squared, Revised Matern, and Rational Quadratic kernels to analyze pharmaceutical sales data. Bayesian optimization was used to identify optimal kernel weights: 0.76 for Exponential Squared, 0.21 for Revised Matern, and 0.13 for Rational Quadratic. The ensemble kernel demonstrated superior performance in predictive accuracy, achieving an R² score near 1.0, and significantly lower values in MSE, MAE, and RMSE. These findings highlight the efficacy of ensemble kernels in GPR for predictive analytics in complex pharmaceutical sales datasets.

Keywords: Gaussian process regression, ensemble kernels, bayesian optimization, pharmaceutical sales analysis, time series forecasting, data analysis

Procedia PDF Downloads 29
34627 Adoption of Big Data by Global Chemical Industries

Authors: Ashiff Khan, A. Seetharaman, Abhijit Dasgupta

Abstract:

The new era of big data (BD) is influencing chemical industries tremendously, providing several opportunities to reshape the way they operate and help them shift towards intelligent manufacturing. Given the availability of free software and the large amount of real-time data generated and stored in process plants, chemical industries are still in the early stages of big data adoption. The industry is just starting to realize the importance of the large amount of data it owns to make the right decisions and support its strategies. This article explores the importance of professional competencies and data science that influence BD in chemical industries to help it move towards intelligent manufacturing fast and reliable. This article utilizes a literature review and identifies potential applications in the chemical industry to move from conventional methods to a data-driven approach. The scope of this document is limited to the adoption of BD in chemical industries and the variables identified in this article. To achieve this objective, government, academia, and industry must work together to overcome all present and future challenges.

Keywords: chemical engineering, big data analytics, industrial revolution, professional competence, data science

Procedia PDF Downloads 54
34626 A Clustering Algorithm for Massive Texts

Authors: Ming Liu, Chong Wu, Bingquan Liu, Lei Chen

Abstract:

Internet users have to face the massive amount of textual data every day. Organizing texts into categories can help users dig the useful information from large-scale text collection. Clustering, in fact, is one of the most promising tools for categorizing texts due to its unsupervised characteristic. Unfortunately, most of traditional clustering algorithms lose their high qualities on large-scale text collection. This situation mainly attributes to the high- dimensional vectors generated from texts. To effectively and efficiently cluster large-scale text collection, this paper proposes a vector reconstruction based clustering algorithm. Only the features that can represent the cluster are preserved in cluster’s representative vector. This algorithm alternately repeats two sub-processes until it converges. One process is partial tuning sub-process, where feature’s weight is fine-tuned by iterative process. To accelerate clustering velocity, an intersection based similarity measurement and its corresponding neuron adjustment function are proposed and implemented in this sub-process. The other process is overall tuning sub-process, where the features are reallocated among different clusters. In this sub-process, the features useless to represent the cluster are removed from cluster’s representative vector. Experimental results on the three text collections (including two small-scale and one large-scale text collections) demonstrate that our algorithm obtains high quality on both small-scale and large-scale text collections.

Keywords: vector reconstruction, large-scale text clustering, partial tuning sub-process, overall tuning sub-process

Procedia PDF Downloads 405
34625 Process Integration: Mathematical Model for Contaminant Removal in Refinery Process Stream

Authors: Wasif Mughees, Malik Al-Ahmad

Abstract:

This research presents the graphical design analysis and mathematical programming technique to dig out the possible water allocation distribution to minimize water usage in process units. The study involves the mass and property integration in its core methodology. Tehran Oil Refinery is studied to implement the focused water pinch technology for regeneration, reuse and recycling of water streams. Process data is manipulated in terms of sources and sinks, which are given in terms of properties. Sources are the streams to be allocated. Sinks are the units which can accept the sources. Suspended Solids (SS) is taken as a single contaminant. The model minimizes the mount of freshwater from 340 to 275m3/h (19.1%). Redesigning and allocation of water streams was built. The graphical technique and mathematical programming shows the consistency of results which confirms mass transfer dependency of water streams.

Keywords: minimization, water pinch, process integration, pollution prevention

Procedia PDF Downloads 297
34624 Social Data-Based Users Profiles' Enrichment

Authors: Amel Hannech, Mehdi Adda, Hamid Mcheick

Abstract:

In this paper, we propose a generic model of user profile integrating several elements that may positively impact the research process. We exploit the classical behavior of users and integrate a delimitation process of their research activities into several research sessions enriched with contextual and temporal information, which allows reflecting the current interests of these users in every period of time and infer data freshness. We argue that the annotation of resources gives more transparency on users' needs. It also strengthens social links among resources and users, and can so increase the scope of the user profile. Based on this idea, we integrate the social tagging practice in order to exploit the social users' behavior to enrich their profiles. These profiles are then integrated into a recommendation system in order to predict the interesting personalized items of users allowing to assist them in their researches and further enrich their profiles. In this recommendation, we provide users new research experiences.

Keywords: user profiles, topical ontology, contextual information, folksonomies, tags' clusters, data freshness, association rules, data recommendation

Procedia PDF Downloads 240
34623 Elderly Health Care Process by Community Participation: A Sub-District in the Lower Northern Region of Thailand

Authors: Amaraporn Puraya, Roongtiva Boonpracom, Somsak Thojampa, Sirikanok Klankhajhon, Kittisak Kumpeera

Abstract:

The objective of this qualitative research was to study the elderly health care process by community participation. Data were collected by quality research methods, including secondary data study, observation, in-depth interviews, and focus group discussions and analyzed by content analysis, reflection and review of information. The research results pointed out that the important elderly health care process by community participation consisted of 2 parts, namely the community participation development process in elderly health care and the outcomes from the participation development process. The community participation development process consisted of 4 steps as follows: 1) Building the leadership team, an important social capital of the community, which started from searching for both formal and informal leaders by giving the opportunity for public participation and creating clear agreements defining roles, duties and responsibilities; 2) investigating the problems and the needs of the community, 3) designing the elderly health care activities under the concept of self-care potential development of the elderly through participation in community forums and meetings to exchange knowledge with common goals, plans and operation and 4) the development process of sustainable health care agreement at the local level, starting from opening communication channels to create awareness and participation in various activities at both individual and group levels as well as pushing activities/projects into the community development plan consistent with the local administration policy. The outcomes from the participation development process were as follows. 1) There was the integration of the elderly for doing the elderly health care activities/projects in the community managed by the elderly themselves. 2) The service system was changed from the passive to the proactive one, focusing on health promotion rather than treating diseases or illnesses. 3) The registered nurses / the public health officers can provide care for the elderly with chronic illnesses through the implementation of activities/projects of elderly health care so that the elderly can access the services more. 4) The local government organization became the main mechanism in driving the elderly health care process by community participation.

Keywords: elderly health care process, community participation, elderly, Thailand

Procedia PDF Downloads 171
34622 Methodology of the Turkey’s National Geographic Information System Integration Project

Authors: Buse A. Ataç, Doğan K. Cenan, Arda Çetinkaya, Naz D. Şahin, Köksal Sanlı, Zeynep Koç, Akın Kısa

Abstract:

With its spatial data reliability, interpretation and questioning capabilities, Geographical Information Systems make significant contributions to scientists, planners and practitioners. Geographic information systems have received great attention in today's digital world, growing rapidly, and increasing the efficiency of use. Access to and use of current and accurate geographical data, which are the most important components of the Geographical Information System, has become a necessity rather than a need for sustainable and economic development. This project aims to enable sharing of data collected by public institutions and organizations on a web-based platform. Within the scope of the project, INSPIRE (Infrastructure for Spatial Information in the European Community) data specifications are considered as a road-map. In this context, Turkey's National Geographic Information System (TUCBS) Integration Project supports sharing spatial data within 61 pilot public institutions as complied with defined national standards. In this paper, which is prepared by the project team members in the TUCBS Integration Project, the technical process with a detailed methodology is explained. In this context, the main technical processes of the Project consist of Geographic Data Analysis, Geographic Data Harmonization (Standardization), Web Service Creation (WMS, WFS) and Metadata Creation-Publication. In this paper, the integration process carried out to provide the data produced by 61 institutions to be shared from the National Geographic Data Portal (GEOPORTAL), have been trying to be conveyed with a detailed methodology.

Keywords: data specification, geoportal, GIS, INSPIRE, Turkish National Geographic Information System, TUCBS, Turkey's national geographic information system

Procedia PDF Downloads 116
34621 Image Ranking to Assist Object Labeling for Training Detection Models

Authors: Tonislav Ivanov, Oleksii Nedashkivskyi, Denis Babeshko, Vadim Pinskiy, Matthew Putman

Abstract:

Training a machine learning model for object detection that generalizes well is known to benefit from a training dataset with diverse examples. However, training datasets usually contain many repeats of common examples of a class and lack rarely seen examples. This is due to the process commonly used during human annotation where a person would proceed sequentially through a list of images labeling a sufficiently high total number of examples. Instead, the method presented involves an active process where, after the initial labeling of several images is completed, the next subset of images for labeling is selected by an algorithm. This process of algorithmic image selection and manual labeling continues in an iterative fashion. The algorithm used for the image selection is a deep learning algorithm, based on the U-shaped architecture, which quantifies the presence of unseen data in each image in order to find images that contain the most novel examples. Moreover, the location of the unseen data in each image is highlighted, aiding the labeler in spotting these examples. Experiments performed using semiconductor wafer data show that labeling a subset of the data, curated by this algorithm, resulted in a model with a better performance than a model produced from sequentially labeling the same amount of data. Also, similar performance is achieved compared to a model trained on exhaustive labeling of the whole dataset. Overall, the proposed approach results in a dataset that has a diverse set of examples per class as well as more balanced classes, which proves beneficial when training a deep learning model.

Keywords: computer vision, deep learning, object detection, semiconductor

Procedia PDF Downloads 107
34620 Covariance of the Queue Process Fed by Isonormal Gaussian Input Process

Authors: Samaneh Rahimirshnani, Hossein Jafari

Abstract:

In this paper, we consider fluid queueing processes fed by an isonormal Gaussian process. We study the correlation structure of the queueing process and the rate of convergence of the running supremum in the queueing process. The Malliavin calculus techniques are applied to obtain relations that show the workload process inherits the dependence properties of the input process. As examples, we consider two isonormal Gaussian processes, the sub-fractional Brownian motion (SFBM) and the fractional Brownian motion (FBM). For these examples, we obtain upper bounds for the covariance function of the queueing process and its rate of convergence to zero. We also discover that the rate of convergence of the queueing process is related to the structure of the covariance function of the input process.

Keywords: queue length process, Malliavin calculus, covariance function, fractional Brownian motion, sub-fractional Brownian motion

Procedia PDF Downloads 25
34619 Design and Characterization of a CMOS Process Sensor Utilizing Vth Extractor Circuit

Authors: Rohana Musa, Yuzman Yusoff, Chia Chieu Yin, Hanif Che Lah

Abstract:

This paper presents the design and characterization of a low power Complementary Metal Oxide Semiconductor (CMOS) process sensor. The design is targeted for implementation using Silterra’s 180 nm CMOS process technology. The proposed process sensor employs a voltage threshold (Vth) extractor architecture for detection of variations in the fabrication process. The process sensor generates output voltages in the range of 401 mV (fast-fast corner) to 443 mV (slow-slow corner) at nominal condition. The power dissipation for this process sensor is 6.3 µW with a supply voltage of 1.8V with a silicon area of 190 µm X 60 µm. The preliminary result of this process sensor that was fabricated indicates a close resemblance between test and simulated results.

Keywords: CMOS process sensor, PVT sensor, threshold extractor circuit, Vth extractor circuit

Procedia PDF Downloads 149
34618 Hierarchical Checkpoint Protocol in Data Grids

Authors: Rahma Souli-Jbali, Minyar Sassi Hidri, Rahma Ben Ayed

Abstract:

Grid of computing nodes has emerged as a representative means of connecting distributed computers or resources scattered all over the world for the purpose of computing and distributed storage. Since fault tolerance becomes complex due to the availability of resources in decentralized grid environment, it can be used in connection with replication in data grids. The objective of our work is to present fault tolerance in data grids with data replication-driven model based on clustering. The performance of the protocol is evaluated with Omnet++ simulator. The computational results show the efficiency of our protocol in terms of recovery time and the number of process in rollbacks.

Keywords: data grids, fault tolerance, clustering, chandy-lamport

Procedia PDF Downloads 302
34617 Business Process Mashup

Authors: Fethia Zenak, Salima Benbernou, Linda Zaoui

Abstract:

Recently, many companies are based on process development from scratch to achieve their business goals. The process development is not trivial and the main objective of enterprise managing processes is to decrease the software development time. Several concepts have been proposed in the field of business process-based reused development, known as BP Mashup. This concept consists of reusing existing business processes which have been modeled in order to respond to a particular goal. To meet user process requirements, our contribution is to mix parts of processes as 'processes fragments' components to build a new process (i.e. process mashup). The main idea of our paper is to offer graphical framework tool for both creating and running processes mashup. Allow users to perform a mixture of fragments, using a simple interface with set of graphical mixture operators based on a proposed formal model. A process mashup and mixture behavior are described within a new specification of a high-level language, language for process mashup (BPML).

Keywords: business process, mashup, fragments, bp mashup

Procedia PDF Downloads 594
34616 Track and Trace Solution on Land Certificate Production: Indonesian Land Certificate

Authors: Adrian Rifqi, Febe Napitupulu, Erdi Hermawan, Edwin Putra, Yang Leprilian

Abstract:

This article focuses on the implementation of the production improvement process of the Indonesian land certificate product that printed in Perum Peruri as the state-owned enterprises. Based on the data obtained, there are several complaints from customers of the 2019 land certificate production. The complaints become a negative value to loyal customers of Perum Peruri. Almost all the complaints are referring to ‘defective printouts and the difference between products in packaging and packaging labels both in terms of type and quantity’. To overcome this problem, we intend to make an improvement to the production process that focuses on complaints ‘there is a difference between products in packaging with packaging labels’. Improvements in the land certificate production process are relying on the technology of the scales and QR code on the packaging label. In addition, using the QR code on the packaging label will facilitate the process of tracking product data. With this method, we hope to reduce the error rate between products in packaging with the packaging label both in terms of quantity, type, and product number on the land certificate and error rate of sending land certificates, which will be sent to many places to 0%. With this solution, we also hope to get precise data and real-time reports on the production of land certificates in the near future, so track and trace implementation can be done as the solution of the land certificate production.

Keywords: land certificates, QR code, track and trace, packaging

Procedia PDF Downloads 129
34615 A Study on the Correlation Analysis between the Pre-Sale Competition Rate and the Apartment Unit Plan Factor through Machine Learning

Authors: Seongjun Kim, Jinwooung Kim, Sung-Ah Kim

Abstract:

The development of information and communication technology also affects human cognition and thinking, especially in the field of design, new techniques are being tried. In architecture, new design methodologies such as machine learning or data-driven design are being applied. In particular, these methodologies are used in analyzing the factors related to the value of real estate or analyzing the feasibility in the early planning stage of the apartment housing. However, since the value of apartment buildings is often determined by external factors such as location and traffic conditions, rather than the interior elements of buildings, data is rarely used in the design process. Therefore, although the technical conditions are provided, the internal elements of the apartment are difficult to apply the data-driven design in the design process of the apartment. As a result, the designers of apartment housing were forced to rely on designer experience or modular design alternatives rather than data-driven design at the design stage, resulting in a uniform arrangement of space in the apartment house. The purpose of this study is to propose a methodology to support the designers to design the apartment unit plan with high consumer preference by deriving the correlation and importance of the floor plan elements of the apartment preferred by the consumers through the machine learning and reflecting this information from the early design process. The data on the pre-sale competition rate and the elements of the floor plan are collected as data, and the correlation between pre-sale competition rate and independent variables is analyzed through machine learning. This analytical model can be used to review the apartment unit plan produced by the designer and to assist the designer. Therefore, it is possible to make a floor plan of apartment housing with high preference because it is possible to feedback apartment unit plan by using trained model when it is used in floor plan design of apartment housing.

Keywords: apartment unit plan, data-driven design, design methodology, machine learning

Procedia PDF Downloads 233
34614 Managing Data from One Hundred Thousand Internet of Things Devices Globally for Mining Insights

Authors: Julian Wise

Abstract:

Newcrest Mining is one of the world’s top five gold and rare earth mining organizations by production, reserves and market capitalization in the world. This paper elaborates on the data acquisition processes employed by Newcrest in collaboration with Fortune 500 listed organization, Insight Enterprises, to standardize machine learning solutions which process data from over a hundred thousand distributed Internet of Things (IoT) devices located at mine sites globally. Through the utilization of software architecture cloud technologies and edge computing, the technological developments enable for standardized processes of machine learning applications to influence the strategic optimization of mineral processing. Target objectives of the machine learning optimizations include time savings on mineral processing, production efficiencies, risk identification, and increased production throughput. The data acquired and utilized for predictive modelling is processed through edge computing by resources collectively stored within a data lake. Being involved in the digital transformation has necessitated the standardization software architecture to manage the machine learning models submitted by vendors, to ensure effective automation and continuous improvements to the mineral process models. Operating at scale, the system processes hundreds of gigabytes of data per day from distributed mine sites across the globe, for the purposes of increased improved worker safety, and production efficiency through big data applications.

Keywords: mineral technology, big data, machine learning operations, data lake

Procedia PDF Downloads 84
34613 Decision Making to Study Abroad among Indonesian Student Migrants in Europe: The Role of Communication Technology

Authors: Inayah Hidayati

Abstract:

Innovation in communication technology has opened up opportunities for student to migrate and study abroad. The increasing number of Indonesian students migrating to study abroad suggests the importance of understanding the reason underline their movements. Objective: This research aims to explain the migration decision-making process of Indonesian student migrants in Europe. In detail, this research will consider the innovation in communication technology in the migration decision-making process of students who emigrated from Indonesia and how they use that in the context of the migration decision-making process. Methods: The data collected included qualitative data from in-depth interviews. An interview guide was formulated to facilitate the in-depth interviews and generate a better understanding of migration behavior. Expectation: 1). Innovation in communication technology help Indonesian student migrants on migration decision making process. 2). Student migrants use communication technology platforms for searching information about destination area. Result: Student migrant in Europe use their communication technology platforms to gain information before they choose that country for study. They use WhatsApp and LINE to making contact with their friends and colleagues in the destination country. WhatsApp and LINE group help Indonesian student to get information about school and daily life.

Keywords: international migration, student, decision making process, communication technology platforms

Procedia PDF Downloads 222
34612 Spatial Rank-Based High-Dimensional Monitoring through Random Projection

Authors: Chen Zhang, Nan Chen

Abstract:

High-dimensional process monitoring becomes increasingly important in many application domains, where usually the process distribution is unknown and much more complicated than the normal distribution, and the between-stream correlation can not be neglected. However, since the process dimension is generally much bigger than the reference sample size, most traditional nonparametric multivariate control charts fail in high-dimensional cases due to the curse of dimensionality. Furthermore, when the process goes out of control, the influenced variables are quite sparse compared with the whole dimension, which increases the detection difficulty. Targeting at these issues, this paper proposes a new nonparametric monitoring scheme for high-dimensional processes. This scheme first projects the high-dimensional process into several subprocesses using random projections for dimension reduction. Then, for every subprocess with the dimension much smaller than the reference sample size, a local nonparametric control chart is constructed based on the spatial rank test to detect changes in this subprocess. Finally, the results of all the local charts are fused together for decision. Furthermore, after an out-of-control (OC) alarm is triggered, a diagnostic framework is proposed. using the square-root LASSO. Numerical studies demonstrate that the chart has satisfactory detection power for sparse OC changes and robust performance for non-normally distributed data, The diagnostic framework is also effective to identify truly changed variables. Finally, a real-data example is presented to demonstrate the application of the proposed method.

Keywords: random projection, high-dimensional process control, spatial rank, sequential change detection

Procedia PDF Downloads 275
34611 Optimization of Manufacturing Process Parameters: An Empirical Study from Taiwan's Tech Companies

Authors: Chao-Ton Su, Li-Fei Chen

Abstract:

The parameter design is crucial to improving the uniformity of a product or process. In the product design stage, parameter design aims to determine the optimal settings for the parameters of each element in the system, thereby minimizing the functional deviations of the product. In the process design stage, parameter design aims to determine the operating settings of the manufacturing processes so that non-uniformity in manufacturing processes can be minimized. The parameter design, trying to minimize the influence of noise on the manufacturing system, plays an important role in the high-tech companies. Taiwan has many well-known high-tech companies, which show key roles in the global economy. Quality remains the most important factor that enables these companies to sustain their competitive advantage. In Taiwan however, many high-tech companies face various quality problems. A common challenge is related to root causes and defect patterns. In the R&D stage, root causes are often unknown, and defect patterns are difficult to classify. Additionally, data collection is not easy. Even when high-volume data can be collected, data interpretation is difficult. To overcome these challenges, high-tech companies in Taiwan use more advanced quality improvement tools. In addition to traditional statistical methods and quality tools, the new trend is the application of powerful tools, such as neural network, fuzzy theory, data mining, industrial engineering, operations research, and innovation skills. In this study, several examples of optimizing the parameter settings for the manufacturing process in Taiwan’s tech companies will be presented to illustrate proposed approach’s effectiveness. Finally, a discussion of using traditional experimental design versus the proposed approach for process optimization will be made.

Keywords: quality engineering, parameter design, neural network, genetic algorithm, experimental design

Procedia PDF Downloads 121
34610 Ix Operation for the Concentration of Low-Grade Uranium Leach Solution

Authors: Heba Ahmed Nawafleh

Abstract:

In this study, two commercial resins were evaluated to concentrate uranium from real solutions that were produced from analkaline leaching process of carbonate deposits. The adsorption was examined using a batch process. Different parameters were evaluated, including initial pH, contact time, temperature, adsorbent dose, and finally, uranium initial concentration. Both resins were effective and selective for uranium ions from the tested leaching solution. The adsorption isotherms data were well fitted for both resins using the Langmuir model. Thermodynamic functions (Gibbs free energy change ΔG, enthalpy change ΔH, and entropy change ΔS) were calculated for the adsorption of uranium. The result shows that the adsorption process is endothermic, spontaneous, and chemisorption processes took place for both resins. The kinetic studies showed that the equilibrium time for uranium ions is about two hours, where the maximum uptake levels were achieved. The kinetics studies were carried out for the adsorption of U ions, and the data was found to follow pseudo-second-order kinetics, which indicates that the adsorption of U ions was chemically controlled. In addition, the reusability (adsorption/ desorption) process was tested for both resins for five cycles, these adsorbents maintained removal efficiency close to first cycle efficiency of about 91% and 80%.

Keywords: uranium, adsorption, ion exchange, thermodynamic and kinetic studies

Procedia PDF Downloads 56
34609 Analysis of Process Methane Hydrate Formation That Include the Important Role of Deep-Sea Sediments with Analogy in Kerek Formation, Sub-Basin Kendeng, Central Java, Indonesia

Authors: Yan Bachtiar Muslih, Hangga Wijaya, Trio Fani, Putri Agustin

Abstract:

Demand of Energy in Indonesia always increases 5-6% a year, but production of conventional energy always decreases 3-5% a year, it means that conventional energy in 20-40 years ahead will not able to complete all energy demand in Indonesia, one of the solve way is using unconventional energy that is gas hydrate, gas hydrate is gas that form by biogenic process, gas hydrate stable in condition with extremely depth and low temperature, gas hydrate can form in two condition that is in pole condition and in deep-sea condition, wherein this research will focus in gas hydrate that association with methane form methane hydrate in deep-sea condition and usually form in depth between 150-2000 m, this research will focus in process of methane hydrate formation that is biogenic process and the important role of deep-sea sediment so can produce accumulation of methane hydrate, methane hydrate usually will be accumulated in find sediment in deep-sea environment with condition high-pressure and low-temperature this condition too usually make methane hydrate change into white nodule, methodology of this research is geology field work and laboratory analysis, from geology field work will get sample data consist of 10-15 samples from Kerek Formation outcrops as random for imagine the condition of deep-sea environment that influence the methane hydrate formation and also from geology field work will get data of measuring stratigraphy in outcrops Kerek Formation too from this data will help to imagine the process in deep-sea sediment like energy flow, supply sediment, and etc, and laboratory analysis is activity to analyze all data that get from geology field work, the result of this research can used to exploration activity of methane hydrate in another prospect deep-sea environment in Indonesia.

Keywords: methane hydrate, deep-sea sediment, kerek formation, sub-basin of kendeng, central java, Indonesia

Procedia PDF Downloads 441
34608 Confidence Intervals for Process Capability Indices for Autocorrelated Data

Authors: Jane A. Luke

Abstract:

Persistent pressure passed on to manufacturers from escalating consumer expectations and the ever growing global competitiveness have produced a rapidly increasing interest in the development of various manufacturing strategy models. Academic and industrial circles are taking keen interest in the field of manufacturing strategy. Many manufacturing strategies are currently centered on the traditional concepts of focused manufacturing capabilities such as quality, cost, dependability and innovation. Process capability indices was conducted assuming that the process under study is in statistical control and independent observations are generated over time. However, in practice, it is very common to come across processes which, due to their inherent natures, generate autocorrelated observations. The degree of autocorrelation affects the behavior of patterns on control charts. Even, small levels of autocorrelation between successive observations can have considerable effects on the statistical properties of conventional control charts. When observations are autocorrelated the classical control charts exhibit nonrandom patterns and lack of control. Many authors have considered the effect of autocorrelation on the performance of statistical process control charts. In this paper, the effect of autocorrelation on confidence intervals for different PCIs was included. Stationary Gaussian processes is explained. Effect of autocorrelation on PCIs is described in detail. Confidence intervals for Cp and Cpk are constructed for PCIs when data are both independent and autocorrelated. Confidence intervals for Cp and Cpk are computed. Approximate lower confidence limits for various Cpk are computed assuming AR(1) model for the data. Simulation studies and industrial examples are considered to demonstrate the results.

Keywords: autocorrelation, AR(1) model, Bissell’s approximation, confidence intervals, statistical process control, specification limits, stationary Gaussian processes

Procedia PDF Downloads 358
34607 Application of Failure Mode and Effects Analysis (FMEA) on the Virtual Process Hazard Analysis of Acetone Production Process

Authors: Princes Ann E. Prieto, Denise F. Alpuerto, John Rafael C. Unlayao, Neil Concibido, Monet Concepcion Maguyon-Detras

Abstract:

Failure Mode and Effects Analysis (FMEA) has been used in the virtual Process Hazard Analysis (PHA) of the Acetone production process through the dehydrogenation of isopropyl alcohol, for which very limited process risk assessment has been published. In this study, the potential failure modes, effects, and possible causes of selected major equipment in the process were identified. During the virtual FMEA mock sessions, the risks in the process were evaluated and recommendations to reduce and/or mitigate the process risks were formulated. The risk was estimated using the calculated risk priority number (RPN) and was classified into four (4) levels according to their effects on acetone production. Results of this study were also used to rank the criticality of equipment in the process based on the calculated criticality rating (CR). Bow tie diagrams were also created for the critical hazard scenarios identified in the study.

Keywords: chemical process safety, failure mode and effects analysis (FMEA), process hazard analysis (PHA), process safety management (PSM)

Procedia PDF Downloads 104
34606 Re-Engineering Management Process in IRAN’s Smart Schools

Authors: M. R. Babaei, S. M. Hosseini, S. Rahmani, L. Moradi

Abstract:

Today, the quality of education and training systems and the effectiveness of the education systems of most concern to stakeholders and decision-makers of our country's development in each country. In Iran this is a double issue of concern to numerous reasons; So that governments, over the past decade have hardly even paid the running costs of education. ICT is claiming it has the power to change the structure of a program for training, reduce costs and increase quality, and do education systems and products consistent with the needs of the community and take steps to practice education. Own of the areas that the introduction of information technology has fundamentally changed is the field of education. The aim of this research is process reengineering management in schools simultaneously has been using field studies to collect data in the form of interviews and a questionnaire survey. The statistical community of this research has been the country of Iran and smart schools under the education. Sampling was targeted. The data collection tool was a questionnaire composed of two parts. The questionnaire consists of 36 questions that each question designates one of effective factors on the management of smart schools. Also each question consists of two parts. The first part designates the operating position in the management process, which represents the domain's belonging to the management agent (planning, organizing, leading, controlling). According to the classification of Dabryn and in second part the factors affect the process of managing the smart schools were examined, that Likert scale is used to classify. Questions the validity of the group of experts and prominent university professors in the fields of information technology, management and reengineering of approved and Cronbach's alpha reliability and also with the use of the formula is evaluated and approved. To analyse the data, descriptive and inferential statistics were used to analyse the factors contributing to the rating of (Linkert scale) descriptive statistics (frequency table data, mean, median, mode) was used. To analyse the data using analysis of variance and nonparametric tests and Friedman test, the assumption was evaluated. The research conclusions show that the factors influencing the management process re-engineering smart schools in school performance is affected.

Keywords: re-engineering, management process, smart school, Iran's school

Procedia PDF Downloads 221
34605 The Impact of Mycotoxins on the Anaerobic Digestion Process

Authors: Harald Lindorfer, Bettina Frauz, Dietmar Ramhold

Abstract:

Next to the well-known inhibitors in anaerobic digestion like ammonia, antibiotics or disinfectants, the number of process failures connected with mould growth in the feedstock increased significantly in the last years. It was assumed that mycotoxins are the cause of the negative effects. The financial damage to plants associated with these process failures is considerable. The aim of this study was to find a way of predicting the failures and furthermore strategies for a fast process recovery. In a first step, mould-contaminated feedstocks causing process failures in full-scale digesters were sampled and analysed on mycotoxin content. A selection of these samples was applied to biological inhibition tests. In this test, crystalline cellulose is applied in addition to the feedstock sample as standard substrate. Affected digesters were also sampled and analytical process data as well as operational data of the plants were recorded. Additionally, different mycotoxin substances, Deoxynivalenol, Zearalenon, Aflatoxin B1, Mycophenolic acid and Citrinin, were applied as pure substances to lab-scale digesters, individually and in various combinations, and effects were monitored. As expected, various mycotoxins were detected in all of the mould-contaminated samples. Nevertheless, inhibition effects were observed with only one of the collected samples, after applying it to an inhibition test. With this sample, the biogas yield of the standard substrate was reduced by approx. 20%. This result corresponds with observations made on full-scale plants. However, none of the tested mycotoxins applied as pure substance caused a negative effect on biogas production in lab scale digesters, neither after application as individual substance nor in combination. The recording of the process data in full-scale plants affected by process failures in most cases showed a severe accumulation of fatty acids alongside a decrease in biogas production and methane concentration. In the analytical data of the digester samples, a typical distribution of fatty acids with exceptionally high acetic acid concentrations could be identified. This typical fatty acid pattern can be used as a rapid identification parameter pointing to the cause of the process troubles and enable a fast implication of countermeasures. The results of the study show that more attention needs to be paid to feedstock storage and feedstock conservation before their application to anaerobic digesters. This is all the more important since first studies indicate that the occurrence of mycotoxins will likely increase in Europe due to the ongoing climate change.

Keywords: Anaerobic digestion, Biogas, Feedstock conservation, Fungal mycotoxins, Inhibition, process failure

Procedia PDF Downloads 99
34604 A Holistic Workflow Modeling Method for Business Process Redesign

Authors: Heejung Lee

Abstract:

In a highly competitive environment, it becomes more important to shorten the whole business process while delivering or even enhancing the business value to the customers and suppliers. Although the workflow management systems receive much attention for its capacity to practically support the business process enactment, the effective workflow modeling method remain still challenging and the high degree of process complexity makes it more difficult to gain the short lead time. This paper presents a workflow structuring method in a holistic way that can reduce the process complexity using activity-needs and formal concept analysis, which eventually enhances the key performance such as quality, delivery, and cost in business process.

Keywords: workflow management, re-engineering, formal concept analysis, business process

Procedia PDF Downloads 382
34603 Big Data Analysis with Rhipe

Authors: Byung Ho Jung, Ji Eun Shin, Dong Hoon Lim

Abstract:

Rhipe that integrates R and Hadoop environment made it possible to process and analyze massive amounts of data using a distributed processing environment. In this paper, we implemented multiple regression analysis using Rhipe with various data sizes of actual data. Experimental results for comparing the performance of our Rhipe with stats and biglm packages available on bigmemory, showed that our Rhipe was more fast than other packages owing to paralleling processing with increasing the number of map tasks as the size of data increases. We also compared the computing speeds of pseudo-distributed and fully-distributed modes for configuring Hadoop cluster. The results showed that fully-distributed mode was faster than pseudo-distributed mode, and computing speeds of fully-distributed mode were faster as the number of data nodes increases.

Keywords: big data, Hadoop, Parallel regression analysis, R, Rhipe

Procedia PDF Downloads 476
34602 A Review on 3D Smart City Platforms Using Remotely Sensed Data to Aid Simulation and Urban Analysis

Authors: Slim Namouchi, Bruno Vallet, Imed Riadh Farah

Abstract:

3D urban models provide powerful tools for decision making, urban planning, and smart city services. The accuracy of this 3D based systems is directly related to the quality of these models. Since manual large-scale modeling, such as cities or countries is highly time intensive and very expensive process, a fully automatic 3D building generation is needed. However, 3D modeling process result depends on the input data, the proprieties of the captured objects, and the required characteristics of the reconstructed 3D model. Nowadays, producing 3D real-world model is no longer a problem. Remotely sensed data had experienced a remarkable increase in the recent years, especially data acquired using unmanned aerial vehicles (UAV). While the scanning techniques are developing, the captured data amount and the resolution are getting bigger and more precise. This paper presents a literature review, which aims to identify different methods of automatic 3D buildings extractions either from LiDAR or the combination of LiDAR and satellite or aerial images. Then, we present open source technologies, and data models (e.g., CityGML, PostGIS, Cesiumjs) used to integrate these models in geospatial base layers for smart city services.

Keywords: CityGML, LiDAR, remote sensing, SIG, Smart City, 3D urban modeling

Procedia PDF Downloads 102