Search results for: process capability indices
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16678

Search results for: process capability indices

16078 Ensuring Quality in DevOps Culture

Authors: Sagar Jitendra Mahendrakar

Abstract:

Integrating quality assurance (QA) practices into DevOps culture has become increasingly important in modern software development environments. Collaboration, automation and continuous feedback characterize the seamless integration of DevOps development and operations teams to achieve rapid and reliable software delivery. In this context, quality assurance plays a key role in ensuring that software products meet the highest quality, performance and reliability standards throughout the development life cycle. This brief explores key principles, challenges, and best practices related to quality assurance in a DevOps culture. This emphasizes the importance of quality transfer in the development process, as quality control processes are integrated in every step of the DevOps process. Automation is the cornerstone of DevOps quality assurance, enabling continuous testing, integration and deployment and providing rapid feedback for early problem identification and resolution. In addition, the summary addresses the cultural and organizational challenges of implementing quality assurance in DevOps, emphasizing the need to foster collaboration, break down silos, and promote a culture of continuous improvement. It also discusses the importance of toolchain integration and capability development to support effective QA practices in DevOps environments. Moreover, the abstract discusses the cultural and organizational challenges in implementing QA within DevOps, emphasizing the need for fostering collaboration, breaking down silos, and nurturing a culture of continuous improvement. It also addresses the importance of toolchain integration and skills development to support effective QA practices within DevOps environments. Overall, this collection works at the intersection of QA and DevOps culture, providing insights into how organizations can use DevOps principles to improve software quality, accelerate delivery, and meet the changing demands of today's dynamic software. landscape.

Keywords: quality engineer, devops, automation, tool

Procedia PDF Downloads 35
16077 Challenges, Chances and Possibilities during the Change Management Process of the National Defence Academy Vienna

Authors: Georg Ebner

Abstract:

The National Defence Academy, an element of the Austrian Ministry of Defence, is undergoing a transition process leading the Academy towards a new target structure that is currently being developed. In so doing, in addition to a subject-oriented approach, also an employee-oriented process was introduced. This process was initiated by the Ministry of Defence and should lead the National Defence Academy into a new constellation. During this process, the National Defence Academy worked in very special adapted World Café sessions. The “change manager” dealed with very different issues. They took the data feedback from the sessions and prepared with the feedback and information from the guidance the next session. So they got various information and a very different picture around the academy. It was very helpful to involve most of the employees of the academy during this process and to take their knowledge and wisdom. The process himself started with very different feelings and ended with great consent. A very interesting part of this process was also that the commander and his deputy worked together during all of this sessions and they answered all questions from the employees in time. The adapted World Café phases were necessary to deal with the information of the staff and to implement these absolutely needful data into this process. In cooperation with the responsible Headquarters, the first items resulting from the World Café phases could already be fed back to the employees and be implemented. The staff-oriented process is currently supported via a point of contact, through which the staff can contribute ideas as well, but also by the active information policy on the part of the Headquarters. The described change process makes innovative innovations possible. So far, in the event of change processes staff members have been entrusted only with the concrete implementation plan and tied into the process when the respective workplaces were to be re-staffed. The procedure described here can be seen as food-for-thought for further change processes. The findings of this process are that a staff oriented process can lead an organisation into a new era of thinking and working. This process has shown, that a lot of innovative ideas can also take place in a ministry. This process can be a background for a lot of change management processes in ministries and governmental and non-governmental organisations.

Keywords: both directions approach, change management, knowledge database, transformation process, World Cafe

Procedia PDF Downloads 176
16076 Importance of New Policies of Process Management for Internet of Things Based on Forensic Investigation

Authors: Venkata Venugopal Rao Gudlur

Abstract:

The Proposed Policies referred to as “SOP”, on the Internet of Things (IoT) based Forensic Investigation into Process Management is the latest revolution to save time and quick solution for investigators. The forensic investigation process has been developed over many years from time to time it has been given the required information with no policies in investigation processes. This research reveals that the current IoT based forensic investigation into Process Management based is more connected to devices which is the latest revolution and policies. All future development in real-time information on gathering monitoring is evolved with smart sensor-based technologies connected directly to IoT. This paper present conceptual framework on process management. The smart devices are leading the way in terms of automated forensic models and frameworks established by different scholars. These models and frameworks were mostly focused on offering a roadmap for performing forensic operations with no policies in place. These initiatives would bring a tremendous benefit to process management and IoT forensic investigators proposing policies. The forensic investigation process may enhance more security and reduced data losses and vulnerabilities.

Keywords: Internet of Things, Process Management, Forensic Investigation, M2M Framework

Procedia PDF Downloads 82
16075 Simplified Measurement of Occupational Energy Expenditure

Authors: J. Wicks

Abstract:

Aim: To develop a simple methodology to allow collected heart rate (HR) data from inexpensive wearable devices to be expressed in a suitable format (METs) to quantitate occupational (and recreational) activity. Introduction: Assessment of occupational activity is commonly done by utilizing questionnaires in combination with prescribed MET levels of a vast range of previously measured activities. However for any individual the intensity of performing a specific activity can vary significantly. Ideally objective measurement of individual activity is preferred. Though there are a wide range of HR recording devices there is a distinct lack methodology to allow processing of collected data to quantitate energy expenditure (EE). The HR index equation expresses METs in relation to relative HR i.e. the ratio of activity HR to resting HR. The use of this equation provides a simple utility for objective measurement of EE. Methods: During a typical occupational work period of approximately 8 hours HR data was recorded using a Polar RS 400 wrist monitor. Recorded data was downloaded to a Windows PC and non HR data was stripped from the ASCII file using ‘Notepad’. The HR data was exported to a spread sheet program and sorted by HR range into a histogram format. Three HRs were determined, namely a resting HR (the HR delimiting the lowest 30 minutes of recorded data), a mean HR and a peak HR (the HR delimiting the highest 30 minutes of recorded data). HR indices were calculated (mean index equals mean HR/rest HR and peak index equals peak HR/rest HR) with mean and peak indices being converted to METs using the HR index equation. Conclusion: Inexpensive HR recording devices can be utilized to make reasonable estimates of occupational (or recreational) EE suitable for large scale demographic screening by utilizing the HR index equation. The intrinsic value of the HR index equation is that it is independent of factors that influence absolute HR, namely fitness, smoking and beta-blockade.

Keywords: energy expenditure, heart rate histograms, heart rate index, occupational activity

Procedia PDF Downloads 282
16074 A Resilience Process Model of Natural Gas Pipeline Systems

Authors: Zhaoming Yang, Qi Xiang, Qian He, Michael Havbro Faber, Enrico Zio, Huai Su, Jinjun Zhang

Abstract:

Resilience is one of the key factors for system safety assessment and optimization, and resilience studies of natural gas pipeline systems (NGPS), especially in terms of process descriptions, are still being explored. Based on the three main stages, which are function loss process, recovery process, and waiting process, the paper has built functions and models which are according to the practical characteristics of NGPS and mainly analyzes the characteristics of deterministic interruptions. The resilience of NGPS also considers the threshold of the system function or users' satisfaction. The outcomes, which quantify the resilience of NGPS in different evaluation views, can be combined with the max flow and shortest path methods, help with the optimization of extra gas supplies and gas routes as well as pipeline maintenance strategies, the quick analysis of disturbance effects and the improvement of NGPS resilience evaluation accuracy.

Keywords: natural gas pipeline system, resilience, process modeling, deterministic disturbance

Procedia PDF Downloads 101
16073 A Gauge Repeatability and Reproducibility Study for Multivariate Measurement Systems

Authors: Jeh-Nan Pan, Chung-I Li

Abstract:

Measurement system analysis (MSA) plays an important role in helping organizations to improve their product quality. Generally speaking, the gauge repeatability and reproducibility (GRR) study is performed according to the MSA handbook stated in QS9000 standards. Usually, GRR study for assessing the adequacy of gauge variation needs to be conducted prior to the process capability analysis. Traditional MSA only considers a single quality characteristic. With the advent of modern technology, industrial products have become very sophisticated with more than one quality characteristic. Thus, it becomes necessary to perform multivariate GRR analysis for a measurement system when collecting data with multiple responses. In this paper, we take the correlation coefficients among tolerances into account to revise the multivariate precision-to-tolerance (P/T) ratio as proposed by Majeske (2008). We then compare the performance of our revised P/T ratio with that of the existing ratios. The simulation results show that our revised P/T ratio outperforms others in terms of robustness and proximity to the actual value. Moreover, the optimal allocation of several parameters such as the number of quality characteristics (v), sample size of parts (p), number of operators (o) and replicate measurements (r) is discussed using the confidence interval of the revised P/T ratio. Finally, a standard operating procedure (S.O.P.) to perform the GRR study for multivariate measurement systems is proposed based on the research results. Hopefully, it can be served as a useful reference for quality practitioners when conducting such study in industries. Measurement system analysis (MSA) plays an important role in helping organizations to improve their product quality. Generally speaking, the gauge repeatability and reproducibility (GRR) study is performed according to the MSA handbook stated in QS9000 standards. Usually, GRR study for assessing the adequacy of gauge variation needs to be conducted prior to the process capability analysis. Traditional MSA only considers a single quality characteristic. With the advent of modern technology, industrial products have become very sophisticated with more than one quality characteristic. Thus, it becomes necessary to perform multivariate GRR analysis for a measurement system when collecting data with multiple responses. In this paper, we take the correlation coefficients among tolerances into account to revise the multivariate precision-to-tolerance (P/T) ratio as proposed by Majeske (2008). We then compare the performance of our revised P/T ratio with that of the existing ratios. The simulation results show that our revised P/T ratio outperforms others in terms of robustness and proximity to the actual value. Moreover, the optimal allocation of several parameters such as the number of quality characteristics (v), sample size of parts (p), number of operators (o) and replicate measurements (r) is discussed using the confidence interval of the revised P/T ratio. Finally, a standard operating procedure (S.O.P.) to perform the GRR study for multivariate measurement systems is proposed based on the research results. Hopefully, it can be served as a useful reference for quality practitioners when conducting such study in industries.

Keywords: gauge repeatability and reproducibility, multivariate measurement system analysis, precision-to-tolerance ratio, Gauge repeatability

Procedia PDF Downloads 242
16072 State of Art in Software Requirement Negotiation Process Models

Authors: Shamsu Abdullahi, Nazir Yusuf, Hazrina Sofian, Abubakar Zakari, Amina Nura, Salisu Suleiman

Abstract:

Requirements negotiation process models help in resolving conflicting requirements of the heterogeneous stakeholders in the software development industry. This is to achieve a shared vision of software projects to be developed by the industry. Negotiating stakeholder agreements is a serious and difficult task in the software development process. There are many requirements negotiation process models that effectively negotiate stakeholder agreements that have been proposed by the research community. Other issues in the requirements negotiation research domain include stakeholder communication, decision-making, lack of negotiation interoperability, and managing requirement changes and analysis. This study highlights the current state of the art in the existing software requirements negotiation process models. The study also describes the issues and limitations in the software requirements negotiations process models.

Keywords: requirements, negotiation, stakeholders, agreements

Procedia PDF Downloads 171
16071 Fabrication of Silicon Solar Cells Using All Sputtering Process

Authors: Ching-Hua Li, Sheng-Hui Chen

Abstract:

Sputtering is a popular technique with many advantages for thin film deposition. To fabricate a hydrogenated silicon thin film using sputtering process for solar cell applications, the ion bombardment during sputtering will generate microstructures (voids and columnar structures) to form silicon dihydride bodings as defects. The properties of heterojunction silicon solar cells were studied by using boron grains and silicon-boron targets. Finally, an 11.7% efficiency of solar cell was achieved by using all sputtering process.

Keywords: solar cell, sputtering process, pvd, alloy target

Procedia PDF Downloads 565
16070 Revolutionizing Financial Forecasts: Enhancing Predictions with Graph Convolutional Networks (GCN) - Long Short-Term Memory (LSTM) Fusion

Authors: Ali Kazemi

Abstract:

Those within the volatile and interconnected international economic markets, appropriately predicting market trends, hold substantial fees for traders and financial establishments. Traditional device mastering strategies have made full-size strides in forecasting marketplace movements; however, monetary data's complicated and networked nature calls for extra sophisticated processes. This observation offers a groundbreaking method for monetary marketplace prediction that leverages the synergistic capability of Graph Convolutional Networks (GCNs) and Long Short-Term Memory (LSTM) networks. Our suggested algorithm is meticulously designed to forecast the traits of inventory market indices and cryptocurrency costs, utilizing a comprehensive dataset spanning from January 1, 2015, to December 31, 2023. This era, marked by sizable volatility and transformation in financial markets, affords a solid basis for schooling and checking out our predictive version. Our algorithm integrates diverse facts to construct a dynamic economic graph that correctly reflects market intricacies. We meticulously collect opening, closing, and high and low costs daily for key inventory marketplace indices (e.g., S&P 500, NASDAQ) and widespread cryptocurrencies (e.g., Bitcoin, Ethereum), ensuring a holistic view of marketplace traits. Daily trading volumes are also incorporated to seize marketplace pastime and liquidity, providing critical insights into the market's shopping for and selling dynamics. Furthermore, recognizing the profound influence of the monetary surroundings on financial markets, we integrate critical macroeconomic signs with hobby fees, inflation rates, GDP increase, and unemployment costs into our model. Our GCN algorithm is adept at learning the relational patterns amongst specific financial devices represented as nodes in a comprehensive market graph. Edges in this graph encapsulate the relationships based totally on co-movement styles and sentiment correlations, enabling our version to grasp the complicated community of influences governing marketplace moves. Complementing this, our LSTM algorithm is trained on sequences of the spatial-temporal illustration discovered through the GCN, enriched with historic fee and extent records. This lets the LSTM seize and expect temporal marketplace developments accurately. Inside the complete assessment of our GCN-LSTM algorithm across the inventory marketplace and cryptocurrency datasets, the version confirmed advanced predictive accuracy and profitability compared to conventional and opportunity machine learning to know benchmarks. Specifically, the model performed a Mean Absolute Error (MAE) of 0.85%, indicating high precision in predicting day-by-day charge movements. The RMSE was recorded at 1.2%, underscoring the model's effectiveness in minimizing tremendous prediction mistakes, which is vital in volatile markets. Furthermore, when assessing the model's predictive performance on directional market movements, it achieved an accuracy rate of 78%, significantly outperforming the benchmark models, averaging an accuracy of 65%. This high degree of accuracy is instrumental for techniques that predict the course of price moves. This study showcases the efficacy of mixing graph-based totally and sequential deep learning knowledge in economic marketplace prediction and highlights the fee of a comprehensive, records-pushed evaluation framework. Our findings promise to revolutionize investment techniques and hazard management practices, offering investors and economic analysts a powerful device to navigate the complexities of cutting-edge economic markets.

Keywords: financial market prediction, graph convolutional networks (GCNs), long short-term memory (LSTM), cryptocurrency forecasting

Procedia PDF Downloads 36
16069 Land Lots and Shannon-Winner Index in Sarpolzahab Agro Ecosystems-Western Iran

Authors: Ashkan Asgari, Korous Khoshbakht, Saeid Soufizadeh

Abstract:

Various factors including land lots can affect biodiversity indices in Agricultural systems. Field study conducted to evaluate factors affecting crop diversity in Sarpolzahab in 2012. Required data were collected through direct observation of farms and filling questionnaires. Total numbers of 140 questionnaires were filled, SAS Software was used to analyse data and Ecological Methodology Program was applied to calculate Shannon-Winner index, subsequently. Results of study indicated that average number of land lots for each farmer was 2.78 and various from 2.2 in Rikhak Olia Village to 4.31 in Golam Kaboud Olia Village which shows small size of land lots due to separating larger lots by children of deceased farmers. The correlation between number of land lots and species biodiversity (0.308**) was significant and Shannon-Winner index was (0.262**). Therefore, according to the mentioned results one can assume that increase in number of land lots results in improving of the target index. Multiple land lots allow farmers to cultivate various crops which results in increasing biodiversity of crops in agro ecosystem. Subsequently, this increase will facilitate economic sustainability of the farmers and distribution of work force in the region throughout the year. The correlation of seasonal workers with biodiversity of crop species (0.256**) and Shannon-Winner (0.286**) was statistically significant and increasing number of seasonal work forces had resulted in improving crop biodiversity and decreasing dominant species or single crop farming systems. Vegetable farms which have a significant diversity, require a significant number of work forces which describes correlation between number of workers and diversity of species.

Keywords: agricultural systems, biodiversity indices, Shannon-Winner index, sustainability, rural

Procedia PDF Downloads 523
16068 Understanding the Importance of Participation in the City Planning Process and Its Influencing Factors

Authors: Louis Nwachi

Abstract:

Urban planning systems in most countries still rely on expert-driven, top-down technocratic plan-making processes rather than a public and people-led process. This paper set out to evaluate the need for public participation in the plan-making process and to highlight the factors that affect public participation in the plan-making process. In doing this, it adopted a qualitative approach based on document review and interviews taken from real-world phenomena. A case study strategy using the Metropolitan Area of Abuja, the capital of Nigeria, as the study sample was used in carrying out the research. The research finds that participation is an important tool in the plan-making process and that public engagement in the process contributes to the identification of key urban issues that are unique to the specific local areas, thereby contributing to the establishment of priorities and, in turn, to the mobilization of resources to meet the identified needs. It also finds that the development of a participation model by city authorities encourages public engagement and helps to develop trust between those in authority and the different key stakeholder groups involved in the plan-making process.

Keywords: plan-making, participation, urban planning, city

Procedia PDF Downloads 80
16067 Quality Based Approach for Efficient Biologics Manufacturing

Authors: Takashi Kaminagayoshi, Shigeyuki Haruyama

Abstract:

To improve the manufacturing efficiency of biologics, such as antibody drugs, a quality engineering framework was designed. Within this framework, critical steps and parameters in the manufacturing process were studied. Identification of these critical steps and critical parameters allows a deeper understanding of manufacturing capabilities, and suggests to process development department process control standards based on actual manufacturing capabilities as part of a PDCA (plan-do-check-act) cycle. This cycle can be applied to each manufacturing process so that it can be standardized, reducing the time needed to establish each new process.

Keywords: antibody drugs, biologics, manufacturing efficiency, PDCA cycle, quality engineering

Procedia PDF Downloads 333
16066 A Greedy Alignment Algorithm Supporting Medication Reconciliation

Authors: David Tresner-Kirsch

Abstract:

Reconciling patient medication lists from multiple sources is a critical task supporting the safe delivery of patient care. Manual reconciliation is a time-consuming and error-prone process, and recently attempts have been made to develop efficiency- and safety-oriented automated support for professionals performing the task. An important capability of any such support system is automated alignment – finding which medications from a list correspond to which medications from a different source, regardless of misspellings, naming differences (e.g. brand name vs. generic), or changes in treatment (e.g. switching a patient from one antidepressant class to another). This work describes a new algorithmic solution to this alignment task, using a greedy matching approach based on string similarity, edit distances, concept extraction and normalization, and synonym search derived from the RxNorm nomenclature. The accuracy of this algorithm was evaluated against a gold-standard corpus of 681 medication records; this evaluation found that the algorithm predicted alignments with 99% precision and 91% recall. This performance is sufficient to support decision support applications for medication reconciliation.

Keywords: clinical decision support, medication reconciliation, natural language processing, RxNorm

Procedia PDF Downloads 267
16065 An Analysis of Packaging Materials for an Energy-Efficient Wrapping System

Authors: John Sweeney, Martin Leeming, Raj Thaker, Cristina L. Tuinea-Bobe

Abstract:

Shrink wrapping is widely used as a method for secondary packaging to assemble individual items, such as cans or other consumer products, into single packages. This method involves conveying the packages into heated tunnels and so has the disadvantages that it is energy-intensive, and, in the case of aerosol products, potentially hazardous. We are developing an automated packaging system that uses stretch wrapping to address both these problems, by using a mechanical rather than a thermal process. In this study, we present a comparative study of shrink wrapping and stretch wrapping materials to assess the relative capability of candidate stretch wrap polymer film in terms of mechanical response. The stretch wrap materials are of oriented polymer and therefore elastically anisotropic. We are developing material constitutive models that include both anisotropy and nonlinearity. These material models are to be incorporated into computer simulations of the automated stretch wrapping system. We present results showing the validity of these models and the feasibility of applying them in the simulations.

Keywords: constitutive model, polymer, mechanical testing, wrapping system

Procedia PDF Downloads 283
16064 Implementation of Deep Neural Networks for Pavement Condition Index Prediction

Authors: M. Sirhan, S. Bekhor, A. Sidess

Abstract:

In-service pavements deteriorate with time due to traffic wheel loads, environment, and climate conditions. Pavement deterioration leads to a reduction in their serviceability and structural behavior. Consequently, proper maintenance and rehabilitation (M&R) are necessary actions to keep the in-service pavement network at the desired level of serviceability. Due to resource and financial constraints, the pavement management system (PMS) prioritizes roads most in need of maintenance and rehabilitation action. It recommends a suitable action for each pavement based on the performance and surface condition of each road in the network. The pavement performance and condition are usually quantified and evaluated by different types of roughness-based and stress-based indices. Examples of such indices are Pavement Serviceability Index (PSI), Pavement Serviceability Ratio (PSR), Mean Panel Rating (MPR), Pavement Condition Rating (PCR), Ride Number (RN), Profile Index (PI), International Roughness Index (IRI), and Pavement Condition Index (PCI). PCI is commonly used in PMS as an indicator of the extent of the distresses on the pavement surface. PCI values range between 0 and 100; where 0 and 100 represent a highly deteriorated pavement and a newly constructed pavement, respectively. The PCI value is a function of distress type, severity, and density (measured as a percentage of the total pavement area). PCI is usually calculated iteratively using the 'Paver' program developed by the US Army Corps. The use of soft computing techniques, especially Artificial Neural Network (ANN), has become increasingly popular in the modeling of engineering problems. ANN techniques have successfully modeled the performance of the in-service pavements, due to its efficiency in predicting and solving non-linear relationships and dealing with an uncertain large amount of data. Typical regression models, which require a pre-defined relationship, can be replaced by ANN, which was found to be an appropriate tool for predicting the different pavement performance indices versus different factors as well. Subsequently, the objective of the presented study is to develop and train an ANN model that predicts the PCI values. The model’s input consists of percentage areas of 11 different damage types; alligator cracking, swelling, rutting, block cracking, longitudinal/transverse cracking, edge cracking, shoving, raveling, potholes, patching, and lane drop off, at three severity levels (low, medium, high) for each. The developed model was trained using 536,000 samples and tested on 134,000 samples. The samples were collected and prepared by The National Transport Infrastructure Company. The predicted results yielded satisfactory compliance with field measurements. The proposed model predicted PCI values with relatively low standard deviations, suggesting that it could be incorporated into the PMS for PCI determination. It is worth mentioning that the most influencing variables for PCI prediction are damages related to alligator cracking, swelling, rutting, and potholes.

Keywords: artificial neural networks, computer programming, pavement condition index, pavement management, performance prediction

Procedia PDF Downloads 122
16063 Flowsheet Development, Simulation and Optimization of Carbon-Di-Oxide Removal System at Natural Gas Reserves by Aspen–Hysys Process Simulator

Authors: Mohammad Ruhul Amin, Nusrat Jahan

Abstract:

Natural gas is a cleaner fuel compared to the others. But it needs some treatment before it is in a state to be used. So natural gas purification is an integral part of any process where natural gas is used as raw material or fuel. There are several impurities in natural gas that have to be removed before use. CO2 is one of the major contaminants. In this project we have removed CO2 by amine process by using MEA solution. We have built up the whole amine process for removing CO2 in Aspen Hysys and simulated the process. At the end of simulation we have got very satisfactory results by using MEA solution for the removal of CO2. Simulation result shows that amine absorption process enables to reduce CO2 content from NG by 58%. HYSYS optimizer allowed us to get a perfect optimized plant. After optimization the profit of existing plant is increased by 2.34 %.Simulation and optimization by Aspen-HYSYS simulator makes available us to enormous information which will help us to further research in future.

Keywords: Aspen–Hysys, CO2 removal, flowsheet development, MEA solution, natural gas optimization

Procedia PDF Downloads 476
16062 Supply Chain Logistics Integration in Bahrain's Construction Industry

Authors: Randolf Von N. Salindo

Abstract:

The study was conducted to measure the logistics integration capabilities of selected companies in the Bahrain construction industry using the Supply Chain 2000 framework; and, determine the extent and direction of influence of these logistics capabilities and integration competencies on the supply chain performance of the firm. A total of 50 executive respondents (from supervisor to managing director level) from 22 construction and construction supplier firms participated in the study from September to November 2014. The results reveal that respondent Bahraini construction firms have significantly lower levels of logistics capabilities, but higher levels of logistics integration competencies compared to international benchmarks. Using stepwise multiple regression analysis, eight logistics capabilities of Bahraini constructions firms were identified to be positively associated with firm performance; with comprehensive metrics as the most positively dominant influential logistics capability. Activity based and total cost methodology is found to be the most negatively dominant influential logistics capability. In terms of logistics integration competencies, the study revealed that that customer integration, internal integration, and, measurement integration are negatively associated with firm performance. There was no logistics integration competency found to be positively associated with the supply chain performance among the companies who participated in the study. The research reveals that there are areas for improvement in supply chain capabilities and logistics integration competencies of the construction firms in the Kingdom of Bahrain to improve their supply chain performance to a global level.

Keywords: comprehensive metrics, customer integration, logistics integration capabilities, logistics integration competencies

Procedia PDF Downloads 624
16061 The Role of Information and Communication Technology to Enhance Transparency in Public Funds Management in the DR Congo

Authors: Itulelo Matiyabu Imaja, Manoj Maharaj, Patrick Ndayizigamiye

Abstract:

Lack of transparency in public funds management is observed in many African countries. The DR Congo is among the most corrupted countries in Africa, and this is due mainly to lack of transparency and accountability in public funds management. Corruption has a negative effect on the welfare of the country’s citizens and the national economic growth. Public funds collection and allocation are the major areas whereby malpractices such as bribe, extortion, embezzlement, nepotism and other practices related to corruption are prevalent. Hence, there is a need to implement strong mechanisms to enforce transparency in public funds management. Many researchers have suggested some control mechanisms in curbing corruption in public funds management focusing mainly on law enforcement and administrative reforms with little or no insight on the role that ICT can play in preventing and curbing the corrupt behaviour. In the Democratic Republic of Congo (DRC), there are slight indications that the government of the DR Congo is integrating ICT to fight corruption in public funds collection and allocation. However, such government initiatives are at an infancy stage, with no tangible evidence on how ICT could be used effectively to address the issue of corruption in the context of the country. Hence, this research assesses the role that ICT can play for transparency in public funds management and suggest a framework for its adoption in the Democratic Republic of Congo. This research uses the revised Capability model (Capability, Empowerment, Sustainability model) as the guiding theoretical framework. The study uses the exploratory design methodology coupled with a qualitative approach to data collection and purposive sampling as sampling strategy.

Keywords: corruption, DR congo, ICT, management, public funds, transparency

Procedia PDF Downloads 323
16060 Analysis of the Performance of State Institutions From 2008-2013 in Pakistan

Authors: Mahrukh Shehzadi

Abstract:

Pakistan is a democratic republic but has spent much time under military rulers; after a few years of independence, Pakistan faced three martial laws in 1958, 1969, and 1977, and the latest in 1999 by General Musharraf. The purpose of this thesis is to analyze the politics, policies and overall performance of Pakistan People’s Party Government from 2008-2013. PPP won a significant victory in the elections of 2008. The co-chairman, Mr. Asif Ali Zardari, announced the end of the fourth dictatorship. It was for the first time in Pakistan’s history that an elected government completed its term (2008-2013). While the completion of its term is an achievement, the performance of the democratically-elected government – federal, provincial and local does not inspire much confidence. Poor governance, persistent confrontational relations between the executive and the judiciary, charges of corruption, and the incompetence of the political leadership to build consensus to combat terrorism continue to cast criticisms on the democratic process and the civilian regime’s capability to sustain democracy. In the present study, the researcher will try to describe and explain the public thinking pattern regarding the policies opted for by the PPP-led government and their impact on the people’s minds of Pakistan.

Keywords: democracy, performance, policies, state, manifesto

Procedia PDF Downloads 43
16059 A Study on FWD Deflection Bowl Parameters for Condition Assessment of Flexible Pavement

Authors: Ujjval J. Solanki, Prof.(Dr.) P.J. Gundaliya, Prof.M.D. Barasara

Abstract:

The application of Falling Weight Deflectometer is to evaluate structural performance of the flexible pavement. The exercise of back calculation is required to know the modulus of elasticity of existing in-service pavement. The process of back calculation needs in-depth field experience for the input of range of modulus of elasticity of bituminous, granular and subgrade layer, and its required number of trial to find such matching moduli with the observed FWD deflection on the field. The study carried out at Barnala-Mansa State Highway Punjab-India using FWD before and after overlay; the deflections obtained at 0 on the load cell, 300, 600, 900,1200, 1500 and 1800 mm interval from the load cell these seven deflection results used to calculate Surface Curvature Index (SCI), Base damage Index (BDI), Base curvature index (BCI). This SCI, BCI and BDI indices are useful to predict the structural performance of in-service pavement and also useful to identify homogeneous section for condition assessment. The SCI, BCI and BDI range are determined for before and after overlay the range of SCI 520 to 51 BDI 294 to 63 BCI 83 to 0.27 for old pavement and SCI 272 to 23 BDI 228 to 28, BCI 25.85 to 4.60 for new pavement. It also shows good correlation with back calculated modulus of elasticity of all the three layer.

Keywords: back calculation, base damage index, base curvature index, FWD (Falling Weight Deflectometer), surface curvature index

Procedia PDF Downloads 316
16058 Evaluation of Reliability, Availability and Maintainability for Automotive Manufacturing Process

Authors: Hamzeh Soltanali, Abbas Rohani, A. H. S. Garmabaki, Mohammad Hossein Abbaspour-Fard, Adithya Thaduri

Abstract:

Toward continuous innovation and high complexity of technological systems, the automotive manufacturing industry is also under pressure to implement adequate management strategies regarding availability and productivity. In this context, evaluation of system’s performance by considering reliability, availability and maintainability (RAM) methodologies can constitute for resilient operation, identifying the bottlenecks of manufacturing process and optimization of maintenance actions. In this paper, RAM parameters are evaluated for improving the operational performance of the fluid filling process. To evaluate the RAM factors through the behavior of states defined for such process, a systematic decision framework was developed. The results of RAM analysis revealed that that the improving reliability and maintainability of main bottlenecks for each filling workstation need to be considered as a priority. The results could be useful to improve operational performance and sustainability of production process.

Keywords: automotive, performance, reliability, RAM, fluid filling process

Procedia PDF Downloads 336
16057 Molecular Topology and TLC Retention Behaviour of s-Triazines: QSRR Study

Authors: Lidija R. Jevrić, Sanja O. Podunavac-Kuzmanović, Strahinja Z. Kovačević

Abstract:

Quantitative structure-retention relationship (QSRR) analysis was used to predict the chromatographic behavior of s-triazine derivatives by using theoretical descriptors computed from the chemical structure. Fundamental basis of the reported investigation is to relate molecular topological descriptors with chromatographic behavior of s-triazine derivatives obtained by reversed-phase (RP) thin layer chromatography (TLC) on silica gel impregnated with paraffin oil and applied ethanol-water (φ = 0.5-0.8; v/v). Retention parameter (RM0) of 14 investigated s-triazine derivatives was used as dependent variable while simple connectivity index different orders were used as independent variables. The best QSRR model for predicting RM0 value was obtained with simple third order connectivity index (3χ) in the second-degree polynomial equation. Numerical values of the correlation coefficient (r=0.915), Fisher's value (F=28.34) and root mean square error (RMSE = 0.36) indicate that model is statistically significant. In order to test the predictive power of the QSRR model leave-one-out cross-validation technique has been applied. The parameters of the internal cross-validation analysis (r2CV=0.79, r2adj=0.81, PRESS=1.89) reflect the high predictive ability of the generated model and it confirms that can be used to predict RM0 value. Multivariate classification technique, hierarchical cluster analysis (HCA), has been applied in order to group molecules according to their molecular connectivity indices. HCA is a descriptive statistical method and it is the most frequently used for important area of data processing such is classification. The HCA performed on simple molecular connectivity indices obtained from the 2D structure of investigated s-triazine compounds resulted in two main clusters in which compounds molecules were grouped according to the number of atoms in the molecule. This is in agreement with the fact that these descriptors were calculated on the basis of the number of atoms in the molecule of the investigated s-triazine derivatives.

Keywords: s-triazines, QSRR, chemometrics, chromatography, molecular descriptors

Procedia PDF Downloads 375
16056 The Mechanism Study on the Difference between High and Low Voltage Performance of Li3V2(PO4)3

Authors: Enhui Wang, Qingzhu Ou, Yan Tang, Xiaodong Guo

Abstract:

As one of most popular polyanionic compounds in lithium-ion cathode materials, Li3V2(PO4)3 has always suffered from the low rate capability especially during 3~4.8V, which is considered to be related with the ion diffusion resistance and structural transformation during the Li+ de/intercalation. Here, as the change of cut-off voltages, cycling numbers and current densities, the process of SEI interfacial film’s formation-growing- destruction-repair on the surface of the cathode, the structural transformation during the charge and discharge, the de/intercalation kinetics reflected by the electrochemical impedance and the diffusion coefficient, have been investigated in detail. Current density, cycle numbers and cut-off voltage impacting on interfacial film and structure was studied specifically. Firstly, the matching between electrolyte and material was investigated, it turned out that the batteries with high voltage electrolyte showed the best electrochemical performance and high voltage electrolyte would be the best electrolyte. Secondly, AC impedance technology was used to study the changes of interface impedance and lithium ion diffusion coefficient, the results showed that current density, cycle numbers and cut-off voltage influenced the interfacial film together and the one who changed the interfacial properties most was the key factor. Scanning electron microscopy (SEM) analysis confirmed that the attenuation of discharge specific capacity was associated with the destruction and repair process of the SEI film. Thirdly, the X-ray diffraction was used to study the changes of structure, which was also impacted by current density, cycle numbers and cut-off voltage. The results indicated that the cell volume of Li3V2 (PO4 )3 increased as the current density increased; cycle numbers merely influenced the structure of material; the cell volume decreased first and moved back gradually after two Li-ion had been deintercalated as the charging cut-off voltage increased, and increased as the intercalation number of Li-ion increased during the discharging process. Then, the results which studied the changes of interface impedance and lithium ion diffusion coefficient turned out that the interface impedance and lithium ion diffusion coefficient increased when the cut-off voltage passed the voltage platforms and decreased when the cut-off voltage was between voltage platforms. Finally, three-electrode system was first adopted to test the activation energy of the system, the results indicated that the activation energy of the three-electrode system (22.385 KJ /mol) was much smaller than that of two-electrode system (40.064 KJ /mol).

Keywords: cut-off voltage, de/intercalation kinetics, solid electrolyte interphase film, structural transformation

Procedia PDF Downloads 284
16055 Understanding Context and Its Effects in the Implementation of Modern Foreign Language Curriculum in Vietnam

Authors: Ngoc T. Bui

Abstract:

The key issue for teachers of a modern foreign language is the creation of a pedagogic environment, and this means that an understanding of context is vital. A pedagogic environment addresses the following: time, feedback, relations with other people, curriculum integration, forms of knowledge, resources and control in the pedagogic relationship. In this light, the multiple case study of the implementation of a modern foreign language curriculum focuses on exploring Vietnamese contexts and participants’ perceptions of factors that may affect their implementation process in order to examine thoroughly how the communicative language teaching (CLT) curriculum is being implemented in second language classrooms. A mixed methods approach is utilized to investigate contextual and personal factors that may affect teachers’ implementation of curriculum and pedagogical reform in Vietnam. This project therefore has the capability to inform stakeholders of useful information and identify further changes and measures to solve potential problems to ensure the achievement of the curriculum goals. The expected outcomes may also lead to intercultural language teaching guidelines to support english as a foreign language (EFL) teachers with curriculum design, planning and how to create pedagogic environment to best implement it.

Keywords: communicative language teaching, context, curriculum implementation, modern foreign language, pedagogic environment

Procedia PDF Downloads 251
16054 A Three-Step Iterative Process for Common Fixed Points of Three Contractive-Like Operators

Authors: Safeer Hussain Khan, H. Fukhar-ud-Din

Abstract:

The concept of quasi-contractive type operators was given by Berinde and extended by Imoru and Olatinwo. They named this new type as contractive-like operators. On the other hand, Xu and Noo introduced a three-step-one-mappings iterative process which can be seen as a generalization of Mann and Ishikawa iterative processes. Approximating common fixed points has its own importance as it has a direct link with minimization problem. Motivated by this, in this paper, we first extend the iterative process of Xu and Noor to the case of three-step-three-mappings and then prove a strong convergence result using contractive-like operators for this iterative process. In general, this generalizes corresponding results using Mann, Ishikawa and Xu-Noor iterative processes with quasi-contractive type operators. It is to be pointed out that our results can also be proved with iterative process involving error terms.

Keywords: contractive-like operator, iterative process, common fixed point, strong convergence

Procedia PDF Downloads 575
16053 QSRR Analysis of 17-Picolyl and 17-Picolinylidene Androstane Derivatives Based on Partial Least Squares and Principal Component Regression

Authors: Sanja Podunavac-Kuzmanović, Strahinja Kovačević, Lidija Jevrić, Evgenija Djurendić, Jovana Ajduković

Abstract:

There are several methods for determination of the lipophilicity of biologically active compounds, however chromatography has been shown as a very suitable method for this purpose. Chromatographic (C18-RP-HPLC) analysis of a series of 24 17-picolyl and 17-picolinylidene androstane derivatives was carried out. The obtained retention indices (logk, methanol (90%) / water (10%)) were correlated with calculated physicochemical and lipophilicity descriptors. The QSRR analysis was carried out applying principal component regression (PCR) and partial least squares regression (PLS). The PCR and PLS model were selected on the basis of the highest variance and the lowest root mean square error of cross-validation. The obtained PCR and PLS model successfully correlate the calculated molecular descriptors with logk parameter indicating the significance of the lipophilicity of compounds in chromatographic process. On the basis of the obtained results it can be concluded that the obtained logk parameters of the analyzed androstane derivatives can be considered as their chromatographic lipophilicity. These results are the part of the project No. 114-451-347/2015-02, financially supported by the Provincial Secretariat for Science and Technological Development of Vojvodina and CMST COST Action CM1105.

Keywords: androstane derivatives, chromatography, molecular structure, principal component regression, partial least squares regression

Procedia PDF Downloads 253
16052 Requirements Management in Agile

Authors: Ravneet Kaur

Abstract:

The concept of Agile Requirements Engineering and Management is not new. However, the struggle to figure out how traditional Requirements Management Process fits within an Agile framework remains complex. This paper talks about a process that can merge the organization’s traditional Requirements Management Process nicely into the Agile Software Development Process. This process provides Traceability of the Product Backlog to the external documents on one hand and User Stories on the other hand. It also gives sufficient evidence that the system will deliver the right functionality with good quality in the form of various statistics and reports. In the nutshell, by overlaying a process on top of Agile, without disturbing the Agility, we are able to get synergic benefits in terms of productivity, profitability, its reporting, and end to end visibility to all Stakeholders. The framework can be used for just-in-time requirements definition or to build a repository of requirements for future use. The goal is to make sure that the business (specifically, the product owner) can clearly articulate what needs to be built and define what is of high quality. To accomplish this, the requirements cycle follows a Scrum-like process that mirrors the development cycle but stays two to three steps ahead. The goal is to create a process by which requirements can be thoroughly vetted, organized, and communicated in a manner that is iterative, timely, and quality-focused. Agile is quickly becoming the most popular way of developing software because it fosters continuous improvement, time-boxed development cycles, and more quickly delivering value to the end users. That value will be driven to a large extent by the quality and clarity of requirements that feed the software development process. An agile, lean, and timely approach to requirements as the starting point will help to ensure that the process is optimized.

Keywords: requirements management, Agile

Procedia PDF Downloads 354
16051 A Study on Stochastic Integral Associated with Catastrophes

Authors: M. Reni Sagayaraj, S. Anand Gnana Selvam, R. Reynald Susainathan

Abstract:

We analyze stochastic integrals associated with a mutation process. To be specific, we describe the cell population process and derive the differential equations for the joint generating functions for the number of mutants and their integrals in generating functions and their applications. We obtain first-order moments of the processes of the two-way mutation process in first-order moment structure of X (t) and Y (t) and the second-order moments of a one-way mutation process. In this paper, we obtain the limiting behaviour of the integrals in limiting distributions of X (t) and Y (t).

Keywords: stochastic integrals, single–server queue model, catastrophes, busy period

Procedia PDF Downloads 627
16050 Investigation of the Operational Principle and Flow Analysis of a Newly Developed Dry Separator

Authors: Sung Uk Park, Young Su Kang, Sangmo Kang, Young Kweon Suh

Abstract:

Mineral product, waste concrete (fine aggregates), waste in the optical field, industry, and construction employ separators to separate solids and classify them according to their size. Various sorting machines are used in the industrial field such as those operating under electrical properties, centrifugal force, wind power, vibration, and magnetic force. Study on separators has been carried out to contribute to the environmental industry. In this study, we perform CFD analysis for understanding the basic mechanism of the separation of waste concrete (fine aggregate) particles from air with a machine built with a rotor with blades. In CFD, we first performed two-dimensional particle tracking for various particle sizes for the model with 1 degree, 1.5 degree, and 2 degree angle between each blade to verify the boundary conditions and the method of rotating domain method to be used in 3D. Then we developed 3D numerical model with ANSYS CFX to calculate the air flow and track the particles. We judged the capability of particle separation for given size by counting the number of particles escaping from the domain toward the exit among 10 particles issued at the inlet. We confirm that particles experience stagnant behavior near the exit of the rotating blades where the centrifugal force acting on the particles is in balance with the air drag force. It was also found that the minimum particle size that can be separated by the machine with the rotor is determined by its capability to stay at the outlet of the rotor channels.

Keywords: environmental industry, separator, CFD, fine aggregate

Procedia PDF Downloads 580
16049 Process Mining as an Ecosystem Platform to Mitigate a Deficiency of Processes Modelling

Authors: Yusra Abdulsalam Alqamati, Ahmed Alkilany

Abstract:

The teaching staff is a distinct group whose impact is on the educational process and which plays an important role in enhancing the quality of the academic education process. To improve the management effectiveness of the academy, the Teaching Staff Management System (TSMS) proposes that all teacher processes be digitized. Since the BPMN approach can accurately describe the processes, it lacks a clear picture of the process flow map, something that the process mining approach has, which is extracting information from event logs for discovery, monitoring, and model enhancement. Therefore, these two methodologies were combined to create the most accurate representation of system operations, the ability to extract data records and mining processes, recreate them in the form of a Petri net, and then generate them in a BPMN model for a more in-depth view of process flow. Additionally, the TSMS processes will be orchestrated to handle all requests in a guaranteed small-time manner thanks to the integration of the Google Cloud Platform (GCP), the BPM engine, and allowing business owners to take part throughout the entire TSMS project development lifecycle.

Keywords: process mining, BPM, business process model and notation, Petri net, teaching staff, Google Cloud Platform

Procedia PDF Downloads 123