Search results for: buying decision process
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17904

Search results for: buying decision process

15984 Organizational Decision to Adopt Digital Forensics: An Empirical Investigation in the Case of Malaysian Law Enforcement Agencies

Authors: Siti N. I. Mat Kamal, Othman Ibrahim, Mehrbakhsh Nilashi, Jafalizan M. Jali

Abstract:

The use of digital forensics (DF) is nowadays essential for law enforcement agencies to identify analysis and interpret the digital information derived from digital sources. In Malaysia, the engagement of Malaysian Law Enforcement Agencies (MLEA) with this new technology is not evenly distributed. To investigate the factors influencing the adoption of DF in Malaysia law enforcement agencies’ operational environment, this study proposed the initial theoretical framework based on the integration of technology organization environment (TOE), institutional theory, and human organization technology (HOT) fit model. A questionnaire survey was conducted on selected law enforcement agencies in Malaysia to verify the validity of the initial integrated framework. Relative advantage, compatibility, coercive pressure, normative pressure, vendor support and perceived technical competence of technical staff were found as the influential factors on digital forensics adoption. In addition to the only moderator of this study (agency size), any significant moderating effect on the perceived technical competence and the decision to adopt digital forensics by Malaysian law enforcement agencies was found insignificant. Thus, these results indicated that the developed integrated framework provides an effective prediction of the digital forensics adoption by Malaysian law enforcement agencies.

Keywords: digital forensics, digital forensics adoption, digital information, law enforcement agency

Procedia PDF Downloads 150
15983 The Impact of Cognition and Communication on the Defense of Capital Murder Cases

Authors: Shameka Stanford

Abstract:

This presentation will discuss how cognitive and communication disorders in the areas of executive functioning, receptive and expressive language can impact the problem-solving and decision making of individuals with such impairments. More specifically, this presentation will discuss approaches the legal defense team of capital case lawyers can add to their experience when servicing individuals who have a history of educational decline, special education, and limited intervention and treatment. The objective of the research is to explore and identify the correlations between impaired executive function skills and decision making and competency for individuals facing death penalty charges. To conduct this research, experimental design, randomized sampling, qualitative analysis was employed. This research contributes to the legal and criminal justice system related to how they view, defend, and characterize, and judge individuals with documented cognitive and communication disorders who are eligible for capital case charges. More importantly, this research contributes to the increased ability of death penalty lawyers to successfully defend clients with a history of academic difficulty, special education, and documented disorders that impact educational progress and academic success.

Keywords: communication disorders, cognitive disorders, capital murder, death penalty, executive function

Procedia PDF Downloads 153
15982 Virtual Metrology for Copper Clad Laminate Manufacturing

Authors: Misuk Kim, Seokho Kang, Jehyuk Lee, Hyunchang Cho, Sungzoon Cho

Abstract:

In semiconductor manufacturing, virtual metrology (VM) refers to methods to predict properties of a wafer based on machine parameters and sensor data of the production equipment, without performing the (costly) physical measurement of the wafer properties (Wikipedia). Additional benefits include avoidance of human bias and identification of important factors affecting the quality of the process which allow improving the process quality in the future. It is however rare to find VM applied to other areas of manufacturing. In this work, we propose to use VM to copper clad laminate (CCL) manufacturing. CCL is a core element of a printed circuit board (PCB) which is used in smartphones, tablets, digital cameras, and laptop computers. The manufacturing of CCL consists of three processes: Treating, lay-up, and pressing. Treating, the most important process among the three, puts resin on glass cloth, heat up in a drying oven, then produces prepreg for lay-up process. In this process, three important quality factors are inspected: Treated weight (T/W), Minimum Viscosity (M/V), and Gel Time (G/T). They are manually inspected, incurring heavy cost in terms of time and money, which makes it a good candidate for VM application. We developed prediction models of the three quality factors T/W, M/V, and G/T, respectively, with process variables, raw material, and environment variables. The actual process data was obtained from a CCL manufacturer. A variety of variable selection methods and learning algorithms were employed to find the best prediction model. We obtained prediction models of M/V and G/T with a high enough accuracy. They also provided us with information on “important” predictor variables, some of which the process engineers had been already aware and the rest of which they had not. They were quite excited to find new insights that the model revealed and set out to do further analysis on them to gain process control implications. T/W did not turn out to be possible to predict with a reasonable accuracy with given factors. The very fact indicates that the factors currently monitored may not affect T/W, thus an effort has to be made to find other factors which are not currently monitored in order to understand the process better and improve the quality of it. In conclusion, VM application to CCL’s treating process was quite successful. The newly built quality prediction model allowed one to reduce the cost associated with actual metrology as well as reveal some insights on the factors affecting the important quality factors and on the level of our less than perfect understanding of the treating process.

Keywords: copper clad laminate, predictive modeling, quality control, virtual metrology

Procedia PDF Downloads 348
15981 Benefit Of Waste Collection Route Optimisation

Authors: Bojana Tot, Goran BošKović, Goran Vujić

Abstract:

Route optimisation is a process of planning one or multiple routes, with the purpose of minimizing overall costs, while achieving the highest possible performance under a set of given constraints. It combines routing or route planning, which is the process of creating the most cost-effective route by minimizing the distance or travelled time necessary to reach a set of planned stops, and route scheduling, which is the process of assigning an arrival and service time for each stop, with drivers being given shifts that adhere to their working hours. The objective of this paper is to provide benefits on the implementation of waste collection route optimisation and thus achieve economic efficiency for public utility companies, better service for citizens and positive environment and health.

Keywords: waste management, environment, collection route optimisation, GIS

Procedia PDF Downloads 155
15980 E-Consumers’ Attribute Non-Attendance Switching Behavior: Effect of Providing Information on Attributes

Authors: Leonard Maaya, Michel Meulders, Martina Vandebroek

Abstract:

Discrete Choice Experiments (DCE) are used to investigate how product attributes affect decision-makers’ choices. In DCEs, choice situations consisting of several alternatives are presented from which choice-makers select the preferred alternative. Standard multinomial logit models based on random utility theory can be used to estimate the utilities for the attributes. The overarching principle in these models is that respondents understand and use all the attributes when making choices. However, studies suggest that respondents sometimes ignore some attributes (commonly referred to as Attribute Non-Attendance/ANA). The choice modeling literature presents ANA as a static process, i.e., respondents’ ANA behavior does not change throughout the experiment. However, respondents may ignore attributes due to changing factors like availability of information on attributes, learning/fatigue in experiments, etc. We develop a dynamic mixture latent Markov model to model changes in ANA when information on attributes is provided. The model is illustrated on e-consumers’ webshop choices. The results indicate that the dynamic ANA model describes the behavioral changes better than modeling the impact of information using changes in parameters. Further, we find that providing information on attributes leads to an increase in the attendance probabilities for the investigated attributes.

Keywords: choice models, discrete choice experiments, dynamic models, e-commerce, statistical modeling

Procedia PDF Downloads 134
15979 Comparative Study of Greenhouse Locations through Satellite Images and Geographic Information System: Methodological Evaluation in Venezuela

Authors: Maria A. Castillo H., Andrés R. Leandro C.

Abstract:

During the last decades, agricultural productivity in Latin America has increased with precision agriculture and more efficient agricultural technologies. The use of automated systems, satellite images, geographic information systems, and tools for data analysis, and artificial intelligence have contributed to making more effective strategic decisions. Twenty years ago, the state of Mérida, located in the Venezuelan Andes, reported the largest area covered by greenhouses in the country, where certified seeds of potatoes, vegetables, ornamentals, and flowers were produced for export and consumption in the central region of the country. In recent years, it is estimated that production under greenhouses has changed, and the area covered has decreased due to different factors, but there are few historical statistical data in sufficient quantity and quality to support this estimate or to be used for analysis and decision making. The objective of this study is to compare data collected about geoposition, use, and covered areas of the greenhouses in 2007 to data available in 2021, as support for the analysis of the current situation of horticultural production in the main municipalities of the state of Mérida. The document presents the development of the work in the diagnosis and integration of geographic coordinates in GIS and data analysis phases. As a result, an evaluation of the process is made, a dashboard is presented with the most relevant data along with the geographical coordinates integrated into GIS, and an analysis of the obtained information is made. Finally, some recommendations for actions are added, and works that expand the information obtained and its geographical traceability over time are proposed. This study contributes to granting greater certainty in the supporting data for the evaluation of social, environmental, and economic sustainability indicators and to make better decisions according to the sustainable development goals in the area under review. At the same time, the methodology provides improvements to the agricultural data collection process that can be extended to other study areas and crops.

Keywords: greenhouses, geographic information system, protected agriculture, data analysis, Venezuela

Procedia PDF Downloads 89
15978 Inter-Communication-Management in Cases with Disabled Children (ICDC)

Authors: Dena A. Hussain

Abstract:

The objective of this project is to design an Information and Communication Technologies (ICT) tool based on a standardized platform to assist the work-integrated learning process of caretakers of disabled children. The tool should assist the intercommunication between caretakers and improve the learning process through knowledge bridging between all involved caretakers. Some children are born with disabilities while others have special needs after an illness or accident. Special needs children often need help in their learning process and require tools and services in a different way. In some cases the child has multiple disabilities that affect several capabilities in different ways. These needs are to be transformed into different learning techniques that the staff or personal (called caretakers in this project) caring for the child needs to learn and adapt. The caretakers involved are also required to learn new learning or training techniques and utilities specialized for the child’s needs. In many cases the number of people caring for the child’s development is rather large; the parents, specialist pedagogues, teachers, therapists, psychologists, personal assistants, etc. Each group of specialists has different objectives and in some cases the merge between theses specifications is very unique. This makes the synchronization between different caretakers difficult, resulting often in low level cooperation. By better intercommunication between professions both the child’s development could be improved but also the caretakers’ methods and knowledge of each other’s work processes and their own profession. This introduces a unique work integrated learning environment for all personnel involve, merging learning and knowledge in the work environment and at the same time assist the children’s development process. Creating an iterative process generates a unique learning experience for all involved. Using a work integrated platform will help encourage and support the process of all the teams involved in the process.We believe that working with children who have special needs is a continues learning/working process that is always integrated to achieve one main goal, which is to make a better future for all children.

Keywords: information and communication technologies (ICT), work integrated learning (WIL), sustainable learning, special needs children

Procedia PDF Downloads 292
15977 Segmentation of Liver Using Random Forest Classifier

Authors: Gajendra Kumar Mourya, Dinesh Bhatia, Akash Handique, Sunita Warjri, Syed Achaab Amir

Abstract:

Nowadays, Medical imaging has become an integral part of modern healthcare. Abdominal CT images are an invaluable mean for abdominal organ investigation and have been widely studied in the recent years. Diagnosis of liver pathologies is one of the major areas of current interests in the field of medical image processing and is still an open problem. To deeply study and diagnose the liver, segmentation of liver is done to identify which part of the liver is mostly affected. Manual segmentation of the liver in CT images is time-consuming and suffers from inter- and intra-observer differences. However, automatic or semi-automatic computer aided segmentation of the Liver is a challenging task due to inter-patient Liver shape and size variability. In this paper, we present a technique for automatic segmenting the liver from CT images using Random Forest Classifier. Random forests or random decision forests are an ensemble learning method for classification that operate by constructing a multitude of decision trees at training time and outputting the class that is the mode of the classes of the individual trees. After comparing with various other techniques, it was found that Random Forest Classifier provide a better segmentation results with respect to accuracy and speed. We have done the validation of our results using various techniques and it shows above 89% accuracy in all the cases.

Keywords: CT images, image validation, random forest, segmentation

Procedia PDF Downloads 308
15976 Impacts of School-Wide Positive Behavioral Interventions and Supports on Student Academics, Behavior and Mental Health

Authors: Catherine Bradshaw

Abstract:

Educators often report difficulty managing behavior problems and other mental health concerns that students display at school. These concerns also interfere with the learning process and can create distraction for teachers and other students. As such, schools play an important role in both preventing and intervening with students who experience these types of challenges. A number of models have been proposed to serve as a framework for delivering prevention and early intervention services in schools. One such model is called Positive Behavioral Interventions and Supports (PBIS), which has been scaled-up to over 26,000 schools in the U.S. and many other countries worldwide. PBIS aims to improve a range of student outcomes through early detection of and intervention related to behavioral and mental health symptoms. PBIS blends and applies social learning, behavioral, and organizational theories to prevent disruptive behavior and enhance the school’s organizational health. PBIS focuses on creating and sustaining tier 1 (universal), tier 2 (selective), and tier 3 (individual) systems of support. Most schools using PBIS have focused on the core elements of the tier 1 supports, which includes the following critical features. The formation of a PBIS team within the school to lead implementation. Identification and training of a behavioral support ‘coach’, who serves as a on-site technical assistance provider. Many of the individuals identified to serve as a PBIS coach are also trained as a school psychologist or guidance counselor; coaches typically have prior PBIS experience and are trained to conduct functional behavioral assessments. The PBIS team also identifies a set of three to five positive behavioral expectations that are implemented for all students and by all staff school-wide (e.g., ‘be respectful, responsible, and ready to learn’); these expectations are posted in all settings across the school, including in the classroom, cafeteria, playground etc. All school staff define and teach the school-wide behavioral expectations to all students and review them regularly. Finally, PBIS schools develop or adopt a school-wide system to reward or reinforce students who demonstrate those 3-5 positive behavioral expectations. Staff and administrators create an agreed upon system for responding to behavioral violations that include definitions about what constitutes a classroom-managed vs. an office-managed discipline problem. Finally, a formal system is developed to collect, analyze, and use disciplinary data (e.g., office discipline referrals) to inform decision-making. This presentation provides a brief overview of PBIS and reports findings from a series of four U.S. based longitudinal randomized controlled trials (RCTs) documenting the impacts of PBIS on school climate, discipline problems, bullying, and academic achievement. The four RCTs include 80 elementary, 40 middle, and 58 high schools and results indicate a broad range of impacts on multiple student and school-wide outcomes. The session will highlight lessons learned regarding PBIS implementation and scale-up. We also review the ways in which PBIS can help educators and school leaders engage in data-based decision-making and share data with other decision-makers and stakeholders (e.g., students, parents, community members), with the overarching goal of increasing use of evidence-based programs in schools.

Keywords: positive behavioral interventions and supports, mental health, randomized trials, school-based prevention

Procedia PDF Downloads 225
15975 Optimization of Hate Speech and Abusive Language Detection on Indonesian-language Twitter using Genetic Algorithms

Authors: Rikson Gultom

Abstract:

Hate Speech and Abusive language on social media is difficult to detect, usually, it is detected after it becomes viral in cyberspace, of course, it is too late for prevention. An early detection system that has a fairly good accuracy is needed so that it can reduce conflicts that occur in society caused by postings on social media that attack individuals, groups, and governments in Indonesia. The purpose of this study is to find an early detection model on Twitter social media using machine learning that has high accuracy from several machine learning methods studied. In this study, the support vector machine (SVM), Naïve Bayes (NB), and Random Forest Decision Tree (RFDT) methods were compared with the Support Vector machine with genetic algorithm (SVM-GA), Nave Bayes with genetic algorithm (NB-GA), and Random Forest Decision Tree with Genetic Algorithm (RFDT-GA). The study produced a comparison table for the accuracy of the hate speech and abusive language detection model, and presented it in the form of a graph of the accuracy of the six algorithms developed based on the Indonesian-language Twitter dataset, and concluded the best model with the highest accuracy.

Keywords: abusive language, hate speech, machine learning, optimization, social media

Procedia PDF Downloads 125
15974 The Effect of Substrate Surface Roughness for Hot Dip Aluminizing of IN718 Alloy

Authors: Aptullah Karakas, Murat Baydogan

Abstract:

The hot dip aluminizing (HDA) process involves immersing a metallic substrate into a molten aluminum bath for several minutes, and removed from the bath and cooled down to room temperature. After the HDA process, various aluminide layers are formed as a result of interdiffusion between the substrate and the molten aluminum and between the aluminide layers. In order to form a uniform aluminide layer, the specimen must be covered and wet well by the molten aluminum. Surface roughness plays an important role in wettability, and thus, surface preparation is an important stage in determining the final surface roughness. In this study, different roughness values were achieved by grinding the surface with emery papers as 180, 320 and 600 grids. After the surface preparation, the HDA process was performed in a molten Al-Si bath at 700 ᴼC for 10 minutes. After the HDA process, a microstructural examination of the coating was carried out to evaluate the uniformity of the coating and adhesion between the substrate and the coating. According to the results, the best adhesion at the interface was observed on the specimen, which was prepared by 320 grid emery paper having a mean surface roughness (Ra) of 0.097 µm.

Keywords: hot-dip aluminizing, microstructure, surface roughness, coating

Procedia PDF Downloads 64
15973 The Implementation of the Lean Six Sigma Production Process in a Telecommunications Company in Brazil

Authors: Carlos Fontanillas

Abstract:

The implementation of the lean six sigma methodology aims to implement practices to systematically improve processes by eliminating defects, making them cheaper. The implementation of projects with the methodology uses a division into five phases: definition, measurement, analysis, implementation, and control. In this process, it is understood that the implementation of said methodology generates benefits to organizations that adhere through the improvement of their processes. In the case of a telecommunications company, it was realized that the implementation of a lean six sigma project contributed to the improvement of the presented process, generating a financial return with the avoided cost. However, such study has limitations such as a specific segment of performance and procedure, i.e., it can not be defined that return under other circumstances will be the same. It is also concluded that lean six sigma projects tend to contribute to improved processes evaluated due to their methodology that is based on statistical analysis and quality management tools and can generate a financial return. It is hoped that the present study can be used to provide a clearer view of the methodology for entrepreneurs who wish to implement process improvement actions in their companies, as well as to provide a foundation for professionals working with lean six sigma projects. After the review of the processes, the completion of the project stages and the monitoring for three months in partnership with the owner of the process to ensure the effectiveness of the actions, the project was completed with the objective reached. There was an average of 60% reduction with the issuance of undue invoices generated after the deactivation and it was possible to extend the project to other companies, which allowed a reduction well above the initially stipulated target.

Keywords: quality, process, lean six sigma, organization

Procedia PDF Downloads 126
15972 Discussion about Frequent Adjustment of Urban Master Planning in China: A Case Study of Changshou District, Chongqing City

Authors: Sun Ailu, Zhao Wanmin

Abstract:

Since the reform and opening, the urbanization process of China has entered a rapid development period. In recent years, the authors participated in some projects of urban master planning in China and found a phenomenon that the rapid urbanization area of China is experiencing frequent adjustment process of urban master planning. This phenomenon is not the natural process of urbanization development. It may be caused by different government roles from different levels. Through the methods of investigation, data comparison and case study, this paper aims to explore the reason why the rapid urbanization area is experiencing frequent adjustment of master planning and give some solution strategies. Firstly, taking Changshou district of Chongqing city as an example, this paper wants to introduce the phenomenon about frequent adjustment process in China. And then, discuss distinct roles in the process between national government, provincial government and local government of China. At last, put forward preliminary solutions strategies for this area in China from the aspects of land use, intergovernmental cooperation and so on.

Keywords: urban master planning, frequent adjustment, urbanization development, problems and strategies, China

Procedia PDF Downloads 360
15971 Deep Reinforcement Learning Model for Autonomous Driving

Authors: Boumaraf Malak

Abstract:

The development of intelligent transportation systems (ITS) and artificial intelligence (AI) are spurring us to pave the way for the widespread adoption of autonomous vehicles (AVs). This is open again opportunities for smart roads, smart traffic safety, and mobility comfort. A highly intelligent decision-making system is essential for autonomous driving around dense, dynamic objects. It must be able to handle complex road geometry and topology, as well as complex multiagent interactions, and closely follow higher-level commands such as routing information. Autonomous vehicles have become a very hot research topic in recent years due to their significant ability to reduce traffic accidents and personal injuries. Using new artificial intelligence-based technologies handles important functions in scene understanding, motion planning, decision making, vehicle control, social behavior, and communication for AV. This paper focuses only on deep reinforcement learning-based methods; it does not include traditional (flat) planar techniques, which have been the subject of extensive research in the past because reinforcement learning (RL) has become a powerful learning framework now capable of learning complex policies in high dimensional environments. The DRL algorithm used so far found solutions to the four main problems of autonomous driving; in our paper, we highlight the challenges and point to possible future research directions.

Keywords: deep reinforcement learning, autonomous driving, deep deterministic policy gradient, deep Q-learning

Procedia PDF Downloads 77
15970 Computational Pipeline for Lynch Syndrome Detection: Integrating Alignment, Variant Calling, and Annotations

Authors: Rofida Gamal, Mostafa Mohammed, Mariam Adel, Marwa Gamal, Marwa kamal, Ayat Saber, Maha Mamdouh, Amira Emad, Mai Ramadan

Abstract:

Lynch Syndrome is an inherited genetic condition associated with an increased risk of colorectal and other cancers. Detecting Lynch Syndrome in individuals is crucial for early intervention and preventive measures. This study proposes a computational pipeline for Lynch Syndrome detection by integrating alignment, variant calling, and annotation. The pipeline leverages popular tools such as FastQC, Trimmomatic, BWA, bcftools, and ANNOVAR to process the input FASTQ file, perform quality trimming, align reads to the reference genome, call variants, and annotate them. It is believed that the computational pipeline was applied to a dataset of Lynch Syndrome cases, and its performance was evaluated. It is believed that the quality check step ensured the integrity of the sequencing data, while the trimming process is thought to have removed low-quality bases and adaptors. In the alignment step, it is believed that the reads were accurately mapped to the reference genome, and the subsequent variant calling step is believed to have identified potential genetic variants. The annotation step is believed to have provided functional insights into the detected variants, including their effects on known Lynch Syndrome-associated genes. The results obtained from the pipeline revealed Lynch Syndrome-related positions in the genome, providing valuable information for further investigation and clinical decision-making. The pipeline's effectiveness was demonstrated through its ability to streamline the analysis workflow and identify potential genetic markers associated with Lynch Syndrome. It is believed that the computational pipeline presents a comprehensive and efficient approach to Lynch Syndrome detection, contributing to early diagnosis and intervention. The modularity and flexibility of the pipeline are believed to enable customization and adaptation to various datasets and research settings. Further optimization and validation are believed to be necessary to enhance performance and applicability across diverse populations.

Keywords: Lynch Syndrome, computational pipeline, alignment, variant calling, annotation, genetic markers

Procedia PDF Downloads 67
15969 The Delone and McLean Model: A Review and Reconceptualisation for Explaining Organisational IS Success

Authors: Probir Kumar Banerjee

Abstract:

Though the revised DeLone and McLean (DM) model of IS success is found to be effective at the individual level of analysis, there is lack of consensus in regard to its effectiveness at the organisational level. This research reviews the DM model in the light of business/IT alignment theory and supporting literature, and suggests its reconceptualization. Specifically, arguments are made for augmenting it with business process quality. Business process quality, it is argued, captures the effect of intent to use, use and user satisfaction interactions, thus eliminating the need to capture their interaction effects in explaining organisational IS success. It is also argued that ‘operational performance’ driven by systems and business process quality, and higher order measures of organisational performance tied to operational performance are appropriate measures of ‘net benefit’. Suggestions are made for reconceptualisation of the other constructs and an adapted model of organisational IS success is proposed.

Keywords: organisational IS success, business/IT alignment, systems quality, business process quality, operational performance, market performance

Procedia PDF Downloads 392
15968 AHP and TOPSIS Methods for Supplier Selection Problem in Medical Devices Company

Authors: Sevde D. Karayel, Ediz Atmaca

Abstract:

Supplier selection subject is vital because of development competitiveness and performance of firms which have right, rapid and with low cost procurement. Considering the fact that competition between firms is no longer on their supply chains, hence it is very clear that performance of the firms’ not only depend on their own success but also success of all departments in supply chain. For this purpose, firms want to work with suppliers which are cost effective, flexible in terms of demand and high quality level for customer satisfaction. However, diversification and redundancy of their expectations from suppliers, supplier selection problems need to be solved as a hard problem. In this study, supplier selection problem is discussed for critical piece, which is using almost all production of products in and has troubles with lead time from supplier, in a firm that produces medical devices. Analyzing policy in the current situation of the firm in the supplier selection indicates that supplier selection is made based on the purchasing department experience and other authorized persons’ general judgments. Because selection do not make based on the analytical methods, it is caused disruptions in production, lateness and extra cost. To solve the problem, AHP and TOPSIS which are multi-criteria decision making techniques, which are effective, easy to implement and can analyze many criteria simultaneously, are used to make a selection among alternative suppliers.

Keywords: AHP-TOPSIS methods, multi-criteria decision making, supplier selection problem, supply chain management

Procedia PDF Downloads 258
15967 Research on the Effect of the System of General Counsel on the Efficiency of M&As in State-Owned Enterprises

Authors: Mao Ju

Abstract:

The system of general counsel (GC) is an important governance structure designed for the construction of state-owned enterprises (SOEs) under the rule of law. This article is based on the setting of mergers and acquisitions (M&As) and takes the efficiency of M&As to examine the implementation effect of the system of GC for SOEs. Research has found that: (1) companies implementing the system of GC for SOEs have higher efficiency in M&As, manifested in better operational and market performance, and this effect depends on the professional ability and power of the GC. This indicates that the GC of SOEs has played a positive role in the decision-making process of M&As, which helps to improve the efficiency of M&As. (2) The impact of the GC of SOEs on the efficiency of M&As is heterogeneous, and this positive effect is mainly reflected in local and commercial SOEs. (3) The path of this impact is that the GC of SOEs can help reduce ineffective M&As in advance, enhance the ability to integrate M&As after the fact and reduce the risk of goodwill impairment and bankruptcy. This article reveals the impact of the construction of SOEs under the rule of law with the system of GC as the core of M&As activities, providing intuitive evidence for the implementation effect of the GC of SOEs. The research conclusion has important practical guiding value for comprehensively deepening the construction of the rule of SOEs under the rule of law and writing a good chapter on the Chinese path to modernization of SOEs.

Keywords: the system of general counsel, merger and acquisition efficiency, state-owned enterprises, mergers and acquisitions

Procedia PDF Downloads 0
15966 Combination of Modelling and Environmental Life Cycle Assessment Approach for Demand Driven Biogas Production

Authors: Juan A. Arzate, Funda C. Ertem, M. Nicolas Cruz-Bournazou, Peter Neubauer, Stefan Junne

Abstract:

— One of the biggest challenges the world faces today is global warming that is caused by greenhouse gases (GHGs) coming from the combustion of fossil fuels for energy generation. In order to mitigate climate change, the European Union has committed to reducing GHG emissions to 80–95% below the level of the 1990s by the year 2050. Renewable technologies are vital to diminish energy-related GHG emissions. Since water and biomass are limited resources, the largest contributions to renewable energy (RE) systems will have to come from wind and solar power. Nevertheless, high proportions of fluctuating RE will present a number of challenges, especially regarding the need to balance the variable energy demand with the weather dependent fluctuation of energy supply. Therefore, biogas plants in this content would play an important role, since they are easily adaptable. Feedstock availability varies locally or seasonally; however there is a lack of knowledge in how biogas plants should be operated in a stable manner by local feedstock. This problem may be prevented through suitable control strategies. Such strategies require the development of convenient mathematical models, which fairly describe the main processes. Modelling allows us to predict the system behavior of biogas plants when different feedstocks are used with different loading rates. Life cycle assessment (LCA) is a technique for analyzing several sides from evolution of a product till its disposal in an environmental point of view. It is highly recommend to use as a decision making tool. In order to achieve suitable strategies, the combination of a flexible energy generation provided by biogas plants, a secure production process and the maximization of the environmental benefits can be obtained by the combination of process modelling and LCA approaches. For this reason, this study focuses on the biogas plant which flexibly generates required energy from the co-digestion of maize, grass and cattle manure, while emitting the lowest amount of GHG´s. To achieve this goal AMOCO model was combined with LCA. The program was structured in Matlab to simulate any biogas process based on the AMOCO model and combined with the equations necessary to obtain climate change, acidification and eutrophication potentials of the whole production system based on ReCiPe midpoint v.1.06 methodology. Developed simulation was optimized based on real data from operating biogas plants and existing literature research. The results prove that AMOCO model can successfully imitate the system behavior of biogas plants and the necessary time required for the process to adapt in order to generate demanded energy from available feedstock. Combination with LCA approach provided opportunity to keep the resulting emissions from operation at the lowest possible level. This would allow for a prediction of the process, when the feedstock utilization supports the establishment of closed material circles within a smart bio-production grid – under the constraint of minimal drawbacks for the environment and maximal sustainability.

Keywords: AMOCO model, GHG emissions, life cycle assessment, modelling

Procedia PDF Downloads 187
15965 A Review on Existing Challenges of Data Mining and Future Research Perspectives

Authors: Hema Bhardwaj, D. Srinivasa Rao

Abstract:

Technology for analysing, processing, and extracting meaningful data from enormous and complicated datasets can be termed as "big data." The technique of big data mining and big data analysis is extremely helpful for business movements such as making decisions, building organisational plans, researching the market efficiently, improving sales, etc., because typical management tools cannot handle such complicated datasets. Special computational and statistical issues, such as measurement errors, noise accumulation, spurious correlation, and storage and scalability limitations, are brought on by big data. These unique problems call for new computational and statistical paradigms. This research paper offers an overview of the literature on big data mining, its process, along with problems and difficulties, with a focus on the unique characteristics of big data. Organizations have several difficulties when undertaking data mining, which has an impact on their decision-making. Every day, terabytes of data are produced, yet only around 1% of that data is really analyzed. The idea of the mining and analysis of data and knowledge discovery techniques that have recently been created with practical application systems is presented in this study. This article's conclusion also includes a list of issues and difficulties for further research in the area. The report discusses the management's main big data and data mining challenges.

Keywords: big data, data mining, data analysis, knowledge discovery techniques, data mining challenges

Procedia PDF Downloads 105
15964 Design of Visual Repository, Constraint and Process Modeling Tool Based on Eclipse Plug-Ins

Authors: Rushiraj Heshi, Smriti Bhandari

Abstract:

Master Data Management requires creation of Central repository, applying constraints on Repository and designing processes to manage data. Designing of Repository, constraints on repository and business processes is very tedious and time consuming task for large Enterprise. Hence Visual Repository, constraints and Process (Workflow) modeling is the most critical step in Master Data Management.In this paper, we realize a Visual Modeling tool for implementing Repositories, Constraints and Processes based on Eclipse Plugin using GMF/EMF which follows principles of Model Driven Engineering (MDE).

Keywords: EMF, GMF, GEF, repository, constraint, process

Procedia PDF Downloads 489
15963 Multivariate Control Chart to Determine Efficiency Measurements in Industrial Processes

Authors: J. J. Vargas, N. Prieto, L. A. Toro

Abstract:

Control charts are commonly used to monitor processes involving either variable or attribute of quality characteristics and determining the control limits as a critical task for quality engineers to improve the processes. Nonetheless, in some applications it is necessary to include an estimation of efficiency. In this paper, the ability to define the efficiency of an industrial process was added to a control chart by means of incorporating a data envelopment analysis (DEA) approach. In depth, a Bayesian estimation was performed to calculate the posterior probability distribution of parameters as means and variance and covariance matrix. This technique allows to analyse the data set without the need of using the hypothetical large sample implied in the problem and to be treated as an approximation to the finite sample distribution. A rejection simulation method was carried out to generate random variables from the parameter functions. Each resulting vector was used by stochastic DEA model during several cycles for establishing the distribution of each efficiency measures for each DMU (decision making units). A control limit was calculated with model obtained and if a condition of a low level efficiency of DMU is presented, system efficiency is out of control. In the efficiency calculated a global optimum was reached, which ensures model reliability.

Keywords: data envelopment analysis, DEA, Multivariate control chart, rejection simulation method

Procedia PDF Downloads 372
15962 Support for Planning of Mobile Personnel Tasks by Solving Time-Dependent Routing Problems

Authors: Wlodzimierz Ogryczak, Tomasz Sliwinski, Jaroslaw Hurkala, Mariusz Kaleta, Bartosz Kozlowski, Piotr Palka

Abstract:

Implementation concepts of a decision support system for planning and management of mobile personnel tasks (sales representatives and others) are discussed. Large-scale periodic time-dependent vehicle routing and scheduling problems with complex constraints are solved for this purpose. Complex nonuniform constraints with respect to frequency, time windows, working time, etc. are taken into account with additional fast adaptive procedures for operational rescheduling of plans in the presence of various disturbances. Five individual solution quality indicators with respect to a single personnel person are considered. This paper deals with modeling issues corresponding to the problem and general solution concepts. The research was supported by the European Union through the European Regional Development Fund under the Operational Programme ‘Innovative Economy’ for the years 2007-2013; Priority 1 Research and development of modern technologies under the project POIG.01.03.01-14-076/12: 'Decision Support System for Large-Scale Periodic Vehicle Routing and Scheduling Problems with Complex Constraints.'

Keywords: mobile personnel management, multiple criteria, time dependent, time windows, vehicle routing and scheduling

Procedia PDF Downloads 322
15961 Mixed Integer Programing for Multi-Tier Rebate with Discontinuous Cost Function

Authors: Y. Long, L. Liu, K. V. Branin

Abstract:

One challenge faced by procurement decision-maker during the acquisition process is how to compare similar products from different suppliers and allocate orders among different products or services. This work focuses on allocating orders among multiple suppliers considering rebate. The objective function is to minimize the total acquisition cost including purchasing cost and rebate benefit. Rebate benefit is complex and difficult to estimate at the ordering step. Rebate rules vary for different suppliers and usually change over time. In this work, we developed a system to collect the rebate policies, standardized the rebate policies and developed two-stage optimization models for ordering allocation. Rebate policy with multi-tiers is considered in modeling. The discontinuous cost function of rebate benefit is formulated for different scenarios. A piecewise linear function is used to approximate the discontinuous cost function of rebate benefit. And a Mixed Integer Programing (MIP) model is built for order allocation problem with multi-tier rebate. A case study is presented and it shows that our optimization model can reduce the total acquisition cost by considering rebate rules.

Keywords: discontinuous cost function, mixed integer programming, optimization, procurement, rebate

Procedia PDF Downloads 255
15960 An Integrated Fuzzy Inference System and Technique for Order of Preference by Similarity to Ideal Solution Approach for Evaluation of Lean Healthcare Systems

Authors: Aydin M. Torkabadi, Ehsan Pourjavad

Abstract:

A decade after the introduction of Lean in Saskatchewan’s public healthcare system, its effectiveness remains a controversial subject among health researchers, workers, managers, and politicians. Therefore, developing a framework to quantitatively assess the Lean achievements is significant. This study investigates the success of initiatives across Saskatchewan health regions by recognizing the Lean healthcare criteria, measuring the success levels, comparing the regions, and identifying the areas for improvements. This study proposes an integrated intelligent computing approach by applying Fuzzy Inference System (FIS) and Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS). FIS is used as an efficient approach to assess the Lean healthcare criteria, and TOPSIS is applied for ranking the values in regards to the level of leanness. Due to the innate uncertainty in decision maker judgments on criteria, principals of the fuzzy theory are applied. Finally, FIS-TOPSIS was established as an efficient technique in determining the lean merit in healthcare systems.

Keywords: lean healthcare, intelligent computing, fuzzy inference system, healthcare evaluation, technique for order of preference by similarity to ideal solution, multi-criteria decision making, MCDM

Procedia PDF Downloads 154
15959 Effect of Rubber Tyre and Plastic Wastes Use in Asphalt Concrete Pavement

Authors: F. Onyango, Salim R. Wanjala, M. Ndege, L. Masu

Abstract:

Asphalt concrete pavements have a short life cycle, failing mainly due to temperature changes, traffic loading and ageing. Modified asphalt mixtures provide the technology to produce a bituminous binder with improved viscoelastic properties which remain in balance over a wider temperature range and loading conditions. In this research, 60/70 penetration grade asphalt binder was modified by adding 2, 4, 6, 8, and 10 percent by weight of asphalt binder following the wet process and the mineral aggregate was modified by adding 1, 2, 3, 4, and 5 percent crumb rubber by volume of the mineral aggregate following the dry process. The LDPE modified asphalt binder Rheological properties were evaluated. The laboratory results showed an increase in viscosity, softening point and stiffness of the binder. The modified asphalt was then used in preparing asphalt mixtures by Marshall Mix design procedure. The Marshall stability values for mixes containing 2% crumb rubber and 4% LDPE were found to be 30% higher than the conventional asphalt concrete mix.

Keywords: crumb rubber, dry process, hot mix asphalt, wet process

Procedia PDF Downloads 364
15958 Artificial Intelligence and Governance in Relevance to Satellites in Space

Authors: Anwesha Pathak

Abstract:

With the increasing number of satellites and space debris, space traffic management (STM) becomes crucial. AI can aid in STM by predicting and preventing potential collisions, optimizing satellite trajectories, and managing orbital slots. Governance frameworks need to address the integration of AI algorithms in STM to ensure safe and sustainable satellite activities. AI and governance play significant roles in the context of satellite activities in space. Artificial intelligence (AI) technologies, such as machine learning and computer vision, can be utilized to process vast amounts of data received from satellites. AI algorithms can analyse satellite imagery, detect patterns, and extract valuable information for applications like weather forecasting, urban planning, agriculture, disaster management, and environmental monitoring. AI can assist in automating and optimizing satellite operations. Autonomous decision-making systems can be developed using AI to handle routine tasks like orbit control, collision avoidance, and antenna pointing. These systems can improve efficiency, reduce human error, and enable real-time responsiveness in satellite operations. AI technologies can be leveraged to enhance the security of satellite systems. AI algorithms can analyze satellite telemetry data to detect anomalies, identify potential cyber threats, and mitigate vulnerabilities. Governance frameworks should encompass regulations and standards for securing satellite systems against cyberattacks and ensuring data privacy. AI can optimize resource allocation and utilization in satellite constellations. By analyzing user demands, traffic patterns, and satellite performance data, AI algorithms can dynamically adjust the deployment and routing of satellites to maximize coverage and minimize latency. Governance frameworks need to address fair and efficient resource allocation among satellite operators to avoid monopolistic practices. Satellite activities involve multiple countries and organizations. Governance frameworks should encourage international cooperation, information sharing, and standardization to address common challenges, ensure interoperability, and prevent conflicts. AI can facilitate cross-border collaborations by providing data analytics and decision support tools for shared satellite missions and data sharing initiatives. AI and governance are critical aspects of satellite activities in space. They enable efficient and secure operations, ensure responsible and ethical use of AI technologies, and promote international cooperation for the benefit of all stakeholders involved in the satellite industry.

Keywords: satellite, space debris, traffic, threats, cyber security.

Procedia PDF Downloads 67
15957 Determination of the Economic Planning Depth for Assembly Process Planning

Authors: A. Kampker, P. Burggräf, Y. Bäumers

Abstract:

In order to be competitive, companies have to reduce their production costs while meeting increasing quality requirements. Therefore, companies try to plan their assembly processes as detailed as possible. However, increasing product individualization leading to a higher number of variants, smaller batch sizes and shorter product life cycles raise the question to what extent the effort of detailed planning is still justified. An important approach in this field of research is the concept of determining the economic planning depth for assembly process planning based on production specific influencing factors. In this paper, first solution hypotheses as well as a first draft of the resulting method will be presented.

Keywords: assembly process planning, economic planning depth, planning benefit, planning effort

Procedia PDF Downloads 500
15956 Governance and Public Policy: The Perception of Civil Society Participation in Brazil and South Africa

Authors: Paulino V. Tavares, Ana L. Romao

Abstract:

Public governance, in general, is essential to qualify and educate, pedagogically, the decision-making process of the government in relation to the management of resources and the provision of public services, with transparency and active participation of individuals and citizens for the development of a more democratic environment, besides stimulating control and social empowerment, aiming at the development of the collectivity. In this context, the participation of society in the elaboration, execution, and control of public policies is prominent to strengthen public governance itself. With this, using a multidimensional approach with the application of two questionnaires to a universe of twenty Counselors of the Courts of Auditors (Brazil), twenty professionals of public administration (Brazil), twenty Government/Provincial Counselors (South Africa), and twenty South African professionals of public administration, the preliminary results indicate that the participation of civil society, for both countries, is very low in the elaboration, execution, and control of public policies. At the same time, about 70% of the answers obtained indicate, on average, three possible paths to increase the participation of civil society. With this, it is delineated that developing new horizons to strengthen both public policies how social participation is necessary, but, for both, it is important that governments and civil society, in their respective countries, have an awareness of the effective importance of this interaction.

Keywords: Brazil, civil society, participation, South Africa

Procedia PDF Downloads 140
15955 Defining Methodology for Multi Model Software Process Improvement Framework

Authors: Aedah Abd Rahman

Abstract:

Software organisations may implement single or multiple frameworks in order to remain competitive. There are wide selection of generic Software Process Improvement (SPI) frameworks, best practices and standards implemented with different focuses and goals. Issues and difficulties emerge in the SPI practices from the context of software development and IT Service Management (ITSM). This research looks into the integration of multiple frameworks from the perspective of software development and ITSM. The research question of this study is how to define steps of methodology to solve the multi model software process improvement problem. The objective of this study is to define the research approach and methodologies to produce a more integrated and efficient Multi Model Process Improvement (MMPI) solution. A multi-step methodology is used which contains the case study, framework mapping and Delphi study. The research outcome has proven the usefulness and appropriateness of the proposed framework in SPI and quality practice in Malaysian software industries. This mixed method research approach is used to tackle problems from every angle in the context of software development and services. This methodology is used to facilitate the implementation and management of multi model environment of SPI frameworks in multiple domains.

Keywords: Delphi study, methodology, multi model software process improvement, service management

Procedia PDF Downloads 257